Parametric vs Nonparametric Learning¶
Statistical learning methods in quantitative finance can be broadly classified into parametric and nonparametric approaches. This distinction reflects a fundamental trade-off between structural assumptions and flexibility, with profound implications for model stability, data efficiency, and out-of-sample performance.
The Learning Problem¶
Consider the general supervised learning setup. We observe data \((X_1, Y_1), \ldots, (X_n, Y_n)\) drawn from some unknown joint distribution \(\mathbb{P}_{X,Y}\) and seek to estimate a function \(f^*\) that minimizes expected loss:
where \(\mathcal{F}\) is the hypothesis class and \(L\) is a loss function (e.g., squared error \(L(y, \hat{y}) = (y - \hat{y})^2\)).
The choice of \(\mathcal{F}\) determines whether our approach is parametric or nonparametric.
Parametric Learning¶
Definition¶
A parametric model assumes that the true function \(f^*\) belongs to a family indexed by a finite-dimensional parameter:
The learning problem reduces to estimating \(\theta\) from data.
Examples in Finance¶
Linear factor models. Asset returns are modeled as:
where \(\theta = (\alpha_i, \beta_{i1}, \ldots, \beta_{ik})\) has dimension \(k+1\).
GARCH volatility models. The conditional variance follows:
with parameter vector \(\theta = (\omega, \alpha_1, \ldots, \alpha_p, \beta_1, \ldots, \beta_q) \in \mathbb{R}^{p+q+1}\).
Parametric option pricing. Under Black–Scholes, the single parameter \(\sigma\) (volatility) determines all option prices. Under Heston, we have \(\theta = (v_0, \kappa, \bar{v}, \sigma_v, \rho) \in \mathbb{R}^5\).
Maximum Likelihood Estimation¶
For parametric models with density \(p(y|x; \theta)\), the maximum likelihood estimator is:
Under regularity conditions, MLE achieves asymptotic efficiency:
where \(\mathcal{I}(\theta)\) is the Fisher information matrix:
Convergence Rates¶
Parametric estimators typically achieve the parametric rate of convergence:
This rate is independent of the input dimension, making parametric methods highly data-efficient when the model is correctly specified.
Model Misspecification¶
The critical vulnerability of parametric methods is misspecification. If \(f^* \notin \mathcal{F}_{\text{param}}\), then even with infinite data, the best we can achieve is the pseudo-true parameter:
where \(D_{\text{KL}}\) denotes Kullback–Leibler divergence. The approximation error \(\|f(\cdot; \theta^\dagger) - f^*\|\) does not vanish with more data.
Nonparametric Learning¶
Definition¶
Nonparametric models do not assume a fixed functional form. Instead, the hypothesis class grows with sample size:
Common smoothness assumptions include Hölder continuity, bounded derivatives, or membership in a Sobolev space.
Kernel Regression (Nadaraya–Watson Estimator)¶
Given observations \((X_i, Y_i)_{i=1}^n\), the kernel regression estimator is:
where \(K_h(\cdot) = h^{-d} K(\cdot/h)\) is a scaled kernel (e.g., Gaussian) and \(h > 0\) is the bandwidth.
Bias-variance decomposition. For the Nadaraya–Watson estimator:
where \(\mu_2(K) = \int u^2 K(u) du\), \(R(K) = \int K(u)^2 du\), \(p(x)\) is the density of \(X\), and \(\sigma^2(x) = \text{Var}(Y|X=x)\).
Local Polynomial Regression¶
Local polynomial regression fits a polynomial of degree \(p\) locally around each point \(x\):
The estimator is \(\hat{f}(x) = \hat{\beta}_0(x)\). Local linear regression (\(p=1\)) has superior boundary behavior compared to Nadaraya–Watson.
Regression Splines¶
Cubic splines approximate \(f\) using piecewise polynomials with continuity constraints at knots \(\xi_1 < \cdots < \xi_K\):
where \((u)_+ = \max(u, 0)\).
Smoothing splines solve:
where \(\lambda > 0\) controls the smoothness penalty.
k-Nearest Neighbors¶
The k-NN estimator averages responses from the \(k\) closest training points:
where \(N_k(x)\) denotes the indices of the \(k\) nearest neighbors of \(x\).
Convergence Rates and Stone's Theorem¶
Theorem (Stone, 1982). Let \(f^*\) belong to the Hölder class \(\mathcal{H}(\beta, L)\) of functions with bounded derivatives up to order \(\lfloor \beta \rfloor\) and \(\beta\)-Hölder continuous highest derivative. Then for any nonparametric regression estimator:
This minimax optimal rate reveals the curse of dimensionality: as dimension \(d\) increases, convergence slows dramatically.
Example. For twice-differentiable functions (\(\beta = 2\)):
- \(d = 1\): Rate is \(n^{-4/5}\)
- \(d = 5\): Rate is \(n^{-4/9}\)
- \(d = 10\): Rate is \(n^{-2/7}\)
- \(d = 20\): Rate is \(n^{-1/6}\)
With 1,000 observations, effective estimation in 20 dimensions requires smoothness equivalent to \(1000^{1/6} \approx 3\) observations in one dimension.
Semi-Parametric Models¶
Definition¶
Semi-parametric models combine parametric components for primary structure with nonparametric components for nuisance functions:
where \(\beta \in \mathbb{R}^p\) is finite-dimensional and \(f\) is nonparametric.
Partially Linear Models¶
The partially linear model:
is estimated via Robinson's differencing:
- Estimate \(\mathbb{E}[Y|Z]\) and \(\mathbb{E}[X|Z]\) nonparametrically
- Compute residuals: \(\tilde{Y} = Y - \hat{\mathbb{E}}[Y|Z]\), \(\tilde{X} = X - \hat{\mathbb{E}}[X|Z]\)
- Estimate \(\beta\) by OLS on residuals: \(\hat{\beta} = (\tilde{X}^\top \tilde{X})^{-1} \tilde{X}^\top \tilde{Y}\)
Under regularity conditions, \(\hat{\beta}\) achieves the parametric rate \(\sqrt{n}\)-consistency despite the nonparametric component.
Single-Index Models¶
The single-index model:
reduces a high-dimensional regression to estimating a direction \(\beta\) and a univariate link function \(g\).
Average derivative estimation. Under suitable conditions:
allowing consistent estimation of \(\beta\) (up to scale) using kernel density estimates.
Additive Models¶
Additive models assume:
The backfitting algorithm iteratively estimates each \(f_j\):
- Initialize \(\hat{f}_j^{(0)} = 0\) for all \(j\)
- For \(m = 1, 2, \ldots\):
-
For \(j = 1, \ldots, d\):
\(\(\hat{f}_j^{(m)} = \mathcal{S}_j\left(Y - \hat{\alpha} - \sum_{k \neq j} \hat{f}_k^{(m-1)}\right)\)\)
where \(\mathcal{S}_j\) is a univariate smoother 3. Iterate until convergence
Additive models achieve the univariate rate \(n^{-2\beta/(2\beta+1)}\) rather than the multivariate rate, circumventing the curse of dimensionality under the additivity assumption.
Financial Applications and Considerations¶
Option Pricing: Parametric vs Nonparametric¶
Parametric approach (risk-neutral). Assume dynamics under \(\mathbb{Q}\):
with parametric volatility \(\sigma(S, t; \theta)\). Option prices are computed via PDE or Monte Carlo.
Nonparametric approach. Directly estimate the option pricing function:
from observed prices, without specifying dynamics.
Trade-offs: - Parametric: Ensures no-arbitrage (if model is arbitrage-free), extrapolates consistently, but may misspecify dynamics - Nonparametric: Flexible fit to observed prices, but may violate no-arbitrage, poor extrapolation
Volatility Surface Estimation¶
The implied volatility surface \(\sigma_{\text{imp}}(K, T)\) can be estimated:
Parametrically: SVI parameterization (see Chapter 8):
where \(k = \log(K/F)\) is log-moneyness.
Nonparametrically: Kernel smoothing of observed implied volatilities:
Semi-parametrically: Parametric term structure with nonparametric smile:
where \(\sigma_{\text{ATM}}(T)\) is parametric and \(g\) is estimated nonparametrically.
Return Prediction¶
For predicting returns \(R_{t+1}\) from predictors \(X_t\):
Parametric (linear predictability):
Campbell and Shiller's dividend-price ratio regression is a classic example.
Nonparametric:
estimated via kernel methods or trees.
Empirical finding: Due to low signal-to-noise ratios in returns (often \(R^2 < 0.05\)), nonparametric methods frequently underperform simple linear models out-of-sample. The flexibility that allows fitting complex patterns also fits noise.
Model Selection¶
Information Criteria¶
For parametric models with \(d\) parameters and maximized log-likelihood \(\ell_n(\hat{\theta})\):
Akaike Information Criterion:
Bayesian Information Criterion:
BIC penalizes complexity more heavily and is consistent (selects true model as \(n \to \infty\) if it's in the candidate set).
Cross-Validation¶
For nonparametric methods, cross-validation estimates out-of-sample error:
where \(\hat{f}_{-i}^\lambda\) is fitted without observation \(i\) using tuning parameter \(\lambda\).
Leave-one-out CV is computationally expensive but has low variance. K-fold CV (typically \(K = 5\) or \(10\)) balances bias and variance.
Time Series Cross-Validation¶
Standard CV assumes i.i.d. data. For time series, use rolling-window or expanding-window validation:
where \(w\) is window size and \(h\) is forecast horizon.
Statistical Properties Summary¶
| Property | Parametric | Nonparametric |
|---|---|---|
| Convergence rate | \(O(n^{-1/2})\) | \(O(n^{-2\beta/(2\beta+d)})\) |
| Dimension dependence | None | Severe |
| Misspecification | Inconsistent | Consistent |
| Interpretability | High | Low |
| Data efficiency | High | Low |
| Extrapolation | Good (if correct) | Poor |
Key Takeaways¶
-
Parametric models assume finite-dimensional structure, achieving fast \(\sqrt{n}\) convergence when correctly specified but suffering from misspecification bias.
-
Nonparametric models let data determine the functional form, achieving consistency under weak assumptions but at slower rates that degrade with dimension.
-
Semi-parametric models combine parametric structure for interpretable components with nonparametric flexibility for nuisance functions.
-
In finance, low signal-to-noise ratios and limited effective sample sizes often favor simpler parametric models, despite potential misspecification.
-
Model selection via cross-validation (appropriately adapted for time series) is essential for choosing complexity levels.
Further Reading¶
- Hastie, Tibshirani & Friedman, The Elements of Statistical Learning, Chapters 2–6
- Tsybakov, Introduction to Nonparametric Estimation
- Fan & Gijbels, Local Polynomial Modelling and Its Applications
- Aït-Sahalia & Lo (1998), "Nonparametric Estimation of State-Price Densities"
- Robinson (1988), "Root-N-Consistent Semiparametric Regression"
Exercises¶
Exercise 1. Consider a GARCH(1,1) model \(\sigma_t^2 = \omega + \alpha r_{t-1}^2 + \beta \sigma_{t-1}^2\).
(a) How many parameters does this model have? What is the convergence rate for estimating them with \(n\) observations?
(b) If the true conditional variance follows a regime-switching process, explain why the GARCH MLE converges to a pseudo-true parameter rather than the truth. What does this imply for risk forecasts?
Solution to Exercise 1
(a) The GARCH(1,1) model \(\sigma_t^2 = \omega + \alpha r_{t-1}^2 + \beta \sigma_{t-1}^2\) has 3 parameters: \(\theta = (\omega, \alpha, \beta)\). Since this is a parametric model with a finite-dimensional parameter vector \(\theta \in \mathbb{R}^3\), the convergence rate for estimating these parameters (via MLE or quasi-MLE) is the standard parametric rate:
This rate is independent of the dimension of the input and depends only on having \(n\) observations and the regularity of the likelihood.
(b) If the true conditional variance follows a regime-switching process, e.g.,
then the single-regime GARCH(1,1) is misspecified: the true DGP does not belong to \(\mathcal{F}_{\text{param}}\). The MLE converges not to the true parameters but to the pseudo-true parameter:
This pseudo-true parameter is the single-regime GARCH that is closest in KL divergence to the true regime-switching model. The approximation error \(\|f(\cdot; \theta^\dagger) - f^*\|\) does not vanish as \(n \to \infty\).
Implications for risk forecasts: The GARCH model will produce volatility estimates that are a weighted average of the two regimes. In the high-volatility regime, GARCH will underestimate risk (its estimate is pulled down by the low-regime component), and in the low-volatility regime, it will overestimate risk. This systematic bias is particularly dangerous for tail risk measures (VaR, Expected Shortfall), where underestimation in crisis periods leads to inadequate capital reserves.
Exercise 2. For the Nadaraya–Watson kernel regression estimator with Gaussian kernel and bandwidth \(h\), the bias is \(O(h^2)\) and the variance is \(O(1/(nh^d))\).
(a) Derive the optimal bandwidth \(h^*\) that minimizes the mean integrated squared error by balancing bias and variance.
(b) Compute \(h^*\) and the resulting MISE rate for \(d = 1\) and \(d = 10\) with \(n = 1000\). Interpret the difference.
Solution to Exercise 2
(a) The mean integrated squared error (MISE) for the Nadaraya-Watson estimator is:
Using \(\text{Bias}^2 = O(h^4)\) and \(\text{Var} = O((nh^d)^{-1})\):
To minimize, take the derivative with respect to \(h\) and set it to zero:
Solving for \(h\):
Therefore:
Substituting back into the MISE expression:
(b) For \(d = 1\) with \(n = 1000\):
For \(d = 10\) with \(n = 1000\):
Interpretation: In one dimension, the MISE is approximately 0.004, indicating good estimation quality. In ten dimensions, the MISE is approximately 0.096---about 24 times larger. The optimal bandwidth \(h^*\) in 10 dimensions is 0.632, meaning the "local" neighborhood spans 63% of the data range in each direction, which is hardly local at all. This illustrates the curse of dimensionality: with the same 1000 observations, moving from \(d = 1\) to \(d = 10\) drastically degrades nonparametric estimation quality.
Exercise 3. A partially linear model for returns is
(a) Describe Robinson's differencing procedure to estimate \(\beta\) at the parametric rate \(\sqrt{n}\).
(b) Why is it advantageous to model the volatility effect nonparametrically while keeping the dividend yield effect linear?
Solution to Exercise 3
(a) Robinson's differencing procedure for the partially linear model \(R_{t+1} = \beta \cdot \text{DivYield}_t + f(\text{Volatility}_t) + \varepsilon_{t+1}\) proceeds in three steps.
Step 1: Nonparametric regression of \(R_{t+1}\) on \(\text{Volatility}_t\). Estimate \(m_R(v) = \mathbb{E}[R_{t+1} | \text{Volatility}_t = v]\) using kernel regression (e.g., Nadaraya-Watson):
Step 2: Nonparametric regression of \(\text{DivYield}_t\) on \(\text{Volatility}_t\). Similarly estimate \(m_D(v) = \mathbb{E}[\text{DivYield}_t | \text{Volatility}_t = v]\):
Step 3: OLS on residuals. Compute the residuals:
Then estimate \(\beta\) by OLS:
The key insight is that by taking conditional expectations of the original equation:
Subtracting eliminates the nonparametric component \(f\):
Under regularity conditions, \(\hat{\beta}\) achieves \(\sqrt{n}\)-consistency because the nonparametric estimation errors in Steps 1 and 2 converge fast enough (at rate \(n^{-2/(4+1)} = n^{-2/5}\)) that they do not affect the asymptotic distribution of \(\hat{\beta}\).
(b) There are several reasons to model the volatility effect nonparametrically while keeping dividend yield linear:
-
Nonlinear volatility-return relationship: The relationship between volatility and future returns is known to be nonlinear. At low volatility, expected returns may be moderate; at high volatility, there is a risk premium but also potential mean-reversion. A linear specification would miss these effects.
-
Linear dividend yield predictability: The dividend-price ratio is a well-established linear predictor of returns (Campbell and Shiller, 1988). Economic theory (the present-value identity) supports a log-linear relationship.
-
Interpretability of \(\beta\): Keeping the dividend yield effect parametric yields a single interpretable coefficient \(\beta\) that measures the marginal predictive effect of dividend yield, controlling for volatility.
-
Data efficiency: The parametric component is estimated at rate \(\sqrt{n}\) regardless of the nonparametric component. If both were nonparametric, estimation would be slower and require more data.
-
Avoiding the curse of dimensionality: With two nonparametric components, we would need to estimate a bivariate surface, facing the curse of dimensionality. The semi-parametric approach reduces this to one parametric direction and one univariate nonparametric function.
Exercise 4. You are estimating the Black–Scholes implied volatility surface from 200 observed option prices.
(a) A parametric approach uses the SVI parameterization with 5 parameters per maturity slice. With 4 maturities, how many total parameters are estimated? What is the convergence rate?
(b) A nonparametric approach uses kernel smoothing in the \((K, T)\) plane. Using Stone's rate with \(\beta = 2\) and \(d = 2\), what is the convergence rate with 200 observations?
(c) Which approach would you recommend and why?
Solution to Exercise 4
(a) The SVI parameterization has 5 parameters per maturity slice: \((a, b, \rho, m, \sigma)\). With 4 maturities, the total number of parameters is:
Since this is a parametric model with \(d = 20\) parameters, the convergence rate is the parametric rate:
(b) For the nonparametric kernel smoothing approach in the \((K, T)\) plane, we have \(d = 2\) input dimensions. Using Stone's minimax rate with \(\beta = 2\) (twice-differentiable functions):
Numerically:
For comparison, the parametric rate gives \(200^{-1} \approx 0.005\) for MSE (since parametric MSE is \(O(n^{-1})\)). The nonparametric rate is about 6 times worse.
(c) With only 200 observations, the parametric SVI approach is preferable for the following reasons:
-
Better convergence rate: The parametric rate \(O(n^{-1})\) for MSE substantially outperforms the nonparametric rate \(O(n^{-2/3})\) at \(n = 200\).
-
No-arbitrage constraints: The SVI parameterization can be constrained to satisfy static no-arbitrage conditions (e.g., non-negative butterfly spreads, non-negative calendar spreads). The nonparametric estimator may produce arbitrage-violable surfaces.
-
Extrapolation: The SVI formula provides meaningful extrapolation to strikes and maturities not observed in the data. The kernel estimator degrades severely outside the data support.
-
Smoothness: The parametric surface is automatically smooth, while the nonparametric estimator may be noisy with only 200 data points spread across two dimensions.
-
Parsimony: 20 parameters for 200 observations gives a ratio of 1:10, which is reasonable. The nonparametric approach implicitly has an effective number of parameters comparable to \(n\), leading to overfitting risk.
The main caveat is that if the SVI functional form is a poor approximation to the true surface, the parametric approach introduces misspecification bias that cannot be reduced with more data.
Exercise 5. Compare AIC and BIC for selecting between a 3-factor and a 5-factor linear return model estimated on \(n = 120\) monthly observations. If both models have log-likelihoods \(\ell_3 = -340\) and \(\ell_5 = -335\):
(a) Compute AIC and BIC for each model.
(b) Which model does each criterion select? Explain why they may disagree.
Solution to Exercise 5
(a) For the 3-factor model (\(d = 3\) parameters: intercept + 3 betas, so \(d = 4\); but typically in this context \(d\) counts the factor loadings plus intercept, so let us use \(d = 3\) for factors + 1 for intercept = 4. However, the problem states \(d = 3\) and \(d = 5\) directly as model parameters):
With \(d_3 = 3\) (including intercept: actually the problem states 3-factor and 5-factor, so the number of parameters is \(d_3 = 4\) and \(d_5 = 6\) respectively if we count the intercept. Let us follow the convention that a \(k\)-factor model has \(k + 1\) parameters).
Actually, re-reading the problem, it says "3-factor" and "5-factor" models with log-likelihoods given. The standard convention is that a \(k\)-factor model has \(d = k + 1\) parameters (intercept + \(k\) slopes). So \(d_3 = 4\) and \(d_5 = 6\).
AIC:
BIC:
(b) Both criteria prefer lower values.
-
AIC selects the 5-factor model (\(\text{AIC}_5 = 682 < 688 = \text{AIC}_3\)). The improvement in log-likelihood (\(\Delta \ell = 5\)) is large enough to justify the 2 additional parameters under AIC's penalty of \(2 \times 2 = 4\).
-
BIC selects the 5-factor model as well (\(\text{BIC}_5 = 698.72 < 699.15 = \text{BIC}_3\)), though the margin is very narrow.
In general, AIC and BIC can disagree because BIC penalizes complexity more heavily (penalty \(d \log n\) vs. \(2d\)). For \(n = 120\), \(\log(120) \approx 4.79\), so BIC penalizes each additional parameter by about 4.79 vs. AIC's penalty of 2. Here the log-likelihood improvement of 5 per additional parameter is large enough that even BIC selects the 5-factor model, but barely. With a smaller improvement (e.g., \(\ell_5 = -337\)), BIC would select the 3-factor model while AIC might still prefer the 5-factor model. BIC is consistent (selects the true model as \(n \to \infty\)) while AIC tends to select slightly overfit models, but AIC is asymptotically efficient (minimizes prediction error).
Exercise 6. Explain why nonparametric return prediction models frequently underperform simple linear models out-of-sample in finance, despite being more flexible. Your answer should reference the signal-to-noise ratio, the curse of dimensionality, and the bias–variance trade-off.
Solution to Exercise 6
Nonparametric models frequently underperform simple linear models out-of-sample in financial return prediction due to the interaction of three factors:
1. Signal-to-noise ratio. Financial returns have an extremely low signal-to-noise ratio. Daily equity returns exhibit \(R^2 \approx 0.25\%\), meaning that the predictable component \(\mu(X_t)\) is tiny relative to the noise \(\varepsilon_{t+1}\):
Any estimator's prediction error is \(\text{MSE} = \text{Bias}^2 + \text{Variance} + \sigma^2\), and the irreducible error \(\sigma^2\) dominates. The tiny signal means that even small amounts of variance in the estimator can overwhelm the bias reduction from using a more flexible model.
2. Curse of dimensionality. Nonparametric methods suffer convergence rates that degrade with input dimension \(d\). By Stone's theorem, the minimax rate is:
With typical financial predictors (\(d = 5\) to \(20\)) and sample sizes (\(n \approx 250\) to \(5000\)), the effective sample size is very small. For example, with \(\beta = 2\), \(d = 10\), and \(n = 1000\), the effective sample size is \(n_{\text{eff}} \approx 8\)---far too few to reliably estimate any nonlinear structure.
By contrast, the parametric linear model achieves \(\text{MSE}_{\text{param}} = O(n^{-1})\) regardless of \(d\) (assuming correct specification or mild misspecification).
3. Bias-variance trade-off. In the decomposition \(\text{MSE} = \text{Bias}^2 + \text{Variance}\):
- Linear model: Has potentially higher bias (if the true function is nonlinear) but very low variance due to few parameters. With \(d + 1\) parameters and \(n\) observations, variance is \(O(d/n)\).
- Nonparametric model: Has lower bias (can approximate any smooth function) but much higher variance because it effectively estimates many local parameters.
When SNR is low, the bias from using a linear model is small in absolute terms (the nonlinear component of the signal is even smaller than the already-tiny linear signal). Meanwhile, the variance penalty of the nonparametric model is large. The net effect is:
In summary, the flexibility of nonparametric models is a liability in finance: the additional degrees of freedom fit noise rather than the weak signal, producing worse out-of-sample predictions. This is why simple linear models, ridge regression, and other heavily regularized approaches tend to dominate in empirical asset pricing.