Misspecification vs Estimation Error¶
Introduction¶
In quantitative finance and econometrics, model uncertainty arises from two fundamentally different sources:
- Misspecification Error: The true data-generating process lies outside the considered model class
- Estimation Error: The true model belongs to the model class, but parameters are estimated with finite-sample error
Understanding the distinction between these error sources and their interaction is crucial for: - Robust model selection - Proper uncertainty quantification - Risk management - Statistical inference under model uncertainty
This chapter provides rigorous mathematical treatment of both error types, their decomposition, and methods for handling their combined effects.
Mathematical Framework¶
1. Setup and Notation¶
Data-Generating Process: Let \(P^*\) denote the true (unknown) probability measure governing observed data \(\{X_i\}_{i=1}^n\).
Model Class: Consider a parametric family \(\mathcal{M} = \{P_{\theta}: \theta \in \Theta\}\) where \(\Theta \subseteq \mathbb{R}^d\) is the parameter space.
Well-Specified Model: The model is correctly specified if:
Misspecified Model: The model is misspecified if:
2. Pseudo-True Parameter¶
Even under misspecification, we can define a best approximation within the model class.
Definition (Pseudo-True Parameter): The pseudo-true parameter \(\theta_0\) is defined as:
where \(D_{\text{KL}}(P \| Q)\) is the Kullback-Leibler divergence.
Alternative Characterizations:
- Maximum Likelihood:
$$ \theta_0 = \arg\max_{\theta \in \Theta} \mathbb{E}{P^*}[\log p(X)] $$
- Minimum Hellinger Distance:
$$ \theta_0 = \arg\min_{\theta \in \Theta} \int \left( \sqrt{p^*} - \sqrt{p_{\theta}} \right)^2 $$
Properties: - When well-specified: \(\theta_0 = \theta^*\) - When misspecified: \(\theta_0\) is the "closest" parameter in the KL sense - \(\theta_0\) is uniquely defined under identifiability conditions
3. Estimator and Estimation Error¶
Maximum Likelihood Estimator (MLE):
Estimation Error: The deviation between estimator and pseudo-true parameter:
Asymptotic Distribution: Under regularity conditions:
where \(\Sigma(\theta_0)\) is the asymptotic covariance matrix.
4. Misspecification Error¶
Approximation Error: The distance between the pseudo-true model and the truth:
Decomposition: For any \(\theta \in \Theta\):
This decomposition shows that any parameter \(\theta \neq \theta_0\) incurs both misspecification error and within-model error.
Asymptotic Theory¶
1. Consistency Under Misspecification¶
Theorem (Consistency): Under standard regularity conditions:
even when the model is misspecified.
Proof Sketch: By the law of large numbers:
uniformly in \(\theta\). The MLE \(\hat{\theta}_n\) maximizes the LHS, which converges to the RHS, which is maximized at \(\theta_0\) by definition.
Interpretation: The MLE consistently estimates the pseudo-true parameter \(\theta_0\), not necessarily the true parameter \(\theta^*\) (which may not exist in \(\mathcal{M}\)).
2. Asymptotic Normality¶
Theorem (Asymptotic Distribution): Under regularity conditions:
where: - Hessian:
$$ H(\theta) = -\mathbb{E}{P^*}\left[ \nabla^2 \log p(X) \right] $$
- Outer Product of Scores (OPG):
$$ J(\theta) = \mathbb{E}{P^*}\left[ \nabla \log p(X)^\top \right] $$}(X) \nabla \log p_{\theta
Sandwich Formula: The asymptotic covariance is:
often called the robust or sandwich covariance matrix.
Well-Specified Case: When \(P^* = P_{\theta_0}\):
the Fisher information, and:
the Cramér-Rao bound.
3. Rate of Convergence¶
Parametric Rate: Both estimation error and convergence rate are \(O_p(n^{-1/2})\):
Misspecification: The misspecification error \(\Delta_{\text{mis}}\) does not vanish as \(n \to \infty\):
Implication: For large \(n\): - Estimation error becomes negligible: \(O(n^{-1/2})\) - Misspecification error dominates: \(O(1)\)
Decomposition of Prediction Error¶
1. Expected Loss Framework¶
Consider prediction error measured by a loss function \(\ell(y, \hat{y})\).
Expected Loss: For predictor \(\hat{f}_n\) based on \(n\) samples:
Decomposition: We can write:
where: - \(f^*\) is the Bayes-optimal predictor under \(P^*\) - \(f_{\theta_0}\) is the best predictor in model class \(\mathcal{M}\) - \(\hat{f}_n\) is the estimated predictor
2. Bias-Variance-Misspecification Tradeoff¶
For squared error loss \(\ell(y, \hat{y}) = (y - \hat{y})^2\):
Mean Squared Error:
Classical Decomposition (Well-Specified):
Extended Decomposition (Misspecified):
Interpretation: 1. Misspecification: Cannot be reduced by collecting more data 2. Bias: Decreases with more flexible models 3. Variance: Decreases with more data but increases with model complexity
Model Selection Under Misspecification¶
1. Information Criteria¶
Information criteria balance goodness-of-fit against model complexity.
2. Akaike Information Criterion (AIC)¶
Definition:
where \(d = \dim(\Theta)\) is the number of parameters.
Asymptotic Justification: AIC estimates:
Key Property: AIC is consistent for selecting the model that minimizes KL divergence to \(P^*\), even under misspecification.
3. Bayesian Information Criterion (BIC)¶
Definition:
Asymptotic Property: BIC is consistent for model selection when the true model is in the candidate set (well-specified case).
Under Misspecification: BIC tends to select simpler models than AIC due to larger penalty (\(\log n\) vs \(2\)).
4. Takeuchi Information Criterion (TIC)¶
Robust Variant accounting for misspecification:
where \(H\) and \(J\) are estimates of \(H(\theta_0)\) and \(J(\theta_0)\).
Interpretation: The trace term \(\text{tr}(H^{-1} J)\) reduces to \(d\) when well-specified, generalizing AIC.
5. Cross-Validation¶
Cross-validation directly estimates prediction error without assuming correct specification.
\(K\)-Fold Cross-Validation:
where \(\hat{f}_{-k}\) is trained on all data except fold \(k\).
Asymptotic Behavior:
Advantages: - Valid under misspecification - No assumption about model form - Directly relevant for prediction
Disadvantages: - Computationally intensive - High variance for small \(n\)
6. Model Averaging¶
Instead of selecting a single model, average predictions across models.
Weighted Average:
Optimal Weights: Minimize expected loss:
subject to \(\sum_m w_m = 1\), \(w_m \geq 0\).
Bayesian Model Averaging (BMA):
Under Misspecification: BMA may not be optimal; frequentist stacking often performs better.
Inference Under Misspecification¶
1. Hypothesis Testing¶
Null Hypothesis: \(H_0: \theta = \theta_0\) (pseudo-true value).
Wald Test Statistic:
Asymptotic Distribution:
using the sandwich estimator \(\hat{\Sigma}\).
Critical Issue: Standard tests assume well-specification. Using incorrect covariance (e.g., \(\mathcal{I}(\theta_0)^{-1}\) instead of sandwich) leads to: - Incorrect size - Invalid p-values - Misleading inference
2. Confidence Intervals¶
Standard CI (assuming well-specification):
Robust CI (allowing misspecification):
where \(\hat{\Sigma}\) is the sandwich estimator.
Coverage: - Standard CI: May have coverage \(< 1-\alpha\) under misspecification - Robust CI: Maintains nominal coverage asymptotically
3. Quasi-Likelihood Methods¶
Quasi-Likelihood: Specify only first two moments:
Estimating Equation:
Advantage: Robust to distributional misspecification; only requires correct specification of mean and variance functions.
Consistency: \(\hat{\theta}_n \to \theta_0\) where:
Misspecification Tests¶
1. Moment Tests¶
Idea: Check if moments implied by the model match empirical moments.
Test Statistic: For moment conditions \(\mathbb{E}[g(X, \theta)] = 0\):
where \(\bar{g}_n(\theta) = n^{-1} \sum_{i=1}^n g(X_i, \theta)\) and \(W\) is a weight matrix.
Asymptotic Distribution:
where \(r\) is the number of overidentifying restrictions.
Rejection: Large \(Q_n\) provides evidence of misspecification.
2. Hausman Test¶
Setup: Two estimators \(\hat{\theta}_1\) and \(\hat{\theta}_2\) where: - \(\hat{\theta}_1\) is consistent under both \(H_0\) and \(H_1\) but inefficient under \(H_0\) - \(\hat{\theta}_2\) is consistent and efficient under \(H_0\) but inconsistent under \(H_1\)
Test Statistic:
Asymptotic Distribution:
Application: Test endogeneity, omitted variables, or other forms of misspecification.
3. Omnibus Tests¶
Goodness-of-Fit Tests:
- Kolmogorov-Smirnov:
$$ D_n = \sup_x |\hat{F}n(x) - F(x)| $$
- Cramér-von Mises:
$$ W_n^2 = n \int_{-\infty}^{\infty} [\hat{F}n(x) - F(x) $$}(x)]^2 dF_{\theta_0
- Anderson-Darling:
$$ A_n^2 = n \int_{-\infty}^{\infty} \frac{[\hat{F}n(x) - F(x) $$}(x)]^2}{F_{\theta_0}(x)[1 - F_{\theta_0}(x)]} dF_{\theta_0
Power: These tests can detect various forms of misspecification but may have low power against specific alternatives.
Robustness to Misspecification¶
1. Robust Estimators¶
M-Estimators: Minimize a robust loss function:
where \(\rho\) is chosen to downweight outliers.
Example (Huber Loss): For regression:
Breakdown Point: Fraction of contamination an estimator can handle before breaking down.
2. Influence Functions¶
Influence Function: Measures sensitivity to infinitesimal contamination:
where \(T\) is a statistical functional and \(\delta_x\) is a point mass at \(x\).
Gross Error Sensitivity:
Interpretation: Smaller \(\gamma^*\) indicates greater robustness to outliers.
For MLE: Influence function is unbounded, making it sensitive to outliers and misspecification.
3. Empirical Likelihood¶
Idea: Nonparametric approach that assigns probabilities \(p_i\) to observations.
Optimization:
Advantage: Does not require specifying a parametric likelihood; robust to distributional misspecification.
Empirical Likelihood Ratio:
subject to moment constraints.
Asymptotic Distribution:
Applications to Quantitative Finance¶
1. Option Pricing Under Misspecification¶
Setup: True stock dynamics unknown; use Black-Scholes as working model.
Black-Scholes Model:
True Dynamics (e.g., with jumps):
where \(J_t\) is a jump process.
Pricing Error:
Components: 1. Misspecification: Ignoring jumps creates systematic bias 2. Estimation Error: Volatility \(\sigma\) estimated from data 3. Interaction: Estimated \(\sigma\) may partially compensate for model misspecification
2. Risk Management¶
Value-at-Risk (VaR):
Under Misspecified Model: Using \(P_{\theta_0}\) instead of \(P^*\):
Model Risk:
Backtesting: Check if actual violations match predicted:
Should be close to \(\alpha\) if model is well-specified.
3. Portfolio Optimization¶
Mean-Variance Problem:
Parameter Estimation: - \(\hat{\mu}\) estimated from sample mean - \(\hat{\Sigma}\) estimated from sample covariance
Estimation Error Impact:
Misspecification: Returns may not be i.i.d. normal: - Fat tails - Time-varying moments - Dependencies
Robust Approach: Use shrinkage estimators or robust covariance estimators to reduce impact of both estimation error and misspecification.
4. Factor Models¶
Model: Returns generated by factors:
Misspecification: - Missing factors - Nonlinear factor loadings - Time-varying betas
Estimation Error: - Estimated \(\hat{\beta}_i\) from regression - Standard error: \(O(n^{-1/2})\)
Joint Impact:
when well-specified. Under misspecification, first term includes structural error.
Advanced Theoretical Results¶
1. Semiparametric Efficiency¶
Semiparametric Model: Parametric component \(\theta\) and infinite-dimensional nuisance component \(\eta\).
Efficient Score: The orthogonal component to the nuisance tangent space:
Efficient Influence Function:
Semiparametric Efficiency Bound:
Under Misspecification: Efficiency bounds still apply, but target is pseudo-true \(\theta_0\) rather than true \(\theta^*\).
2. Double Robustness¶
Setup: Two working models: 1. Outcome regression: \(m(X; \alpha)\) 2. Propensity score: \(\pi(X; \beta)\)
Doubly Robust Estimator:
Property: \(\hat{\tau}\) is consistent if either (not necessarily both): - \(m(X; \alpha)\) is correctly specified, or - \(\pi(X; \beta)\) is correctly specified
Interpretation: Provides insurance against misspecification of one component.
3. High-Dimensional Asymptotics¶
Setup: Dimension \(d = d_n \to \infty\) as \(n \to \infty\).
Regime: \(d/n \to \kappa \in (0, 1)\) (proportional asymptotics).
Estimation Error: In high dimensions:
rather than \(O_p(n^{-1})\) in fixed-dimensional case.
Misspecification: Can be exacerbated in high dimensions: - More parameters to estimate - Greater potential for overfitting - Model selection more difficult
Regularization: Methods like LASSO add penalty to control both estimation error and effective model complexity:
Practical Guidelines¶
1. Detection Strategy¶
Step 1: Diagnostic Checks - Plot residuals vs. fitted values - Q-Q plots for normality - ACF plots for independence
Step 2: Formal Tests - Run omnibus goodness-of-fit tests - Check moment conditions - Perform Hausman-type specification tests
Step 3: Cross-Validation - Compare in-sample vs. out-of-sample performance - Large discrepancy suggests overfitting or misspecification
2. Remedial Measures¶
When Misspecification Detected:
- Expand Model Class: Add more flexible features
- Interaction terms
- Nonlinear transformations
-
Regime-switching components
-
Use Robust Methods:
- Sandwich covariance for inference
- M-estimators for estimation
-
Empirical likelihood
-
Model Averaging: Combine multiple models to hedge misspecification risk
-
Regularization: Add penalties to prevent overfitting while allowing for misspecification
3. Reporting Standards¶
Best Practices:
-
Always report sandwich standard errors when model specification uncertain
-
Conduct sensitivity analysis: Report results under multiple model specifications
-
Validate out-of-sample: Use holdout data or cross-validation
-
Document assumptions: Be explicit about modeling choices and their implications
-
Report information criteria: AIC, BIC, or TIC to facilitate model comparison
4. Balancing Act¶
Trade-off Triangle:
Complexity
/\
/ \
/ \
/ \
/ \
/ \
/ \
/______________\
Misspecification Estimation
Error Error
Optimal Model: Minimizes total error:
Guidelines: - Small \(n\): Use simpler models to reduce estimation error - Large \(n\): Can afford more complex models; focus on reducing misspecification - High \(d\): Regularization crucial to control estimation error
Summary and Key Insights¶
1. Fundamental Distinctions¶
- Nature of Error:
- Misspecification: Structural, does not vanish with more data
-
Estimation: Statistical, decreases at rate \(O(n^{-1/2})\)
-
Asymptotic Behavior:
- As \(n \to \infty\): Estimation error → 0, Misspecification error persists
-
Dominance switches from estimation (small \(n\)) to misspecification (large \(n\))
-
Inference:
- Misspecification requires sandwich covariance
- Standard errors based on Fisher information are invalid under misspecification
2. Practical Implications¶
-
Model Selection: Use information criteria (AIC/BIC) or cross-validation that account for both error sources
-
Uncertainty Quantification: Sandwich estimators provide valid inference under misspecification
-
Robustness: M-estimators and quasi-likelihood methods offer protection
-
High Dimensions: Regularization helps control estimation error; model selection critical for misspecification
3. Research Frontiers¶
-
Optimal Model Averaging: Theory for optimal weights under joint misspecification and estimation error
-
Adaptive Inference: Methods that automatically adjust for detected misspecification
-
Deep Learning: Understanding misspecification vs. estimation error in neural networks
-
Causal Inference: Double robustness and related techniques for handling model uncertainty
The interplay between misspecification and estimation error is central to statistical modeling. A deep understanding of both error sources and their interaction is essential for robust quantitative finance applications, from derivative pricing to risk management to portfolio optimization.
Exercises¶
Exercise 1. Consider a true data-generating process \(P^*\) that is a mixture of two normals: \(P^* = 0.7 \cdot N(0.05, 0.04) + 0.3 \cdot N(-0.10, 0.09)\). A practitioner fits a single Gaussian model \(P_\theta = N(\mu, \sigma^2)\). Compute the pseudo-true parameter \(\theta^* = \arg\min_\theta D_{\text{KL}}(P^* \| P_\theta)\) and the misspecification error. Explain why the fitted model fails to capture the left tail of \(P^*\).
Solution to Exercise 1
The true distribution is \(P^* = 0.7 \cdot N(0.05, 0.04) + 0.3 \cdot N(-0.10, 0.09)\).
Step 1: Compute the moments of \(P^*\). The mean is:
The second moment is:
So the variance is \(\sigma^{*2} = 0.05975 - 0.005^2 = 0.059725\).
Step 2: Find the pseudo-true parameter. For a Gaussian model \(P_\theta = N(\mu, \sigma^2)\), the KL divergence \(D_{\text{KL}}(P^* \| P_\theta)\) is minimized when \(P_\theta\) matches the first two moments of \(P^*\) (this is a well-known property of the Gaussian family as the maximum entropy distribution for given mean and variance).
The pseudo-true parameters are:
Step 3: Compute the misspecification error. The KL divergence is:
Since \(P_{\theta_0} = N(\mu_0, \sigma_0^2)\) matches the mean and variance of \(P^*\):
Therefore \(D_{\text{KL}}(P^* \| P_{\theta_0}) = -H(P^*) + \frac{1}{2}\log(2\pi e \sigma_0^2)\), where \(H(P^*)\) is the differential entropy of the mixture.
The entropy of a Gaussian with variance \(\sigma_0^2\) is \(H_{\text{Gauss}} = \frac{1}{2}\log(2\pi e \sigma_0^2)\), and since a Gaussian maximizes entropy for given variance, we have \(H(P^*) \leq H_{\text{Gauss}}\), so \(D_{\text{KL}}(P^* \| P_{\theta_0}) \geq 0\). The misspecification error is strictly positive because \(P^*\) is bimodal while \(P_{\theta_0}\) is unimodal.
Step 4: Left tail failure. The left tail of \(P^*\) is heavier than that of \(P_{\theta_0}\) because the mixture component \(N(-0.10, 0.09)\) contributes significant mass in the far left. The fitted Gaussian, while matching the overall variance, distributes its tails symmetrically. In particular:
This is because the mixture's second component (with mean \(-0.10\) and standard deviation \(0.3\)) places substantial probability at extreme negative values. The single Gaussian cannot capture this asymmetric, heavy-tailed structure, leading to systematic underestimation of left-tail risk.
Exercise 2. For a correctly specified model \(P^* = P_{\theta_0}\) with \(n = 100\) observations, the MLE \(\hat{\theta}_n\) has asymptotic variance \(I(\theta_0)^{-1}/n\). Compute the estimation error contribution to the total pricing error \(|V(\hat{\theta}_n) - V(\theta_0)|\) using the delta method when \(V(\theta) = \text{BS}(S_0, K, \sigma(\theta), T)\) is a Black-Scholes call price. Express the result in terms of the vega and the Fisher information.
Solution to Exercise 2
Setup: The model is correctly specified, so \(P^* = P_{\theta_0}\), \(\hat{\theta}_n\) is the MLE with \(n = 100\), and \(V(\theta) = \text{BS}(S_0, K, \sigma(\theta), T)\) is a Black-Scholes call price.
Delta method: For a smooth function \(V(\theta)\) of the parameter, the pricing error satisfies:
The variance of this pricing error is:
Single-parameter case (\(\theta = \sigma\)): If the only uncertain parameter is volatility, then \(\nabla_\sigma V = \mathcal{V}\) (the Black-Scholes vega), and in the well-specified case:
where \(I(\sigma_0)\) is the Fisher information for the volatility parameter.
Therefore the estimation error contribution to the pricing error is:
with standard deviation:
For \(n = 100\):
Interpretation: The pricing error is proportional to the vega (sensitivity of the option price to volatility) and inversely proportional to \(\sqrt{n I(\sigma_0)}\). Options with high vega (e.g., ATM options with long maturity) are most sensitive to estimation error. The Fisher information \(I(\sigma_0)\) captures how informative the data is about \(\sigma\): higher \(I(\sigma_0)\) means more precise estimation and smaller pricing error.
Multi-parameter extension: If \(\theta = (\mu, \sigma)^\top\), the pricing error variance is:
Since the Black-Scholes call price under the risk-neutral measure does not depend on \(\mu\), we have \(\partial V / \partial \mu = 0\) in the risk-neutral framework, and the result reduces to the vega-based formula above.
Exercise 3. Explain the decomposition of total model risk into misspecification and estimation components:
In which regime (small \(n\) vs. large \(n\)) does each term dominate? Illustrate with a concrete example of fitting a geometric Brownian motion to data with stochastic volatility.
Solution to Exercise 3
Decomposition: The total model risk in pricing a derivative \(V\) is:
where \(V^*\) is the true price, \(\theta^*\) is the pseudo-true parameter, and \(\hat{\theta}_n\) is the MLE.
Asymptotic behavior:
-
Misspecification error \(|V(\theta^*) - V^*| = O(1)\): This is a fixed bias that does not vanish with sample size. It reflects the structural inability of the model to capture the true dynamics.
-
Estimation error \(|V(\hat{\theta}_n) - V(\theta^*)| = O_p(n^{-1/2})\): By the delta method and asymptotic normality of the MLE, this vanishes as \(n \to \infty\).
Regime analysis:
-
Small \(n\): Estimation error dominates. With few observations, parameter estimates are noisy, and \(|V(\hat{\theta}_n) - V(\theta^*)|\) is large. The misspecification error may be relatively small in comparison. Using a simpler model (lower misspecification within a simpler class, or equivalently, less estimation error) may be preferable.
-
Large \(n\): Misspecification error dominates. The estimation error shrinks to zero, but the structural bias remains. Increasing sample size does not help reduce total error; only improving the model specification can help.
Concrete example: GBM fitted to Heston data. Suppose the true dynamics follow the Heston model:
and we fit a GBM: \(dS_t = \mu S_t \, dt + \sigma S_t \, dW_t\) (constant volatility).
The pseudo-true \(\sigma^*\) is essentially the average integrated volatility: \(\sigma^{*2} \approx \bar{v}\). The misspecification error arises because:
- The GBM cannot produce the implied volatility smile (different strikes are mispriced systematically).
- For an OTM put, the GBM underestimates the true price because it misses the fat left tail generated by stochastic volatility and leverage effect (\(\rho < 0\)).
With \(n = 50\) daily returns, estimation error in \(\hat{\sigma}\) might be \(\pm 5\%\), leading to pricing errors of order \(\mathcal{V} \times 0.05\). With \(n = 5000\), estimation error drops to \(\pm 0.5\%\), but the misspecification bias (e.g., 10-15% for deep OTM puts) persists.
Exercise 4. The sandwich (Huber-White) variance estimator provides valid standard errors under model misspecification. For a quasi-MLE \(\hat{\theta}\), the sandwich variance is \(\hat{V} = A^{-1} B (A^{-1})^\top\) where \(A = -\frac{1}{n}\sum_{i=1}^n \nabla^2 \ell_i(\hat{\theta})\) and \(B = \frac{1}{n}\sum_{i=1}^n \nabla \ell_i(\hat{\theta}) \nabla \ell_i(\hat{\theta})^\top\). Explain why \(A \neq B\) indicates misspecification and how this affects confidence intervals compared to the standard MLE variance \(A^{-1}/n\).
Solution to Exercise 4
Sandwich estimator: The sandwich (Huber-White) variance is:
where:
Why \(A \neq B\) indicates misspecification: Under correct specification, the information matrix equality holds:
This is because \(\mathbb{E}[\nabla \ell] = 0\) under the true model, and differentiating this identity gives \(\mathbb{E}[\nabla^2 \ell] + \mathbb{E}[\nabla \ell \nabla \ell^\top] = 0\). So \(A = B\) in population when well-specified.
When misspecified, the score \(\nabla \ell(\theta_0)\) does not integrate to zero under \(P^*\) in the same way, and the second Bartlett identity fails. Therefore \(A \neq B\), and the discrepancy \(\|A - B\|\) is a measure of the degree of misspecification.
Impact on confidence intervals: The standard MLE variance uses \(A^{-1}/n\) (equivalent to the inverse Fisher information). The correct asymptotic variance under misspecification is \(A^{-1}BA^{-1}/n\).
-
If \(B > A\) (in the matrix ordering sense), then \(A^{-1}BA^{-1} > A^{-1}\), meaning the standard CI is too narrow. This leads to under-coverage: the nominal 95% CI may have actual coverage well below 95%, giving false confidence in the estimates.
-
If \(B < A\), the standard CI is too wide (conservative), which is less dangerous but still inefficient.
In practice, misspecification typically increases variance relative to the well-specified case, so standard CIs tend to be anti-conservative. The sandwich estimator corrects this by using the empirically observed score variability rather than the theoretically expected variability under the assumed model.
Exercise 5. A modeler compares Black-Scholes (1 parameter: \(\sigma\)) against Heston (5 parameters) for pricing a set of 20 vanilla options. Using AIC, derive the penalty for each model and explain when the simpler Black-Scholes model might be preferred despite higher in-sample calibration error. How does this relate to the bias-variance trade-off?
Solution to Exercise 5
AIC for Black-Scholes (BS): With 1 parameter (\(d_{\text{BS}} = 1\)), fitting 20 options:
The penalty term is \(2d = 2\).
AIC for Heston: With 5 parameters (\(d_H = 5\): \(v_0, \kappa, \bar{v}, \xi, \rho\)):
The penalty term is \(2d = 10\).
When BS might be preferred: The AIC difference is:
BS is preferred (lower AIC) when \(\text{AIC}_{\text{BS}} < \text{AIC}_H\), i.e.:
This occurs when the Heston model's improvement in log-likelihood over BS is less than 4. With only 20 options, Heston's 4 extra parameters may not sufficiently improve the fit to justify the complexity penalty.
Connection to bias-variance trade-off:
-
Bias (misspecification): BS has higher bias since it cannot capture the volatility smile. Heston has lower bias as it models stochastic volatility.
-
Variance (estimation error): BS has lower estimation variance since it has only 1 parameter estimated from 20 data points. Heston has higher variance since 5 parameters must be estimated from the same 20 data points, with a ratio of \(d/n = 5/20 = 0.25\) (high).
-
Total error: \(\text{MSE} = \text{Bias}^2 + \text{Variance}\). For small datasets (\(n = 20\)), the variance reduction from using BS may outweigh its bias increase, making the simpler model preferable. AIC formalizes this trade-off through the penalty \(2d\).
This is why, in practice, traders sometimes use Black-Scholes with a volatility smile adjustment rather than calibrating a full stochastic volatility model when the dataset is small.
Exercise 6. Consider the White (1982) information matrix test for misspecification. The null hypothesis is \(H_0: A = B\) (where \(A\) and \(B\) are defined as in Exercise 4). Describe how to implement this test for a GARCH(1,1) model fitted to daily returns, and interpret rejection of \(H_0\) in terms of what aspects of the return distribution the GARCH model fails to capture.
Solution to Exercise 6
White's information matrix test: The null hypothesis is \(H_0: A(\theta_0) = B(\theta_0)\), where:
Under correct specification, \(A = B\) (the information matrix equality).
Implementation for GARCH(1,1): The GARCH(1,1) model is:
with parameters \(\theta = (\mu, \omega, \alpha, \beta)^\top\) and \(z_t \sim N(0,1)\).
Step 1: Estimate \(\hat{\theta}\) by MLE.
Step 2: Compute for each observation \(t\):
where \(\ell_t(\theta) = -\frac{1}{2}\log(2\pi) - \frac{1}{2}\log\sigma_t^2 - \frac{r_t - \mu)^2}{2\sigma_t^2}\).
Step 3: Form the sample estimates:
Step 4: Vectorize \(\hat{A} - \hat{B}\) (taking unique elements due to symmetry; for \(d = 4\) parameters, there are \(d(d+1)/2 = 10\) unique elements). Construct the test statistic:
where \(\hat{\Omega}\) is a consistent estimate of the asymptotic covariance of \(\text{vech}(\hat{A} - \hat{B})\).
Under \(H_0\): \(T \xrightarrow{d} \chi^2_{10}\).
Step 5: Reject \(H_0\) if \(T > \chi^2_{10, \alpha}\).
Interpretation of rejection: Rejection of \(H_0\) (i.e., \(A \neq B\)) indicates that the GARCH(1,1) model with Gaussian innovations is misspecified. Common reasons include:
-
Non-normal innovations: Returns may have heavier tails than Gaussian (Student-\(t\) or GED innovations would be more appropriate). This manifests as excess kurtosis in standardized residuals \(\hat{z}_t = \hat{\varepsilon}_t / \hat{\sigma}_t\).
-
Asymmetric volatility response: GARCH(1,1) treats positive and negative shocks symmetrically, while empirically negative returns increase volatility more (leverage effect). This is captured by GJR-GARCH or EGARCH models.
-
Long memory in volatility: GARCH(1,1) has exponentially decaying autocorrelation in \(\sigma_t^2\), while realized volatility often exhibits long memory (slow hyperbolic decay), suggesting FIGARCH or rough volatility models.
-
Regime changes: The assumption of constant parameters \((\omega, \alpha, \beta)\) may fail during structural breaks or regime switches.
Exercise 7. In a simulation study, generate \(n = 500\) observations from a stochastic volatility model (Heston) and fit both a Black-Scholes model and a Heston model. For each fitted model, compute the pricing error for an out-of-the-money put option (\(K = 0.9 S_0\)). Decompose the Black-Scholes error into misspecification and estimation components. For the Heston model, argue that the misspecification error is zero and all error comes from estimation. Under what conditions might the misspecified Black-Scholes model produce a smaller total error?
Solution to Exercise 7
Setup: The true DGP is Heston with parameters \((\kappa, \bar{v}, v_0, \xi, \rho, \mu)\). We observe \(n = 500\) returns and fit:
- Model A (BS): \(dS_t = \hat{\mu} S_t \, dt + \hat{\sigma} S_t \, dW_t\) (1 parameter for pricing: \(\hat{\sigma}\))
- Model B (Heston): All 5 parameters estimated
Black-Scholes error decomposition: The total pricing error for an OTM put (\(K = 0.9 S_0\)) is:
-
Misspecification: \(V_{\text{BS}}(\sigma^*)\) uses the pseudo-true volatility \(\sigma^* = \sqrt{\bar{v}}\) (approximately). For an OTM put, BS systematically underprices because it cannot capture the left-skew in the return distribution generated by negative \(\rho\) (leverage effect). The implied volatility smile means that the OTM put's market-implied volatility exceeds \(\sigma^*\). This error is \(O(1)\) and does not vanish with more data.
-
Estimation: \(|V_{\text{BS}}(\hat{\sigma}) - V_{\text{BS}}(\sigma^*)| \approx |\mathcal{V}| \cdot |\hat{\sigma} - \sigma^*|\). With \(n = 500\), \(|\hat{\sigma} - \sigma^*| = O_p(n^{-1/2}) \approx O_p(0.045)\). This is relatively small.
Heston error decomposition: Since the true DGP is Heston, the model is correctly specified:
The misspecification error is exactly zero. All error comes from estimation. However, with 5 parameters and \(n = 500\), the estimation error may be substantial because:
- Parameters like \(\xi\) (vol-of-vol) and \(\rho\) (correlation) are hard to estimate precisely from return data alone.
- The pricing function \(V_H(\theta)\) may have high sensitivity to certain parameters (e.g., \(\xi\) affects the smile shape strongly).
- The estimation error is \(O_p(n^{-1/2})\) but with a potentially large constant due to the high-dimensional Fisher information structure.
When BS might win: The misspecified BS can produce smaller total error when:
This can happen when:
- The option is near ATM: BS misspecification error is small for near-ATM options where the smile effect is minimal.
- Heston parameters are poorly identified: When \(\kappa\) and \(\bar{v}\) are confounded, or when \(\xi\) is difficult to estimate, the Heston estimation error is inflated.
- Small \(n\) or short observation period: With limited data, the 5-parameter Heston model overfits, producing larger out-of-sample errors than the parsimonious BS model.
- Low vol-of-vol regime: When \(\xi\) is small, the Heston model is close to BS, so misspecification is mild but Heston's extra parameters add unnecessary estimation noise.
This exemplifies the bias-variance trade-off: BS has higher bias but lower variance, while Heston has zero bias but higher variance. For small to moderate \(n\), the simpler model can be preferred overall.