Richardson Extrapolation¶
Richardson extrapolation combines numerical solutions computed on two or more grids to cancel leading-order error terms, producing a more accurate approximation without redesigning the numerical scheme. It is a powerful and general technique for improving convergence order.
The Error Expansion¶
Asymptotic Form¶
Suppose a numerical method with mesh parameter \(h\) produces an approximation \(V(h)\) to the exact value \(V^*\). If the method has order \(p\), the error admits an asymptotic expansion:
where the constants \(c_p, c_{p+1}, \ldots\) are independent of \(h\).
Key assumption: The error is a smooth function of \(h\) with a power series structure. This holds when the PDE solution is sufficiently smooth.
Example: Crank-Nicolson¶
Crank-Nicolson for the Black-Scholes equation has second-order accuracy:
If we refine both \(\Delta S\) and \(\Delta\tau\) proportionally (using a single mesh parameter \(h\)):
The Extrapolation Formula¶
Eliminating the Leading Error Term¶
Compute \(V(h)\) on a grid with spacing \(h\) and \(V(h/r)\) on a refined grid with spacing \(h/r\) (typically \(r = 2\)). The errors are:
Eliminating \(c h^p\) from these two equations:
Richardson Extrapolation Formula
The extrapolated value is:
This has accuracy \(O(h^{p+1})\), one order higher than the original scheme.
Special Case: \(r = 2\) (Grid Doubling)¶
For the common choice of doubling the number of grid points (\(r = 2\)):
| Original order \(p\) | Extrapolated order | Formula |
|---|---|---|
| 1 (Euler) | 2 | \(2V(h/2) - V(h)\) |
| 2 (Crank-Nicolson) | 4 | \(\frac{4V(h/2) - V(h)}{3}\) |
Application to First-Order Methods¶
Backward Euler (Implicit Scheme)¶
The implicit scheme for the heat equation has error:
where \(h\) represents the time step \(\Delta\tau\) (assuming spatial error is negligible or handled separately).
Compute with \(N\) and \(2N\) time steps:
This cancels the \(O(h)\) term, yielding a second-order approximation from a first-order scheme.
Numerical Example¶
Consider pricing a European call with \(K = 100\), \(S = 100\), \(T = 1\), \(\sigma = 0.2\), \(r = 0.05\). Using the implicit scheme with \(M = 200\) spatial points:
| Time steps \(N\) | FDM price \(V(h)\) | Error |
|---|---|---|
| 50 | 10.4312 | \(2.0 \times 10^{-2}\) |
| 100 | 10.4412 | \(1.0 \times 10^{-2}\) |
| 200 | 10.4462 | \(5.0 \times 10^{-3}\) |
Extrapolation (\(p=1\), \(r=2\)):
The exact Black-Scholes price is \(V^* = 10.4506\), so the extrapolated error (\(6 \times 10^{-4}\)) is roughly 17 times smaller than the \(N=100\) error.
Application to Second-Order Methods¶
Crank-Nicolson¶
With second-order convergence (\(p = 2\)):
This produces a fourth-order approximation.
Numerical Example (Continued)¶
Using Crank-Nicolson with \(M = 200\):
| Time steps \(N\) | FDM price \(V(h)\) | Error |
|---|---|---|
| 25 | 10.4496 | \(1.0 \times 10^{-3}\) |
| 50 | 10.4504 | \(2.5 \times 10^{-4}\) |
| 100 | 10.4505 | \(6.3 \times 10^{-5}\) |
Extrapolation (\(p=2\), \(r=2\)):
The extrapolated error (\(\sim 10^{-5}\)) is roughly 25 times smaller than the \(N=50\) error.
Repeated Richardson Extrapolation¶
The Romberg Table¶
Repeated extrapolation can further increase the order. Starting from solutions on grids \(h, h/2, h/4, \ldots\):
Level 0 (original solutions):
Level \(k\) (extrapolated):
where \(p\) is the original order and the effective order at level \(k\) is \(p + k\).
Example Table (Crank-Nicolson, \(p = 2\))¶
| \(V^{(0)}\) (order 2) | \(V^{(1)}\) (order 4) | \(V^{(2)}\) (order 6) | |
|---|---|---|---|
| \(h\) | \(V(h)\) | ||
| \(h/2\) | \(V(h/2)\) | \(\frac{4V(h/2) - V(h)}{3}\) | |
| \(h/4\) | \(V(h/4)\) | \(\frac{4V(h/4) - V(h/2)}{3}\) | \(\frac{16V^{(1)}_1 - V^{(1)}_0}{15}\) |
Each column increases the effective order by 2 (since only even powers appear in the Crank-Nicolson error expansion).
Convergence Order Verification¶
Richardson extrapolation provides a robust way to verify the convergence order of a numerical method without knowing the exact solution.
The Richardson Error Estimator¶
From the error expansion \(V(h) = V^* + ch^p + O(h^{p+1})\):
The Richardson ratio is:
Convergence Order Test
If \(R \approx 2\), the method is first-order (\(p = 1\)). If \(R \approx 4\), the method is second-order (\(p = 2\)). If \(R \approx 8\), the method is third-order (\(p = 3\)).
Extracting the Order¶
Solving for \(p\) directly:
This is the observed convergence order, which should match the theoretical order when the solution is smooth and the asymptotic regime has been reached.
Spatial and Temporal Extrapolation¶
Independent Extrapolation¶
When spatial and temporal errors are controlled independently, extrapolation can be applied in each direction separately.
Step 1: Fix \(\Delta\tau\) and extrapolate in space:
Step 2: Fix \(\Delta S\) and extrapolate in time:
Step 3: Apply both extrapolations (requires four solutions):
Proportional Refinement¶
Alternatively, refine \(\Delta S\) and \(\Delta\tau\) proportionally (setting \(h = \Delta S = \alpha \Delta\tau\) for some fixed \(\alpha\)). Then a single extrapolation handles both errors simultaneously.
When Richardson Extrapolation Fails¶
Non-Smooth Data¶
The error expansion \(V(h) = V^* + ch^p + O(h^{p+1})\) assumes smooth dependence on \(h\). This breaks down when:
- Payoff kinks: The payoff \((S-K)^+\) has a discontinuous derivative at \(S = K\). The error expansion may contain fractional powers of \(h\) or logarithmic terms
- Barrier options: Discontinuities in the boundary condition disrupt the error expansion
- American options: The free boundary introduces additional non-smoothness
In these cases, the Richardson ratio \(R\) may deviate significantly from \(2^p\), and the extrapolated value may be less accurate than the fine-grid solution alone.
Remedies¶
- Rannacher smoothing: Apply a few implicit steps first to smooth the solution, restoring the regular error expansion
- Payoff smoothing: Replace the kink with a smooth approximation over a small interval
- Grid alignment: Place grid points exactly at the strike \(K\) so the kink coincides with a node
After smoothing, Richardson extrapolation typically recovers its full effectiveness.
Pre-Asymptotic Regime¶
Extrapolation also fails if the grid is too coarse for the asymptotic expansion to be valid. Always verify that the Richardson ratio is close to \(2^p\) before trusting the extrapolated value.
Summary¶
| Aspect | Detail |
|---|---|
| Purpose | Cancel leading error term, gain one order |
| Requirement | Smooth error expansion in powers of \(h\) |
| For Euler (\(p=1\)) | \(V_{\text{ext}} = 2V(h/2) - V(h)\) gives order 2 |
| For C-N (\(p=2\)) | \(V_{\text{ext}} = \frac{4V(h/2) - V(h)}{3}\) gives order 4 |
| Order verification | Richardson ratio \(R \approx 2^p\) confirms order \(p\) |
| Limitation | Requires smooth solutions; fails near kinks and free boundaries |
Richardson extrapolation is a cost-effective way to improve accuracy: computing two solutions on different grids and combining them is often cheaper than solving once on a very fine grid. Combined with Rannacher smoothing, it provides a practical route to high-accuracy option pricing.
Exercises¶
Exercise 1. An implicit scheme with \(N = 50\) time steps gives a price of \(V(h) = 10.4312\), and with \(N = 100\) gives \(V(h/2) = 10.4412\). Assuming first-order convergence (\(p = 1\)), compute the Richardson-extrapolated value \(V_{\text{ext}} = 2V(h/2) - V(h)\). If the exact price is \(10.4506\), compare the errors of \(V(h/2)\) and \(V_{\text{ext}}\).
Solution to Exercise 1
Using the Richardson extrapolation formula for a first-order method (\(p = 1\), \(r = 2\)):
The errors are:
- Error of \(V(h/2) = 10.4412\): \(|10.4412 - 10.4506| = 0.0094\)
- Error of \(V_{\text{ext}} = 10.4512\): \(|10.4512 - 10.4506| = 0.0006\)
The extrapolated error is approximately \(0.0006 / 0.0094 \approx 15.7\) times smaller than the fine-grid error. Richardson extrapolation has effectively converted a first-order scheme into a second-order result using two relatively coarse grids.
Exercise 2. For a second-order method (Crank-Nicolson), three grid solutions give: \(V(h) = 10.4496\), \(V(h/2) = 10.4504\), \(V(h/4) = 10.4505\). Compute the Richardson ratio \(R = (V(h) - V(h/2))/(V(h/2) - V(h/4))\). Is \(R \approx 4\) consistent with second-order convergence?
Solution to Exercise 2
The Richardson ratio is:
For a second-order method, we expect \(R \approx 2^p = 2^2 = 4\). The observed value \(R = 8\) is significantly larger than 4.
This is not consistent with pure second-order convergence. The ratio \(R = 8 = 2^3\) would suggest third-order convergence. This can happen when:
- The leading second-order error term has a very small coefficient, so the next-order term (e.g., \(O(h^3)\) or \(O(h^4)\)) dominates at these grid sizes
- The solution happens to be at a special point where the second-order error coefficient nearly vanishes
- The grids are not yet in the asymptotic regime
Further refinement (computing \(V(h/8)\)) would help determine whether the true asymptotic ratio is 4 or whether the method genuinely exhibits higher-order convergence for this particular problem.
Exercise 3. Construct the Romberg table for a Crank-Nicolson solver using solutions at \(h\), \(h/2\), and \(h/4\) with values \(V(h) = 10.44\), \(V(h/2) = 10.449\), \(V(h/4) = 10.4504\). Compute the two level-1 extrapolated values (order 4) and the single level-2 value (order 6).
Solution to Exercise 3
Level 0 (original values, order 2):
Level 1 (order 4, using \(p = 2\), \(r = 2\), so the formula is \((4V_{\text{fine}} - V_{\text{coarse}})/3\)):
Level 2 (order 6). Since the Crank-Nicolson error expansion contains only even powers of \(h\), the level-1 values have error \(O(h^4)\), and we eliminate this with \(r^{p'} = 2^4 = 16\):
The Romberg table:
| Order 2 | Order 4 | Order 6 | |
|---|---|---|---|
| \(h\) | 10.4400 | ||
| \(h/2\) | 10.4490 | 10.4520 | |
| \(h/4\) | 10.4504 | 10.4509 | 10.4508 |
Exercise 4. Explain why Richardson extrapolation fails when the payoff has a kink at \(S = K\). Specifically, what happens to the error expansion \(V(h) = V^* + ch^p + \cdots\) when the solution is not smooth? What does the Richardson ratio look like in this case?
Solution to Exercise 4
The Richardson extrapolation formula \(V_{\text{ext}} = (r^p V(h/r) - V(h))/(r^p - 1)\) relies on the error expansion:
This expansion requires the error to be a smooth (analytic) function of the mesh parameter \(h\), which in turn requires the PDE solution to be sufficiently smooth.
When the payoff has a kink at \(S = K\), the solution near maturity has a discontinuous second derivative. This non-smoothness causes the error expansion to contain non-integer powers of \(h\):
or possibly logarithmic terms like \(h^2 \ln h\). The regular power series structure is destroyed.
When we apply the standard formula assuming \(p = 2\) to cancel \(ch^2\), we are actually leaving behind the dominant \(O(h^{1/2})\) and \(O(h)\) terms, which are not canceled. The extrapolated value may be no better than (or even worse than) the fine-grid solution.
The Richardson ratio in this case will not be close to \(2^p = 4\). Instead, it will fluctuate or converge to a value like \(2^{1/2} \approx 1.41\) or \(2^1 = 2\), reflecting the actual (reduced) convergence order. Seeing \(R \neq 2^p\) is a clear diagnostic sign that the smooth error expansion assumption has failed.
Exercise 5. For independent spatial and temporal extrapolation, four PDE solutions are needed: \(V(\Delta S, \Delta\tau)\), \(V(\Delta S/2, \Delta\tau)\), \(V(\Delta S, \Delta\tau/2)\), and \(V(\Delta S/2, \Delta\tau/2)\). Write down the combined extrapolation formula that eliminates leading-order errors in both space and time for a second-order scheme. What is the effective order of the result?
Solution to Exercise 5
For a second-order scheme, the error has the form:
We have four solutions:
- \(V_{11} = V(\Delta S, \Delta\tau) = V^* + a(\Delta S)^2 + b(\Delta\tau)^2\)
- \(V_{21} = V(\Delta S/2, \Delta\tau) = V^* + a(\Delta S)^2/4 + b(\Delta\tau)^2\)
- \(V_{12} = V(\Delta S, \Delta\tau/2) = V^* + a(\Delta S)^2 + b(\Delta\tau)^2/4\)
- \(V_{22} = V(\Delta S/2, \Delta\tau/2) = V^* + a(\Delta S)^2/4 + b(\Delta\tau)^2/4\)
Step 1: Extrapolate in space (eliminate \(a(\Delta S)^2\)):
Step 2: Extrapolate in time (eliminate \(b(\Delta\tau)^2\)):
Substituting:
The effective order is \(O((\Delta S)^4 + (\Delta\tau)^4)\), i.e., fourth-order in both space and time.
Exercise 6. A practitioner wants to improve the accuracy of an implicit (first-order) scheme without switching to Crank-Nicolson. By running the implicit scheme on two grids and applying Richardson extrapolation, they can obtain a second-order result. Compare the total computational cost of this approach to running Crank-Nicolson on the fine grid alone.
Solution to Exercise 6
Richardson-extrapolated implicit approach: Run the implicit (backward Euler) scheme on two grids:
- Coarse grid: \(M\) spatial points, \(N\) time steps. Cost: \(O(MN)\) (tridiagonal solve per step)
- Fine grid: \(M\) spatial points, \(2N\) time steps. Cost: \(O(2MN)\)
- Total cost: \(O(3MN)\)
The extrapolated result \(V_{\text{ext}} = 2V(h/2) - V(h)\) has second-order accuracy in time.
Crank-Nicolson on the fine grid: Run Crank-Nicolson with \(M\) spatial points and \(N\) time steps (Crank-Nicolson is already second-order, so it needs fewer steps for the same accuracy):
- Cost: \(O(MN)\) (tridiagonal solve per step, same cost as implicit)
Comparison: The Richardson-extrapolated implicit approach costs roughly \(3\times\) as much as Crank-Nicolson on a single grid for the same second-order accuracy. Crank-Nicolson is the better choice if available, since it achieves second-order convergence in a single solve.
However, the Richardson approach has practical advantages:
- Simplicity: Implementing backward Euler is easier than Crank-Nicolson (no averaging of time levels)
- No oscillations: Backward Euler damps all modes monotonically, avoiding the oscillatory artifacts that Crank-Nicolson can produce with non-smooth data
- Code reuse: If an implicit solver already exists, Richardson extrapolation can improve its accuracy without modifying the core solver
For these reasons, the Richardson-extrapolated implicit scheme is a viable alternative when implementation simplicity or robustness to non-smooth data is prioritized over raw efficiency.