Learn how finite difference methods tackle the Black–Scholes PDE by discretizing time and price into a grid. Explore explicit, implicit, and Crank–Nicolson schemes, boundary conditions, implementation tips, advantages, and pitfalls.
Well, let’s start with a tiny personal anecdote. The first time I tried coding a finite difference method (FDM) for option pricing, I spent a solid hour wondering why my results were bouncing all over the place—turns out I’d reversed the sign on a boundary condition, and the entire grid solution went haywire. Not my proudest moment, but it taught me how crucial setup can be when we discretize partial differential equations (PDEs). In the context of option pricing, FDMs are among the most systematic numerical methods we can use to solve PDEs (particularly the Black–Scholes PDE) to find how an option’s price evolves over time.
Where does this all fit into the grand scheme of derivatives pricing? In earlier sections (see 10.3 Binomial Option Pricing Model and 10.4 Risk-Neutral Valuation), we introduced discrete-time approximations and continuous-time frameworks. Finite difference techniques serve as a bridge between a purely continuous PDE approach (like the Black–Scholes–Merton model) and the more discrete lattice-based binomial or trinomial trees. FDMs let us handle various complexities like dividends, early exercise, or time-dependent volatility.
Below, we’ll explore the key elements of FDMs for option pricing: from why we take the PDE approach, to boundary conditions, to the three main solution methods. Our aim is to help you gain both intuitive and technical mastery.
The starting point for FDM pricing is the continuous-time Black–Scholes PDE. For a non-dividend-paying underlying stock, the PDE for a European call option V(S, t) can be written as:
where:
• S is the underlying asset price.
• t is time.
• σ is the volatility of the underlying.
• r is the risk-free interest rate.
• V(S, t) is the option’s value at price S and time t.
For an American-style option, the PDE is the same in the region where the option is alive (i.e., it’s not yet optimal to exercise). However, we incorporate an early exercise condition that states:
In other words, if you can do better by exercising than holding onto the option, you’d typically choose to exercise for an American option. The tricky part is weaving that condition into our numerical method.
Finite difference methods start by chopping up both the price axis (S) and the time axis (t) into small increments. Then, we approximate the derivatives appearing in the PDE by using finite difference expressions. Eventually, we solve a grid of equations to get the option price at each (S, t) node.
We define a grid for S from 0 to some maximum price Smax. Time runs backward from T (the option’s maturity) to 0. Our grid might look something like:
• ΔS: the step size (increments) in the asset price dimension.
• Δt: the step size in the time dimension.
When t = T (the final time in backward sense), we know the “terminal condition” (the payoff). For a call option:
For a put option:
At S = 0 or S = S_\mathrm{max}, we define boundary conditions that reflect how an option’s value should behave at those extremes.
Below is a simplified illustration of a finite difference grid in time (vertical axis) and stock price (horizontal axis). Each node represents the approximate option value at that node’s (S, t). We solve from top (terminal conditions at T) down to bottom (option value at t=0).
graph LR A["Time = T <br/> Known Payoff"] --> B["Next Step in Time"] B --> C["Next Step in Time"] C --> D["t=0: Value we seek"] A --> A2["(S=0) boundary"] D --> D2["(S=0) boundary"] A2 --> D2 A --> A3["(S=Smax) boundary"] D --> D3["(S=Smax) boundary"] A3 --> D3
At each discrete time slice, we move backward in time from the known payoff at T toward 0. For American options, we might do an extra step at each node to enforce the early-exercise condition.
Now that our domain is chunked into a grid, how do we handle the PDE? Finite difference approximations replace derivatives with differences between adjacent grid points. Let’s loosely define them:
The explicit method basically says:
“Take your PDE, rearrange it so that the next step in time is a function of the current step’s known values.”
This yields a direct formula for Vᵏ₊₁ (the value at the next time step) in terms of Vᵏ at the current time. The advantage? It’s super easy to implement: no complicated matrix inversions needed. One might code it in just a few lines in Python. The disadvantage? It can suffer from numerical instability if Δt and ΔS are not chosen carefully. Often, you get a condition like Δt ≤ C(ΔS)² for some constant C if you want to keep your solution from blowing up.
The implicit method rearranges the PDE so that:
“Current step in time is a function of the unknown next step values.”
Yes, that means we have a system of linear equations to solve each time step. That’s more computationally expensive—but the upside is better numerical stability. You can typically take larger time steps without everything going bananas. Implicit schemes often come in handy in multi-factor PDEs or when you want to handle large S ranges effectively.
The Crank–Nicolson method is a classic blend—well, a 50% explicit, 50% implicit compromise:
It’s more stable than the explicit approach and often yields better accuracy than the fully implicit method for many problems in finance. However, you still have to solve a linear system at each time step (the implicit portion demands it). It’s quite popular for pricing, especially if your main worry is second-order accuracy in both time and S.
Boundary conditions can make or break your solution. You usually define them at:
• S = 0: For a call, the value is about 0 if it’s worthless to buy the underlying at 0. For a put, you might want to reflect that V(0, t) = K·e⁻ʳᵗ (the approximate PV of the strike if you can exercise before maturity, though more precisely for American puts, it’s often near K).
• S = S_\mathrm{max}: If S_\mathrm{max} is large enough, the call behaves like S - (present value of K), but we often just fix a large boundary that ensures the option payoff is trivially in-the-money.
And at t = T, you know the terminal payoff from the option’s definition. For American options, each time step you do:
If the intrinsic value is higher, you override your solution with that immediate exercise payoff.
• Natural handling of early exercise. Whether it’s an American put or a Bermuda-style call, you can slip in the “take the max of hold or exercise” condition at each node.
• Flexible with dividends. You can incorporate discrete or continuous dividend yields by adjusting the PDE or the payoff at ex-dividend times.
• Varying volatility structures or local volatility models can be inserted by changing σ = σ(S, t).
• Well-established convergence theory: as ΔS, Δt → 0, your solution typically converges to the exact PDE solution.
• Numerical stability can be a headache—especially for the explicit method.
• Complex boundary conditions or multiple underlying factors can lead to large memory requirements.
• It can be slower than closed-form solutions (where those exist) or some advanced approximation methods.
• A poor choice of Smax can cause errors, particularly if the underlying can jump or if it’s quite volatile.
• If you implement an early exercise condition incorrectly, you can get big, systematic mispricings.
I recall messing up early exercise logic once in a project: I forgot to apply the “max” condition at each node for an American put, so the solution was basically the European put’s price—way off for deep in-the-money scenarios.
Below is a concise example of building an explicit finite difference approach for a simple American put option. This skeleton is by no means production-ready, but it’ll give you a sense of typical steps.
1import numpy as np
2
3S_max = 200.0
4K = 100.0
5r = 0.05
6sigma = 0.20
7T = 1.0
8num_S_steps = 200
9num_t_steps = 200
10
11dS = S_max / num_S_steps
12dt = T / num_t_steps
13
14S_vals = np.linspace(0, S_max, num_S_steps + 1)
15V = np.zeros((num_t_steps + 1, num_S_steps + 1))
16
17V[-1, :] = np.maximum(K - S_vals, 0)
18
19for i in range(num_t_steps + 1):
20 # At S=0, put is worth K (un-discounted in naive approach)
21 V[i, 0] = K
22 # At S=S_max, put is almost worthless
23 V[i, -1] = 0
24
25for n in reversed(range(num_t_steps)):
26 for i in range(1, num_S_steps):
27 a = 0.5 * dt * (sigma**2 * i**2 - r * i)
28 b = 1 - dt * (sigma**2 * i**2 + r)
29 c = 0.5 * dt * (sigma**2 * i**2 + r * i)
30
31 V[n, i] = a * V[n+1, i-1] + b * V[n+1, i] + c * V[n+1, i+1]
32
33 # Early exercise condition
34 exercise_val = max(0, K - (i * dS))
35 V[n, i] = max(V[n, i], exercise_val)
Points to note:
• This is a purely explicit scheme with potential numerical stability constraints.
• For real production or exam contexts, you might prefer Crank–Nicolson to reduce error.
• The boundary condition at S=0 is set directly to K, ignoring discounting—some folks discount it, some prefer a more advanced approach.
When you bring dividends, time-varying rates, or local volatility surfaces into the picture, the PDE changes shape. You’ll have extra or modified terms in the PDE, but the same finite difference discretization steps apply:
For American options on dividend-paying stocks, you might adjust the PDE to account for a continuous dividend yield \( q \):
Everything else remains conceptually the same.
• Not making Smax large enough. If an option is deep in-the-money for calls, the boundary “fakes out” the PDE solution, leading to a mispricing.
• Overly large Δt. The explicit scheme can blow up if you’re not satisfying the stability condition.
• Forgetting to apply the American early-exercise condition at each time step. This leads to European-like values.
• Inconsistent boundary conditions: mixing up the boundary formula for calls vs. puts.
• Not refining the grid enough. If the grid is too coarse, you get big discretization errors.
• On constructed-response questions, you might be asked to demonstrate the FDM approach on a mini-grid. Make sure you remember how to approximate ∂V/∂t, ∂²V/∂S².
• If a question references stability, recall that an explicit scheme is conditionally stable; implicit and Crank–Nicolson are generally more robust.
• They could present a scenario with an American call on a dividend-paying stock and require an explanation of how the PDE changes or how early exercise might be triggered around ex-dividend dates.
• Time management: these questions can get heavy with calculations, but you can usually outline the method rather than compute dozens of grid nodes by hand.
• Make sure to remember boundary conditions. Some test-takers forget them entirely, overshadowed by PDE details.
Finite difference methods add an incredibly flexible tool to your derivatives pricing toolkit. Yes, you have to roll up your sleeves and handle a grid approach, but the trade-off is a powerful method that works for a broad array of exotic features, including local volatility, early exercise, and path-dependent payoffs (with some modifications).
At first, you might find PDE-based approaches intimidating. But once you get the hang of it, it can feel like filling out a big spreadsheet: each cell (node) depends on a few neighbors, and you iterate through. If you’ve struggled even a bit reading through all this: that’s normal. Practice a small-coded example, or even a hand-drawn grid, and watch how the method unfolds step by step. It’s a fun puzzle—well, maybe not always “fun” per se, but definitely rewarding when your final solution converges nicely to a known benchmark.
Important Notice: FinancialAnalystGuide.com provides supplemental CFA study materials, including mock exams, sample exam questions, and other practice resources to aid your exam preparation. These resources are not affiliated with or endorsed by the CFA Institute. CFA® and Chartered Financial Analyst® are registered trademarks owned exclusively by CFA Institute. Our content is independent, and we do not guarantee exam success. CFA Institute does not endorse, promote, or warrant the accuracy or quality of our products.