Professional Documents
Culture Documents
American MC PriceIt
American MC PriceIt
where Γ is the set of all stopping times with values in {T1 , T2 , . . . , TN }, and Q, and QN ,
are the risk neutral and, respectively, forward neutral probabilities.
As τ ∈ Γ can also be referred to as an Exercise Strategy (or exercise Boundary), the
expression above means that the price of a Bermudan option is its expected value given
the best Exercise Strategy.
2. The computed exercise boundary is used for forward pricing using classical Monte
Carlo (once known, the exercise boundary can be used to price the option like a
trigger option).
This methodology can be explained easily: as shown, the price of a Bermudan option is
given by its expected value provided its best exercise strategy (or stopping time). Hence,
a Bermudan option can be priced by first approximating its best exercise strategy and
then using it to price the option as a trigger.
The difference between these two numerical methods (L&S and Andersen) lies in the
way exercise boundaries approximations are computed.
400
51.2 Description of the Methods 401
Now, the point is about the way to find out those (a∗1 , . . . , a∗N ).
A first method would be to compute the expression
h I (x ) i
τ τ
τa∗1 ,...,a∗N = ArgMax E QN
τ ∈ΓN B(τ, TN )
by approximating
h I (x ) i
τ τ
E QN
B(τ, TN )
with
M N (i)
1 XX In (xn )
1τ (x(i) )=n ,
M B(Tn , TN )
i=1 n=1
x(1) , . . . , x(M )
where are M simulated paths, and by solving
M X
N (i)
X In (xn )
τa∗1 ,...,a∗N = ArgMax 1τ (x(i) )=n .
τ ∈ΓN B(Tn , TN )
i=1 n=1
2. Set a∗N = 0. This reflects the fact that on the last exercise date, the option is
exercised if and only if its exercise value if positive.
3. Assume that (a∗n+1 , . . . , a∗N ) are computed. We would like to compute a∗n . Define
fn (a) as
M (i)
X In (xn )
fn (a) = 1I (x(i) )≥a + Cn (xn(i) )1I (x(i) )<a ,
B(Tn , TN ) n n n n
i=1
(i)
where Cn (xn ) is defined as
Cn (x(i) (i)
n ) = Iq (xq )
with
∗
q = inf p > n | Ip (x(i)
p ) ≥ ap .
Now solve
a∗n = ArgMax fn (a).
a
(i)
For a better understanding of the method, we can say that Cn (xn ) can be interpreted as
(i)
the value of the option at date tn in the state of the world xn in case it is note exercised
at time tn . Hence, fn (an ) is the value of the option at date Tn provided that its holder
exercise if In (xn ) ≥ an .
This method is continued until n = 0.
M (i)
1 X In(i) (xn(i) )
P = B(0, TN ) ,
M B(Tn(i) , TN )
i=1
with
(i)
n Ip (xp ) o
n(i) = inf p ≥ 0 ≥ a∗p .
B(Tp , TN )
2. Define CN∗ (x ) = 0, coming from the fact that the decision to exercise at the last
N
∗ (x ) ≥ 0.
exercise date is taken iff IN N
M (i)
1 X In(i) (xn(i) )
P = B(T0 , TN )
M B(Tn(i) , TN )
i=1
where
(i)
n Ip (xp ) o
n(i) = inf p ≥ 0 ≥ Cp∗ (x(i) ) .
p
B(Tp , TN )
Basis used in Price-it® The Longstaff & Schwartz algorithm works with a basis. As
is well known, the choice of the basis used in the regression is critical as it changes the
numerical results substantially.
The intuitive idea is to take a basis whose first terms represent well the continuation
value function. Ideally, the optimal basis would be the one that has as the first term the
continuation value function Cp∗ (xp ) as in this case, the regression will be optimal with
only one single term and the projection of the continuation value function as a linear
combination of the Longstaff & Schwartz will keep all the information. However, this
optimal basis cannot be used in practice as we are precisely trying to infer numerically
the continuation value function itself and hence this is an unknown.
Another important characteristic in the choice of the basis is to use one whose projection
of the continuation value function converges rapidly to the continuation value function.
By rapidly, we mean a polynomial of degree 4 or 5. In Price-it® , we decided to use for
the Longstaff & Schwartz basis the basis given by 1, F (S(ti )), [F (S(ti )]2 , [F (S(ti ))]3
and [F (S(ti ))]4 where F (S(ti )) is the intrinsic value of the option. This basis is very
interesting as it makes the Longstaff & Schwartz very similar to the Andersen method
making the two methods highly compatible.
To make the problem easily writable, let’s introduce some other notations:
• u(tN , StN ) = f (tN , StN )
• u(tn , Stn ) = max f (tn , Stn ) , E(u(tn+1 , Stn+1 )|Ftn ) for n = 0, . . . , N − 1
n o
• τ0∗ = inf tn u(tn , Stn ) = f (tn , Stn )
We make the hypothesis that f (tn , Stn )0≤n≤N is a martingale w.r.t. the filtration
(Ftn )0≤n≤N
405
52.3 Some Results 406
Theorem 1.
sup E (f (τ, Sτ )) = E f (τ0∗ , Sτ0∗ ) .
u(t0 , S0 ) = (52.1)
τ ∈Tt0 ;tN
n o
If we now introduce τj∗ = inf tn ≥ tj u(tn , Stn ) = f (tn , Stn ) with j = 1, ..., N − 1 we
get a similar result
Corollary 1.
u(tj , Stj ) = sup E (f (τ, Sτ )) = E f (τj∗ , Sτj∗ )Ftj (52.2)
τ ∈Ttj ;tN
where Ttj ;tN the set of stopping time with real values in {tj , . . . , tN }.
Theorem 2. (u(tn , Stn ))0≤n≤N is the smallest super-martingale above (f (tn , Stn ))0≤n≤N .
Then
E u(tn+1 , Stn+1 )|Ftn ≤ u(tn , Stn ) n ∈ {0, . . . , N − 1}
and for n ∈ {0, . . . , N }
τj∗ = tj 1 + τ ∗ 1
j+1
(i)
f (tj ,Stj )=u(tj ,Stj ) f (tj ,Stj )<u(tj ,Stj )
We explain this point: by definition of τj∗ , if f (tj , Stj ) = u(tj , Stj ), it is clear that
τj∗ = tj . Suppose now we have f (tj , Stj ) < u(tj , Stj ). Then τj∗ > tj so τj∗ ≥ tj+1 which
implies τj∗ ≥ τj+1
∗ since
n o
∗
τj+1 = inf tn ≥ tj+1 u(tn , Stn ) = f (tn , Stn ) .
As τj∗ ≤ τj+1
∗ by definition of the 2 stopping time, it is now evident that considering the
event {f (tj , Stj ) < u(tj , Stj )}, τj∗ = τj+1
∗ .
τj∗ = tj 1 + τ ∗ 1
j+1
(ii)
Zj =Uj Zj <Uj
Actually, we just have to use the definition of Uj = max{Zj , E(Uj+1 |Ftj )}. As Zj < Uj ,
we get Zj < E(Uj+1 |Ftj ) and conversely, considering Zj < E(Uj+1 |Ftj ) we have Zj < Uj .
(it is important to note that the case Zj > Uj is impossible for all j = 0, . . . , N .)
(ii) becomes
τj∗ = tj 1 + τ ∗ 1
j+1
(v)
Zj ≥E(Uj+1 |Ftj ) Zj <E(Uj+1 |Ftj )
Thanks to Corollary 1, Uj+1 ∗
= E f (τj+1 , Sτj+1
∗ )|Ft ; using the properties of a filtration,
j+1
in particular Ftj ⊂ Ftj+1 and the properties of the conditional expectation with embedded
filtrations we get the result
!
∗ ∗
E(Uj+1 )|Ftj ) = E E f (τj+1 ∗ )Ft
, Sτj+1 Ft = E(f (τj+1 , Sτj+1
∗ )|Ft )
j+1 j j
τj∗ = tj 1 ∗ ,S ∗
+ τ ∗ 1
j+1 ∗ ,S ∗
(vi)
f (tj ,Stj )≥E(f (τj+1 τ )|Ftj ) f (tj ,Stj )<E(f (τj+1 τ )|Ftj )
j+1 j+1
that is
τj∗ = tj 1 + τ ∗ 1
j+1
Zj ≥E(Uj+1 |Ftj ) Zj <E(Uj+1 |Ftj )
• Set τN = tN = T
• Considering that τj+1
∗ has already been computed in previous steps, one sets
τj∗ = tj 1Zj ≥E(Uj+1 |Ftj ) + τj+1
∗
1Zj <E(Uj+1 |Ftj )
and we do this until j reaches 0 i.e. until we obtain τ0∗ . Note that if τj+1
∗ has been
computed, Uj+1 is easily obtained as we have the relation, thanks to the corollary
∗ , S ∗ ).
of Theorem 1, Uj+1 = f (τj+1 τj+1
Now we need to be more precise to see how we can compute the (τj∗ )0≤j≤N −1 . One
first must remark that to compute τj∗ knowing the next one τj+1
∗ , it suffices to know how
∗ , S ∗ )|F ). So in order to go on, we are going
to compute E(Uj+1 |Ftj ) that is E(f (τj+1 τj+1 tj
to remind some definition about Hilbert Spaces and Projection.
Characterization of a Hilbert Space Consider an Hilbert space H with the scalar prod-
uct < ., . >. Then it exists a countable base of elements (hi )i≥1 of H such that
- the adherence of {hi : i ≥ 1} is equals to H that is to say {hi : i ≥ 1} is dense in H (this
is a way to define an Hilbert Space).
- all elements g of H can be written as follows
∞
X
g= αi hi with αi =< g, hi > .
i=1
where (bgi )i≥1 is a base of G. Remember that for i ≥ 1 we have αei =< h, gbi > but they can
be difficult to calculate using this relation. The previous proceed to find the projection of
h on G is a least square minimization. Moreover to find the (e αi )i≥1 is to find the Hilbert
decomposition of h on G and we have
X ∞
X
gh = p ⊥ 2
G (h) = arg min ||h − g|| = arg min ||h − αi gbi ||2 = α
ei gbi .
g∈G (αi )i≥1
i≥1 i=1
where ϕ is a borel function, σ(Stj ) is the σ-algebra generated by the random variable
Stj and it is also an Hilbert space included in L2 (Ω). Let us remark that all variable Y
σ(Stj )-measurable depend in fact of Stj which can seen as the relation Y = φ(Stj ) where
φ is a borel function. The space of projection is here σ(Stj ) ⊂ L2 (Ω).
The computation of the conditional expectation is then done by least square minimiza-
tion. So we now suppose that we have a truncated base (as it is impossible to work with
∞) of random variables (gk (Stj ))1≤k≤K σ(Stj )-measurable to compute the conditional
expectation. We note by K the number used to truncate the Hilbert decomposition of
αkj )1≤k≤K the coefficients of its decomposition.
E(u(tj+1 , Stj+1 )|Ftj ) and by (e
0 = t0 , t1 , . . . , tN = T.
(m) (m) (m) (m)
2. Set τN = T and U (m) = f (τN , S (m) ) for m = 1, . . . , M
τN τN
(m) (m)
At this sub-state, (U (m) )1≤m≤M , (gk (StN −1 ))1≤m≤M are all known so we just
τN
have to compute the least square minimization.
Remark that even though we use the truncated sum
K
(m)
X
ekN −1 gk (StN −1 ),
α
k=1
c) set
n K o
(m) (m) (m)
X
U (m)
m
= max f (τN −1 , S (m) ), ekN −1 gk (StN −1 )
α
τN −1 τN −1
k=1
d) calculate
M K0
( 2 )!
1 X (m)
X (m)
αkN −2 )1≤k≤K 0 = Arg min
(e U (m) − αk gk0 (StN −2 )
α M τN −1
m=1 k=1
where (gk0 )1≤k≤K 0 is the Hilbert truncated base of σ(StN −2 ) as at this step we
need to compute
∗
E(u(tN −1 , StN −1 )|FtN −2 ) = E(f (τN −1 , SτN
∗
−1
)|StN −2 )
e) set
(m)
τN −2 = tN −2 1 (m) P 0 (m)
f (tN −2 ,St )≥ K
k=1 ekN −2 hk (St
α )
N −2 N −2
(m)
+τN −1 1 (m) P 0 −2 (m)
f (tN −2 ,St )< K
k=1 eN
α k hk (St )
N −2 N −2
and so on ...
(m)
4. the backward induction is repeated until the τ0 ’s have been found form =
1, . . . , M .
5. Finally, it suffices to make a simple Monte Carlo evaluation to find the price i.e.
M
1 X (m)
U0 = u(t0 , St0 ) = f (τ0 , Sτ (m) ).
M 0
m=1