Professional Documents
Culture Documents
N 4804
April 2003
ISSN 0249-6399
ISRN INRIA/RR--4804--FR+ENG
THME 4
apport
de recherche
, Lucia Caramellino
, Antonino Zanette
Abstract:
Following the pioneering papers of Fourni, Lasry, Lebouchoux, Lions and Touzi (see
[12] and [13]), an important work concerning the applications of the Malliavin calculus in numerical methods for mathematical nance has come after. One is concerned with two problems:
computation of a large number of conditional expectations on one hand and computation of Greeks
(sensitivities) on the other hand. A signicant test of the power of this approach is given by its
application to pricing and hedging American options.
Our paper gives a global and simplied presentation of this topic including the reduction of variance
techniques based on localization and control variables.
implementation, number of numerical tests are presented and their performances are carefully
discussed.
Key-words:
The authors are also grateful to Bruno Bouchard and Herv Regnier for useful and interesting discussions.
The work of Antonino Zanette was partially supported by the research project Progetto di Ricerca Modelli
matematici innovativi per lo studio dei rischi nanziari e assicurativi(Regione Autonoma Friuli-Venezia Giulia,
Legge Regionale 3/1998).
Equipe Statistiques et Processus, Universit du Maine, and INRIA Rocquencourt, Project Math, Domaine de
Voluceau, Rocquencourt, BP 105, 78153 Le Chesnay Cedex, France
Dipartimento di Matematica, Universit di Roma
, via della Ricerca Scientica, I-00133 Roma,
Tor Vergata
Italy
Dipartimento di Finanza dell'Impresa e dei Mercati Finanziari, Universit di Udine, Via Tomadini 30/A, I-33100
Udine, Italy
A la suite des premiers articles de Fourni, Lasry, Lebouchoux, Lions et Touzi (voir [12]
et [13]), un travail considrable sur les applications du calcul de Malliavin pour les mthodes numriques en nance a t dvelopp. Deux problmes se posent en particulier : le calcul d'esprances
conditionnelles d'une part, les sensitivits et le calcul de la couverture de l'autre. Un test signicatif de la puissance de cette approche est donn par l'application au calcul de prix d'options
amricaines.
Notre papier donne une prsentation globale et simplie, avec l'utilisation de techniques de rduction de la variance, base sur la localisation et l'usage de variables de contrle.
Un intrt
particulier est donn l'implmentation concrte ; plusieurs tests numriques ont t raliss et
discuts.
Mots-cls :
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach3
1 Introduction
The theory of Markov Processes and the Stochastic Calculus have provided a probabilistic interpretation for the solutions of linear partial dierential equations (shortly, linear PDE's) by means
of the Feynman-Kac formula. One of the most striking applications is the emergence of the Monte
Carlo method as an alternative to deterministic numerical algorithms for solving PDE's.
method is slower than the analytical ones, but as the dimension increases (more than
3),
This
it is well
known that the analytical methods do no more work and so the Monte Carlo method remains the
only alternative.
But one has to notice that both the Feynman-Kac formula and the Monte Carlo method are specic to linear problems and collapse when dealing in the non linear case. In the last decay much
work has been done in order to extend the probabilistic methods in a non linear frame. On one
hand, in a set of papers Pardoux and Peng (see [20], [21]) introduced the Backward Stochastic
Dierential Equations (BSDE's in short), which generalize the probabilistic representation given
by the Feynman-Kac formula to non linear problems. In [14] (see also [1]) this representation was
obtained for obstacle problems as well (by means of reected BSDE's). Applications to mathematical nance have been discussed in [15]. So the theoretical background has been prepared. The
second step is to obtain ecient probabilistic algorithms for solving such problems. Recall that
one specic point in the Monte Carlo method is that it does not employ grids. More precisely, in
order to compute the solution
u(0, x0 )
(t + L)u(t, x) = 0,
(t, x) [0, T ] Rd
u(T, x) = f (x),
one represents
u(0, x0 )
order to compute this expectation. On the contrary, an analytical method constructs a time-space
grid and employs a dynamical programing algorithm in order to compute the solution. This means
that in order to compute the solution in one point (0, x0 ), the analytical method is obliged to
(tk , xik ) while the Monte Carlo method provides directly the solution
in (0, x0 ). This is a big advantage because in large dimension grids become dicult to handle.
Think now that instead of the above linear problem, one has to solve a PDE of the type
compute the solution on a whole grid for the simple reason that one has to know the value of
u(t, x)
which comes on in
essential. At this stage, a probabilistic approach is in the same position as an analytical one: one
has to construct a grid, then to solve at each time step a linear problem and nally to add the
input
a conditional expectation instead of saying that he solves a linear PDE. But the problem remains
the same: computing a large number of conditional expectations, and this is not trivial.
The question is now if a probabilistic point of view provides a new approach to this kind of
problems. And the answer is armative. These last years, essentially motivated by mathematical
nance problems - and the most striking one is pricing American options in large dimension several new ideas appeared in this eld. Roughly speaking they may be divided in three families.
In the rst one, a tree is built up in order to obtain a discretization of the underlying diusion on a
grid. This family includes Broadie and Glasserman algorithm (see [10]), the quantization algorithm
(see [2], [3]) as well as Chevance's algorithm (see [11]). The second idea is to use regression on a
2
truncated basis of L in order to compute the conditional expectations. This is done in Longsta
RR n 4804
and Schwartz [18] and in Tsisiklis and Van Roy [22]. Finally in [12], [13], Fourni et al. obtain
a representation of the conditional expectation using Malliavin's calculus and then employ this
representation in order to perform a Monte Carlo method. The specicity of this last approach is
that it appears as a pure Monte Carlo method despite the non linearity.
Except for solving PDE's (which amounts to pricing an option in the nancial frame), a second
problem of interest is to compute the sensitivity of the solution to some parameter (hedging and
Greeks, in nancial language). It seems that Malliavin calculus is an especially promising tool for
solving such a problem. It has been used by Lions and Reigner [17] which follow the third method,
as well as in [4] where quantization algorithm is employed.
The present paper deals with the last method based on Malliavin calculus. Although this approach
works for a large class of non linear problems, we focus on the American option pricing, which
amounts to solve obstacle problems for PDE's. This seems to be a signicant test on the eciency
of the method and several papers related to this subject appeared in the last time. Our aim is
to give a general view on this topic (with simplied proofs), to present concrete examples and to
discuss the numerical results. A special interest is given to the reduction of variance techniques
based on localization and control variables and to their impact in concrete simulation.
It worth to mention that as long as one restricts himself to the Black Scholes model (which of course
is of interest in mathematical nance) the computations related to Malliavin calculus are rather
explicit and elementary.
accessible to readers who are not familiar with the rather heavy machinery of Mallavin calculus.
In the case of a general diusion the same reasonings work as well, but of course one has to employ
the Malliavin calculus in all the generality.
available in the Balck Scholes model become heavier (in particular, they involve the inverse of the
Malliavin covariance matrix). The presentation that we give here is based on one more remark:
the results obtained in the one dimensional case extend easily (using an elementary geometrical
argument) to the multidimensional case. This signicantly simplies the proofs.
The paper is organized as follows. In Section 2 we present the problem and in Section 3 we give
the results (representation formulas for conditional expectation and for the strategy) in the one
dimensional case.
This section also contains the Malliavin calculus approach in the elementary
frame which is the ours. In Section 4 we present the multidimensional case and in Section 5 we
discuss optimal localization. In Section 6 we give the algorithm based on this approach and we
discus the numerical tests.
T.
Let
Xs
T,
process:
(1)
X0 = x
b and denote a vector and a matrix eld on Rd and W is a d-dimensional Brownian motion.
(Xs ) denote the cash-ow associated with the option. The price as seen at time t of such an
where
Let
R
P (t, Xt ) = sup E e t rs ds (X ) Ft
Tt,T
where
Tt,T
stands for the set of all the stopping times taking values on
[t, T ]
and
denotes the
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach5
The solution of this optimal stopping problem has been provided by using the theory of the Snell
envelopes: the rst optimal stopping time is given by
P (t, x),
Lg(x) =
being
a = ,
P (t, x) solves
X , that is
d
d
X
1 X
2g
g
aij (x)
(x) +
bi (x)
(x)
2 i,j=1
xi xj
xi
i=1
then
P
(t, x) + LP (t, x) r(t)P (t, x) = 0
t
P (t, x) > (x), with the nal condition P (T, x) = (x). A
whenever
rigorous reasoning turns out
1
to be much more dicult because generally we have not sucient regularity for P (that is, C in t
2
and C in x) on one hand and on the other one, the behavior of the solution P in a neighborhood of
the free boundary
term in the PDE). This leads to a weak formulation of the problem. The rst one was given in
terms of variational inequalities by Bensoussan and Lions [6]; in El Karoui et al. [14], one can nd
a formulation in viscosity sense and in [1], Bally et al. give a formulation in Sobolev sense.
Thus, the problem of the pricing of an American option is a strongly nonlinear problem, and there
is non hope to nd closed formulas. In order to numerically compute the price of an American
option, one can use either a deterministic or a stochastic algorithm.
Concerning the deterministic methods, they can be based on nite elements methods for variational
inequalities. But here, we are interested in a stochastic approach, that is in a Monte Carlo algorithm
for the pricing of American options. It turns out from a dynamic programming principle, but let
us rstly start by formalizing better the framework.
Let
and set
(2)
Tt,T
P (0, x),
0, it is possible to set up a Bellman dynamic programming principle.
Indeed, let 0 = t0 < t1 < . . . < tn = T be a discretization of the time interval [0, T ], with step size
kt ' Xkt .
kt )k=0,1,...,n an approximation of (Xt )t[0,T ] , that is X
equal to t = T /n, and let (X
kt ) can be approximated by means of the quantity Pkt (X
kt ), given by the
The price P (kt, X
where we have supposed that the spot rate is constant. In order to numerically evaluate
RR n 4804
Theorem 2.1
Given
1, n 2, . . . , 1, 0,
t = T /n (0, 1),
dene
kt ) = (X
nt )
Pnt (X
k = n
kt .
kt ) = max (X
kt ) , ert E P(k+1)t (X
(k+1)t ) X
Pkt (X
Then
The above statement is heuristic and a rigorous formulation supposes to precise the hypothesis on
the diusion coecients and on the regularity of the obstacle. This is done by Bally and Pags
[3]. Roughly speaking, the error is of order
if more
regularity holds.
As a consequence, one can numerically evaluate the delta
(t, x)
the derivative of the price with respect to the initial value of the underlying asset price:
x P (t, x).
(t, x) =
Recall that this is important since it gives the sensibility of the price with respect to
the initial underlying asset price and also for the hedging of the option. By considering the case
t = 0,
Proposition 2.2
For any
t = T /n (0, 1),
0 (x)
of
(0, x)
can be stated.
set
t = { Rd ; Pt () < ()},
where
Pt ()
t = .
2t ) X
Pt () = max () , ert E P2t (X
Then, by setting
kt = 1c
(k+1)t ) X
()
= () 1t + ert E P(k+1)t (X
t
0 (x) = Ex (Xt )
where
and
0 (x)
(0, x) '
Such an assertion is heuristic and a rigorous statement, including error bound, turns out to be a
more dicult problem. One may nd in Bally et al. [4] such a bound but it is given in a weaker
L2 ([0, T ], dt)).
sense (in
P (0, x)
(0, x),
Xt ,
[4], the algorithm by Broadie and Glassermann [10] or the one by Barraquand and Martineau [5].
Another Monte Carlo approach makes use of regression methods to perform the approximation of
the conditional expectation, as made by Longsta and Schwartz [18] or by Tsitsiklis and VanRoy
[22]
Also in order to overcome the problem of the discretization of the path space, another method can
be used. It has been introduced by Lions and Regnier [17] and uses formulas allowing to represent
conditional expectations like
E(F (Xt ) | Xs = )
E(F (Xt ) | Xs = )
written in
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach7
E(F (X ) )
t
s
E F (Xt ) Xs = =
E(s )
E(F (X ) 1, )E( ) E(F (X ) )E( 1, )
t
t
s
s
s
s
E F (Xt ) Xs = =
E(s )2
being
and
s1,
(3)
suitable weights, which could also depend on suitable localizing functions. Such
representations can be proved by using Malliavin calculus techniques. The proofs providing the
formulas as in (3) can be found in Section 3 and 4, in the framework of the Black and Scholes
model; the original reference is the paper by Lions and Regnier [17] but here we develop a little
dierent approach, mainly for the multidimensional case, allowing to simplify considerably the
proofs. Moreover, in Section 5 we discuss an optimality criterium for the choice of the localizing
functions, which play an important role for practical purposes, as already studied and observed by
Kohatsu-Higa and Petterson [16] and by Bouchard, Ekeland and Touzi [7].
Let us point out that we could employ the strong Malliavin calculus for our proofs (that is,
Malliavin derivatives, Skorohod integrals etc. - words which do not appear in this paper) but we
have opted for a more friendly approach, in order to make it readable and comprehensible also
to people who are not familiar with the topic.
Representations (3) can be used for the practical purpose of the pricing of American options
1,
can be written explicitly, expectations like
as follows. In fact, since the weights s and s
1,
E(f (Xt ) s ) or E(f (Xt ) s ) can be approximated through the associated empirical means and
used to numerically compute the price
P (0, x)
(0, x)
Proposition 2.2, thus avoiding the problem of the approximation of the transition density and of the
discretization of the path space. This plan gives also the considerable gain to provide a Monte Carlo
algorithm for the evaluation of
P (0, x)
and
(0, x)
trajectories for the computation of any conditional expectation appearing in Theorem 2.1 and in
Proposition 2.2. Let us remark that, using this approach, the valuation of the delta is not made
through nite dierence approximations but it is performed by means of representation formulas
written in terms of expectations. We postpone to Section 6 (cfr. 6.1 and 6.2) a comprehensive
presentation of such an algorithm. Finally, some numerical results can be found in 6.3.
Finally, here we consider price and delta, then representation formulas for the conditional expectation and its gradient.
It is worth to point out that one could also attack the problem of the
numerical evaluation of other greeks. Indeed, similar techniques could be applied in order to obtain
representation formulas for other derivatives of the conditional expectation: the second derivative
(for the gamma), the derivative w.r.t.
w.r.t. the volatility
be a single underlying asset price process, driven by the Black and Scholes model, i.e. it
RR n 4804
where:
of the option;
>0
(4)
E((Xt ) | Xs = )
where
0 < s < t, R+
and
Eb (R) = {f M(R) :
where
there exist
M(R) = {f : R R : f
C>0
and
mN
such that
0 s < t, Eb
where
being
and
H(Xs )
Ts,t [f ]() = E f (Xt )
Ws,t
s(t s)Xs
H() = 10 , R,
(5)
and
where
Ts,t [f ]
H(Xs ) h (Ws,t )2
t i
+
W
.
Rs,t [f ]() = E f (Xt )
s,t
s(t s)Xs2 s(t s)
f,
the operator
Rs,t [f ]()
where
(6)
s,t L2 ()
E (Xt )0 (Xs ) = E((Xt )H(Xs )s,t )
stands for the Dirac measure in
associated with
0 ).
(notice that
Obviously, the Dirac mass is something very irregular". In some sense, the
Malliavin integration by parts formula allows to regularize it, thus to overcome the problem of
handling a Dirac mass, but it is worth to point out that this procedure provides an high variance
because of the presence of the Heaviside function
H.
method for the computation of conditional expectations, as made in the following Theorem 3.3
(Lions and Regnier [17], Lemme 2.1.1) by introducing the localizing function
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach9
Lemma 3.2
Ts,t
true with
R
: R [0, +) be such that R ()d = 1. Then formulas (5)
and Rs,t replaced by Ts,t = Ts,t and Rs,t = Rs,t respectively, where
Let
H(Xs ) (Xs )
T
[f
]()
=
E(f
(X
)(X
))
+
E
f
(X
)
W
t
s
t
s,t
s,t
s(t s)Xs
and
denotes the
E f (Xt )(Xs )
Ws,t
+
s(t s)Xs
H(Xs ) (Xs ) h (Ws,t )2
t i
+
W
.
E f (Xt )
s,t
s(t s)Xs2
s(t s)
Ry
probability distribution function associated with : (y) = ()d .
R
s,t [f ]() =
where
(7)
(8)
By using the localized version for the operators, we can immediately set up localized representation
formulas:
: R [0, +),
such that
i ()d = 1,
For any
0 s < t, F E b ,
one has
T []()
s,t
E (Xt ) Xs = =
Ts,t [1]()
and
R []()T [1]() T []()R [1]()
s,t
s,t
s,t
s,t
E (Xt ) Xs = =
,
2
Ts,t [1]()
T
s,t [f ]()
and
R
s,t [f ]()
We give now another localization result due to Kohatsu-Higa and Pettersson [16] in the onedimensionale case and to Bouchard, Ekeland and Touzi [7] in the multidimensional setting.
Cb1 (R)
such that
()
=1
Let
0 s t, E b
and let
>0
be xed.
then
J []()
s,t
E (Xt ) Xs = =
J
s,t [1]()
where
J
s,t [f ]() =
i
1
H(Xs ) h
E f (Xt )
(Xs )Ws,t + 0 (Xs ) Xs s(t s)
s(t s)
Xs
We shall give here an elementary approach to the computation of conditional expectations allowing
to prove Theorems 3.1, 3.3 and 3.4. We start next section by proving some results on Malliavin
Calculus allowing to prove the above Theorems, whose proofs can be found in the section after the
next one.
RR n 4804
10
3.1
The representation formulas stated in the previous section are based on Malliavin Calculus techniques.
In the frame of the Black and Scholes model (on which we focus in this paper), these
techniques reduce to elementary reasonings based on (standard) integration by parts and so are
available to readers who are not familiar with this theory. Therefore, we present in this section
the elementary counterpart of Malliavin's integration by parts formula and the way in which one
employs this formula for computing conditional expectations. In the case of a general diusion,
the strategy is the same and people familiar with this calculus may easily adapt the arguments.
Let us start by a simple denition.
Denition 3.5
Let
and
Cb (R)
for any
F (G)
IP (F, G)
such that
Cb (R)
(9)
One has
Lemma 3.6
probability density
IP (F, 1)
holds.
has a
given by
IP (F, 1)
and
IP (F, G)
hold. Then
E(G | F = ) =
E(H(F ) F (G))
,
E(H(F ) F (1))
E(H(F ) F (1)) = 0.
Proof.
(i)
negative function
(x) = (x/)/
e
1
[0,+) (x) = 1(0,+) (x) + 0 (x)/2. Take now G as a random variable, independent of F , whose
law is (x) dx. Clearly lim0 E(f (F G )) = E(f (F )), for any continuous function f . Then we
can write
Z Z
E(f (F G )) =
f (u v) (v) dv dP F 1 (u)
R R
Z
Z
Z
f (z) (u z) dP F 1 (u)dz =
f (z)E( (F z)) dz.
=
R
Since the
IP (F, 1)
Z
lim E(f (F G )) =
e
f (z)E(1
[0,+) (F z) F (1)) dz.
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach11
we have
Z
E(f (F )) =
which allows to state that the law of
e
f (z)E(1
[0,+) (F z) F (1)) dz,
e
p(z) = E(1
[0,+) (F
e
replace 1[0,+) (F z) with
Let us denote
() =
E(H(F ) F (G))
.
E(H(F ) F (1))
We only have to check that for any bounded and continuous function
E(f (F ) (F )). By
IP (F, G) property
continuous).
it holds that
E(f (F ) G) =
using again the regularization function introduced above, the validity of the
and what proved in part
(i)
Z
E(f (F ) G) = E G lim
f (z) (F z) dz = lim
f (z) E(G (F z)) dz
0 R
0 R
Z
Z
f (z) E(G 0 (F z)) dz = lim
f (z) E( (F z) F (G)) dz
= lim
0 R
0 R
Z
Z
f (z) E(1[0,+) (F z) F (G)) dz =
f (z) (z) p(z) dz = E(f (F ) (F )),
=
R
2
We give now a localized version of the result presented above, which will be used in order to prove
Theorem 3.4.
Denition 3.7
Let
and
for any
Cb (R)
such that
and let
D R
be an
supp(0 ) D.
(10)
E(G | F = ) = ()
E(f (F ) G) = E((F ) G)
for
for any
such that
supp(f ) D.
if
f Cb (R)
such that
supp(f ) D.
With these localized denitions, we obtain the exact analogous of Lemma 3.6, whose proof is
dropped since it is actually identical to the proof of Lemma 3.6:
Lemma 3.8
on
RR n 4804
has a
12
IPD (F, 1)
and
IPD (F, G)
E(G | F = ) =
hold. Then
E(H(F ) FD (G))
,
E(H(F ) FD (1))
IP
E(H(F ) FD (1)) = 0.
and
IPD
properties in a
Lemma 3.9
Let
N(0, ).
f, g C 1 ,
one has
X = xe+ ,
(b) Setting
one has
g(X)
+ g 0 (X) ,
E(f 0 (X) g(X)) = E f (X)
X
with
f, g C 1 .
(a)
Proof.
E(f () g()) =
Z
1
2
1
y
y 2 /(2)
e
ey /(2) dy
f (y) g(y)
dy = f (y) g(y) g 0 (y)
2
2
0
= E f () g() g () .
f1 () and
g 0 (X) X .
g ()
g1 ()
g1 ()
1
= E f1 ()
E(f 0 (X) g(X)) = E f10 ()
X
X
X
g(X)
0
+ g (X) .
= E f (X)
X
2
We can then state the following
Proposition 3.10
(i) The
Let
IP (Xs , g(Xt ))
s<t
and
g : RR
Cb (R),
with
s [g](Xt ) = g(Xt )
being
1
Ws,t ,
s(t s) Xs
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach13
> 0,
Cb1 (R)
let
such
with
> 0.
B ()
s [g](Xt ) =
g(Xt )
Ws,t ,
s(t s) Xs
being
s ) tWs sWt + s(t s) + 0 (Xs ) Xs s(t s)
Ws,t = (X
s )Ws,t + 0 (Xs ) Xs s(t s).
= (X
In particular, the
s,t
IP (Xs , g(Xt ) s(ts)X
)
s
(iii) The
Cb (R),
Ws,t
s [g](Xt ))
= E((Xs )
s(t s)Xs
(ii)
and
(W )2
1
t
s,t
+
W
.
s,t
s(t s) Xs2 s(t s)
(iii):
the statement in
(ii).
(i)
(ii).
Xt = Xs Y ,
(ii), is that 0 = 0 ,
s ) g(Xt ) .
E 0 (Xs ) g(Xt ) = E 0 (Xs ) (X
so that
Since
with
s [g](Xt ) = g(Xt )
Proof.
E 0 (Xs ) g(Xt )
for any
D = B (),
Y = eh(ts)+(Wt Ws )
independent of Xs , we can
s ) g(Xt )) = E E(0 (Xs ) (X
s ) g(y Xs ))
E(0 (Xs ) (X
with
write
.
y=Y
Now,
Xs = x ehs+Ws ,
so that by using
(b)
s ) g(y Xs ))
E(0 (Xs ) (X
(X
s ) g(y Xs ) Ws
s ) y g 0 (y Xs )
+ 0 (Xs ) g(y Xs ) (X
= E (Xs )
Xs
s
(X
s ) g(y Xs ) Ws
s ) g 0 (y Xs ) .
+ 0 (Xs ) g(y Xs ) E y (Xs ) (X
= E (Xs )
Xs
s
By using again the independence between
Xs
and
Y,
one has
s ) g(Xt ))
E(0 (Xs ) (X
(X
s ) g(Xt ) Ws
s ) g 0 (Xt ) .
+ 0 (Xs ) g(Xt ) E Y (Xs ) (X
= E (Xs )
Xs
s
RR n 4804
14
Let us consider the second addendum of the r.h.s.: we will use the integration by parts formula
based on
write
g.
Xs ,
one can
s ) g 0 (Xt ) = E E (z) (z)
Y g 0 (z Y )
E Y (Xs ) (X
z=Xs
0
.
= E (z) (z)E
Y g (z Y )
z=Xs
By using
(b)
g(z Y ) W W
g(z Y ) Y W W
t
s
t
s
+ 1 =E
E g 0 (z Y ) Y = E
z
Y
ts
z
ts
and by inserting this in the above equality one obtains
s ) g 0 (Xt ) = E (Xs ) (X
s ) g(Xt ) Wt Ws .
E Y (Xs ) (X
Xs
ts
We can nally write
s ) g(Xt ))
E(0 (Xs ) (X
(X
s ) g(Xt ) Ws
s ) g(Xt ) Wt Ws
+ 0 (Xs ) g(Xt ) E (Xs ) (X
= E (Xs )
Xs
s
Xs
ts
and straightforward computations allow to conclude. Let us nally observe that to achieve this
1
representation one has implicitly assumed that g is regular (C ), which is not true in general. But
this is not really a problem: one can regularize
(iii).
v(Xs , Xt ) =
Then by using
(b)
Xt = y Xs , g(Xt ) = g(y Xs )
and
Ws,t = v(Xs , Xt ),
where
s
t s
ln Xs ln x hs + 2 s
ln Xt ln Xs h(t s) .
h g(X ) v(X , X )
t
s
t
Ws,t
E 0 (Xs )g(Xt ) s(ts)X
)
(Ws + s)
=
E
(X
s
s
2 s2 (t s)Xs2
vx (Xs , Xt ) + Y vy (Xs , Xt ) g(Xt ) v(Xs , Xt ) i
v(Xs ,Xt )
g(X
)
+
Y g 0 (Xt ) s(ts)X
t
s
s(t s)Xs
s(t s)Xs2
v(Xs ,Xt )
0
0
Consider now the term containing g , that is E (Xs )Y g (Xt )
s(ts)Xs : by conditioning rst
on Xs , one has to evaluate:
(x )
v(x, xY )
=
E(Y g 0 (xY ) v(x, xY )).
E (x )Y g 0 (xY )
s(t s)x
s(t s)x
By recalling that
Y = exp(h(t s) + (Wt Ws ))
(b)
has
g(xY ) h Y v(x, xY ) W W
i
t
s
+ v(x, xY ) Y vy (x, xY ) x
E
x
Y
ts
g(xY )
v(x, xY ) (Wt Ws ) Y g(xY ) vy (x, xY )
E
(t s)x
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach15
so that
v(Xs , Xt )
E (Xs )Y g 0 (Xt )
s(t s)Xs
i
g(Xt ) h v(Xs , Xt )
(Wt Ws ) Y vy (Xs , Xt ) .
= E (Xs )
s(t s)Xs (t s)Xs
By inserting such a quantity in the expression above, straightforward computations allow to obtain
Ws,t
E 0 (Xs )g(Xt ) s(ts)X
=
E
(Xs )
s
g(Xt )
[v(Xs , Xt )(t s)(Ws + s)
s)Xs2
v(Xs , Xt )s(Wt Ws ) + v(Xs , Xt )s(t s) vx (Xs , Xt )s(t s)Xs ] .
Now, since
vx (Xs , Xt ) =
2 s2 (t
t
Xs , one can rewrite the above expression as
E 0 (Xs )g(Xt )
Ws,t
(Xs ) h (Ws,t )2
t i
+ Ws,t
.
= E g(Xt )
2
s(t s)Xs
s(t s)Xs s(t s)
Remark 3.11
The key point in the above lemma was that we used two integration by parts forXs and the second with respect to Y = Xt /Xs = eh(ts)+(Wt Ws ) .
From a numerical point of view, it is the integration by parts with respect to Y which is dangermulas: the rst with respect to
ous. In fact, the noise contained in Y is Wt Ws , so, if t s is small, we have few noise.
As a consequence, the weight which appears in the integration by parts contains 1/ t s, which
blows up if
t = tk ,
t s is small. Think now to our dynamical programming principle, where s = tk1 and
t s = tk tk1 . If one wants to take a very rich time-grid (which is necessary in
so that
order to get a good approximation of an American option with a Bermudean one), then one has to
face the problem of
1/ tk tk1 .
3.2
Proofs
Concerning part
ii),
Part
i) is
H(Xs ) h (Ws,t )2
t i
+
W
,
Ts,t [f ]() = Rs,t [f ]() = E f (Xt )
s,t
s(t s)Xs2 s(t s)
for any
f Eb (R).
as a
converges,
H (Xs )
Ts,t [f ]() = E f (Xt )
Ws,t .
s(t s)Xs
Obviously,
as
0,
RR n 4804
16
Ts,t [f ]().
One has
Ts,t [f ]() = E f (Xt ) h (Xs )
By using part
(iii)
Ws,t
.
s(t s)Xs
H (Xs ) h (Ws,t )2
ti
+
W
Ts,t [f ]() = E f (Xt )
s,t
s(t s)Xs2 s(t s)
(recall that
h = H0 )
0,
2
Proof of Lemma 3.2.
F = Xs ,
G = f (Xt ),
R (t) dt = 1
F (G) = f (Xt )
1
Ws,t ,
s(t s) Xs
Ts,t [f ]() =
R xE(H(F ) F (G)), where H(x) = 1x0 .
(x) = (t) dt. Since the IP (F, G) holds (recall
and set
(11)
Take
0
(i)
(11) and
F, G
and
F (G)
(H )(Xs )
Ts,t [f ]() = E f (Xt ) (Xs ) + E f (Xt )
Ws,t = T
s,t [f ]().
s(t s) Xs
Concerning
Rs,t ,
we set
(W )2
1
1
t
s,t
= f (Xt )
+
W
,
Ws,t , F (G)
s,t
s(t s) Xs
s(t s) Xs2 s(t s)
(12)
R
. Take 0 such that
so thatRs,t [f ]() = E(H(F ) F (G))
(t)
dt
=
1
and set (x) =
R
Rx
(t) dt. By using the IP (F, G) property proved in (iii) of Proposition 3.10, one has
= f (Xt )
F = Xs , G
= E((F ) G)
+ E(H(F ) F (G))
E(0 (F ) G)
E(H(F ) F (G))
F, G
and
F (G)
as in (12).
stronger condition
We recall that
B ()
= 1,
for some
is xed and
> 0.
()
= 1.
Then, by using
(ii)
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach17
that the
IPB () (F, G)
G = (Xt ),
E H(Xs ) X
((Xs ))
s
,
E((Xt ) | Xs = ) =
for
E H(Xs ) X
(1)
s
obtain
F = Xs
and
any
B (),
(13)
= .
()
= 1. Then we may construct a sequence {n }n such that
c
0
0
n = 1
on
any and
n
X
(1) X
(1) as
s
s
for such a
n ,
be the underlying asset price process, driven by the Black and Scholes model, i.e. it solves
dXt = (
r )Xt dt + Xt dWt
X0 = x
where:
x Rd+ ;
r, Rd ,
with
ri = r
for any
i = 1, . . . , d,
being
the
denotes the
is a
dd
d-dimensional
Xt
i < j,
and that
is a standard
can be written as
i
X
Xti = xi exp hi t +
ij Wtj ,
i = 1, . . . , d
j=1
where from now on we set
1X 2
,
2 j=1 ij
i
hi = ri i
i = 1, . . . , d.
E((Xt ) | Xs = )
RR n 4804
(14)
18
0 < s < t, Rd+ and Eb (Rd ) the class of the measurable functions with polynomial
m
growth, i.e |(y)| C(1 + |y| ), in order to derive representation formulas in dimension d using
the ones in dimension d = 1.
e with independent components
In few words, to this goal it suces to consider an auxiliary process X
where
for which a formula for the conditional expectation immediately follows as a product. In a second
step, such a formula can be adapted to the original process
giving
e.
X
`t = (`1t , . . . , `dt )
be a xed
C1
e i = xi exp hi t + `i + ii W i ,
X
t
t
t
i = 1, . . . , d
(15)
et = (
et dt +
et dWt
dX
r + `0t )X
X
e0 = x
X
being
Lemma 4.1
For any
t0
Let
t, `, x
ii , i = 1, . . . , d.
Ft () : Rd+ Rd+
et )
Xt = Ft (X
Proof.
e
X
(16)
such that
Ft
X:
is inversible and
et = Ft1 (Xt ).
X
Wti =
ei
1 X
ln it hi t `it .
ii
x
i
i e j /
Y
X
Xt ij jj
ij
Xti = xi exp hi t
(hj t + `jt )
xj
j=1 jj
j=1
Thus, by setting
eij =
and by using the notation
ij
,
jj
ln = (ln 1 , . . . , ln d ),
i, j = 1, . . . , d.
for
satises
ln Ft (y) = e
`t +
e ln y + (I
e)(ln x + ht)
and, by setting
b=
e
ln Ft1 (z) = `t +
b ln z + (I
b )(ln x + ht).
2
Let us summarize for later use the previous result. We set
eij =
ij
, i, j = 1, . . . , d,
jj
and
b=
e1
(17)
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach19
b is easy to
compute because
et )
Xt = Ft (X
Ft
et = Gt (Xt )
X
and
are given by
Fti (y) = e
Pi
j=1
eij `jt
yi
i1
Y
j=1
and
i
i1
Y
z
j=1
By using the process
e,
X
xj
yj hj t eij
e
,
xj
ehj t
bij
i = 1, . . . , d, y Rd+
(18)
i = 1, . . . , d, z Rd+ .
(19)
mainly the fact that its components are independent, we can easily obtain
a rst formula for the conditional expectation starting from the one-dimensional one.
0s<t
Eb (Rd )
and
Rd+ ,
one has
T []()
s,t
E (Xt ) | Xs = =
Ts,t [1]()
where
being
d
Y
ei
H(X
ei )
s
i
Ts,t [f ]() = E f (Xt )
Ws,t
ei
i=1 ii s(t s)Xs
es = Gs (Xs )
X
and
e = Gs () [Gs
i
Ws,t
= (t s)(Wsi + ii s) s(Wti Wsi ),
ii) Let
0s<t
Eb (Rd ),
H() = 10 , R,
(20)
and
i = 1, . . . , d.
j = 1, . . . , d,
j
X
Rs,t;k []()Ts,t [1]() Ts,t []()Rs,t;k [1]()
ek
j E (Xt ) | Xs = =
bkj
,
j
Ts,t [1]()2
k=1
where
Ts,t [f ]()
Rs,t;k [f ]() =
Remark 4.3
k = 1, . . . , d,
E f (Xt )
k 2
esk
)
H(X
t i
ek ) h (Ws,t
k
+ Ws,t
esk )2 kk s(t s)
kk
kk s(t s)(X
d
Y
esi
H(X
ei )
i
Ws,t
.
ei
i=1,i6=k ii s(t s)Xs
is
Iddd . Thus, the sum appearing for the evaluation of j E(F (Xt ) | Xs = )
j
reduces to the single term with k = j , with coecient
ej /j = e`s , which in turn is equal to 1
whenever ` = 0.
diagonal, then
RR n 4804
b =
(21)
20
Since
et )
Xt = Ft (X
(recall that
i)
Let us set
e
e t (y) (y)
= Ft (y), y Rd+ ,
being
Ft
dened in (18).
for any
Gs = Fs1 ).
Thus, setting
es
e = Gs (),
T
e s,t [](e
e )
es =
e X
et ) X
E (
e =
e s,t [1](e
T
)
where
e s,t [f ](e
T
) =
Now, let us rstly suppose that
of
d functions
(22)
d
Y
ei
1
H(X
ei )
s
i
et )
E f (X
Ws,t
ei
s(t s)
X
s
i=1 ii
i=1
d
Y
e
e 1 (y1 )
e can be separated in the product
e d (yd ), that is
(y)
=
Eb (R). In such a case,
d
Y
e
esi =
e
e
e i (X
eti ) X
E (Xt ) Xs =
e =
E
ei .
i=1
ei
X
t
e i is
Now, it is easy to see that for each
Theorem 3.1 can be applied (indeed, here the drift of X
t
i ei
i
of the form t Xt while the process handled in the one-dimensional case has t = const., but this
is not a diculty). Thus, recalling that
e
X
d
d
Y
Y
e i ](e
Tis,t [
i )
es =
ei =
e X
et ) X
e i (X
e i ) X
E (
e =
E
e
=
i
t
s
i
T
[1](e
i)
s,t
i=1
i=1
being
Tis,t [g](e
i ) =
e i ei )
1
i
eti ) H(Xs
E g(X
Ws,t
.
esi
ii s(t s)
X
e s,t [](e
e ) =
T
d
Y
e i ](e
Tis,t [
i )
and
e,
X
e s,t [1](e
T
) =
i=1
d
Y
Tis,t [1](e
i ),
i=1
e
e 1 (y1 )
e d (yd ). In the general case, the statement holds by
(y)
=
d
e Eb (R ) there exists a sequence of functions {
e n }n Eb (Rd )
using a density argument: for any
e n (X
e n is a linear combination of functions which
et ) (
e X
et ) in L2 and such that each
such that
e n , it nally holds for
e
separate the variables as above. Since representation (22) holds for any
so that (22) holds when
ii)
bkj
ek /j = j Gks () = j
ek . Thus, by considering Ts,t [f ]()
e
as
d
Y
esi
H(X
ei )
i
e s,t [f ](e
et )
Rs,t;k [f ]() = ek T
) = ek E fe(X
Ws,t
,
ei
i=1 ii s(t s)Xs
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach21
k th
of
et
X
except for
d
Y
e k ek )
ei
H(X
ei )
s
k
i
e s,t [f ](e
e k ) H(Xs
T
) = E E fek (X
W
Ws,t
,
t
s,t
j
e ,j6=k
ek
ei
x
ej =X
kk s(t s)X
t
s
i=1,i6=k ii s(t s)Xs
being
etk ) = fe(e
etk , x
x1 , . . . , x
ek1 , X
ek+1 , . . . x
ed ). Thus,
fek (X
k
k )
k
e s,t [f ](e
etk ) H(Xes e
ek T
) = E ek E fek (X
W
s,t
k
e
s(ts)X
kk
d
Y
i=1,i6=k
e j ,j6=k
x
ej =X
t
ei
H(X
ei )
s
i
Ws,t
.
ei
ii s(t s)X
s
Now, by using the one dimensional results (see Theorem 3.1 and (6) in particular), one has
k
e k ek )
e k ek ) h Ws,t
t i
k
k
e k ) H(Xs
e k ) H(Xs
+Ws,t
ek E fek (X
Ws,t
= E fek (X
t
t
ek
e k )2 kk s(t s)
kk s(t s)X
kk s(t s)(X
s
s
By rearranging everything, the formula holds.
2
Let us give some commentary remarks about the signicance of the drift-function
Qit (z)
=e
`it
Git (z)
= zi
i1
Y
j=1
`t .
Setting
zj hj t bij
e
,
xj
then
H(Gis (Xs ) Gis ()) = 1Gis (Xs )Gis () = 1Qis (Xs )Qis () = H(Qis (Xs ) Qis ()),
so that we obtain
Ts,t [f ]() = e
Pd
i=1
`is
d
Y
1
H(Qis (Xs ) Qis ())
i
E f (Xt )
Ws,t
.
i
s(t s)
Qs (Xs )
i=1 ii
i=1
d
Y
(23)
i
Now, by recalling that Qt (z) does no more depend on `, the only way in which `t comes out in our
Pd
i
formula is
i=1 `t , so that we have just one degree of freedom. The problem is how to choose `,
let us now discuss about this.
The main application of the formula is addressed to Monte Carlo simulations, therefore, as usual,
one way is to optimize (here, minimize) the variance or the integrated variance of the Monte Carlo
`.
m() = E((Xt ) | Xs = )
mN () =
where
TN
s,t [f ]() =
d
Y
N
d
Y
esi,(k)
1 X
1
H(X
ei )
(k)
i,(k)
f (Xt )
Ws,t
i,(k)
s(t
s)
N
e
Xs
i=1 ii
i=1
k=1
RR n 4804
TN
s,t []()
TN
s,t [1]()
22
or equivalently
TN
s,t [f ]() =
d
Y
N
d
(k)
Y
1 X
1
H(Gis (Xs ) Gis ())
(k)
i,(k)
Ws,t ,
f (Xt )
i (X (k) )
s(t
s)
N
ii
G
s
i=1
i=1
s
k=1
where
(W (k) , X (k) ),
as
k = 1, . . . , N
(W, X).
By using (23), we
TN
s,t [f ]() = e
Pd
i=1
`is
d
Y
N
d
(k)
Y
1 X
1
H(Qis (Xs ) Qis ())
(k)
i,(k)
Ws,t ,
f (Xt )
(k)
i
s(t
s)
N
Qs (Xs )
i=1 ii
i=1
k=1
TN
s,t [f ]() is the empirical mean, up to a constant
`s
Hs,t
[f ]() = e
Pd
i=1
d
Y
d
Y
1
H(Qis (Xs ) Qis ())
i
f (Xt )
Ws,t
.
i (X )
s(t
s)
Q
ii
s
s
i=1
i=1
`is
Thus, one could ask for `s to be set in order to minimize the variance of
`s
Hs,t
[f ](), or its integrated
variance, that is
`s
v[f ](`s ) = Var Hs,t
[f ]()
By recalling that the
Qis 's
Z
v[f ](`s ) = Var
or
do not depend on
Rd
`,
we can write
v[f ](`s ) = e2
where
v[f ]
`s
Hs,t
[f ](d) .
Pd
i=1
`i (s)
v[f ]
is independent of
do not depend on
`,
is a nonsense in this
context.
Proposition 4.4
`i (s)
Gs
if and only if
` = ` ,
where
` = ` ,
where
is
given by
j
bij hj s ln
,
xj
j=1
i1
X
`i
Gs () = ,
`t = ` t
then
Gs () =
1 j
bij hj ln
,
s xj
j=1
i1
X
i = 1, . . . , d,
if and only if
i = 1, . . . , d.
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach23
Proof.
hs),
thus
= Gs ()
ln Gs (z) = `s +
b ln z + (I
b)(ln x +
if and only if
`s = `s =
b I h s ln
x
where the symbol
ln x
ith
entry is given by
ln xii .
2
Remark 4.5
Fti (y)
e=I
= yi
so that
i1
Y
j=1
moreover, since
yj xj t/s
xj j
eij
,
ln Gs (z) = `s +
b ln z + (I
b)(ln x + hs),
conclude that
Git (z)
= zi
i1
Y
j=1
zj xj t/s
xj j
i = 1, . . . , d;
(24)
bij
,
i = 1, . . . , d.
(25)
Let us now discuss about the connection with the formula given by Lions and Regnier [17], which
takes into account the process
b
X
dened as
i1
X
bt = xi exp hi t +
X
ij wtj + ii Wti
j=1
where (as usual,
P0
j=1 ()
:= 0
i
X
and)
wt
ik wtk = ln i ln xi hi t,
i = 1, . . . , d.
(26)
k=1
It is worth remarking that
b
X
e.
X
Ls,t [f ]() =
then
d
Y
bsi i )
1
H(X
i
E f (Xt )
Ws,t
bsi
s(t s)
X
i=1 ii
i=1
d
Y
L []()
s,t
E (Xt ) | Xs = =
Ls,t [1]()
(27)
e and
Thus, both formulas depend on an auxiliary process with independent components (X
respectively), which in turn is determined by the introduction of a new drift (` and
the latter being dened through (26)). Furthermore, the formulas giving the operators
RR n 4804
b
X
respectively,
Ts,t
and
24
Ls,t , allowing to write down the conditional expectation, are really very similar except for the point
. More precisely, both operators can be rewritten as
d
Y
1
H(Y i gi ())
i
E f (Xt )
As,t [f ]() =
W
,
s,t
s(t s)
Yi
i=1 ii
i=1
d
Y
in which
choose
choose
es
Y =X
bs
Y =X
and
and
1
Thus: what is the connection between the two approaches? It is simple: if one sets `t = 0 and for
P
i1
j
i = 2, . . . , d, `it = j=1 ij wt , then it is straightforward to see that condition (26) equals to the
condition studied in Proposition 4.4, that is
e = Gt () = . This means that if `s is xed in order
to have
Gs () = ,
as it has been made in Proposition 4.4, the two formulas are actually identical.
Therefore, the approach studied here is some more general than the one developed by Lions and
Regnier.
Let us resume the above observations in the following
Remark 4.6
` = 0:
e;
X
`,
Ts,t [f ]() =
where
d
Y
1
E f (Xt )
s(t s)
i=1 ii
i=1
d
Y
Hi (Xs , )
i
W
,
s,t
b
Qi1 Xsj h s ij
i
j
Xs j=1 xj e
Hi (y, ) = 1Qi
b ij
1 .
j=1 (Xs /j )
` = ` ,
with
expectation in point of fact identical to the one provided by Lions and Reigner and the operator
Ts,t [f ]()
is here
d
Y
e`i s
E f (Xt )
Ts,t [f ]() =
s(t s)
i=1 ii
i=1
d
Y
Hi (Xs )
i
W
,
s,t
bij
Q
Xsj hj s
Xsi i1
e
j=1
xj
Ts,t [f ]() =
Qi1
d
Y
Hi (Xs )
i
E f (Xt )
W
.
Qi
s,t
j
bij
s(t s)
j=1 (Xs )
i=1 ii
i=1
d
Y
j=1
j ij
Let us now discuss formulas involving localization functions. If we restrict our attention to producttype localizing function, then we can rst state a localized formula for the operators
Rs,t;j [f ]()
R
Qd
(x) = i=1 i (xi ), x = (x1 , . . . , xd ) Rd , with i 0 and R i ()d = 1.
operators Ts,t and Rs,t;j , dened in (20) and (21) respectively, can be localized as follows:
Lemma 4.7
Then the
and then for the conditional expectation and its gradient. In fact, one rst has
Let
Ts,t [f ]() = T
s,t [f ]()
and
Rs,t;k [f ]() = R
s,t;j [f ](),
k = 1, . . . , d,
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach25
where
d h
i
Y
esi
esi
H(X
ei ) i (X
ei )
i
T
[f
]()
=
E
f
(X
)
Ws,t
i (Xs )) +
t
s,t
esi
ii s(t s)X
i=1
(28)
and
h
esk
R
[f
]()
=
E
f
(X
)
k (X
ek )
t
s,t;k
k
Ws,t
+
esk
kk s(t s)X
k 2
ek
ek
)
t i
ek ) k (X
ek ) (Ws,t
H(X
s
s
k
+ Ws,t
+
e k )2
kk s(t s)
kk
kk s(t s)(X
s
d
h
i
Y
ei
ei
H(X
ei ) i (X
ei )
s
s
i
ei
ei ) +
Ws,t
i (X
.
s
ei
ii s(t s)X
s
i=1,i6=k
where
Proof.
i : i (y) =
et
et ). By conditioning w.r.t. all the coordinates of X
et ) = f Ft (X
fe(X
e has independent components, one has
by recalling that X
Set
one, and
Ry
(29)
i ()d .
d
Y
e 1 e1 )
esi
H(X
ei )
1
i
et1 ) H(Xs
Ts,t [f ]() = E E fe1 (X
Ws,t
W
s,t ,
e j ,j6=1
es1
ei
x
ej =X
11 s(t s)X
t
i=1,i6=1 ii s(t s)Xs
where we have set
etk ) = fe(X
et1 , x
e2 , . . . x
ed ).
fe1 (X
h
i
e 1 e1 )
e1
(H )(X
e1 )
s
1
1
e 1 ) H(Xs
e 1 ) 1 (X 1
E fe1 (X
Ws,t
e1 ) +
Ws,t
= E fe1 (X
,
t
t
s
e1
e1
11 s(t s)X
11 s(t s)X
s
s
so that
d
h
i Y
e1
ei
(H )(X
e1 )
H(X
ei )
s
s
1
i
Ts,t [f ]() = E f (Xt ) 1 (Xs1
e1 ) +
Ws,t
Ws,t
.
e1
ei
11 s(t s)X
s
i=1,i6=1 ii s(t s)Xs
Now, by considering the conditioning w.r.t.
et1 , X
et3 , . . . , X
etd ,
X
one has
2 h
d
i Y
Y
ej
ei
(H )(X
ej )
H(X
ei )
j
s
s
i
Ts,t [f ]() = E f (Xt )
j )+
W
W
j (Xsj e
.
s,t
s,t
esj
ei
jj s(t s)X
j=1
i=1,i6=1,2 ii s(t s)Xs
By iterating a similar procedure on the remaining components, one arrives to the end.
Concerning
Rs,t;k ,
the statement can be proved in the same way, by using the one dimensional
k th
Rs,t
in Lemma 3.2.
h
ek
Rs,t;k [f ]() = E f (Xt ) k (X
ek )
s
k
Ws,t
+
ek
kk s(t s)X
s
k 2
esk
esk
)
t i
ek ) k (X
ek ) (Ws,t
H(X
k
+ Ws,t
+
esk )2
kk s(t s)
kk
kk s(t s)(X
d
i
Y
esi
H(X
ei )
i
Ws,t
.
ei
i=1,i6=k ii s(t s)Xs
RR n 4804
26
Ts,t
j th
one, with
j 6= k ,
and by using
2
By using the localized version for the operators, the localized representation formulas for the
conditional expectation and its gradient immediately follows:
Ld ,
For any
0 s < t, F E b ,
one has
T [F ]()
s,t
E F (Xt ) Xs = =
Ts,t [1]()
and, as
j = 1, . . . , d,
j
X
R
ek
s,t;k [F ]()Ts,t [1]() Ts,t [F ]()Rs,t;k [1]()
bkj
,
j E F (Xt ) Xs = =
2
j
T
s,t [1]()
k=1
Remark 4.9
T
s,t [f ]()
and
R
s,t;k [f ]()
In principle, one could take dierent localizing functions for each operator, that is:
E F (Xt ) Xs = =
j E F (Xt ) Xs = =
1
T
s,t [F ]()
2
T
s,t [1]()
j
X
bkj
k=1
4
5
6
3
R
ek
s,t;k [F ]()Ts,t [1]() Ts,t [F ]()Rs,t;k [1]()
.
7
2
j
T
s,t [1]()
See next section for a discussion on localizing functions. Furthermore, what observed in Remark
k=
E((Xt ) | Xs = )
h
i
H(Xs ) (Xs )
T
Ws,t ,
s,t [f ]() = E f (Xt ) (Xs ) +
s(t s)Xs
with
f =
and
f = 1.
T
s,t [f ]() '
N
(q)
(q)
h
i
1 X
H(Xs ) (Xs )
(q)
(q)
f (Xt ) (Xs(q) ) +
W
s,t .
(q)
N q=1
s(t s)Xs
I1f ()
h
i2
H(Xs ) (Xs )
E f 2 (Xt ) (Xs ) +
Ws,t
d,
s(t s)Xs
R
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach27
Ts,t [f ]().
term
Proposition 5.1
Setting
then
and
R
R
T
s,t [f ]() =
(t) dt = 1},
L1
where
= (), R,
() =
Proof.
||
e
,
2
with
2 1/2
1
2
s(ts) Xs Ws,t
E f (Xt )
= [f ] =
.
2
E f (Xt )
Let us set
s,t =
Ws,t
.
s(t s)Xs
L1 (R)
R and let
dt, one has
(t)
Now, set
Rx
then
+ L1 .
Setting
(x)
=
= lim 1 I f ( + )
I f ()
(I1f )0 ()()
1
1
0
Z
() + (H )()s,t d.
()
= 2E f 2 (Xt ) ()
s,t
R
L1
such that
=0
(I1f )0 ( )()
for any
satisfying
the conditions above. Consider the rst term of the (last) r.h.s.: by using the standard integration
0 = ), we can write
by parts formula (recall that
()
f 2 (Xt ) () + (H )()s,t (1) d
R
+ Z
2
f 2 (Xt ) () + (H )()s,t d
()
= () f (Xt ) () + (H )()s,t
Now, since
()
0
as
and
() + (H )() 0 as +, the rst term nullies. Moreover, () +
(H )()s,t = 0 () ()s,t for almost any . Thus, we obtain
Z
f 0
2
()
f 2 (Xt ) 0 () ()s,t + () s,t + (H )()s,t
(I1 ) ()() = 2E
d
R
Z
2
E f 2 (Xt ) 0 () + (H )()s,t
d.
= 2 ()
in particular
RR n 4804
28
Let us denote
2
E(f 2 (Xt ) s,t
)
2
E(f (Xt ))
a2 =
and
v() = (),
so that
v 0 () = ()
and
v 00 () = 0 ().
v 00 () a2 v() + a2 H() = 0,
whose solution is simply
v() =
with
c1 , . . . , c4 R
and
a=
a2 .
c1 ea + c2 ea + 1
c3 ea + c4 ea
if
if
>0
< 0,
v(+) = 1, v() = 0
and
v 0 ()
has
v () =
1 12 ea
1 a
2 e
if
if
or equivalently
() = v () =
>0
< 0,
a a||
e
.
2
minimum,
where
(x) =
Rx
(z) dz .
Since
(I1f )0 ( )() = 0
I( + ) I( ) = E
which proves that
for any
L1 (R),
then
Remark 5.2
(recall that
corresponding
to
1 hs+2 s
[1] = x
t + 2 s(t s)
.
2 s(t s)
[1] =
1/2
1
E Xs2 (Ws,t )2
,
s(t s)
Xs ,
it holds
2
E Xs2 (Ws,t )2 = x2 e2hs E e2Ws (t s)Ws s(Wt Ws ) + s(t s)
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach29
2
= x2 e2hs E E e2Ws (t s)Ws s y + s(t s)
y=Wt Ws
E(e2Ws ) = e2
Now, since
and
one has
= e2
By inserting
2
E e2Ws (t s)Ws s y + s(t s)
s(t s)2 (1 + 4 2 s) + s2 ((t s) y)2 4s(t s)((t s) y) .
so that
1 hs+2 s
[1] = x
that
t + 2 s(t s)
.
2 s(t s)
The above optimization criterium has been introduced by Kohatsu-Higa and Petterson [16].
In
principle, one could consider a measure more general than the Lebesgue one, namely to replace d
f
with (d) in the expression for I1 (), but in such a case it is not possible to write down explicitly
the optimal localizing function.
Such an approach can be generalized to the multidimensional case, where the functional
I1f
has to
be obviously replaced by
Z
Idf () =
d h
i2
Y
ei
(H i )(X
ei )
s
i
esi
E f 2 (Xt )
ei ) +
Ws,t
i (X
de
.
ei
ii s(t s)X
Rd
s
i=1
Proposition 5.3
i},
Setting
Ld = { : Rd [0, +) ; (x) =
then
inf I f ()
Ld d
Qd
i=1
j 2
j
Ws,t
i2 Y h
esj
jj s(t s)X
i : i6=j
Y h 2
E f 2 (Xt )
i +
i : i6=j
where
i L1 ,
for any
= Idf ( )
Qd
where () =
j=1 j (j ), = (1 , . . . , d ) R , with j (j )
i (xi ),
i 2 +
i
Ws,t
j
2
esi
ii s(t s)X
2 i
W i
ej |j | , j R
and
j =
2 i
,
j = 1, . . . , d.
s,t
ei
ii s(t s)X
s
fe12 (x1 ) =
RR n 4804
d h
i2
Y
esi
(H i )(X
ei )
2
2
d
i
e
e
e
esi
de
2 . . . de
d E f (x1 , Xt , . . . , Xt )
ei ) +
Ws,t
i (X
esi
ii s(t s)X
Rd1
i=2
30
then
Z
Idf () =
h
i2
e1
(H 1 )(X
e1 )
s
1
e 1 ) 1 (X
e1
de
1 E fe12 (X
e1 ) +
Ws,t
= I1f1 (1 ).
t
s
1
e
11 s(t s)Xs
R
where
Idf () b
I f1 (1 ) b
(1 , 0, . . . , 0) = 1
(1 )
b1
b1
Z
1
h
i2
Ws,t
b 1 () E fe12 (X
et1 ) 10 () + (H 1 )()
,
= 2 d
e1
11 s(t s)X
R
s
R
b 1 (y) = y b1 () d . We solve the equation
Idf () b
(1 , 0, . . . , 0) = 0,
b1
since
b1
is arbitrary,
b1
for any
b1 :
h
e 1 ) 0 () + (H 1 )()
E fe12 (X
t
1
i2
1
Ws,t
e1
11 s(t s)X
s
= 0,
for any
This gives rise to an ordinary dierential equation which has been studied in the proof of Proposition 5.1: the solution is then
1 () =
1 1 ||
e
,
2
R,
with
h
i2
1
Ws,t
et1 )
E fe12 (X
1
e
s(ts)X
11 s
1 2 = 1 (f1 )2 =
2
e
et1 )
E f1 (X
h
i2 Q
h
i2
1
R
e i e
Ws,t
(Hi )(X
d
2
i
s i )
ei
de
.
.
.
de
E
f
(X
)
(
X
e
)
+
W
2
d
t
i
i
d1
s
s,t
1
i
i=2
e
e
R
11 s(ts)Xs
ii s(ts)Xs
=
h
i2
R
Q
e i e
(H
)(
X
)
d
i
i
esi
2 . . . de
d E f 2 (Xt ) i=2 i (X
ei ) + s(ts)sXe i i Ws,t
Rd1 de
ii
Now, since everything is symmetric, we obtain similar equations for the remaining coordinates
2, . . . , d.
Now, if we want all these equations to hold simultaneously, we obtain the following
system:
j 2
j
2
ej || ,
R,
h
i2 Q
h
i2
j
R
e i i )
Ws,t
(H
ei
i
i )(Xs e
de
j E f 2 (Xt ) s(ts)
(
X
e
)
+
W
i
j
s
s,t
i
i
:
i6
=
j
i
e
e
Rd1
Xs
ii s(ts)Xs
jj
=
h
i2
R
)(X
Q
e i e
(H
i )
s
i
i
2
i
e
de
E
f
(X
)
(
X
e
)
+
W
j
t
i
s
s,t
i
i : i6=j
ei
Rd1
s(ts)X
j () =
ii
ej .
de
j
means that one has to integrate with respect to all the variables except
It remains now to give a better representation of the above integrals. They can be rewritten
s,t;i =
i
Ws,t
esi
ii s(t s)X
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach31
Z
Rd1
i2
Y h
2
esi
esi
de
j E f 2 (Xt ) s,t;j
ei ) + (H i )(X
ei )s,t;i
i (X
i : i6=j
Z
2
2
= E f (Xt ) s,t;j
Rd1
with
that
Y h
de
j
ei
ei
ei ) + (H i )(X
ei )s,t;i
i (X
s
s
i2
,
i : i6=j
s,t;j = s,t;j or s,t;j = 1. Let us consider the integral inside the expectation: by recalling
i : i6=j
Z
Rd1
i2
Y h
2
ei
ei
de
j E f 2 (Xt ) s,t;j
ei ) + (H i )(X
ei )s,t;i
i (X
s
s
i : i6=j
Y
2
= E f 2 (Xt ) s,t;j
i : i6=j
1 2
2
[
+
]
i
s,t;i
4i
Remark 5.4
` = 0 (see Remark 4.6 for details), for f = 1 the corresponding optimal values
j are given by (recall that x1 , . . . , xd are the starting underlying asset prices)
s
2 s(t s)
t + jj
2
hj s+jj s
j = 1, . . . , d.
j [1] = x1
e
j
2 s(t s) ,
jj
If
the parameters
of
of minimizing the variance coming out from the expectation giving the operator Rs,t [f ]() and
R
s,t;k [f ]() (see (7) and (28)). To this purpose, let us recall that
h
i
k
k
ek
ek
es
R
=
E
f
(X
)
(
X
e
)
+
(H
)(
X
e
)
((
X
e
)
)
,
t
k
s,t;k
k
s,t;k
s,t;k
k
s
s
s,t;k
where
s,t;k =
k
Ws,t
ek
kk s(t s)X
s
and
es
s,t;k ((X
e)k ) =
s,t;k =
d
Y
i=1,i6=k
RR n 4804
(W k )2
t
s,t
k
+ Ws,t
e k )2 kk s(t s)
kk
kk s(t s)(X
s
1
ei
ei
H(X
ei ) i (X
ei )
s
s
i
ei
e
)
+
W
i (X
i
s
s,t
ei
ii s(t s)X
s
(30)
32
Rd , the notation k means the vector on Rd1 having the same coordinates of
th one. In the one dimensional case, one simply has to drop the subscript k , take
except for the k
s
s
where, for
Rd
Proposition 5.5
i},
Setting
Ld = { : Rd [0, +) ; (x) =
then
Qd
i=1
i (xi ),
where
i L1 ,
for any
Ld
Qd
where ;k () =
j=1 j;k (j ),
= (1 , . . . , d ) Rd ,
= j = 1),
with
j;k
(j ) =
j;k
2
ej;k |j | , j R
and
one has:
1/2
E f 2 (Xt ) 2s,t
= [f ] =
E f 2 (Xt ) 2s,t
where
s,t
and
s,t
are as in (30);
k = 1, . . . , d,
the
j = 1, . . . , d:
j;k 's
h
i
Q
E f 2 (Xt ) 2s,t;k di=1,i6=k (i;k )2 + 2s,t;j
h
i and for j 6= k :
(k;k )2 =
Qd
E f 2 (Xt ) 2s,t;i i=1,i6=k (i;k )2 + 2s,t;i
h
i
Q
E f 2 (Xt ) (k;k )2 2s,t;k + 2s,t;k 2s,t;j di=1,i6=j,k (i;k )2 + 2s,t;i
Q
h
i
(j;k )2 =
d
)2 + 2
E f 2 (Xt ) (k;k )2 2s,t;k + 2s,t;k
(
s,t;i
i;k
i=1,i6=j,k
where
s,t;k
and
s,t;k
In the proof, the same technique as in the proof Proposition 5.1 and 5.3 is used. Since it is also
quite tedious, it is postponed to the Appendix.
Anyway, for practical purposes, numerical evidences show that the choice
= 1/ t s
works
good enough (thus avoiding to weight the algorithm with the computation of further expectations).
Finally, let us conclude with a short consideration. For simplicity, let us consider the one dimensional case. The main problem is a good estimate of E((Xt ) | Xs = ), which can be written as
the ratio between Ts,t []() and Ts,t [1]() but also in the following way:
s,t
,
E((Xt ) | Xs = ) = E (Xt )
E(s,t
)
s,t
= (Xs ) +
H(Xs ) (Xs )
Ws,t
s(t s)Xs
(one could also complicate things by considering two dierent localizing functions in the above
ratio...). So, another reasonable way to proceed might take into account the variance coming out
from the weight s,t . But since it is written in terms of a ratio, at this stage it does not seem
reasonably feasible to obtain results giving the associated optimal localizing function
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach33
6.1
(0, x)
of
and maturity
T,
P (0, x)
evolves following the Black-Scholes model, that is as in (14). It has been briey described in the
Introduction, let us now go into the details.
0 = t0 < t1 < . . . < tn = T be a discretization of the time interval [0, T ], with step size equal to
= T /n. By using Theorem 2.1 and Proposition 2.2, the price P (0, x) is approximated by means
0 (x), where Pk (Xk ), as k = 0, 1, . . . , n, is iteratively dened as:
of P
Let
Pn (Xn ) = (Xn
n ) (XT )
o
k = n 1, . . . , 1, 0 :
and the delta
(0, x)
()
er E P2 (X2 ) X =
=X
0 (x) = Ex (X ) .
) =
(X
setting
then
)|=X
(31)
=X
E(P(k+1) (X(k+1) ) | Xk )
if
P (X ) < (X )
if
P (X ) > (X )
(32)
E(P2 (X2 ) | X =
will be computed through the formulas given in the previous section, by means of suitable
Remark 6.1
simulated paths.
k
X
tk = k.
can be exactly
of Xk , and thus we write directly Xk . Furthermore, it is worth remarking that the algorithm
allows to use the same sample in order to simulate all the involved conditional expectations, as it
will follows from the next description.
As
k = n, n 1, . . . , 0
we need
1
i
Xk
= xi e(ri 2
In order to have
Xk ,
we simply need
Pi
j=1
Wk .
2
ij
)k+
Pi
j=1
j
ij Wk
i = 1, . . . , d.
simulation we consider the following backward approach, which uses the Brownian bridge law.
Indeed, at time
Wn =
which gives
T = n,
we can simulate
n Un ,
with
Xn .
Wn
Un = (Un1 , . . . , Und ),
RR n 4804
W(n1)
given the
independent,
34
Ws given
s(t s)/t I .
of
that
Wt = y
for
0<s<t
Thus,
W(n1)
n1
Wn +
=
n
1
d
i
Un1 = (Un1
, . . . , Un1
), Un1
N(0, 1), i = 1, . . . , d, are all
proceed similarly for the simulation of Wk , as k = n 2, . . . , 1:
r
k
k
W(k+1) +
Uk ,
Wk =
k+1
k+1
with
and variance
n1
Un1
n
with
can
s/t y
independent.
independent. Obviously, we
U = {Uki,q ; k = 1, . . . , n
(time),
i = 1, . . . , d
(dimension),
q = 1, . . . , N
(sample)}
(33)
1, . . . , N ,
for
for
i,q
(Wk
)i=1,...,d; k=1,...,n , q = 1, . . . , N :
k=n:
i,q
Wn
=
n Uni,q ,
i,q
Wk
k = n 1, . . . , 1 :
i = 1, . . . , d,
k
W i,q
=
+
k + 1 (k+1)
k
Uki,q ,
k+1
i,q
(Xk
)i=1,...,d; k=0,...,n , q = 1, . . . , N :
k = n, n 1, . . . , 1
set for
i,q
= xi e(ri 2
Xk
and
q=
and
1, . . . , N ,
set
X0i,q = xi , i = 1, . . . , d.
time=1
time=0
Figure 1
Pi
j=1
2
ij
)k+
Pi
j=1
i,q
ij Wk
i = 1, . . . , d
q =
(35)
X.
Xq
........ .
.
......
.......
................1
....
..................
...............................................
....... ................................ .
......
.
..
........................................................................................................
.........
............
.........
..........................
....
............
.... ....................................................................................................................
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
...
........ ......
.....
........
....
................
....
....................................
...........................................................................
.............
....
..
......... .................................................. ................................
......
... ....... ......
............
............. ........... ...........................................................
.......
.......... .....
....
.
.
.................. ................... .................................................
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
............................. ...
........
................................................
..
...............
....
.....
..................... ........ ........... ..........................
...................................
................................. ...................................................... ..................
.....
..........................
..........
.........
.............................................................................................................................
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
. ..
.........................................
.... ..........
..................... ................................. ........................
. .
...................... ......... ....... .......................................................................................................
. .. .
...
..
........................................................................................................................................
.............
................. ...... ..... ............................
..
.....
......... ... .........
...............................................................................................................................................................
...................................................
............................
3. Computation of
(34)
i = 1, . . . , d.
10
on
[0, 1].
introduce
which could vary according to the time interval of interest [k, (k + 1)], that is
i
one has something like `k , for k = 1, . . . , n 1 (time) and i = 1, . . . , d (dimension). As a rst
i
1
d
stage, the `k 's can be chosen arbitrarily. But since one could also choose `k = (`k , . . . , `k )
th
depending on the position of X at time k, the weight `k could also depend on the q
sample,
the drift
`,
L = {`i,q
k ; k = 1, . . . , n 1
(time),
i = 1, . . . , d
(dimension),
q = 1, . . . , N
(sample)}.
(36)
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach35
Once
xed sample
q = 1, . . . , N ,
set for
k = n, n 1, . . . , 1
e i,q = xi e(ri 12
X
k
and
Pi
j=1
i,q
2
ij
)k+`i,q
k k+ii Wk
for any
i = 1, . . . , d,
(37)
e i,q = xi , i = 1, . . . , d.
X
0
As a consequence,
e 1,q , . . . ,
X
k
ii.
e i,q
X
k
`ik
i,q
Xk
.
dened as
Wki,q := Wti,q
= (tk+1 tk )Wti,q
tk (Wti,q
Wti,q
) + tk (tk+1 tk )ii .
k tk+1
k
k+1
k
So, for any xed sample
q = 1, . . . , N ,
set for
k = n 1, . . . , 1
i,q
i,q
i,q
k(W(k+1)
Wk
) + k2 ii ,
Wki,q = Wk
1, . . . , d.
Once we have all the previous ingredients, we can proceed to the computation of
(38)
Pk (Xk )
given
q
Pk [F ](Xk
) = E F (X(k+1) ) Xk =
where
q
1,q
d,q
Xk
= (Xk
, . . . , Xk
) = the q th
q
=Xk
which has to be considered here as a datum. Notice that for each xed (time) k , we have a random
q
1,q
d,q
q
d
space grid Xk = (Xk , . . . , Xk ) R and we compute Pk [F ](Xk ) for each point of the grid.
q
By Theorem 4.2, Pk [F ](Xk ) is given by
d
Y
ei
H(X
e
)
i
i
k
E F (X(k+1) )
Wk
(k+1)
ei
X
k
i=1
q
Pk [F ](Xk ) =
d
Y
ei
H(X
e
)
i
i
k
E
Wk (k+1)
i
e
X
k
i=1
eq
e =X
k
where, for
R, H() = 1
if
and
h() = 0
otherwise.
expectations are computed and thus replaced by the associated empirical means, that is we set in
practice
N
X
q
Pk [F ](Xk
)=
q0 =1
q0
F (X(k+1)
)
i=1
d
N Y
X
e i,q X
e i,q )
0
H(X
k
k
Wki,q
0
i,q
e
X
0
q =1 i=1
RR n 4804
d
Y
e i,q X
e i,q )
0
H(X
k
k
Wki,q
0
i,q
e
X
(39)
36
Here,
e i,q
X
k
and
Wki,q
are computed through (37) and (38) respectively. Finally, we can set up
for
k = n 1, . . . , 1, 0
Obviously,
q
by u0 (X0 )
then
q
q
un (Xn
) = (Xn
), q = 1, . . . , N
q
q
q
uk (Xk
) = max (Xk
), Pk [uk+1 ](Xk
) , q = 1, . . . , N.
uk gives the Monte Carlo estimate for Pk in (31) and nally the price P0 is approximated
u0 (x) = max((x), P0 [u1 ](Xq )), where in practice we set
P0 [u1 ](x) =
N
1 X
u1 (Xq ).
N q=1
This procedure gives the price. Concerning the delta, everything starts at the nal steps, that is
when time
0 (x) = (
0;1 (x), . . . ,
0;d (x))
N
1 X
v0;j (x) =
v1;j (Xq )
N q=1
v1;j (Xq ) = j ()|=Xq if u (Xq ) < (Xq ) and v1;j (Xq ) = er j E(u2 (X2 ) | X =
)|=Xq if u (Xq ) > (Xq ) (recall that u2 is the estimate for P2 ). The gradient () should
where
obviously be considered as given in input. Concerning the gradient of the conditional expectation,
by using Theorem 4.2 one has to evaluate
Hj [u2 ](Xq ) j E u2 (X2 ) X =
=Xq
j
X
bmj
m=1
where
Ts,t
and
em,q
X
Xj,q
Rs,t;m
Rs,t;m [u2 ](Xq )Ts,t [1](Xq ) Ts,t [u2 ](Xq )Rs,t;m [1](Xq )
Ts,t [1](Xq )2
(40)
by means of the associated empirical mean: by taking into account (20) and (21), we write, for
q0
q0
q0
f = u2 or f = 1 (more precisely, by replacing f (X2
) = u2 (X2
) or f (X2
) = 1 in the formulas
below),
Ts,t [f ](Xq )
N
d
Y
e i,q0 X
e i,q )
0
1 X
H(X
q0
f (X2 )
W1i,q
i,q0
2
e
N 0
ii X
i=1
q =1
Rs,t;m [f ](Xq )
0
N
em,q0 X
em,q ) h (W m,q )2
1 X
H(X
t i
q0
m,q0
1
+
W
f (X2 )
0
1
em,q )2 mm s(t s)
N 0
mm
mm s(t s)(X
q =1
d
Y
i=1,i6=m
ei,q0 X
ei,q )
H(X
i,q0
0 W1
i,q
e
ii s(t s)X
(41)
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach37
changes except for the choice of the localizing functions, for which we refer to the discussion in
Section 5.
Finally, let us observe that one could use a further technique allowing to reduce the variance: the
introduction of a control variable. Unfortunately, there is not a standard way to proceed in this
direction. For example, one could use as a control variable the price of the associated European
option. The idea is the following. For a xed initial time
P am (t, x)
and
same payo
P eu (t, x)
x,
let us set
and maturity
T.
We dene
b X ) Ft
P (t, Xt ) = sup E er(t) (,
Tt,T
where
Tt,T
stands for the set of all the stopping times taking values on
[t, T ]
and
is dened by
b x) = (x) P eu (t, x)
(t,
(notice the obstacle
0).
b x)
(t,
P (0, x),
in point of fact identical to the one previously described, provided that the obstacle
by the new obstacle
b x).
(t,
b
(T,
x) =
P0 (x)
and delta
0 (x)
is replaced
are computed,
the approximation of the price and delta of the American option is then given by
and
eu
am
respectively. Notice that the new obstacle has to be evaluated at each time step: in order to set up
this program, it should be possible to compute the price/delta of an European option on
This
happens for some call or put options, for which prices and deltas are known in closed form. But
one could think also to proceed by simulation for their computation, by using the formulas given
in Theorem 3.1 and Theorem 4.2.
6.2
The algorithm itself can be stated as follows. We refer here to the simplest case: we do not consider
localizing functions (but we underline in footnote where they should be) and control variables.
maturity
Choose
T,
the payo
and set
i ,
= T /n.
STEP n
RR n 4804
Produce
q
Wn
, q = 1, . . . , N .
q
Xn
, q = 1, . . . , N .
Initialization: set
q
q
un (Xn
) = (Xn
).
38
Produce
Using
Using
Do the following .
* Choose `qk , q = 1, . . . , N . For example:
q
: `k = 0;
q
q
choice 1
choice 2
`1,q
k = 0
and for
i = 2, . . . , d:
q
= Xk
`i,q
k =
and
s = k):
X j,q
1
ln k .
bij hj
k
xj
j=1
i1
X
e q , q = 1, . . . , N .
X
k
q
Using (38), compute Wk , q = 1, . . . , N .
q
), q = 1, . . . , N .
Using (39)2 , compute Pk [uk+1 ](Xk
Compute for any q = 1, . . . , N ,
q
q
q
uk (Xk
) = max (Xk
), er Pk [uk+1 ](Xk
)
If
k=1
*
*
4
Using (41) and (40), compute
Compute for any
Hj [u2 ](Xq ), j = 1, . . . , d, q = 1, . . . , N .
j = 1, . . . , d and q = 1, . . . , N ,
v1;j (Xq ) = j (Xq ) 1{u1 (Xq )<(Xq )} + er Hj [u2 ](Xq ) 1{u1 (Xq )>(Xq )} .
Remark.
For
q
q
q
q eq
q
q = 1, . . . , N , Uk+1
, W
(k+1) , X(k+1) , `k , Xk , Wk
and
q
uk+1 (Xk
),
will not
be employed anymore.
STEP 0:
Compute
P0 [u1 ](x) =
u0 (x) = max (x), er P0 [u1 ](x) ,
and set
N
1 X
u1 (Xq ).
N q=1
P (0, x).
j = 1, . . . , d
v0;j (x) =
j th
N
1 X
v1;j (Xq ),
N q=1
(0, x).
d = 1).
i)
of
associated probability distribution functions and after use the localized version of (41), coming from part
Theorem 4.8 (or Theorem 3.3 if
d = 1).
ii)
of
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach39
6.3
Numerical results
In this section, the Malliavin-Monte Carlo method is numerically illustrated on three test problems.
We rst study a one-dimensional and two dimensional American put problem in order to assess the
numerical behavior of the algorithm. We then investigate the methodology by pricing and hedging
American put options on the geometric mean of ve stocks.
(x) = (K x)+ ,
exercise price
r = ln(1.1)
K = 100,
volatility
= 0.2,
= 0.
maturity
T = 1
year,
reference price and delta, the one issued of the Binomal Black Scholes Richardson extrapolation
tree-method [9] (BBSR-P and BBSR- in the next tables) with
for problems with
iterations.
geometric mean of
this payo, the simulation results are benchmarked with the one-dimensional American BBSR
tree-method [9] with
1000 step.
periods
in Table 3.
Tables of results.
and
denotes
Europen option has been used as a control variable. Moreover, also the localization has been taken
into account, as a Laplace-type probability density function (see Section 5). In order to relax the
computational executing time, we have always used the parameter of the localizing function equal
to
being
= T /n,
where
maturity and
both
`=0
and
(see Remark 4.6 for details). Actually, the results are not inuenced by the two possibilities: we
obtain really similar outcomes.
For the sake of comparison with other existing Monte Carlo methods, in the case of the standard
American put (Table 1) also the price and delta provided by the Barraquand and Martineau [5] and
Longsta-Schwartz [18] algorithms
5 It is worth noticing that the Barraquand and Martineau algorithm takes into account a space-grid, while
the Longsta and Schwartz algorithm uses least-squares approximation techniques: we have considered the ones
implemented in the software
RR n 4804
40
Nmc
BBSR-P
BM-P
LS-P
5.611
4.914
-0.378
5.147
4.714
-0.387
5.038
4.710
-0.384
BM-
LS-
-0.316
-0.284
-0.325
-0.232
-0.287
-0.258
-0.387
-0.279
-0.254
4.816
-0.388
-0.270
0.265
5.469
5.025
-0.398
-0.284
-0.237
5.198
4.903
-0.385
-0.261
-0.251
5.086
4.765
-0.385
-0.281
-0.256
4.873
5.076
4.843
-0.392
-0.272
-0.262
20000
4.875
5.057
4.868
-0.387
-0.278
-0.269
500
4.936
5.347
5.165
-0.384
-0.261
-0.258
1000
4.951
5.904
5.118
-0.384
-0.270
-0.3018
500
4.807
10 time
1000
4.795
periods
5000
4.804
10000
4.818
4.973
4.747
20000
4.823
4.928
500
4.896
20 time
1000
4.896
periods
5000
4.864
10000
50 time
periods
4.918
4.918
4.918
BBSR-
-0.387
-0.387
5000
4.897
5.184
4.880
-0.386
-0.283
-0.263
10000
4.904
5.181
4.859
-0.392
-0.387
-0.273
-0.263
20000
4.910
5.149
4.907
-0.389
-0.281
-0.278
Nmc
VZ-P
-0.288
-0.296
-0.295
-0.294
-0.295
-0.297
500
10.298
10 time
1000
10.283
periods
5000
10.271
10000
10.277
-0.297
-0.297
20000
10.263
-0.296
-0.295
500
10.388
-0.302
-0.294
20 time
1000
10.371
-0.293
-0.292
periods
5000
10.350
-0.295
-0.296
10000
10.349
-0.296
-0.297
20000
10.327
-0.296
-0.297
500
10.511
-0.306
-0.287
50 time
1000
10.443
-0.288
-0.289
periods
5000
10.416
-0.300
-0.299
10000
10.424
-0.297
-0.299
20000
10.403
-0.298
-0.297
10.306
10.306
10.306
VZ-
-0.295
-0.295
-0.295
Table 2: American put on the minimum of 2 assets (here, VZ- denotes the common value of VZ-1 and
VZ-2 ).
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach41
Nmc
BBSR-P
-0.0752
-0.0741
-0.0824
-0.0767
-0.0708
-0.0743
-0.0739
-0.0744
-0.0687
-0.0784
-0.0738
-0.0794
-0.0750
-0.0727
0.0761
-0.0759
-0.0780
500
1.525
5 time
1000
1.535
periods
5000
1.528
10000
1.537
-0.0775
-0.0765
-0.0786
20000
1.541
-0.0777
-0.0777
-0.0797
0.0779
-0.0783
500
1.716
-0.0781
-0.0766
-0.0868
-0.0934
-0.0708
10 time
1000
1.684
-0.0763
-0.0756
-0.0839
-0.0786
-0.0823
periods
5000
1.740
-0.0878
-0.0864
-0.0870
-0.0825
-0.0863
10000
1.737
-0.0864
-0.0842
-0.0872
-0.0865
-0.0878
20000
1.727
-0.0838
0.0871
-0.0867
-0.084
-0.0842
500
1.800
-0.0777
-0.0754
-0.0681
-0.0827
-0.0643
20 time
1000
1.846
-0.0851
-0.0872
-0.0753
-0.0720
-0.0768
periods
5000
1.907
-0.0864
-0.0833
0.0919
-0.0933
-0.0983
10000
1.880
0.0918
-0.0882
-0.0890
-0.0829
0.0849
20000
1.879
-0.0847
-0.0913
-0.0911
-0.0907
-0.0872
1.583
1.583
1.583
BBSR-
-0.0755
-0.0755
-0.0755
Table 3: American put on the geometric mean of 5 assets (here, BBSR- denotes the common value of
BBSR-1 , . . . ,BBSR-5 ).
RR n 4804
42
It is worth noticing that this holds for a not high number of both time periods
(10) and Monte Carlo iterations (500, 1000). Moreover, from Table 3 one may deduce that, as
the dimension increases, the convergence (in terms of the number of iterations) is slower when the
number of time periods is high.
which at the moment there are no results. However, Bouchard and Touzi in [8] (Theorem 6.3) have
proved that, in a simulation scheme very similar to the one developed here (in particular, using
localizing functions which generalize the one given in Theorem 3.4 in the one dimensional case),
p
the maximum of all the L distances between the true conditional expectations and the associated
d/(4p)
regression estimators is of order
N 1/(2p) as 0 (recall that is the discretization
time-step,
a dependence on the dimension as well). In particular, the maximum of the variances is of order
d/4 N 1/2 as 0. This would suggest that as the dimension increases, one should increase
also the number of Monte Carlo iterations in order to achieve good results.
We also tested what happens if one does not consider the localization function and/or the control
variable. The results are shown in Table 4, which refers to the simplest case of the one dimensional
American put previously introduced.
Nmc
P :
with loc.
without loc.
without loc.
500
5.056
4.615
1000
4.873
4.725
4.272
5000
4.717
10.417
124.148
10000
4.773
14.330
10437.562
20000
4.829
10.224
219.259
10
First of all, it is worth to point out that the algorithm numerically does not work if the localization
is not taken into account.
without any control variables, gives unstable but rather reasonable prices.
tables 1, 2, 3 and 4 allow to deduce that the introduction of the control variable together with the
localization then brings to more stable results, with quite satisfactory precision both for prices and
deltas also with a small number of Monte Carlo iterations (500, 1000).
Let us analyze the CPU time arising from the running of the algorithm. Our computations have
been performed in double precision on a PC Pentium IV
1.8
GHz with
256
Mb of RAM. Figure 2
shows the computational time spent for the pricing of the one dimensional American put option
as in the above example. As expected, this Monte Carlo procedure has a rather high CPU time
cost, which empirically comes out to be quadratic with respect to the number of simulations.
Figure 2 allows to conclude that this method is interesting for further developments but is uncompetitive with respect to other Monte Carlo methods, such as Barraquand-Martineau [5] and
Longsta-Schwartz [18], in terms of computing time.
procedures (which do not provide an ad hoc way for the Greeks: they are computed simply and
poorly by nite dierences), this method allows to eciently obtain the delta.
Therefore, as a
further research development, one could think to combine this procedure with some other one in
order to achieve better results, mainly cheaper in terms of CPU time costs, as it has been suggested
by H. Regnier in private communications.
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach43
3200
2800
2400
2000
1600
1200
800
400
CPU (seconds)
..
......
......
......
......
......
.
.
.
.
.
.
......
......
......
20 time
......
......
.
.
.
.
.
..
periods
......
......
.
.
.
.
.
..
...
......
..........
......
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
......
...........
......
..........
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
....
......
..........
10 time
......
..........
......
..........
.......
...........
periods
.............
..........
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.......
......
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
.
..............
...................
.............................................................
.
.
.
.
.
.
.
.
.
.
.
.
.
.
...................................
MC iterations
.......................................
1000 2000 3000 4000 5000 6000 7000 8000 9000 10000
Figure 2
Appendix
Proof of Proposition 5.5.
We give here only a sketch of the proof, since it is quite similar to the
(x) = (t) dt, one has (as in the proof of Proposition 5.1)
= 2E
(J1f )0 ()()
= 2
ZR
R
then
+ L1 .
Setting
f 2 (Xt ) ()
()s,t + (H )()s,t d.
s,t ()s,t
()
E f 2 (Xt ) 0 ()2s,t + (H )()2s,t d.
By setting
2 =
v() = (), one has to solve the ordinary dierential equation v 00 () a2 v() + a2 H() = 0.
||
This gives () = v() = e
/2. Now, it is immediate to see that gives the minimum,
and
d > 1.
in (18).
First, we compute the derivative of
fe12 (x1 ) =
Jdf ;1 ()
in the direction
(1 , 0, . . . , 0).
then
Jdf ;1 () =
d h
i2
Y
ei
(H i )(X
ei )
s
i
et2 , . . . , X
etd)
esi
de
1 E fe2 (x1 , X
ei ) +
Ws,t
i (X
ei
ii s(t s)X
Rd1
s
i=2
Z
R
h
i2
e
et1 ) 1 (X
es1
es1
de
1 E fe12 (X
e1 )s,t;1 + (H 1 )(X
e1 )s,t;1
= J1f1 (1 ).
1 () =
RR n 4804
Setting
1 1 ||
e
,
2
R,
(1 , 0, . . . , 0),
44
with
1 2
de
1 E f 2 (Xt ) 2s,t;1
=R
de
E
f 2 (Xt ) 2s,t;1
d1
1
R
R
Rd1
e 1 ) 2
E fe12 (X
t
s,t;1
=
et1 ) 2
E fe12 (X
s,t;1
Qd h
e i ei ) +
i=2 i (Xs
Qd h
e i ei ) +
i=2 i (Xs
fe22 (x2 ) =
Z
Rd1
e i e
(Hi )(X
s i )
ei
ii s(ts)X
s
i
Ws,t
e i e
(Hi )(X
s i )
ei
ii s(ts)X
s
i
Ws,t
(0, 2 , 0, . . . , 0).
i2
i2
Setting
h
i2
et1 , x2 , . . . , X
etd ) 1 (X
es1
es1
de
2 E fe2 (X
e1 )s,t;1 + (H 1 )(X
e1 )s,t;1
d h
i2
Y
ei
(H i )(X
ei )
s
i
ei
e
)
+
W
i (X
i
s
s,t
ei
ii s(t s)X
s
i=3
then
Jdf ;1 ()
where
=
R
I1f ()
h
e 2 ) 2 (X
e2
e2
de
2 E fe22 (X
e2 ) + (H 2 )(X
e2 )
t
s
s
2
Ws,t
es2
22 s(t s)X
i2
= I1f2 (2 ),
is the one handled in Proposition 5.1. By using similar arguments, one obtains
2 () =
2 2 ||
e
,
2
R,
with
h
i2
2
Ws,t
e 2)
E fe22 (X
t
e2
s(ts)X
22 s
2 2 =
2
2
e )
E fe2 (X
t
h
i2 Q
h
i2
2
R
e i e
Ws,t
d
(Hi )(X
i
s i )
e1
ei
de
2 E f 2 (Xt ) (X
e1 )2 s(ts)
(
X
e
)
+
W
i
i
s
s
s,t
2
i
i=3
e
e
Rd1
Xs
ii s(ts)Xs
22
=
i2
R
Qd h
e i e
(Hi )(X
i
s i )
2
1
2
i
e
e
de
E
f
(X
)(
X
e
)
(
X
e
)
+
W
2
t
1
i
i
d1
s
s
s,t
i=3
ei
R
s(ts)X
ii
where
es1
es1
es1
es1
(X
e1 ) 1 (X
e1 ) = 1 (X
e1 )s,t;1 + (H 1 )(X
e1 )s,t;1
Now, for the remaining coordinates one can apply similar arguments. Thus, by summarizing, one
Qd
j |j | ,
has that () =
j R and
j=1 j (j ), = (1 , . . . , d ) R , with j (j ) = 2 e
i2
Qd h
(H
i
i )(i )
d1 E f 2 (Xt ) 2s,t;1 i=2 i (i ) + s(ts)
W
s,t
ei
X
ii
s
= R
h
i2 and for j = 2, . . . , d :
)( )
Q
(H
d
i
( ) +
i
i
2 (X ) 2
d
E
f
W
1
t
i
d1
s,t
s,t;1
i
i
i=2
e
R
ii s(ts)Xs
h W j
i2 Q
h
i2
R
d
(H
s,t
2
i
i )(i )
i=2,i6=j i (i ) + ii s(ts)X
esj
e i Ws,t
Rd1 dj E f (Xt ) (1 ) jj s(ts)X
s
=
h
i2
R
)( )
Q
(H
d
i
i
i
dj E f 2 (Xt ) (1 )2 i=2,i6=j i (i ) + s(ts)
W
s,t
i
e
Rd1
X
R
1 2
j 2
Rd1
ii
INRIA
Pricing and hedging American options by Monte Carlo methods using a Malliavin calculus approach45
interesting representation for the 's, it is easy to show that (see also
where
5.3)
Z
R
so that the
1 2
j 2
's
i (i ) +
i
h
2
i2
Ws,t
(H i )(i )
1
i
Ws,t
di = i 2 +
,
ei
ei
4i
ii s(t s)X
ii s(t s)X
s
s
Z
1 2 2
2
(1 )2 d1 =
,
1
s,t;1
s,t;1
41
R
h
i2 i
i
Qd h
Ws,t
E f 2 (Xt ) 2s,t;1 i=2 i 2 + s(ts)
ei
X
ii
s
=
h W i
h
i2 i and for j = 2, . . . , d :
Q
d
2
s,t
2
2
E f (Xt ) s,t;1 i=2 i + s(ts)Xe i
ii
s
h
h
i2 Q
h
i2 i
j
i
Ws,t
Ws,t
d
2
2 2
2
E f (Xt ) 1 s,t;1 + s,t;1 s(ts)Xe j
i 2 + s(ts)
i
e
i=2,i6
=
j
Xs
jj
ii
s
=
.
h W i
Q
h
i2 i
d
2
2
s,t
2
2
2
E f (Xt ) 1 s,t;1 + s,t;1
i=2,i6=j i + s(ts)X
ei
ii
References
[1] V. Bally, M.E. Caballero, B. Fernandez, N. El Karoui (2002) Reected BSDE's, PDE's and
variational inequalities. Preprint INRIA.
[2] V. Bally and G. Pags (2000). A quantization algorithm for solving multi-dimensional optimal stopping problems. Prpublication du Laboratoire de Probabilits & Modles Alatoires, n. 628, Universit Paris VI.
[3] V. Bally and G. Pags (2002). Error analysis of a quantization algorithm for obstacle
problems. To appear on Stoch. Proc. and their Appl..
[4] V. Bally, G. Pags and J. Printems (2003). First-orders Schemes in the Numerical Quantization Method. Mathematical Finance,
13,1-16.
[5] J. Barraquand and D. Martineau (1995). Numerical valuation of high dimensiona multivariate American securities. Journal of Finance and Quantitative Analysis,
30, 383-405.
[6] A. Bensoussan and J.L. Lions (1984). Impulse Control and Quasivariational Inequalities.
Gauthiers-Villars.
[7] B. Bouchard, I. Ekeland and N. Touzi (2002). On the Malliavin approach to Monte Carlo approximation of conditional expectations. Prpublication du Laboratoire de Probabilits & Modles Alatoires, n. 709; available at
preprints/index.html
http://www.proba.jussieu.fr/mathdoc/
[8] B. Bouchard and N. Touzi (2002). Discrete time approximation and Monte-Carlo simulation
of Backward Stochastic Dierential Equations. Preprint.
RR n 4804
46
[9] M. Broadie and J. Detemple (1996). American option valuation: new bounds, approximations, and a comparison of existing methods securities using simulation. The Review of
Financial Studies,
9, 1221-1250.
[10] M. Broadie and P. Glassermann (1997). Pricing American-style securities using simulation.
21, 1323-1352.
avin calculus to Monte carlo methods in Finance. Finance & Stochastics, , 391-412.
[13] . Fourni, J.M. Lasry, J. Lebouchoux, P.L. Lions (2001). Applications of Malliavin calculus
Probability,
25, 702-73.
[15] N. El Karoui, S. Peng and M.C. Quenez (1994). Backward stochastic dierential equations
in nance. Mathematical Finance,
25, 702-73.
[16] A. Kohatsu-Higa and R. Pettersson (2001). Variance reduction methods for simulation of
4, 431-450.
[17] P.L. Lions and H. Regnier (2001). Calcul du prix et des sensibilits d'une option amricaine
par une mthode de Monte Carlo. Preprint.
[18] F.A. Longsta and E.S. Schwartz (2001). Valuing American options by simulations:
simple least squares approach. The Review of Financial Studies,
14, 113-148.
[19] D. Nualart (1995). The Malliavin Calculus and Related Topics. Springer Verlag, Berlin.
[20] E. Pardoux and S. Peng (1990). Adapted solution of a backward stochastic dierential
equation. Systems and Control Letters,
14, 55-61.
[21] E. Pardoux and S. Peng (1990). Backward stochastic dierential equations and quasilinear
parabolic partial dierential equations and their applications. In Lect. Notes Control Inf.
44, 1840-1851.
[23] S. Villeneuve and A. Zanette (2002). Parabolic ADI methods for pricing American options
on two stocks. Mathematics of Operations Research,
27, 121-149.
INRIA
diteur
INRIA - Domaine de Voluceau - Rocquencourt, BP 105 - 78153 Le Chesnay Cedex (France)
http://www.inria.fr
ISSN 0249-6399