You are on page 1of 14

Detailed Mathematical Workings

Applied Stochastic Calculus

The diagram above shows the ‡ucatuations in the price of apple stock over a 35 year period. This
uncertainty that is quite conspicuous is the most important feature of …nancial modelling. Because there
is so much randomness, the most successful mathematical models of …nancial assets have a probabilistic
foundation.
The evolution of …nancial assets is random and depends on time. They are examples of stochastic processes
which are random variables indexed (parameterized) with time. If the movement of an asset is discrete it
is called a random walk. A continuous movement is called a di¤usion process. We will consider the asset
price dynamics to exhibit continuous behaviour and each random path traced out is called a realization.
If we used a to denote the price of apple stock at time t; we could express this as at : This would be an
example of a stochastic process.
Hence the need for de…ning a robust set of properties for the randomness observed in an asset price
realization, which is Brownian Motion. This was named after the Scottish Botanist who in the summer of
1827, while examining grains of pollen of the plant Clarkia pulchella suspended in water under a microscope,
observed minute particles, ejected from the pollen grains, executing a continuous highly irregular …dgety
motion. Further study showed that …ner particles moved more rapidly, and that the motion was stimulated
by heat and by a decrease in the viscosity of the liquid. The …ndings were published in A Brief Account
of Microscopical Observations Made in the Months of June, July and August 1827 There is little doubt
that as well as the most well known (and famous) stochastic process, Brownian motion is the most widely
used.
The origins of quantitative …nance can be traced back to the start of the twentieth century. Louis Jean-
Baptiste Alphonse Bachelier (March 11, 1870 - April 28, 1946) is credited with being the …rst person to
derive the price of an option where the share price movement was modelled by Brownian motion, as part
of his PhD at the Sorbonne, entitled Théorie de la Spéculation (published 1900).
1
Thus, Bachelier may be considered a pioneer in the study of …nancial mathematics and one of the earliest
exponents of Brownian Motion. Five years later Einstein used Brownian motion to study di¤usions; which
described the microscopic transport of material and heat. In 1920 Norbert Wiener, a mathematician at
MIT provided a mathematical construction of Brownian motion together with numerous results about the
properties of Brownian motion - in fact he was the …rst to show that Brownian motion exists and is a
well-de…ned entity. Hence Wiener process is also used as a name for this, and denoted W; W (t) or Wt .

Construction of Brownian Motion


Brownian Motion can be constructed as a carefully scaled limit of a symmetric random walk, in the context
of a simple gambling game. Consider the coin tossing experiment

where we de…ne the random variable


+1 if H
Ri =
1 if T
and examine the statistical properties of Ri :

Firstly the mean


1 1
E [Ri ] = (+1) + ( 1) = 0
2 2
and secondly the variance

V [Ri ] = E Ri2 E2 [Ri ]


| {z }
=0
= E Ri2 =1

Suppose we now wish to keep a score of our winnings after the nth toss - we introduce a new random
variable
Xn
Wn = Ri
i=1

This allows us to keep a track of our total winnings. This represents the position of a marker that starts
o¤ at the origin (no winnings). So starting with no money means

W0 = 0

2
Now we can calculate expectations of Wn
" #
X
n X
n
E [Wn ] = E Ri = E [Ri ] = 0
i=1 i=1

E [Xn2 ] =

E Xn2 = E R12 + R22 + ::::Rn2 + 2R1 R2 + ::::::: + 2Rn 1 Rn


2 3
" n #
X 6X n Xn
7 X n Xn X
n
= E Ri2 + E 6 4 R R 7
i j5 = E R 2
i + E [Ri ] E [Rj ]
i=1 i=1 j=1 i=1 i=1 j=1
j6=i j6=i
= n 1+0 0=n

A Note on Variations
t
Consider a function ft ; where ti = i ; we can de…ne di¤erent measures of how much ft varies over time as
n
X
n
N
N
V = fti fti 1
i=1

The cases N = 1; 2 are important.

X
n
V = fti fti 1
total variation of trajectory - sum of absolute changes
i=1

X
n
2
V2 = fti fti 1
quadratic variation - sum of squared changes
i=1

Now look at the quadratic variation of the random walk. After each toss, we have won or lost $1. That is

Wn Wn 1 = 1 =) jWn Wn 1 j = 1

Hence
X
n
(Wi Wi 1 )2 = n
| {z }
i=1 =1

Let’s now extend this by introducing time dependence.


p Perform six tosses of a coin in a time
p t: So each toss
must be performed in time t=6; and a bet size of t=6 (and not $1); i.e. we win or lose t=6 depending
on the outcome.

Let’s examine the quadratic variation for this experiment

X
6
(Wi Wi 1 )2
i=1
X
6
p 2
= t=6
i=1
t
= 6 =t
6

3
p
Now speed up the game. So we perform n tosses within time t with each bet being t=n: Time for each
toss is t=n. p
Wi Wi 1 = t=n
The quadratic variation is
X
n
p 2
(Wi Wi 1 )2 = n t=n
i=1
= t

As n becomes larger and larger, time between subsequent tosses decreases and the bet sizes become smaller.
The time and bet size decrease in turn like
1
time decrease O
n
1
bet size O p
n

The diagram below shows a series of coin tossing experiments.

The scaling we have used has been chosen carefully to both keep the random walk …nite and also not
becoming zero. i.e. In the limit n ! 1; the random walk stays …nite. It has an expectation conditional
on a starting value of zero, of
" #
Xn
E [Wt ] = E lim Ri
n!1
i=1
X
n
= lim E [Ri ] = n 0
n!1
i=1
Mean of Wt = 0

4
" #
X
n
E Wt2 = E lim Ri2
n!1
i=1
X
n
p 2
= lim E Ri2 = lim n t=n
n!1 n!1
i=1
V [Wt ] = E Wt2 =t

This limiting process as dt tends to zero is called Brownian Motion and denoted Wt .

Alternative notation for Brownian motion/Wiener process is Xt or Bt :

Another construction of a Brownian motion


Here we provide an alternative construction. Consider the coin tossing experiment with payo¤

+1 if H
Z=
1 if T

Let
X
n
Xn = Zi
i=1

this is the position of a random walker starting at the origin and after each toss moves one unit up or down
with equal probability of a half. We know
X
n
E [Xn ] = E [Zi ] = 0
i=1

X
n
V [Xn ] = V [Zi ] = n
i=1

Now introduce time t; steps N: That is, perform N tosses in time t. Can we …nd a continuous time limit?
Each step takes t = t=N: If we let N ! 1; steps of size 1 would become in…nite (or may possibly
collapse to zero). We require a …nite random walk. So look for a suitable scaling of the time-step (away
from 1), keeping in mind the Central Limit Theorem. Let

Y = NZ

for some suitable N and let


fXn : n = 0; 1; :::; N g
be the path of the random walk with steps of size N:

Thus
E [XN ] = 0 8N

V [XN ] = E (XN )2
2 !2 3 " #
X
N X
N
= E4 Yi 5 = E 2
N Zi2
i=1 i=1

2 t 2
= NN = N
t

5
2 2
We need N= t O (1) : Choose N= t = 1: Therefore

E (XN )2 = V [XN ] = t:

As N ! 1; the random walk


fXt : t 2 [0; 1)g
converges to Brownian Motion.

Quadratic Variation
Consider a function ft ; which has at most a …nite number of jumps or discontinuities. De…ne the Quadratic
Variation
X1
N
2
Q [f ] = lim fti+1 fti
N !1
i=0

Take the time period [0; T ] with N partitions so dt = T =N ; ti = idt: For Brownian Motion Wt on the
interval [0; T ] we have

X1
N
2
Q [Wt ] = lim Wti+1 Wti
N !1
i=0
X1
N
p 2
= lim dt = lim N dt
N !1 N !1
i=0
= T

So the Quadratic Variation Q [Wt ] = T:

The Variation given

X1
N
V [Wt ] = lim Wti+1 Wti
N !1
i=0
X1
N
p
= lim dt
N !1
i=0
p
= lim N dt ! 1
N !1

6
Properties of a Wiener Process/Brownian motion
A stochastic process fWt : t 2 R+ g is de…ned to be Brownian motion (or a Wiener process ) if

Brownian motion starts at zero, i.e. W0 = 0 (with probability one):

Continuity - paths of Wt are continuous (no jumps) a.s. (di¤erentiable nowhere)

Brownian motion has independent Gaussian increments, with zero mean and variance equal to the
temporal extension of the increment. That is for each t > 0 and s > 0, Wt Ws is normal with mean
0 and variance jt sj ;
i.e.
Wt Ws N (0; jt sj) :
Coin tosses are Binomial, but due to a large number and the Central Limit Theorem we have a
distribution that is normal. Wt Ws (= x) has a pdf given by

1 x2
p (x) = p exp
2 jt sj 2 jt sj

More speci…cally Wt+s Wt is independent of Wt : This means if

0 t0 t1 t2 :::::

dW1 = W1 W0 is independent of dW2 = W2 W1


dW3 = W3 W2 is independent of dW4 = W4 W3
and so on

Also called standard Brownian motion if the above properties hold. More importantly is the result (in
stochastic di¤erential equations)
dW = Wt+dt Wt N (0; dt)

7
Mean Square Convergence
Consider a function F (X) : If
E (F (X) l)2 !0
then we say that F (X) = l in the mean square limit, also called mean square convergence. We present a
full derivation of the mean square limit. Starting with the quantity:

2 !2 3
X
n
E4 (W (tj ) W (tj 1 ))2 t 5
j=1

jt
where tj = = j t:
n
Hence we are saying that up to mean square convergence,
dW 2 = dt:
This is the symbolic way of writing this property of a Wiener process, as the partitions t become smaller
and smaller.
Developing the terms inside the expectation
First, we will simplify the notation in order to deal more easily with the outer (right most) squaring. Let
Y (tj ) = (W (tj ) W (tj 1 ))2 , then we can rewrite the expectation as:

2 !2 3
X
n
E4 Y (tj ) t 5
j=1

Expanding we have:
E [(Y (t1 ) + Y (t2 ) + : : : + Y (tn ) t) (Y (t1 ) + Y (t2 ) + : : : + Y (tn ) t)]
The term inside the Expectation is equal to
Y (t1 )2 + Y (t1 )Y (t2 ) + : : : + Y (t1 )Y (tn ) Y (t1 )t
+Y (t2 )2 + Y (t2 )Y (t1 ) + : : : + Y (t2 )Y (tn ) Y (t2 )t
..
.
+Y (tn )2 + Y (tn )Y (t1 ) + : : : + Y (tn )Y (tn 1 ) Y (tn )t
tY (t1 ) tY (t2 ) : : : tY (tn ) + t2
Rearranging
Y (t1 )2 + Y (t2 )2 + : : : + Y (tn )2
2Y (t1 )Y (t2 ) + 2Y (t1 )Y (t3 ) + : : : + 2Y (tn 1 )Y (tn )
2Y (t1 )t 2Y (t2 )t : : : 2Y (tn )t
+t2
We can now factorize to get
X
n X
n X X
n
2
Y (tj ) + 2 Y (ti )Y (tj ) 2t Y (tj ) + t2
j=1 i=1 j<i j=1

8
Substituting back Y (tj ) = (W (tj ) W (tj 1 ))2 and taking the expectation, we arrive at:
X
n
E[ (W (tj ) W (tj 1 ))4
j=1
Xn X
+2 (W (ti ) W (ti 1 ))2 (W (tj ) W (tj 1 ))2
i=1 j<i
Xn
2t (W (tj ) W (tj 1 ))2
j=1
2
+t ]

Computing the expectation


By linearity of the expectation operator, we can write the previous expression as:
X
n
E (W (tj ) W (tj 1 ))4
j=1
X
n X
+2 E (W (ti ) W (ti 1 ))2 (W (tj ) W (tj 1 ))2
i=1 j<i
Xn
2t E (W (tj ) W (tj 1 ))2
j=1
2
+t
t
Now, since Z(tj ) = W (tj ) W (tj 1 ) follows a Normal distribution with mean 0 and variance (= dt) ; it
2
n
follows (standard result) that its fourth moment is equal to 3 nt 2 . We will show this shortly.
Firstly we know that Z(tj ) N 0; nt ; i.e.
t
E [Z(tj )] = 0; V [Z(tj )] =
n
2
therefore we can construct its PDF. For any random variable N( ; ) its probability density is given
by !
1 1( )2
p ( ) = p exp 2 2
2
hence for Z(tj ) the PDF is
1 1 z2
p (z) = p p exp 2
t=n 2 t=n

E (W (tj ) W (tj 1 ))4 = E Z4


t2
= 3 2 for j = 1; : : : ; n
n
So
Z
4
E Z = Z 4 p (z) dz
rR Z
n z2
= Z 4 exp 1
2
dz
2t R t=n

9
now put p
z
u= p ! du = n=tdz
t=n
Our integral becomes !4
r Z r r
n t 1 2 t
u exp 2
u du
2t R n n

r
Z
1 t2
= u4 exp 1 2
2
u du
2 n2 R
r Z
t2 1
= 2
: u4 exp 1 2
2
u du
n 2 R
2
t
= 2
:E u4 :
n
So the problem reduces to …nding the fourth moment of a standard normal random variable. Here we do
not have to explicitly calculate any integral. Two ways to do this.

Either use the MGF as we did earlier and obtained the fourth moment to be three.

Or the other method is to make use of the fact that the kurtosis of the standardised normal distribution
is 3.

That is " # " #


( )4 ( 0)4
E 4
=E = 3:
14
t2
Hence E [u4 ] = 3 and we can …nally write 3 :
n2
and
t
E (W (tj ) W (tj 1 ))2 = for j = 1; : : : ; n
n

Because of the single summation, the fourth moment and the variance multiplied by t actually recur n
times. Because of the double summation, the product of variances occurs n(n2 1) times.

We can now conclude that the expectation is equal to:

t2 t2 t
+
3nn(n 1) 2tn + t2
n2 n2 n
t2 t2 t2
= 3 + t2 2t2 + t2 = 2
n n n
1
= O( )
n

So, as our partition becomes …ner and …ner and n tends to in…nity, the quadratic variation will tend to t
in the mean square limit.

10
Wiener Process Trajectory

120

100

80

60
W(t)

40

20

0
0 1 2 3 4 5 6 7 8 9
-20

-40
t

11
Taylor Series and Itô
If we were to do a naive Taylor series expansion of F (Wt ), completely disregarding the nature of Wt , and
treating dWt as a small increment in Wt , we would get
dF 1 d2 F
F (Wt + dWt ) = F (Wt ) + dWt + dWt2 ;
dWt 2 dWt2
ignoring higher-order terms. We could argue that F (Wt + dWt ) F (Wt ) was just the ‘change in’F and so
dF 1 d2 F
dF = dWt + dWt2 :
dWt 2 dWt2

This is almost correct.


Because of the way that we have de…ned Brownian motion, and have seen how the quadratic variation
behaves, it turns out that the dWt2 term isn’t really random at all. The dWt2 term becomes (as all time
steps become smaller and smaller) the same as its average value, dt. Taylor series and the ‘proper’Itô are
very similar. The only di¤erence being that the correct Itô’s lemma has a dt instead of a dWt2 .
You can, with little risk of error, use Taylor series with the ‘rule of thumb’
dWt2 = dt:
and in practice you will get the right result.
We can now answer the question, “If F = Wt2 what is dF ?”In this example
dF d2 F
= 2Wt and = 2:
dWt dWt2
Therefore Itô’s lemma tells us that
dF = dt + 2Wt dWt :
This is an example of a stochastic di¤erential equation (SDE), written more generally as
dF = A (Wt ) dt + B (Wt ) dWt :

There are two parts. The part before the plus, A (Wt ) dt; is the deterministic bit. The random component
follows B (Wt ) dWt . More importantly A (Wt ) is the drift; B (Wt ) is the di¤usion.
Now consider a slight extension. A function of a Wiener Process f = f (t; Wt ) ; so we can allow both t and
Wt to change, i.e.
t ! t + dt
Wt ! Wt + dWt :
Using Taylor as before
2
@f @f 1 @ f
f (t + dt; Wt + dWt ) = f (t; Wt ) + dt + dWt + 2 2
dWt2 + ::
@t @Wt @Wt
df = f (t + dt; Wt + dWt ) f (t; Wt )
This gives another form of Itô:
@f @2f @f
df = + 12 dt + dWt : (*)
@t @Wt2 @Wt
This is also a SDE.

12
Examples:

@f @f @2f
1. Obtain a SDE for f = teWt : We need @t
= eWt ; @Wt
= teWt = @Wt2
, then substituting in ( )

df = eWt + 21 teWt dt + teWt dWt :

We can factor out teWt and rewrite the above as


df 1 1
= t
+ 2
dt + dWt :
f

2. Consider the function of a stochastic variable f = t2 Wtn

@f @f @2f
= 2tWtn ; = nt2 Wtn 1 ; = n (n 1) t2 Wtn 2 ;
@t @Wt @Wt2

in ( ) gives
df = 2tWtn + 21 n (n 1) t2 Wtn 2
dt + nt2 Wtn 1 dWt :

Itô multiplication table:


dt dWt
dt dt2 = 0 dtdWt = 0
dWt dWt dt = 0 dWt2 = dt

A Formula for Stochastic Integration


If we take the 2D form of Itô given by ( ) ; rearrange and integrate over [0; t] ; we obtain a very nice formula
for integrating functions of the form f (t; W (t)) :
Z t Z t
@f @f @2f
@W
dW = f (t; W (t)) f (0; W (0)) @
+ 12 @W 2 d
0 0

This was derived in class during lecture 1.4 (30 January 2020).

Example: Show that


Z t Z t
W W
t+e dW = tW + e 1 W + 21 eW d :
0 0

@f
Comparing this to the stochastic integral formula above, we see that dW
t + eW =) f = tW + eW : Also
@2f @f
@W 2
= eWt ; @t
= Wt :

Substituting all these terms in to the formula and noting that f (0; W (0)) = 1 veri…es the result.

Naturally if f = f (W (t)) then the integral formula simply collapses to


Z t Z t
df 1 d2 f
dW
dW = f (W (t)) f (W (0)) 2 dW 2d
0 0

13
Example
Itô’s lemma can be used to deduce the following formula for stochastic di¤erential equations and stochastic
integrals Z t Z t
@F @F 1 @2F
dW ( ) = F (W (t) ; t) F (W (0) ; 0) + d
0 @W 0 @ 2 @W 2
for a function F (W ( ); ) where dW ( ) is an increment of a Brownian motion.

If W (0) = 0 evaluate Z t
2
sin W dW ( ):
0
@F
# = t2 sin W ! F = t2 cos W #
@W2
@ F @F
2
= t2 cos W = 2t cos W
@W @t
and substitute into the integral formula
Z t Z t
2 2 1 2
sin W dW ( ) = t cos W 2 cos W + cos W d
0 0 2

14

You might also like