Professional Documents
Culture Documents
Laura Mayoral
IAE and Barcelona GSE
Barcelona, Winter 2021
Goal
Still consistent?
Still consistent?
y1 , y2 , . . . , yt , . . . , yT ,
strict and
This means that the graphs over two equal-length time intervals
of a realisation of the time series should exhibit similar statistical
characteristics.
i) γ (0) ≥ 0
ii) |γ (h)| ≤ γ (0) for all h ∈ Z
iii) γ (−h) = γ (h) for all h ∈ Z
The autocorrelation function
γX (h)
ρX ( h ) = = Corr (Xt+h , Xt ), for all t, h ∈ Z.
γX (0)
The relation between Stationary and Strict Stationarity
Strict stationarity implies weak stationarity, provided the first
and second moments of the variables exist, but the converse of
this statement is not true in general.
That is, for sufficiently large n, Yi and Yi+n are nearly indepen-
dent. A more formal definition is provided below.
P∞
• If j =0 |γ (j )| < ∞, then then {Yt } is ergodic for second mo-
ments. (Proof: Brockwell and Davis, p. 220)
• Furthermore,
P∞ if {Yt } is a stationary Gaussian process and
j =0 γj < ∞, then the process is ergodic of all moments.
Some examples of stationary processes
Exercise:
Show that if E (εt ) = 0 and the second order moments exist then,
Xt = εt + θεt−1 ,
where {εt } is a white noise process and is stationary for any value
of θ.
Xt = εt + θ1 εt−1 · · · + θq εt−q ,
where {εt } is a white noise process and is stationary for any value
of θ1 . . . θq
Example 7 Autoregression of order 1.
The process {Xt } is called an autoregressive process of order
1 if {Xt } is defined as
Xt = φXt−1 + εt ,
Xt = φ1 Xt−1 + . . . φp Xt−p + εt ,
-1
-2
-3
-4
0 50 100 150 200 250 300
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
-0.1
0 5 10 15 20 25
AR(1) process, φ = 0.8: plot and autocorrelation function
-2
-4
-6
0 50 100 150 200 250 300
0.5
-0.5
0 5 10 15 20 25
MA(1) process, θ = 0.8: plot and autocorrelation function
-2
-4
0 50 100 150 200 250 300
0.5
0.4
0.3
0.2
0.1
-0.1
0 5 10 15 20 25
Some examples of non-stationary processes
Xt = βt + εt ,
Xt = Xt−1 + εt , t≥0
Xt = εt , t < k
Xt = µ + εt , t ≥ k
where µ 6= 0.
And some more graphs
Random walk: plot and autocorrelation funciton
40
30
20
10
-10
0 50 100 150 200 250 300
0.8
0.6
0.4
0.2
0
0 5 10 15 20 25
Trend-stationary process: plot and autocorrelation funciton
60
40
20
-20
0 50 100 150 200 250 300
0.8
0.6
0.4
0.2
0
0 5 10 15 20 25
The Lag Operator
L (L (Lxt )) = L3 xt = xt−3 .
Let {yt } be a stationary time series with E (yt ) = µ and var (yt ) < ∞.
A fundamental decomposition result is the following
Wold Representation theorem:
∞
X
yt = µ + ψj εt−j (1)
j =0
With:
ψ0 = 1
P∞ 2 <∞
(square-summability) ψ
j =0 j
mean: E (yt ) = µ
P∞
variance: var (yt ) = σ2 2
j =0 ψj
P∞ 2 <∞
Square-summability is a stationarity condition: ψ
j =0 j
P∞
Absolute-summability: j =0 ψj < ∞
∞
X
γj < ∞
j =0
P∞ P∞
It can be shown that j =0 ψj < ∞ implies j =0 γj < ∞
P∞
Thus, a stationary process with j =0 ψj < ∞ is ergodic for
second moments.
MA(q) processes
The Wold representation theorem allows us to write any sta-
tionary process as a (potentially infinite) linear combination of a
white noise process: MA(∞).
α(L) = α0 + α1 L + α2 L2 + . . . .
∞
X
= αj xt−j
j =0
Inversion
−φ + α1 = 0 =⇒ α1 = φ,
−φα1 + α2 = 0 =⇒ α2 = φ2 .
...
Therefore
∞
(1 − φL)−1 = 1 +
X
φj Lj , provided |φ| < 1.
j =1
P∞ j j
It is easy to check that 1 + j =1 φ L is the inverse of (1 − φL)
since:
k
X
(1 − φL) 1 + φj Lj = 1 − φk+1 Lk+1 → 1 as k → ∞
j =1
Example 12 Let p=2. Find the inverse of φ1 (L) = (1 − φL) −
φ2 L2 ).
1 + φ1 L + φ2 L2 = (1 − λ1 L) (1 − λ2 L)
2 −1
= (1 − λ1 L)−1 (1 − λ2 L)−1
1 + φ1 L + φ2 L
∞ ∞
λj1 Lj λj2 Lj
X X
=
j =0 j =0
AR(∞) processes
xt = ψ ( L ) εt
where ψ (L) = 1 + ψ1 L + ψ2 L2 . . . .
xt = ψ ( L ) εt ⇒
xt = φ1 xt−1 + φ2 xt−2 + · · · + εt ,
φ ( L ) xt = εt
Then,
xt = xt−1 + εt−1
yt = yt−1 + εt
The MA(∞) and AR(∞) processes are not very useful in prac-
tice: both have an infinite number of coefficients!
xt = ψ ( L ) εt
The following approximation writes the MA polynomial (with
an infinite number of terms) as a ratio of two finiter-order polyno-
mials:
φp (L)
ψ (L) ∼
θq (L)
Then,
φp (L)xt = θq (L)εt
xt is an ARMA(p, q) process.
T
X
ȳT = T −1 yt
t=1
The Law of Large Numbers and the Central Limit Theorem are
the most important results for computing the limit of this sequence.
εt is assumed to be i.i.d.
Remark 2
Alternative versions of the C.L.Ts for dependent processes ex-
ist. See Hayashi, Theorem 6.10
∞
X
LRV = γ (j )
j =−∞
∞
X
LRV = γ0 + 2 γ (j )
j =1
∞
X
= σ 2 ψ (1) = σ 2 ( ψj )2
j =0
1. How to pick q?
p
2. q must grow with T in order for the LRVq → LRV
\
3. In finite samples, L
\ RVq might have bad properties, it can even
be negative.
Kernel-based estimators
Recall that our goal is to compute the CLT for the sample
mean of t εt which is a k × 1 vector!
gt = Ψ(L)ηt , η ∼ i.i.d.(0, Σ)
Then,
T
1 X p
xt εt → 0
T t=1
T
1 X d
√ xt εt → N (0, LRV )
T t=1
With
∞
X
S = LRV = Γ0 + (Γj + Γ0j ) = Ψ(1)ΣΨ(1)0
j =1
Γ0 = E (gt gt0 ) = E (xt x0t ε2t ), Γt = E (gt gt0 ) = E (xt x0t εt εt−j )
Asymptotics with autocorrelated errors
OLS with autocorrelated errors
Bottom line:
Example:
yt = αyt−1 + εt
εt = βεt−1 + ηt
ηt
εt =
1 − βL
Which implies:
(1 − βL)yt = (1 − βL)yt−1 + ηt
Try to specify the model in such a way that the residuals look
uncorrelated (you can test for this)
0 c
(δ̂ − δ0 ) = (Sxz W Sxz )−1 Sxz
0 c
W ḡ
p
GMM is consistent (ḡ → 0 by the LNN)