Professional Documents
Culture Documents
This chapter is to introduce the most commonly used stationary linear time
series models–the autoregressive moving average (ARMA) models. These models
have assumed great importance in modeling real-world processes.
where σ 2 = var(Zt ).
1
(c) Its covariance functions are
q
∑ q
∑
cov(Xt−k , Xt ) = cov( θj Zt−k−j , θl Zt−l )
j=0 l=0
q ∑
∑ q q−k
∑
= θj θl cov(Zt−k−j , Zt−l ) = θj+k θj σ 2
j=0 l=0 j=0
( ) 2
= θk + θ1 θk+1 + · · · + θq−k θq σ for k ≤ q.
(e) Generally, the ACF for a MA(q) time series cut-off after lag q
(f) Since the mean and covariance function are both independent of time t we
conclude that MA(q) is stationary.
Xt = Zt + θ1 Zt−1 + θ2 Zt−2
2
DEFINITION 2 Let
∞
∑
Xt = ψj Zt−j
j=1
(a)
E(Xt ) = 0
(b)
∞
∑
V ar(Xt ) = ψj2 σ 2
j=0
(c)
{
1,
∑
if k = 0
∞
ρ(k) = j=0 ψj ψh+j
∑ ∞ 2 , if k ̸= 0
j=0 ψj
EXAMPLE 3 Let
∞
∑
Xt = ϕj Zt−j
j=1
∴ Xt = ϕXt−1 + Zt
3
or
Xt − ϕXt−1 = Zt
Xt is called an autoregressive process of order one (or AR(1)). For AR(1), we
further have
σ2
V ar(Xt ) =
1 − ϕ2
and
σ 2 ϕk
γ(k) =
1 − ϕ2
and
γ(k)
∴ ρ(k) = = ϕk .
γ(0)
2 Autoregressive models
DEFINITION 3 : Assume that {Zt } is white noise. Then the process {Xt } defined
by
∑p
Xt = ϕi Xt−i + Zt
i=1
for i = 1, 2, . . ..
Define the pth degree polynomial,
ϕ(z) = 1 − ϕ1 z − · · · − ϕp zp .
ϕ(B)Xt = Zt ,
∑p j.
where ϕ(B) = 1 − j=1 ϕj B It is called the AR characteristic polynomial.
4
THEOREM 1 AR(p) is stationary if all the roots of the polynomial ϕ(z) are
outside the unit circle.
(a) EXt = 0
PROOF Fact (d) may be obtained by first multiplying Xt−k on both sides of AR(p)
model, then taking expectation on both sides and eventually using (b) and the fact
that γ(k) = Cov(Xt , Xt−k ) = E(Xt Xt−k ) because EXt = 0. Fact (c) can be proved
similarly. Fact (b) can be seen from (2.1).
5
EXAMPLE 4 (AR(1) process continued): Consider an AR(1) model of the form
Xt = ϕ1 Xt−1 + Zt , Zt ∼ W N(0, σ 2 ).
SOLUTION : (i)
Part a. If |ϕ1 | < 1, then it has been shown that {Xt } is stationary (refer to
Example 3).
(ii) Referring to Example 3, we have as k Ï ∞
−1 )
|ρ(k)| = |ϕ1 |k = e−k log(|ϕ1 | Ï0
exponentially. This fact is very useful in identifying an AR(1) model. Note that
the exponential behaviour of the ACF of AR(1).
Alternatively, by the Yule-Walker equations
ρk = ϕ1 ρk−1 .
This yields
ρk = ϕ1k ρ0 = ϕ1k .
6
MTH451/MH4500 TIME SERIES ANALYSIS
Chapter 3 (Part 2): Stationarity linear time series
This chapter is to introduce the most commonly used stationary linear time
series models–the autoregressive moving average (ARMA) models. These models
have assumed great importance in modeling real-world processes.
Xt − ϕ1 Xt−1 − · · · − ϕp Xt−p
= Zt + θ1 Zt−1 + · · · + θq Zt−q , (1.1)
where Zt ∼ W N(0, σ 2 ).
ϕ(z) = 1 − ϕ1 z − · · · − ϕp zp
and
θ(z) = 1 + θ1 z + · · · + θq zq
and B is the backward shift operator defined by
ϕ(B)Xt = Zt
1
and the process is said to be an autoregressive process of order p (or AR(p)).
This model can also be written as
Xt − ϕ1 Xt−1 = Zt + θ1 Zt−1 .
Write
(1 − ϕ1 B)Xt = Zt + θ1 Zt−1 .
If |ϕ1 | < 1, we have
∑ j ∞
1
Xt = (Zt + θ1 Zt−1 ) = ϕ1 Bj (Zt + θ1 Zt−1 )
(1 − ϕ1 B)
j=0
∞
∑ j
= ϕ1 (Zt−j + θ1 Zt−1−j ).
j=0
i.e.
and
2
4
2
0
ts.sim
−2
−4
−6
Time
Series ts.sim
1.0
0.8
0.6
ACF
0.4
0.2
0.0
−0.2
0 5 10 15 20
Lag
3
Conclusion: ACF of ARMA(1,1) dies down (without cut-off).
Alternatively, as in AR(p), one may use the Yule-Walker equation to obtain
AVCFs of ARMA(1,1) and here we ignore the details.
This suggests another approach of verifying that a process is stationary by
converting it to a MA(∞).
Xt = Xt−1 + Zt .
Note that ϕ(z) = 1 − z. Its root is one which does not lie outside the unit circle. It
follows that it is not stationary.
Xt = ϕXt−1 + Zt .
Then
γ(2) = cov(Xt , Xt−2 ) = cov(ϕXt−1 + Zt , Xt−2 )
= cov(ϕ2 Xt−2 + ϕZt−1 + Zt , Xt−2 ) = ϕ2 γ(0),
because Zt and Zt−1 are both uncorrelated with Xt−2 (so called causal property).
The correlation between Xt and Xt−2 is not equal to zero because Xt is dependent
on Xt−2 through Xt−1 . Suppose we break this chain of dependence by removing
Xt−1 . That is, consider the correlation between Xt − ϕXt−1 and Xt−2 − ϕXt−1
because the correlation between Xt and Xt−2 with the linear dependence of each
on Xt−1 removed. In this way, we have broken the dependence chain between
Xt and Xt−2 . Indeed,
cov(Xt − ϕXt−1 , Xt−2 − ϕXt−1 )
= cov(Zt , Xt−2 − ϕXt−1 ) = 0.
4
2.2 Definition of PACF
The aim of introducing PACF is to detect the relation between xt and xt−h with
the effect of xt−1 , · · · , xt−h+1 removed.
xt = ϕ1 xt−1 + · · · + ϕp xt−p + at ,
ϕ1,1 = ρ(1),
ϕp,p = ϕp ,
ϕk,k = 0 for k > p.
When k > p, the linear combination minimizing the mean square linear pre-
diction error is
f1,p = ϕ1 xk−1 + · · · + ϕp xk−p
(obtaining such an expression is complicated and beyond the scope of the course).
Then PACF at lag k > p is
because Zk is uncorrelated with the past values of Xk−j with j > 0. The calculation
of ϕp,p = ϕp is also beyond the scope of the course.
5
THEOREM 2 Generally, the PACF for an AR(p) time series cut-off after lag p.
ϕh = Γh−1 gh
where the matrix Γh = [γ(i − j)]hi,j=1 and gh = [γ(1), γ(2), · · · , γ(h)]T (this formula
is for your reference only).
Summary
Identification of ARMA model
EXAMPLE 6 Suppose that {Xt } is a stationary process. Let Yt = Xt +c1 +c2 t +c3 t 2 .
Is {Yt } stationary? (1 − B)Yt ? (1 − B)2 Yt ?
Answer: (1 − B)2 Yt is stationary.
t
∑
(1 − B)Xt = Zt , or Xt = X0 + Zi ,
i=1
where Zt is white noise. The above special case is also called random walk. It
can be classified as ARIMA(0,1,0).
6
EXAMPLE 8 (ARIMA(0, 1, 1)): Consider a time series model of the form
Xt = Xt−1 + Zt + θZt−1 . (3.1)
where {Zt } is a white noise.
Model (3.1) is an ARIMA(0, 1, 1) since
▽Xt = Zt + θZt−1
defines {▽Xt } an ARMA(0, 1)=MA(1).
20
10
0
Time
Series ts.sim
1.0
0.8
0.6
ACF
0.4
0.2
0.0
0 5 10 15 20
Lag
7
1.0445, -0.1338, 0.6706, 0.3755, -0.5110, -0.2352, 0.1595, 1.6258, -1.6739, 2.4478,
-3.1019, 2.6860, -0.9905, 1.2113, -0.0929, 0.9905, 0.5213, -0.1139, -0.4062, 0.5438
2
1
0
x
−1
−2
−3
5 10 15 20
We cannot make any “good” prediction using trend and seasonality model.
Suppose we try the follow model:
xt = 0.40 − 0.82xt−1
This is a AR(1) model (we will explain how we get 0.4 and -0.82 in the subse-
quent courses).
The fitted values are (time: prediction) 2: -0.4569, 3: 0.5130, 4: -0.1491, 5:
0.0938, 6: 0.8235, 7: 0.5965, 8: 0.2716, 9: -0.9354, 10: 1.7809, 11: -1.6121,
12: 2.9564, 13: -1.8082, 14: 1.2183, 15: -0.5942, 16: 0.47939, 17: -0.4124, 18:
-0.0262, 19: 0.4966, 20: 0.73730
which can make reasonable prediction.
◃ ACF; PACF?
8
2
1
0
x
−1
−2
−3
5 10 15 20
Series ts.sim
1.0
0.8
0.6
ACF
0.4
0.2
0.0
0 5 10 15 20
Lag
9
For stationary TS (time series), the SAC will die down very quickly or
even cuts off. Otherwise, the time series is not stationary.
Given observations of a time series, one approach to the fitting of a model
to the data is to match the sample ACF and sample PACF of the data with the
ACF and PACF of the model, respectively. Suppose that z1 , · · · , zn is a realization
(sample, observations) from a stationary time series. The sample ACF (SACF) at
lag h is
∑n−h
(zt − z̄)(zt+h − z̄)
rh = t=1∑n 2
t=1 (zt − z̄)
r1,1 = r1
∑k−1
rk − j=1 rk−1,j rk−j
rk,k = ∑k−1 , for k = 2, 3, · · ·
1 − j=1 rk−1,j rj
rk,j = rk−1,j − rk,k rk−1,k−j ,
for j = 1, 2, · · · , k − 1
∑
where z̄ = ni=1 zt /n.
The standard error of rh is
( ∑h−1 )1/2
1+2 j=1 rj2
srh = .
n
Recall that
10
How do we classify ACF as negligible ? If (with probability approximately 0.95)
|rh | ≤ 1.96srh ,
then SACFs are considered as negligible. In practice, we frequently use the more
√
stringent values 1.96/ n as the bound.
0.5
0.8
0.4
0.6
0.3
Partial ACF
0.2
0.4
ACF
0.1
0.2
0.0
0.0
−0.1
−0.2
0 5 10 15 20 5 10 15 20
Lag Lag
11
ACF
ACF
ACF
−0.4 −0.2 0.0 0.2 0.4 0.6 0.8 1.0
−0.4 −0.2 0.0 0.2 0.4 0.6 0.8 1.0
−0.2 0.0 0.2 0.4 0.6 0.8 1.0
0
0
5
5
10
10
10
Lag
15
Lag
Lag
Series 1
Series 1
Series 1
15
15
20
25
20
20
30
Partial ACF Partial ACF
12
Partial ACF
−0.4 −0.2 0.0 0.2 −0.2 −0.1 0.0 0.1 0.2 0.3 0.4
−0.5 0.0 0.5
0
Time series 4
Time series 3
Time series 2
5
5
5
10
10
10
15
Lag
Lag
Lag
Series x
Series x
Series x
15
15
20
25
20
20
30
Time series 5
Series 1 Series x
1.0
0.8
0.8
0.6
0.6
0.4
Partial ACF
ACF
0.2
0.4
0.0
0.2
−0.2
0.0
−0.4
0 5 10 15 20 25 30 0 5 10 15 20 25 30
Lag Lag
Time series 6
Series 1 Series x
1.0
0.0
0.5
−0.2
Partial ACF
−0.4
ACF
0.0
−0.6
−0.5
−0.8
0 5 10 15 20 25 30 0 5 10 15 20 25 30
Lag Lag
Time series 7
Series 1 Series x
1.0
0.8
0.8
0.6
0.6
Partial ACF
ACF
0.4
0.4
0.2
0.2
0.0
0.0
0 20 40 60 80 100 0 20 40 60 80 100
Lag Lag
13
Time series 8
Series 1 Series x
0.6
1.0
0.8
0.4
0.6
0.2
Partial ACF
0.4
ACF
0.2
0.0
0.0
−0.2
−0.2
0 20 40 60 80 100 0 20 40 60 80 100
Lag Lag
Time series 9
Series 1 Series x
0.8
1.0
0.8
0.6
0.6
Partial ACF
0.4
ACF
0.4
0.2
0.2
0.0
0.0
0 20 40 60 80 100 0 20 40 60 80 100
Lag Lag
14