Professional Documents
Culture Documents
t(t + k)
= (1 +
k
t
)
1/2
. (102)
This expression indicates that if t is large, the coecients of the autocorrelation
function will be close to one and will decay in an approximately linear form with k.
Indeed, if we assume that the process starts in the distant past such that k/t is
small, then the function (1 + k/t)
1/2
can be approximated using a rst order
Taylor expansion such as:
(t, t + k) 1
k
2t
. (103)
This equation indicates that assuming xed t, if we look at the autocorrelation
as a function of k we obtain a straight line with slope (1/2t).
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 17 / 51
The random walk
Taking logarithms in (102) we have
log (t, t + k) =
1
2
log(1 +
k
t
)
k
2t
and we see that if we express the correlations in logarithms they also decay linearly
with lag k.
This behavior of the autocorrelation function of an integrated process can be
thought of as an extreme case of an AR(1) when is close to the unit.
Indeed, if = 1 , with small , we have
k
= (1 )
k
1 k
for small .
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 18 / 51
Example 50
The gure shows a graph of the autocorrelation function of a simulation of 200
observations that follow a random walk. Here, we can clearly see the linear decay
of the ACF.
C
o
r
r
e
l
o
g
r
a
m
o
f
R
W
1
D
a
t
e
:
0
1
/
3
1
/
0
8
T
i
m
e
:
1
3
:
0
9
S
a
m
p
l
e
:
1
2
0
0
I
n
c
l
u
d
e
d
o
b
s
e
r
v
a
t
i
o
n
s
:
2
0
0
A
u
t
o
c
o
r
r
e
l
a
t
i
o
n
P
a
r
t
i
a
l
C
o
r
r
e
l
a
t
i
o
n
A
C
P
A
C
Q
-
S
t
a
t
P
r
o
b
1
0
.
9
7
2
0
.
9
7
2
1
9
1
.
7
8
0
.
0
0
0
2
0
.
9
4
2
-
0
.
0
5
2
3
7
2
.
7
6
0
.
0
0
0
3
0
.
9
1
2
-
0
.
0
0
8
5
4
3
.
3
2
0
.
0
0
0
4
0
.
8
8
1
-
0
.
0
4
3
7
0
3
.
1
8
0
.
0
0
0
5
0
.
8
5
3
0
.
0
5
2
8
5
3
.
9
3
0
.
0
0
0
6
0
.
8
2
7
0
.
0
1
8
9
9
6
.
5
2
0
.
0
0
0
7
0
.
8
0
4
0
.
0
2
0
1
1
3
1
.
8
0
.
0
0
0
8
0
.
7
7
9
-
0
.
0
3
9
1
2
5
9
.
5
0
.
0
0
0
9
0
.
7
6
0
0
.
0
8
5
1
3
8
1
.
5
0
.
0
0
0
1
0
0
.
7
4
0
-
0
.
0
2
0
1
4
9
7
.
9
0
.
0
0
0
1
1
0
.
7
1
7
-
0
.
0
5
2
1
6
0
7
.
9
0
.
0
0
0
1
2
0
.
6
9
8
0
.
0
4
2
1
7
1
2
.
7
0
.
0
0
0
1
3
0
.
6
7
5
-
0
.
0
8
0
1
8
1
1
.
0
0
.
0
0
0
1
4
0
.
6
5
9
0
.
1
3
8
1
9
0
5
.
3
0
.
0
0
0
1
5
0
.
6
4
4
-
0
.
0
1
1
1
9
9
5
.
8
0
.
0
0
0
1
6
0
.
6
2
8
-
0
.
0
2
9
2
0
8
2
.
3
0
.
0
0
0
1
7
0
.
6
1
4
0
.
0
3
5
2
1
6
5
.
5
0
.
0
0
0
1
8
0
.
6
0
6
0
.
1
0
2
2
2
4
6
.
9
0
.
0
0
0
1
9
0
.
5
9
6
-
0
.
0
4
8
2
3
2
6
.
1
0
.
0
0
0
2
0
0
.
5
8
4
-
0
.
0
2
7
2
4
0
2
.
6
0
.
0
0
0
2
1
0
.
5
7
2
-
0
.
0
1
1
2
4
7
6
.
5
0
.
0
0
0
2
2
0
.
5
6
4
0
.
0
9
0
2
5
4
8
.
7
0
.
0
0
0
2
3
0
.
5
5
5
-
0
.
0
0
9
2
6
1
9
.
0
0
.
0
0
0
2
4
0
.
5
4
6
-
0
.
0
3
8
2
6
8
7
.
4
0
.
0
0
0
2
5
0
.
5
3
5
-
0
.
0
3
0
2
7
5
3
.
3
0
.
0
0
0
2
6
0
.
5
2
3
0
.
0
0
6
2
8
1
6
.
9
0
.
0
0
0
2
7
0
.
5
0
7
-
0
.
0
8
2
2
8
7
6
.
9
0
.
0
0
0
2
8
0
.
4
9
0
-
0
.
0
2
9
2
9
3
3
.
2
0
.
0
0
0
2
9
0
.
4
7
4
0
.
0
1
5
2
9
8
6
.
2
0
.
0
0
0
3
0
0
.
4
6
2
0
.
0
7
3
3
0
3
6
.
9
0
.
0
0
0
3
1
0
.
4
5
0
0
.
0
0
2
3
0
8
5
.
2
0
.
0
0
0
3
2
0
.
4
4
0
0
.
0
1
6
3
1
3
1
.
8
0
.
0
0
0
3
3
0
.
4
3
5
0
.
0
6
7
3
1
7
7
.
7
0
.
0
0
0
3
4
0
.
4
2
9
-
0
.
0
2
4
3
2
2
2
.
6
0
.
0
0
0
3
5
0
.
4
2
1
-
0
.
0
3
0
3
2
6
5
.
9
0
.
0
0
0
3
6
0
.
4
1
2
-
0
.
0
1
4
3
3
0
7
.
7
0
.
0
0
0
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 19 / 51
Example 51
The gure shows the autocorrelation function of series in Example 47. Notice that
the correlations decline linearly, indicating a non-stationary process.
C
o
r
r
e
l
o
g
r
a
m
o
f
L
S
G
I
D
a
t
e
:
0
1
/
3
1
/
0
8
T
i
m
e
:
1
3
:
2
1
S
a
m
p
l
e
:
1
9
8
8
M
0
1
2
0
0
3
M
0
5
I
n
c
l
u
d
e
d
o
b
s
e
r
v
a
t
i
o
n
s
:
1
8
5
A
u
t
o
c
o
r
r
e
l
a
t
i
o
n
P
a
r
t
i
a
l
C
o
r
r
e
l
a
t
i
o
n
A
C
P
A
C
Q
-
S
t
a
t
P
r
o
b
1
0
.
9
8
9
0
.
9
8
9
1
8
3
.
7
6
0
.
0
0
0
2
0
.
9
7
6
-
0
.
0
4
7
3
6
3
.
9
6
0
.
0
0
0
3
0
.
9
6
6
0
.
0
8
9
5
4
1
.
3
8
0
.
0
0
0
4
0
.
9
5
5
-
0
.
0
6
3
7
1
5
.
6
2
0
.
0
0
0
5
0
.
9
4
4
0
.
0
4
5
8
8
7
.
0
1
0
.
0
0
0
6
0
.
9
3
6
0
.
0
6
7
1
0
5
6
.
2
0
.
0
0
0
7
0
.
9
2
6
-
0
.
0
5
9
1
2
2
2
.
9
0
.
0
0
0
8
0
.
9
1
6
0
.
0
1
9
1
3
8
7
.
0
0
.
0
0
0
9
0
.
9
0
8
0
.
0
1
6
1
5
4
8
.
9
0
.
0
0
0
1
0
0
.
8
9
7
-
0
.
0
9
2
1
7
0
7
.
8
0
.
0
0
0
1
1
0
.
8
8
4
-
0
.
0
8
0
1
8
6
3
.
0
0
.
0
0
0
1
2
0
.
8
6
9
-
0
.
1
1
6
2
0
1
3
.
9
0
.
0
0
0
1
3
0
.
8
5
3
-
0
.
0
4
0
2
1
6
0
.
2
0
.
0
0
0
1
4
0
.
8
3
6
-
0
.
0
6
5
2
3
0
1
.
7
0
.
0
0
0
1
5
0
.
8
2
0
-
0
.
0
2
7
2
4
3
8
.
4
0
.
0
0
0
1
6
0
.
8
0
3
-
0
.
0
1
5
2
5
7
0
.
4
0
.
0
0
0
1
7
0
.
7
8
9
0
.
0
7
1
2
6
9
8
.
5
0
.
0
0
0
1
8
0
.
7
7
5
0
.
0
1
6
2
8
2
2
.
8
0
.
0
0
0
1
9
0
.
7
6
2
0
.
0
3
8
2
9
4
3
.
8
0
.
0
0
0
2
0
0
.
7
4
9
0
.
0
3
6
3
0
6
1
.
5
0
.
0
0
0
2
1
0
.
7
3
8
0
.
0
7
4
3
1
7
6
.
4
0
.
0
0
0
2
2
0
.
7
2
4
-
0
.
1
0
5
3
2
8
7
.
6
0
.
0
0
0
2
3
0
.
7
0
8
-
0
.
0
2
5
3
3
9
4
.
8
0
.
0
0
0
2
4
0
.
6
9
1
-
0
.
1
1
4
3
4
9
7
.
3
0
.
0
0
0
2
5
0
.
6
7
1
-
0
.
1
0
0
3
5
9
4
.
6
0
.
0
0
0
2
6
0
.
6
5
0
-
0
.
1
1
4
3
6
8
6
.
4
0
.
0
0
0
2
7
0
.
6
2
8
-
0
.
0
8
9
3
7
7
2
.
8
0
.
0
0
0
2
8
0
.
6
0
7
-
0
.
0
3
7
3
8
5
3
.
9
0
.
0
0
0
2
9
0
.
5
8
7
-
0
.
0
0
5
3
9
3
0
.
2
0
.
0
0
0
3
0
0
.
5
6
8
-
0
.
0
2
5
4
0
0
2
.
1
0
.
0
0
0
3
1
0
.
5
4
8
-
0
.
0
2
5
4
0
6
9
.
6
0
.
0
0
0
3
2
0
.
5
2
7
-
0
.
0
6
9
4
1
3
2
.
3
0
.
0
0
0
3
3
0
.
5
0
3
-
0
.
1
1
5
4
1
8
9
.
8
0
.
0
0
0
3
4
0
.
4
7
9
0
.
0
7
4
4
2
4
2
.
5
0
.
0
0
0
3
5
0
.
4
5
6
0
.
0
0
6
4
2
9
0
.
4
0
.
0
0
0
3
6
0
.
4
3
2
0
.
0
2
2
4
3
3
3
.
8
0
.
0
0
0
This behavior is observed both in the autocorrelations of the series as well as in
the logarithms of the autocorrelations.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 20 / 51
The simple exponential smoothing process
A non-stationary process that has been widely used for forecasting and which
can be considered a generalization of the random walk is:
z
t
= c + (1 B) a
t
. (104)
In this process if we make = 0 we obtain a random walk. It can also be seen as
an extreme case of the ARMA(1,1) when the autoregressive coecient is the unit.
For simplicity we assume c = 0 and || < 1 and inverting the MA part, this
process is written:
z
t
= (1 )z
t1
+ (1 )z
t2
+
2
(1 )z
t3
+ ... + a
t
. (105)
In the usual case where 0 < < 1, all the coecients of the above AR() are
positive and add up to the unit. These coecients form an innite geometric
progression with the rst term (1 ) and ratio , hence their sum is one.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 21 / 51
The simple exponential smoothing process
The equation (105) indicates that the observed value at each time is the
weighted mean of the values of the series at earlier times, with coecients that
decay geometrically with the lag:
If is close to the unit the mean is calculated using many coecients, but
each of them with little weight.
For example, if = .9 the weights of the lags are .1, .09, .081, . . . and the
weights decay very slowly.
If is close to zero, the mean is calculated by weighting only the last values
observed.
For example, if = .1 the weights of the lags are .9, .09, .009,. . . and
practically only the the last lags are taken into account, with very dierent
weights.
This is known as the simple exponential smoothing process and was
introduced intuitively in the rst session.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 22 / 51
The simple exponential smoothing process
To calculate the descriptive measures of the process we assume that it started
at t = 0, with z
0
= a
0
= 0 and c = 0. Thus:
z
1
= a
1
z
2
= z
1
+ a
2
a
1
= a
2
+ (1 ) a
1
.
.
.
.
.
.
.
.
.
z
t
= a
t
+ (1 ) a
t1
+ ... + (1 ) a
1
.
The mean of the process is zero, as in the random walk when c = 0. Its
variance is:
Var (z
t
) =
2
(1 + (t 1)(1 )
2
)
and the process is non-stationary in the variance, since the function depends on
the time.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 23 / 51
The simple exponential smoothing process
The covariances are calculated as:
Cov(t, t + k) = E(z
t
z
t+k
) =
= E
(a
t
+ (1 )
t1
i =1
a
ti
)(a
t+k
+ (1 )
t+k1
i =1
a
t+ki
)
(1 + (t 1)(1 )
2
) (1 + (t + k 1)(1 )
2
)
.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 24 / 51
The simple exponential smoothing process
Assuming that t is large, such that t 1 t, and that is not very close to
one, such that 1 + t(1 )
2
t(1 )
2
, then:
(t, t + k) =
t
t(t + k)
That is the same autocorrelation function (102) obtained for the random walk,
i.e., the coecients show approximately linear decay when the lag increases.
Nevertheless, if is close to one, the coecients do not necessarily have to be
large, although we should always observe a linear decay.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 25 / 51
Example 52
The gure shows a series generated using = .9. Notice that although the
autocorrelations are no longer close to one, they show the expected characteristic
of linear decay.
0 100 200 300
-2
-1
0
1
2
3
4
5
6
5 10 15 20 25 30
-0.2
-0.1
0
0.1
0.2
0.3
0.4
0.5
0.6
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 26 / 51
Integrated process of order two
Many real series with trend are represented with a model of two dierences, or
integrated of order two. A simple model that appears in many applications is:
2
z
t
= (1 B)a
t
. (106)
To justify this model, we assume a random walk but where the drift changes
with time, that is, the process:
z
t
= c
t
+ u
t
.
This equation indicates that z
t
has a mean, which is the growth of z
t
, that
evolves over time. Successively substituting in the above equation and assuming
that the process starts at t = 0 and that z
0
= u
0
= 0, we obtain:
z
t
= (c
t
+ ... + c
1
) + u
t
+ ... + u
1
.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 27 / 51
Integrated process of order two
Let us assume that the evolution of the growth coecient at each time, c
t
, is
smooth, and such that
c
t
= c
t1
+
t
,
where
t
is a white noise process, independent of u
t
.
Thus:
2
z
t
= z
t
z
t1
= c
t
+ u
t
(c
t1
+ u
t1
) =
=
t
+ u
t
u
t1
= (1 B)a
t
,
since the sum of white noise and a non-invertible MA(1) will be an invertible
MA(1).
We conclude, therefore, that process (106) is a generalization of the random
walk that allows the drift to vary smoothly over time.
In general, integrated processes of order two can be seen as a generalization of
integrated processes of order one but where the slope of the growth line, instead
of being xed, varies over time.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 28 / 51
ARIMA processes
The two processes discussed, RW and IMA, were obtained by accepting that
the root of the AR part of the AR(1) and ARMA(1,1) processes is the unit. This
idea can be generalized for any ARMA process, allowing one or various roots of
the AR operator to be the unit:
(1
1
B ...
p
B
p
)(1 B)
d
z
t
= c + (1
1
B ...
q
B
q
)a
t
which we call ARIMA (p, d, q) processes.
p is the order of the autoregressive stationary part.
d is the number of unit roots (order of integration of the process).
q is the order of the moving average part.
Using the dierence operator, = 1 B, the above process is usually written
as:
p
(B)
d
z
t
= c +
q
(B)a
t
. (107)
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 29 / 51
ARIMA processes
ARIMA stands for Autoregressive Integrated Moving Average, where
integrated indicates that letting
t
=
d
z
t
denote the stationary process, z
t
is
obtained as a sum (integration) of
t
. Indeed, if
t
= (1 B)z
t
since, dening
(1 B)
1
= 1 + B + B
2
+ B
3
+ ...
results in:
z
t
= (1 B)
1
t
=
t
j =
t
.
We have seen two examples of ARIMA processes in two earlier examples: the
random walk is an ARIMA(0,1,0) model and the simple exponential smoothing is
the ARIMA(0, 1, 1) or IMA(1, l).
Both are characterized by the fact that the autocorrelation function has slowly
decaying coecients.
All non-stationary ARIMA processes have this general property.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 30 / 51
ARIMA processes
To prove it, recall that the correlogram of an ARMA(p,q) satises the equation
for k > q:
p
(B)
k
= 0 k > q
whose solution is of type:
k
=
p
i =1
A
i
G
k
i
where G
1
i
are the roots of
p
(B) = 0 and |G
i
| < 1.
If one of these roots G
i
is very close to the unit, writing G
i
= 1 , with very
small , for large k the terms A
j
G
k
j
will be zero due to the other roots (since G
k
j
0) and we have that, approximately:
k
= A
i
(1 )
k
A
i
(1 k) for large k
As a result, the ACF will have positive coecients that will fade out
approximately linearly and may be dierent from zero for high values of k.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 31 / 51
This property of persistence of positive values in the correlogram (even though
they are small) and linear decay characterizes the non-stationary processes.
Example 53
The gure shows a realization of the series (1 .4B)z
t
= (1 .8B)a
t
and its
estimated autocorrelation function. The linear decay of the autocorrelation
coecients is again observed.
0 100 200 300
-8
-6
-4
-2
0
2
4
5 10 15 20 25 30
-0.2
0
0.2
0.4
0.6
0.8
Lag
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 32 / 51
Integrated processes and trends
Not all non-stationary processes are integrated, but integrated processes cover
many interesting cases that we nd in practice.
It is easy to prove that any process which is the sum of a polynomial trend and
a stationary process will be integrated. For example, let us take the case of a
deterministic linear growth process:
z
t
= b + ct + u
t
(108)
where u
t
is white noise.
This process is integrated of order one or I (1), since taking the rst dierence
of the series z
t
:
t
= z
t
z
t1
= b + ct + u
t
(b + c (t 1) + u
t1
)
= c + u
t
u
t1
since c is constant and u
t
is a stationary process (being the dierence of two
stationary processes, see section 3.4), process
t
is stationary as well.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 33 / 51
Integrated processes and trends
Therefore, z
t
dened by (108), is integrated of order one, and its rst dierence
follows model (104) with a
t
= u
t
and parameter equal to the unit.
The process is not invertible, and this property tells us that we have a
deterministic linear component in the model.
We observe that by dierentiating the process the remaining constant is the
slope of the deterministic linear growth.
This result is valid for processes generated as a sum of a polynomial trend of
order h and any stationary process, u
t
:
z
t
=
t
+ u
t
(109)
where
t
= a + bt + ct
2
+ ... + dt
h
.
If we dierentiate this process h times a stationary process is obtained with a
non-invertible moving average part.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 34 / 51
Integrated processes and trends
For example, let us assume a quadratic trend. Then, taking two dierences in
the process, we have:
2
z
t
= c +
2
u
t
and the process
t
=
2
u
t
has a non-invertible moving average.
For example, if u
t
is an AR(1), (1 B)u
t
= a
t
, we have
t
= (1 B)
1
2
a
t
and
t
is a non-invertible ARMA(1,2) process.
This example shows us that if a series has a deterministic polynomial trend and
it is modelled by an ARIMA process we have: (1) non-invertible moving average
components; (2) a stationary series with a mean dierent from zero and a
constant that is the highest order coecient in the deterministic trend.
Again, we see the importance of constants in integrated processes, since they
represent deterministic eects.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 35 / 51
Integrated processes and trends
Processes with a deterministic linear trend are infrequent in practice, but we
can generalize the above model somewhat allowing the slope at each time to vary
a little with respect to the previous value. We can write the process with a linear
trend but with variable components, as:
z
t
=
t
+ v
t
. (110)
With this model the value of the series at each point in time is the sum of its
level,
t
, and a white noise process, v
t
.
We assume that the level evolves as:
t
=
t1
+ c +
t
. (111)
The process has a trend c, because the level at time t,
t
is obtained from the
previous time adding a constant c, which is the slope, plus a small random
component,
t
.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 36 / 51
Integrated processes and trends
In this process the rst dierence is:
z
t
= c +
t
+v
t
and since the sum of an MA(1) process and white noise is an MA(1), we again
obtain process (104), but now with a parameter smaller than the unit, as in
(104).
The above model can be further generalized by making the deterministic slope
c, change with time, but with certain inertia. To do that, we replace expression
(111) with:
t
=
t1
+
t
+
t
and
t
=
t1
+
t
where the processes
t
and
t
are independent white noises. Now the level grows
linearly, but with a slope,
t
, that changes over time.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 37 / 51
Integrated processes and trends
In this representation if the variance of
t
is very small the series has an almost
constant trend, and we return to the above model. If we take the dierence in the
original series, we have
z
t
=
t
+
t
+u
t
and taking a second dierence
2
z
t
=
t
+
t
+
2
u
t
= (1
1
B
2
B
2
)a
t
(112)
since if we sum MA processes a new MA process is obtained.
Depending on the variances of the three white noise processes we can have
dierent situations: (i) if V(u
t
) is much smaller than that of the other noise
processes the term
2
u
t
may not be taken into account in the right hand side of
(112) and the sum will be an MA(1); (ii) if the term
t
is the dominant one and
we do not take the other two into account, the series is approximately the model
2
z
t
=
t
, and (iii) if V(
t
) is zero, then
t
=
t1
= and we return to the
model which has a deterministic trend.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 38 / 51
Integrated processes and trends
It is important to understand the dierence between the two types of models
presented here:
Model (108) is pure deterministic, since at each time, t, the expected value
of the series is determined, and is b +ct. Moreover, the expected value of the
series at z
t
knowing the value at z
t1
is still the same, b + ct. In this model,
knowing the previous values does not modify the predictions.
Model (110) is more general, because although at each point in time it also
has linear growth equal to c, the level of the series at each point in time is
not determined from the beginning and can vary randomly. This means that
the expected value of the series at z
t
knowing the value at z
t1
is z
t1
+ c,
and depends on the observed value.
To summarize, the ARIMA models, by including dierences, incorporate
stochastic trends and, in extreme cases, deterministic trends. In this case, the
dierentiated series will have a mean dierent from zero and non-invertible
moving average terms may appear.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 39 / 51
Long-memory processes
There is a class of stationary processes that are easily confused with
non-stationary integrated processes. These are long-memory processes, which
are characterized by having many autocorrelation coecients with small
coecients and which decay very slowly.
This property has been observed in some stationary meteorological or
environmental series (but also in nancial series). The slow decay of the
autocorrelations could lead to modelling these series as if they were integrated of
order one, by taking a dierence.
Nevertheless the decay is dierent from that of integrated processes: it is much
faster for the rst lags than in an integrated process, but slower for high lags.
This structure means that in a long-memory process we observe many small
autocorrelation coecients and they decay very slowly for high lags, as opposed to
what usually happens with integrated processes, where the decay is linear.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 40 / 51
Long-memory processes
The simplest long-memory process is obtained by generalizing the random walk
using:
(1 B)
d
z
t
= a
t
(113)
where now the parameter d instead of being an integer is a real number.
To dene the operator (1 B)
d
when d is not an integer, we start from the
expression of Newtons binomial theorem. When d is a positive integer, it is
veried that:
(1 B)
d
=
d
i =0
d
i
(1)
di
(B)
i
=
d
i =0
i
B
i
, (114)
where the coecients of this sum are dened by
i
=
d
i
(1)
i
=
d!
(d i )!i !
(1)
i
=
(d + 1)
(i + 1)(d i + 1)
(1)
i
,
where (a + 1) = a(a) is the Gamma function which for integer values coincides
with the factorial function.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 41 / 51
Long-memory processes
To generalize this denition when d is not an integer, we can take advantage of
the fact that the Gamma function is dened for any real number, and for real d
write:
(1 B)
d
=
i =0
i
B
i
, (115)
where the coecients of this innite expansion are given by
i
=
(d + 1)
(i + 1)(d i + 1)
(1)
i
.
Moreover, it can be proved that when d is an integer this denition is
consistent with (114), since all terms greater than d in this innite summation are
zero, such that for an integer d (115) is reduced to (114).
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 42 / 51
Long-memory processes
We observe that for i = 0 the result is
0
= 1, and using the properties of the
Gamma function it is shown that:
i
=
0<j i
j 1 d
j
, for i = 1, 2, ...
Using this expression we can write the fractional process (113) by means of the
AR() representation developing the operator (1 B)
d
, to obtain:
z
t
=
i =0
i
z
ti
+ a
t
. (116)
It can be proved that this process also admits an MA() representation of the
form
z
t
=
i =0
i
a
ti
,
where the coecients verify Wolds representation of a stationary MA() series
with nite variance, that is,
2
i
< .
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 43 / 51
Long-memory processes
These coecients are obtained using
i
=
(d + i )
(i + 1)(d)
.
It is proved that the autocorrelation function of the long-memory fractional
process (113) is
(k) =
(d + k)(1 d)
(k + 1 d)(d)
.
Particularly for k = 1, we obtain
(1) =
d(d)(1 d)
(1 d)(1 d)(d)
=
d
1 d
,
for k = 2
(2) =
d(d + 1)
(2 d)(1 d)
= (1)
(d + 1)
(2 d)
and, in general
(k) = (k 1)
(d + k + 1)
(k d)
.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 44 / 51
Long-memory processes
For large k, we can approximate the correlations by:
(k)
(1 d)
(d)
|k|
2d1
and taking logarithms
log (k) a + (2d 1) log |k| .
Therefore, if we represent log (k) with respect to log|k| for high values of k we
obtain a line with slope 2d 1.
This property dierentiates a long-memory process from a non-stationary
integrated process, where we have seen that log (k) decays linearly with k and
not with log k.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 45 / 51
Example 54
The left gure shows the coecients -
i
of the AR representation given by (116)
for several values of d and right gure the autocorrelation function for those
values of d.
5 10 15 20 25 30
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
d = 0.4
d = 0.3
d = 0.1
5 10 15 20 25 30
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
d = 0.4
d = 0.3
d = 0.1
Notice that the AR coecients quickly become very small, and that the AC function
has many non-null coecients that decay slowly.
These properties make it easy to confuse this process with a high order AR(p), and if
d is close to .5 the high values and slow decay of the AC function can suggest an I(1)
process.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 46 / 51
Long-memory processes
Generalizing on the above, we can consider stationary ARMA processes, but
ones that include long-memory as well, dening
(1
1
B ...
p
B
p
)(1 B)
d
z
t
= c + (1
1
B ...
q
B
q
)a
t
where d is not necessarily an integer.
If d < 0.5 the process is stationary, whereas if d 0.5 it is non-stationary.
These models are called ARFIMA(p, d, q), autoregressive fractionally
integrated moving average.
If the process is stationary it has the property whereby for high lags, when the
ARMA structure disappears, the long-memory characteristic will appear. If it is not
stationary, taking a dierence will convert it to stationary, but with long-memory.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 47 / 51
Long-memory processes
Long-memory processes may appear when short-memory series are added under
certain circumstances. Granger (1980) proved that if we add N AR(1)
independent processes as a limit we obtain a long-memory process.
We known that the aggregation of independent AR processes leads to ARMA
processes, and this result is expected with few summands.
Grangers results indicate that in the limit we will have a long-memory process
and suggest that these processes can be approximated by an ARMA with orders p
and q that are high and similar.
With series that are not very long it is dicult to dierentiate a long-memory
process from an ARMA. Nevertheless, if we have a series with a large number of
observations, as is typical in meteorological or nancial data, in some cases LM
processe can provide better t than the short-memory ARMA processesn.
However, as we will see in the next chapter, the dierences in the prediction are
usually small.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 48 / 51
Long-memory processes - Nile river example
Example 55
The gure shows a series of yearly data for minimum levels of the Nile River
between the years 622 and 1284, which can be found in the le nilomin.dat.
9
10
11
12
13
14
15
0700 0800 0900 1000 1100 1200
Minimum yearly levels of Nile river, 622 - 1284
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 49 / 51
The gure shows its autocorrelation function, and it is observed that the
correlations decay very slowly.
5 10 15 20 25 30 35 40
-0.1
0
0.1
0.2
0.3
0.4
0.5
0.6
Lag
As a reference its also included the AC function of a fractional process with
d = 0.35.
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 50 / 51
The gure shows the logarithm of the AC compared to the logarithm of the lag,
and we see that the relationship is approximately linear.
As a comparison we have also in-
cluded in this gure the Madrid Stock
Exchange index data.
In this integrated process for small
values of k the AC decay more slowly
in the integrated series than in the
long-memory one, but for larger k the
coecients for this series, which fol-
low an ARIMA model, decay faster
than that which would correspond to
a long-memory process.
0 0.5 1 1.5 2 2.5 3 3.5 4
-2.5
-2
-1.5
-1
-0.5
0
log(k)
l
o
g
(
(
k
)
)
Nile river
Madrid Stock Index
Alonso and Garca-Martos (UC3M-UPM) Time Series Analysis June July, 2012 51 / 51