Professional Documents
Culture Documents
2 (2015) 247
ON THE PERFORMANCE OF AUTOREGRESSIVE MOVING AVERAGE
POLYNOMIAL DISTRIBUTED LAG MODEL
ABSTRACT
This study focused on the performance of Autoregressive Moving Average Polynomial Distributed Lag Model
among all other distributed lag models. Four models were considered; Distributed Lag (DL) model, Polynomial
Distributed Lag (PDL) model, Autoregressive Polynomial Distributed Lag (ARPDL) model and Autoregressive
Moving Average Polynomial Distributed Lag (ARMAPDL) model. The parameters of these models were
estimated using least squares and Newton Raphson iterative methods. To determine the order of the models,
Akaike Information Criterion (AIC) and Bayesian Information Criterion (BIC) were used. To determine the best
model, the residual variances attached to these models were studied and the model with the minimum residual
variance was considered to perform better than others. Using numerical example, DL, PDL, ARPDL and
ARMAPDL models were fitted. Autoregressive Moving Average Polynomial Distributed Lag Model
(ARMAPDL) model performed better than the other models.
Keywords: Distributed Lag Model, Selection Criterion, Parameter Estimation, Residual Variance.
Yt is an endogenous variable and Xt is exogenous on the original variable X. OLS method is used to
variable, a is the intercept, b0 is the distributed lag estimate the coefficient of the model since the
weight, et is the error term. The parameters of the assumptions of the disturbance term is satisfied.
model can be estimated using least squares
method. The coefficients of d0, d1, d2 can be estimated by di
Assumptions of the model are: = (Z Z)-1 Z Y (6)
v The model is linear in parameters: Thereafter the estimate the coefficients of β can
be estimated from the original model by equation
yt = b0 + b1xt1 + . . .+ bjxtj + et
3.
v There is the need to make a zero conditional
mean assumption: E(et |X) = 0, t = 1, 2, …, n. Autoregressive Polynomial Distributed Lag
v The Xs are strictly exogenous Model
v et is independent with mean zero and variance
2
of s The model can be defined as:
v There is no serial correlation: Corr. (et es |
X)=0 for t ¹ s
(7)
Polynomial Distributed Lag Model
Polynomial Distributed Lag Model is obtained where bj is approximated by polynomial in the lag
from a finite distributed lag given as k as
(2)
where
is approximated by polynomial of lower (8)
degree.
(5)
where
The sequence of random deviation (et ) can be
estimated by:
Where Zi are constructed from the original lagged To obtain the unknown parameters of the model,
variable Xt, Xt-1, Xt-2 and Xt-3. Therefore Y is we make some assumption that random error is
regressed on the constructed variable Zi and not independently and identically distributed with
Ojo and Aiyebutaju: On the Performance of Autoregressive Moving Average 249
2
mean zero and variance of s . where i =1, 2, …, R, m=1,2,…, R. (14)
Minimizing the likelihood function, with respect
to the parameters (j1 , j 2 , d 0 , d 1 , d 2 ) we can Where the partial derivatives satisfy the recursive
obtain estimate of the parameters of the model equations
using least squares method; Chen (2010) and ¶S (G )
= -2å (v t )Yt -1
subsequently, we obtain the parameters of β. ¶j1
¶S (G )
Autoregressive Moving Average Polynomial = -2å (v t )Yt -i
Distributed Lag Model ¶j i where i = 2, …, p (15)
The model is defined as ¶S(G)
=- 2å (v t ) X t
¶b 0
(11) ¶S(G)
=- 2å (v t ) X t-j where k =1, 2,…, j (16)
where ¶b k
j1 ,........., j p are the parameters of the autoregressive
¶S (G )
component, f1 ,........., f q are the parameters of the = -2å (v t ) Ît -1
moving average component, β0....βj are the ¶f1
parameters of the polynomial distributed lag ¶S (G )
model, Yt and Xt are the dependent and = -2å (v t ) Ît - r where r = 2,…,q. (17)
¶f r
independent variable respectively, vt is the error The second derivative is given as
term and is assumed to be normally distributed
with mean zero and variance s .
2 ¶ 2 S (G ) én ù
= -2 êå (v t )(0) + (Yt -1 )(-å Yt -1 )ú = 2å Yt 2-1
¶j1j1¢ ë i =1 û
Estimation of parameters of Autoregressive ¶ 2 S (G ) én ù
= -2 êå (v t )(0) + (Yt -1 )(-å Yt -i )ú = 2å Yt 2-i
Moving Average Polynomial Distributed Lag ¶j i j i¢ ë i =1 û
Model where i =2, …, p. (18)
2
¶ S (G ) én ù
= -2 êå (vt )(0) + ( X t -1 )(-å X t -1 )ú = 2å X t2-1
¶b 0 b 0¢ ë i =1 û
We consider Newton Raphson iterative method ¶ 2 S (G ) én ù
= -2 êå (vt )(0) + ( X t -1 )(-å X t - k )ú = 2å X t2- k
using the approach of Ojo (2009) and Pascal ¶b k b k¢ ë i =1 û
(2001) to estimate the parameters of the model. where k =1, 2,…, j. (19)
Representing the mean response as the error
term becomes ¶ 2 S (G ) én ù
= -2 êå (vt )(0) + (Ît -1 )(-å Ît -1 )ú = 2å Ît2-1
(12) ¶f1f1¢ ë i =1 û
The least square estimator of G of G which ¶ 2 S (G ) én ù
= -2 êå (vt )(0) + (Ît -1 )(-å Ît - r )ú = 2å Ît2- r
minimizes the sum of the square of residual is ¶f r f r¢ ë i =1 û
n
S (G ) = å (vt ) 2 where r = 2, …, q. (20)
t =1
2 n
¶ S (G ) é ù
We differentiate S(G) with respect to the = -2 êå (vt )(0) + (Yt -1 )(-å X t )ú = 2å X t Yt -1
¶j1 b 0 ë i =1 û
parameter G ( ) ¶ 2 S (G ) én ù
We shall write = -2 êå (vt )(0) + (Yt -1 )(-å Yt -i )ú = 2å X t - k Yt -i
¶j i b k ë i =1 û
The partial derivatives of S(G) are
where i = 2, …, p and k = 1, 2, …, j. (21)
¶S(G) ¶ (v t)
Gi = = -2å (v t ) ¶ 2 S (G )
¶G ¶Gi én ù
= -2 êå (vt )(0) + (Yt -1 )(-å Ît -1 )ú = 2å Yt -1 Ît -1
¶j1f1 ë i =1 û
where i =1, 2, …, R and R= p+j+q (13)
¶ 2 S (G ) én ù
2
¶ S (G ) én ¶ 2 (v t ) ¶ (v t ) ¶ (v t ) ù = -2 êå (vt )(0) + (Yt -i )(-å Ît - r )ú = 2å Yt -i Ît - r
H= = -2 êå (Vt ) + ¶j i f r ë i =1 û
ú
¶Gi G m ë t =1 ¶Gi ¶G m ¶Gi ¶G m û where i = 2, …, p and r = 2, …, q. (22)
250 Ojo and Aiyebutaju: On the Performance of Autoregressive Moving Average
¶ 2 S (G ) én ù
= -2 êå (vt )(0) + ( X t )(-å Ît -1 )ú = 2å X t Ît -1 Selection of the Length of the Lag
¶b 0f1 ë i =1 û Numerous procedures have been suggested for
¶ 2 S (G ) én ù selecting the length n of a finite distributed lag in
= -2 êå (vt )(0) + ( X t - k )(-å Ît - r )ú = 2å X t - k Ît - r
¶b k f r ë i =1 û Judge et al. (2000).Two goodness-of-fit measures
that are more appropriate are Akaike's
where k = 1, 2, …, j and r = 2, …, q. (23) Information Criterion (AIC)
¶S (G ) SSE n 2( n + 2)
= 0, AIC = ln + . (24)
¶Gi T-N T-N
¶ 2 S (G ) Schwarz criterion known as Bayesian
= 0.
¶Gi ¶G m Information Criterion (BIC)
SSE n (n + 2) ln(T - N )
Set the gradient to be V(G) where SC (n) = ln + . (25)
T-N T-N
¶S (G ) ¶S (G ) ¶S (G )
VG
( )= , ,..., For each of these measures we seek that lag length
¶G1 ¶G2 ¶GR n that minimizes the criterion can be used. Since
*
And the Hessian is represented by H adding more lagged variables reduces SSE, the
second part of each of the criteria is a penalty
¶ 2 S (G ) function for adding additional lags. These
H= .
¶Gi ¶G j measures weigh reductions in sum of squared
errors obtained by adding additional lags against
The approximate mean responses f(Xi,G) for the n the penalty imposed by each. They are useful for
cases by the linear term in the Taylor series comparing lag lengths of alternative models
expansion we obtain estimated using the same number of observations
V(G1)»V(G0)+H(G0)(G1-G0)=0 Ojo (2013). In this study we shall use AIC and
G1-G0= -H-1(G0)V(G0) BIC criteria for selecting best order for the
thereby obtaining the iterative equation given by models under study.
G(k+1)=Gk-H-1(Gk)V(Gk). RESULTS AND DISCUSSION
k th
G is the set of estimates obtained at the k stage
of iteration. The estimates obtained by the above Numerical Example
iterative equation usually converge. For starting To present the application of these models we will
the iteration, we need to have good sets of initial use a real time series dataset, monthly rainfall and
values of the parameters. This can be obtained by temperature series between 1979 and 2008
fitting the best autoregressive moving average obtained from Forestry Research Institute of
model. Nigeria (FRIN), Ibadan, Nigeria (see the
appendix). Rainfall series is the endogenous
Performance of the Model Indicator variable while temperature series is the exogenous
variable. For the fitted model, the estimation
Residual Variance technique in the previous section were used
Residual variance or unexplained variance is part
of the variance of any residual. In analysis of Fitted Distributed Lag Model
variance and regression analysis, residual variance
is that part of the variance which cannot be Ŷ t= 0.434900 - 8.060110Xt - 3.430149Xt-1+
attributed to specific causes. The unexplained 0.126389Xt-2+2.731801Xt-3+5.413812Xt-4+6.769427Xt-5
variance can be divided into two parts. First, the Fitted Polynomial Distributed Lag Model
part related to random, everyday, normal, free will
differences in a population or sample. Among any Ŷ t =-7.9488Xt-3.59344Xt-1+0.05829Xt-2+3.00639Xt-3+
aggregation of data these conditions equal out. 5.25086Xt-4+6.79170Xt-5
Second, the part that comes from some condition Fitted Autoregressive Polynomial Distributed Lag Model
that has not been identified, but that is systematic.
Ŷ t= 0.392048Yt-1 - 0.152001Yt-2 - 0.069668Yt-3 -
That part introduces a bias and if not identified
0.163191Yt-4 -6.31189Xt-3.12314Xt-1-0.25566Xt-2+
can lead to a false conclusion (Ojo et al., 2008). 2.29056Xt-3+4.51657Xt-4+6.41919Xt-5
Ojo and Aiyebutaju: On the Performance of Autoregressive Moving Average 251
Fitted Autoregressive Moving Average Polynomial
Distributed Lag Model
Yˆt = 1.179189Yt -1 - 0.400669Yt - 2 - 0.826939e t -1 - 5.73789 X t - 2.91633 X t -1
- 0.31444 X t - 2 + 2.06779 X t -3 + 4.23038 X t - 4 + 6.17331X t -5
The inclusion of Moving Average term in Autoregressive Polynomial Distributed Lagmodel yielded
reduction in the value of the residual variance which made ARMAPDL the best model.
APPENDIX