Professional Documents
Culture Documents
Under the null hypothesis that successive outcomes are independent and assuming N1> 10
andN2> 10, the number of run is distributed normally with
2 N1 N2
Mean: E ( R )= +1
N
2 2 N 1 N 2 (2 N 1 N 2−N )
Variance: σ R =
N 2 ( N −1)
If the null hypothesis of randomness is sustainable, following the properties of the normal
distribution, we should expect that
That is, the probability is 95% that the preceding interval will include R. So
Do not reject the null hypothesis of randomness with 95% confidence if R, the number of runs,
lies in the preceding confidence interval; reject the null hypothesis if the estimated R lies outside
these limits.
Durbin Watson Test:
The most celebrated test for detecting serial correlation is that developed by statisticians Durbin
and Watson. It is popularly known as the Durbin- Watson d statistic. Null and alternative
hypothesis for the D-W test is
Which is simply the ratio of the sum of squared differences in successive residuals to the RSS.
Assumptions:
The exact sampling or probability distribution of the d statistic given in 1 is difficult to derive
because as the Durbin Watson have shown , it depends in a complicated way on the X values
present in a given samples. Since d is computed from ^μt which are of course dependent on the
given X’s, therefore unlike t, F, or ℵ2 tests there is no unique critical value that will lead to the
rejection or the acceptance of the null hypothesis that there is no first order auto correlation.
Durbin Watson was successful in deriving a lower bound d L and an upper bound du such that if
the computed d lies outside of these critical values, a decision can be made regarding the
presence of auto correlation.
Decision rule:
Y t =β 1+ β2 X t + μ t … … ..(1)
and assume that the error term follows the AR(1) scheme , namely,
μt =ρ μt−1 +ε t−1 ≤ ρ≤ 1
When ρ is known:
If the coefficient of the first order autocorrelation is known, the problem of autocorrelation can
be easily solved. If equation no.1 holds true at time t-1. Hence
Y t −1 =β1 + β 2 X t−1 + μt −1 … … …( 2)
Y ¿t =β ¿1+ β¿2 X ¿t + ε t
Where
Since ρ lies between 0 and ±1, one could start from two extreme positions. In one extreme, one
could assume that ρ=0, that is, no (first order ) serial correlation, and at the other extreme we
could let ρ=±1, that is perfect positive or negative correlation. As a matter of fact, when a
regression is run, one generally assumes that there is no autocorrelation and then lets the Durbin
–Watson or other test show whether this assumption is justified. If however, ρ=+1, the
generalized difference equation (4) reduce to the first difference equation:
Since the error term in (5) is free from autocorrelation, to run the regression (5) all one has to
do is form a first differences of both the regressand and regressor(s) and run the regression on
these first difference.
This transformation may be appropriate if the coefficient of autocorrelation is very high, say in
excess of .8, or the Durbin-Watson d is quite low. Use first difference method form whenever
d<R2 ( proposed by Maddala). An interesting feature of the (5) is that there is no intercept in it.
Hence to estimate (5) one have to use the regression through origin routine. If, however, you
forget to drop the intercept term in the model and estimate the following model that includes the
intercept term
∆ Y t =β 1 + β 2 ∆ X t +ε t
Then the original model must have a trend in it and β1 represents the coefficient of the trend
variable.