Professional Documents
Culture Documents
Iles Gillard Tech Report2
Iles Gillard Tech Report2
for
Linear Regression
with
Errors in both Variables
by
J.W. Gillard and T.C. Iles
1 Introduction 2
1
Chapter 1
Introduction
In the technical report by Gillard and Iles [9], several method of moments based es-
timators for the errors in variables model were introduced and discussed. Broadly
speaking, the method of moments estimators can be divided into two separate classi-
fications. These are:
Restricted parameter space estimators These estimators are based on first and
second order moments, and are derived by assuming that a parameter (or a function
of parameters) is known a priori.
Higher moment based estimators If the data displays sufficient skewness and
kurtosis, one may estimate the parameters of the model by appealing to estimators
based on third and fourth order moments. Although these estimators have the initial
appeal of avoiding a restriction on the parameter space, they must be used with caution.
Details of how these estimators were derived, and how they may be used in practise
were again given by Gillard and Iles [9]. This present technical report aims to provide
the practitioner with further details concerning asymptotic variance and covariances
for both the restricted cases, and higher moment based estimators. In this way, this
report can be viewed as a direct sequel to the technical report by Gillard and Iles [9].
2
Chapter 2
Consider two variables, ξ and η which are linearly related in the form
ηi = α + βξi , i = 1, . . . , n
xi = ξi + δi
yi = ηi + εi = α + βξi + εi
It is assumed that E[δi ] = E[εi ] = 0 and that V ar[δi ] = σδ2 , V ar[εi ] = σε2 for all i. Also
the errors δ and ε are mutually uncorrelated. Thus
Cov[δi , δj ] = Cov[εi , εj ] = 0, i 6= j
Cov[δi , εj ] = 0, ∀ i, j (2.1)
There have been several reviews of errors in variables methods, notably Casella and
Berger [2], Cheng and Van Ness [3], Fuller [7], Kendall and Stuart [14] and Sprent [15].
Unfortunately the notation has not been standardised. This report closely follows the
notation set out by Cheng and Van Ness [3] but for convenience, it has been necessary
to modify parts of their notation.
3
Errors in variables modelling can be split into two general classifications defined by
Kendall [12], [13], as the functional and structural models. The fundamental difference
between these models lies in the treatment of the ξi0 s
The functional model This assumes the ξi0 s to be unknown, but fixed constants µi .
The structural model This model assumes the ξi0 s to be a random sample from a
random variable with mean µ and variance σ 2 .
It is the linear structural model that is the main focus of this technical report.
E[X] = E[ξ] = µ
E[Y ] = E[η] = α + βµ
The method of moments estimating equations can now be found by replacing the
4
population moments with their sample equivalents
x̄ = µ̃ (2.2)
ȳ = α̃ + β̃ µ̃ (2.3)
sxy = β̃ σ̃ 2 (2.6)
Here, a tilde is placed over the symbol for a parameter to denote the method of moments
estimator. From equations (2.4), (2.5) and (2.6) it can be seen that there is a hyperbolic
relationship between the method of moments estimators. This was called the Frisch
hyperbola by van Montfort [16].
This is a useful equation as it relates pairs of estimates (σ̃δ2 , σ̃ε2 ) to the data in question.
One of the main problems with fitting an errors in variables model is that of iden-
tifiability. It can be seen from equations (2.2), (2.3), (2.4), (2.5) and (2.6) a unique
solution cannot be found for the parameters; there are five equations, but six unknown
parameters. A way to proceed with this method is to assume that there is some prior
knowledge of the parameters that enables a restriction to be imposed. The method
of moments equations under this restriction can then be readily solved. Other than
this, additional estimating equations may be derived by deriving equations based on
the higher moments.
There is a comparison with this identifiability problem and the maximum likelihood
approach. The only tractable assumption to obtain a maximum likelihood solution is
to assume that the distributions of ξ, δ and ε are all Normal. Otherwise, the algebraic
manipulation required becomes an enormous task. Further details will be presented
in Gillard [8]. If all the distributions are assumed Normal, this leads to the bivari-
ate random variable (x, y) having a bivariate Normal distribution. This distribution
5
has five parameters, and the maximum likelihood estimators for these parameters are
identical to the method of moments estimators based on the moment equations (2.2),
(2.3), (2.4), (2.5), and (2.6) above. In this case therefore it is not possible to find
unique solutions to the likelihood equations without making an additional assumption,
effectively restricting the parameter space. A fuller treatment of restrictions on the
parameter space and method of moments estimators can be found in Gillard and Iles
[9].
µ̃ = x̄ (2.7)
The estimators for the remaining parameters can be expressed as functions of the
slope, β, and other sample moments. An estimator for the intercept may be found by
substituting (2.2) into (2.3) and rearranging to give
α̃ = ȳ − β̃ x̄ (2.8)
This shows, just as in simple linear regression, that the errors in variables regression
line also passes through the centroid (x̄, ȳ) of the data.
In order to ensure that the estimators for the variances are non negative, admissibil-
ity conditions must be placed on the equations. The straightforward conditions are
included below
More precisely, the estimate of the slope must lie between the slopes of the regression
line of y on x and that of x on y for variance estimators using (2.4), (2.5) and (2.6) to
be non negative.
As stated in Gillard and Iles [9], there is a variety of estimators for the slope. This
technical paper will concentrate on the asymptotic theory regarding some of these
estimators. The estimators whose variances are derived here will be stated later, but
for more detail concerning practical advice and also the derivation of the estimators,
the reader is once again referred to Gillard and Iles [9].
7
Chapter 3
Consider a first order Taylor expansion of a function of a sample moment x, f (x) where
E[x] = µ.
f (x) ' f (µ) + (x − µ)f 0 (µ) (3.1)
Upon taking the expectation of both sides of (3.1) the usual approximation
8
is found. Additionally,
V ar [f (x)] = E {f (x) − E[f (x)]}2 ' {f 0 (µx )}2 E[(x − µx )2 ]
The notation
n ∂f o ∂f
=
∂x ∂x x=E[x]
was introduced by Cramer to denote a partial derivative evaluated at the expected
values of the sample moments.
This can be naturally extended to functions of more than one sample moment. For a
function f (x, y)
n ∂f o2 n ∂f o2 n ∂f on ∂f o
V ar[f (x, y)] ' V ar[x] + V ar[y] + 2 Cov[x, y]
∂x ∂y ∂x ∂y
where
T ∂f ∂f
∇ = ,...
∂x1 ∂xp
is the vector of derivatives with each sample moment substituted for its expected value,
and
V ar[x1 ] Cov[x1 , x2 ] . . . Cov[x1 , xp ]
.. .. ..
V = . . .
Cov[x1 , xp ] Cov[x2 , xp ] . . . V ar[xp ]
is the p×p matrix containing the variances of and covariances between sample moments.
9
Chapter 4
Essentially, use of the method outlined above requires the prior computation of the
variance of each relevant sample moment, and the covariances between each sample
moment. For each of the restricted cases discussed by Gillard and Iles [9], the fol-
lowing variances and covariances are used. The variances and covariances needed to
compute the asymptotics for the higher moment based estimators will be stated later
on in this report.
σ 2 + σδ2
V ar[x̄] ' (4.1)
n
β 2 σ 2 + σε2
V ar[ȳ] ' (4.2)
n
(µξ4 − σ 4 ) + (µδ4 − σδ4 ) + 4σ 2 σδ2
V ar[sxx ] ' (4.3)
n
β 2 (µξ4 − σ 4 ) + σ 2 σε2 + β 2 σ 2 σδ2 + σδ2 σε2
V ar[sxy ] '
n
β 4 (µξ4 − σ 4 ) + (µε4 − σε4 ) + 4β 2 σ 2 σε2
V ar[syy ] '
n
10
The covariances between all first and second order moments are:
βσ 2
Cov[x̄, ȳ] ' (4.4)
n
µξ3 + µδ3
Cov[x̄, sxx ] '
n
βµξ3
Cov[ȳ, sxx ] '
n
βµξ3
Cov[x̄, sxy ] '
n
β 2 µξ3
Cov[ȳ, sxy ] '
n
β 2 µξ3
Cov[x̄, syy ] '
n
3
β µξ3 + µε3
Cov[ȳ, syy ] '
n
β(µξ4 − σ 4 ) + 2βσ 2 σδ2
Cov[sxx , sxy ] ' (4.5)
n
β (µξ4 − σ 4 )
2
Cov[sxx , syy ] '
n
β 3 (µξ4 − σ 4 ) − 2βσ 2 σε2
Cov[sxy , syy ] '
n
Expressions (4.1), (4.2) and (4.4) follow from the definition of the linear structural
model. To show how these may be derived, the algebra underlying expressions (4.3)
and (4.5) shall be outlined. For brevity the notation ξi∗ = ξi − ξ¯ and ηi∗ = ηi − η̄ is
introduced.
The above result also follows from the method of moment estimating equation stated
earlier, sxx = σ 2 + σδ2 .
11
X
n 2
2 1 2
E[(sxx ) ] = E (xi − x̄)
n2 i=1
X
n 2
1 2
= E (ξi∗ ) + (δi∗ )
n2 i=1
X
n
1
= E (ξi∗ )4 + 4(ξi∗ )3 (δi∗ ) + 6(ξi∗ )2 (δi∗ )2 + 4(ξi∗ )(δi∗ )3 + (δi∗ )4
n2 i=1
XX
1
+ E (ξi∗ )2 (ξj∗ )2 + 2(ξi∗ )2 (ξj∗ )(δj∗ ) + (ξi∗ )2 (δj∗ )2 + 2(ξi∗ )(ξj∗ )2 (δi∗ )
n2 i6=j
+ 4(ξi∗ )(ξj∗ )(δi∗ )(δj∗ ) + 2(ξi∗ )(δi∗ )2 (δj∗ ) + (ξj∗ )2 (δi∗ )2 + 2(ξj∗ )(δi∗ )(δj∗ )2 + (δi∗ )2 (δj∗ )2
1 2 2 4 2 2 4
' n(µξ4 + 6σ σδ + µδ4 ) + n(n − 1)(σ + 2σ σδ + σδ )
n2
Xn Xn
1 2
E[sxx sxy ] = 2 E (xi − x̄) × (xi − x̄)(yi − ȳ)
n i=1 i=1
Now, (xi − x̄) = (ξi∗ ) + (δi∗ ) and (yi − ȳ) = β(ξi∗ ) + (ε∗i ). Substituting these into the
12
above summation, and multiplying out leads to
Xn
1
E[sxx sxy ] = E β(ξi∗ )4 + 2β(ξi∗ )3 (δi∗ ) + β(ξi∗ )2 (δi∗ )2 + β(ξi∗ )3 (δi∗ ) + 2β(ξi∗ )2 (δi∗ )2
n2 i=1
+β(ξi∗ )(δi∗ )3 + (ξi∗ )3 (ε∗i ) + 2(ξi∗ )2 (δi∗ )(ε∗i ) + (ξi∗ )(δi∗ )2 (ε∗i ) + (ξi∗ )2 (δi∗ )(ε∗i )
∗ ∗ 2 ∗ ∗ 3 ∗
+2(ξi )(δi ) (εi ) + (δi ) (εi )
XX
1
+ 2E β(ξi∗ )2 (ξj∗ )2 + 2β(ξi∗ )2 (ξj∗ )(δj∗ ) + β(ξi∗ )2 (δj∗ )2 + β(ξi∗ )(ξj∗ )2 (δi∗ )
n i6=j
+2β(ξi∗ )(ξj∗ )(δi∗ )(δj∗ ) + β(ξi∗ )(δi∗ )(δj∗ )2 + (ξi∗ )(ξj∗ )2 (ε∗i ) + 2(ξi∗ )(ξj∗ )(δj∗ )(ε∗i )
∗ ∗ 2 ∗ ∗ 2 ∗ ∗ ∗ ∗ ∗ ∗ ∗ ∗ 2 ∗
+(ξi )(δj ) (εi ) + (ξj ) (δj )(εj ) + 2(ξj )(δi )(δj )(εi ) + (δi )(δj ) (εi )
1 2 2 2 2 4 2 2
' n(βµξ4 + βσ σδ + 2βσ σδ ) + n(n − 1)(βσ + βσ σδ )
n2
Hence,
13
Chapter 5
For each restricted case, and for the estimators of the slope based on the higher mo-
ments a variance-covariance matrix can be constructed. As there are six parameters
in the linear structural model µ, α, β, σ 2 , σδ2 and σε2 the maximum size of the variance
covariance matrix is 6 × 6. If the parameter space is restricted, then the size of the
variance covariance matrix will decrease in accordance with the number of assumed
parameters.
It is possible to use the delta method in order to construct ‘shortcut’ formulae or ap-
proximations to enable quicker calculation of each element of the variance covariance
matrix. Usually, these shortcut formulae depend on the variability of the slope esti-
mator and the covariance of the slope estimator with a first or second order sample
moment. In some cases the variances and covariances do not depend on the slope esti-
mator used, and as a result are robust to the choice of slope estimator. These shortcut
formulae are stated below, and repeating the style of the previous section, example
derivations will be given. For brevity, the notation |Σ| = σδ2 σε2 + β 2 σ 2 σδ2 + σ 2 σε2 is
introduced. This is the determinant of the variance covariance matrix of the bivariate
distribution of x and y.
14
5.1 Shortcut Formulae for Variances
Firstly, the shortcut formulae for the variances will be considered. V ar[α̃] will be the
example derivation provided.
σ 2 + σδ2
V ar[µ̃] =
n
β 2 σδ2 + σε2
V ar[α̃] = µ2 V ar[β̃] + + 2µ(βCov[x̄, β̃] − Cov[ȳ, β̃])
n
σ4 |Σ| + β 2 (µξ4 − σ 4 ) 2σ 2
V ar[σ̃ 2 ] = V ar[β̃] + − 2 Cov[sxy , β̃]
β2 β2n β
4
2 σ |Σ| + β (µδ4 − σδ ) 2σ 2
2 4
Cov[sxy , β̃]
V ar[σ̃δ ] = V ar[β̃] + + Cov[sxx , β̃] −
β2 β 2n β β
β |Σ| + (µε4 − σε4 )
2
V ar[σ̃ε2 ] = β 2 σ 4 V ar[β̃] + 2βσ 2 (βCov[sxy , β̃] − Cov[syy , β̃]) +
n
Derivation of V ar[α̃]
n ∂ α̃ o2 n ∂ α̃ o2 n ∂ α̃ o2
V ar[α̃] = V ar[ȳ − β x̄] = V ar[ȳ]+ V ar[β̃]+ V ar[x̄]
∂ ȳ ∂ β̃ ∂ x̄
n ∂ α̃ on ∂ α̃ o n ∂ α̃ on ∂ α̃ o
+ 2 Cov[ȳ, β̃] + 2 Cov[x̄, β̃]
∂ ȳ ∂ β̃ ∂ x̄ ∂ β̃
n ∂ α̃ on ∂ α̃ o
+ 2 Cov[x̄, ȳ]
∂ x̄ ∂ ȳ
β 2 σδ2 + σε2
= µ2 V ar[β̃] + + 2µ(βCov[x̄, β̃] − Cov[ȳ, β̃])
n
A similar shortcut formula was provided in the paper by Hood et al [11]. As out-
lined in Gillard [8], they assumed that (ξ, δ, ε) follow a trivariate Normal distribution.
They then used the theory of maximum likelihood to obtain the information matri-
ces required for the asymptotic variance covariance matrices for the parameters of the
model. Applying various algebraic manipulations to V ar[α̃] they showed that
β 2 σδ2 + σε2
V ar[α̃] = µ2 V ar[β̃] +
n
The shortcut formula derived above is a generalisation of that derived by Hood et al.
[11] to cope with non-Normal ξ. Indeed, if (ξ, δ, ε) do follow a trivariate Normal distrib-
ution, then as β̃ is a function only of second order moments (or higher), β̃ is statistically
independent of the first order sample moments. As a result Cov[x̄, β̃] = Cov[ȳ, β̃] = 0
15
and the shortcut formula derived above collapses to that suggested by Hood et al. [11]
βσδ2
Cov[µ̃, α̃] = − − µCov[x̄, β̃]
n
Cov[µ̃, β̃] = Cov[x̄, β̃]
µξ3 σ 2
Cov[µ̃, σ̃ 2 ] = − Cov[x̄, β̃]
n β
µδ3 σ 2
Cov[µ̃, σ̃δ2 ] = + Cov[x̄, β̃]
n β
Cov[µ̃, σ̃ε2 ] = 2
−βσ Cov[x̄, β̃]
The shortcut formulae for the covariances of α̃ with all other parameters are listed
here.
The shortcut formulae for the covariances of β̃ with the remaining parameters are listed
here.
1 σ2
Cov[β̃, σ̃ 2 ] = Cov[sxy , β̃] − V ar[β̃]
β β
Cov[sxy , β̃] σ 2
Cov[β̃, σ̃δ2 ] = Cov[sxx , β̃] − + V ar[β̃]
β β
Cov[β̃, σ̃ε ] = Cov[syy , β̃] − βCov[sxy , β̃] − βσ 2 V ar[β̃]
2
16
The shortcut formulae for the covariances of σ̃ 2 with the remaining parameters are
listed here.
2 σ4 σ2 2 |Σ| − 2σ 2 σε2 − 2σδ2 σε2
Cov[σ̃ , σ̃δ2 ]
= − 2 V ar[β̃] + Cov[sxy , β̃] − Cov[sxx , β̃] +
β β β β 2n
σ2 |Σ| − 2β 2 σ 2 σδ2 − 2σδ2 σε2
Cov[σ̃ 2 , σ̃ε2 ] = σ 4 V ar[β̃] − Cov[syy , β̃] +
β n
Finally, the covariance between the error variance estimates is
σ2 |Σ| − 2β 2 σ 2 σδ2 − 2σ 2 σε2
Cov[σ̃δ2 , σ̃ε2 ] = −σ 4 V ar[β̃] + Cov[syy , β̃] − βσ 2 Cov[sxx , β̃] + .
β n
Again, an example derivation is provided.
A first order Taylor expansion of σ˜2 around the expected values of sxy and β̃ is
1 σ2
σ̃ 2 = σ 2 + (sxy − βσ 2 ) − (β̃ − β) .
β β
Hence,
2 2 2 1 σ2
Cov[β̃, σ̃ ] = E[(β̃ − β)(σ̃ − σ )] = Cov[sxy , β̃] − V ar[β̃].
β β
The matrix B gives the additional terms that are necessary if ξ has non zero third
moment and a non zero measure of kurtosis. It can be seen that in most cases the B
17
matrices are sparce, needing only adjustment for the terms for V ar[σ̃ 2 ] and Cov[µ̃, σ 2 ].
The exceptions are the cases where the reliability ratio is assumed known, and slope
estimators involving the higher moments.
The C matrices are additional terms that are needed if the third moments and measures
of kurtosis are non zero for the error terms δ and ε. It is likely that these C matrices
will prove of less value to practitioners than the A and B matrices. It is quite possible
that a practitioner would not wish to assume that the distribution of the variable ξ
is Normal, or even that its third and fourth moments behave like those of a Normal
distribution. Indeed, the necessity for this assumption to be made in the likelihood
approach may well have been one of the obstacles against a more widespread use of
errors in variables methodology. The assumption of Normal like distributions for the
error terms, however, is more likely to be acceptable. Thus, in many applications, the
C matrix may be ignored.
As a check on the method employed the A matrices were checked with those given
by Hood [10] and Hood et al. [11], where a different likelihood approach was used in
deriving the asymptotic variance covariance matrices. In all cases exact agreement was
found, although some simplification of the algebra has been found to be possible. As
discussed earlier, the limitation of the likelihood approach is that it is limited to the
case where all random variables are assumed to be Normally distributed. The moments
approach described by Gillard and Iles [9] does not have this limitation.
18
Chapter 6
This section contains the variance covariance matrices for each of the restricted case
slope estimators outlined by Gillard and Iles [9]. These are
• Intercept α known
σ2
• Reliability ratio κ = σ 2 +σδ2
known
σε2
• Ratio of the error variances λ = σδ2
known
For brevity, the notation U = σ 2 + σδ2 , V = β 2 σδ2 + σε2 , e1 = µδ4 − 3σδ4 , e2 = µε4 − 3σε4
and e3 = βλµδ3 + µε3 shall be used.
19
6.1 Intercept α known
The method of moments estimator for the slope based on this assumption is
ȳ − α
β̃ = .
x̄
Since α is assumed to be known, the variance covariance matrix for µ̃, β̃, σ̃ 2 , σ̃δ2 and σ̃ε2
is required.
βσδ2 σ 2 σδ2 σ 2 σδ2 β 2 σ 2 σδ2
U − µ µ
− µ µ
2
V σ2
− βµ2V
σ2
V − βσ V
nµ2 βµ2 µ2
1
|Σ| σ4 σ4 σ4 2 2
A1 = β2
+ β 2 µ2
V + 2σ 4 − |Σ|
β2
− β 2 µ2
V − 2σ 2 σδ2 −|Σ| + µ2 V + 2σ σε
n
|Σ|
+ σ4
V + 2σδ4 |Σ| − σµ2 V − 2β 2 σ 2 σδ2
4
β2 β 2 µ2
2 β 2 σ4 4
β |Σ| + µ2 V + 2σε
0 0 µξ3 0 0
0 0 0 0
1
B1 = µξ4 − 3σ 4 0
−2σ 2 β 2 µξ3
n
µ
0 0
0
0 0 0 µδ3 0
0 0 −βµδ3 µ3
µ
1 σ2 σ2
C1 = 0 µ
µ δ3
− βµ µε3
n
µδ4 − 3σδ3 − 2 σµ µδ3
2 βσ 2
(βµ + µ 2
µ )
µ δ3 ε3
3 βσ 2
µε4 − 3σε − 2 µ µε3
20
6.2 Error variance σδ2 known
The method of moments estimator for the slope based on this assumption is
sxy
β̃ = .
sxx − σδ2
Since σδ2 is assumed known, the variance covariance matrix for µ̃, α̃, β̃, σ̃ 2 and σ̃ε2 is
required.
U −βσδ2 0 0 0
µ2 2µβσδ2 2µβσδ2
σ4
(|Σ| + 2β 2 σδ4 ) +V − σµ4 (|Σ| + 2β 2 σδ4 ) σ2
U σ2
V
1
A2 = − σ2 U − σ2 V
2
2βσδ 2
2βσδ
1
n σ4
(|Σ| + 2β 2 σδ4 )
2U 2 2 4
2β σδ
2V 2
0 0 0 µξ3 0
0 0 0 0
1
B2 = 0 0 0
n
µξ4 − 3σ 4 0
0
βµµδ3
0 σ2
− βµσ2δ3 µδ3 β 2 µδ3
2 β 2 µδ3
− 2β σµµ δ3
−βµδ3 µε3 − β µδ3 3
2 σ2
1
C2 = β2
3
e − σβ2 e1 − βσ2 e1
n
σ4 1
e1 β 2 e1
e2 + β 4 e1
21
6.3 Error variance σε2 known
The method of moments estimator for the slope based on this assumption is
syy − σε2
β̃ = .
sxy
Since σε2 is assumed known, the variance covariance matrix for µ̃, α̃, β̃, σ̃ 2 and σ̃δ2 is
required. For brevity, the notation W = (β 2 |Σ| + 2σε2 ) + V is introduced.
U −βσδ2 0 0 0
µ2
W − β 2µσ4 W 2µ
(σε2 V + β 4 σ 2 σδ2 )
− 2µσ
2
εV
β 2 σ4 β 3 σ2 β 3 σ2
1
A3 =
2σ 2V
1
W − β 32σ2 (σε2 V + β 4 σ 2 σδ2 ) ε
n
β 2 σ4 3
β σ 2
2 4 2 2 4 4 2 2 2 2 2
β4
(β U + V − 2β σδ ) − β 4 (σε V + 2β σδ σε )
2V 2
β4
0 0 0 µξ3 0
0 0 0 0
1
B3 = 0 0 0
n
µξ4 − 3σ 4 0
0
0 0 0 0 µδ3
− 2µµ ε3
βσ 2
µε3
βσ 2
− µβε32 µε3
β2
− βµδ3
1
C3 = 1
e − β 31σ2 e2 1
e
n
β 2 σ4 2 β 3 σ2 2
1
e − β14 e2
β4 2
1
β4
(β 4 e1 + e2 )
22
σ2
6.4 Reliability ratio κ = σ +σδ2
2 known
The method of moments estimator for the slope based on this assumption is
sxy
β̃ = .
κsxx
The variance covariance matrix for µ̃, α̃, β̃, σ̃ 2 and σ̃δ2 is required. For brevity, the no-
tation $ = 1 − κ is introduced.
U −βσδ2 0 0 0
µ2 |Σ|
σ4
+V −µ |Σ|
σ4
0 2µβ (1−κ)
σ2
|Σ|
1
A4 = |Σ|
0 −2β (1−κ) |Σ|
n
σ4 σ2
2σ 4 −2β 2 κσ 2 σδ2
4β 2 (1 − κ)|Σ| + 2σε4
β2$
0 −µ β$ µ
σ 2 ξ3 σ2
µξ3 µξ3 −β 2 $µξ3
0 0 0 0
1
B4 = β 2 $2
(µ − 3σ 4
) βκ$
(µ − 3σ 4
) − β 3 $2
(µ − 3σ 4
)
n
σ 4 ξ4 σ 2 ξ4 σ 2 ξ4
2 4 2 4
κ (µξ4 − 3σ ) −β κ$(µξ4 − 3σ )
β 4 $2 (µξ4 − 3σ 4 )
0 µ βκ µ
σ 2 δ3
− βκ µ
σ 2 δ3
κµδ3 β 2 κµδ3
2 β2κ
−2µ βσ2κ µδ3 µ
σ 2 δ3
3
−βκµδ3 −β κµδ3 + µε3
1
C4 = β 2 κ2
2 3 2
e − βκ e − βσκ2 e1
n
σ4 1 σ2 1
β 2 κ2 e1 β 2 κ2 e1
β 4 κ2 e1 + e2
23
σε2
6.5 Ratio of the error variances λ = σδ2
known
The method of moments estimator for the slope based on this assumption is
p
(syy − λsxx ) + (syy − λsxx )2 + 4λ(sxy )2
β̃ = .
2sxy
The variance covariance matrix for µ̃, α̃, β̃, σ̃ 2 and σ̃δ2 is required.
U βσδ2 0 0 0
µ2 |Σ| +V −µ |Σ| 2µβ
|Σ| 0
σ4 σ4 (β 2 +λ)σ 2
1
A5 = |Σ| 2β
− (β 2 +λ)σ 2 |Σ| 0
n
σ4
4|Σ| 2σ 2 σ 2
2σ 4 + − (β 2δ+λ)
ε
(β 2 +λ)
2σδ4
0 0 0 µξ3 0
0 0 0 0
1
B5 = 0 0 0
n
µξ4 − 3σ 4 0
0
µλβ λβ β2
λ
0 µ
(β 2 +λ)σ 2 δ3
− (β 2 +λ)σ 2 µδ3 µ
(β 2 +λ) δ3
µ
(β 2 +λ) δ3
µβ β β
−2 (β 2 +λ)σ 2 e3 e
(β 2 +λ)σ 2 3
− (β 2e+λ)
3
e
(β 2 +λ) 3
− βµδ3
1
β 2 e2 +λ2 β 2 e1 2 3
C5 = (β 2 +λ)2 σ 4
− (βe 2 +λ βe1 )
(β 2 +λ)2 σ 2
βe2 −λβ e1
(β 2 +λ)2 σ 2
n
e2 +λ2 e1 (e2 +λβ 2 e1 )
− (β 2 +λ)2
(β 2 +λ)2
e2 +β 4 e1
(β 2 +λ)2
24
6.6 Both variances σδ2 and σε2 known
The method of moments estimator for the slope based on this assumption is
s
syy − σε2
β̃ = sgn(sxy ) .
sxx − σδ2
U −βσδ2 0 0
0 0 0
1
A7 =
n
|Σ|
+
(β 2 σδ2 −σε2 ) βσδ2 2
− σ2 (U + σ )
σ4 2β 2 σ 4
µξ4 − 3σ 4
0 0 0 µξ3
0 0 0
1
B7 =
n
0 0
µξ4 − 3σ 4
µβµδ3
0 2σ 2
− βµ δ3
2σ 2
µδ3
2 β2
1
− βσ2µ µδ3 − µ
µ
βσ 2 ε3
µ
2σ 2 δ3
+ µε3
2βσ 2
−βµδ3
C7 =
n β2
e
4σ 4 1
+ e3
β 2 σ4
− 2σβ 2 e1
e1
25
Chapter 7
The methodology underlying the derivation of the asymptotic variances and covari-
ances for estimators based on higher moments is identical to that outlined previously.
However, the algebraic expressions for the variances and covariances of higher moment
based estimators are longer and more cumbersome than those for the restricted pa-
rameter space. As a result, the full variance covariance matrices for higher moment
estimators will not be reported here. However, the expressions needed to work out the
full variance covariance matrices for the slope estimator based on third moments will
be provided. These expressions can then be substituted into the shortcut formulae to
derive the full variance covariance matrices.
26
between them.
Using the method illustrated in deriving V ar[sxx ] and Cov[sxx , sxy ], the required co-
variances between the second order and third order moments are:
By using the methodology outlined earlier, we can now obtain the variance of our slope
estimator β̃8 , and the covariances of our slope estimator with the first and second order
27
moments.
β 2 µξ4 σε 2 + β 4 µξ4 σδ 2 + 2βµξ3 µε3 + σ 2 µε4 + σδ 2 µε4 − 6β 2 σ 2 σδ 2 σε 2
V ar[β̃8 ] =
β 2 µξ3 2 n
2β 4 µξ3 µδ3 + β 4 σ 2 µδ4 + β 2 µδ4 σε 2 − 2βµδ3 µε3
+
β 2 µ2ξ3 n
σδ 2 σε 2 − 2β 2 σ 2 σδ 2 + σ 2 σε 2
Cov[x̄, β̃8 ] =
βµξ3 n
2σ σε − σδ 2 σε 2 − β 2 σ 2 σδ 2
2 2
Cov[ȳ, β̃8 ] =
µξ3 n
−3β µξ3 σδ − 3β 2 σ 2 µδ3 + µξ3 σε 2 + µδ3 σε 2
2 2
Cov[sxx , β̃8 ] =
βµξ3 n
2βµξ3 σε + σ µε3 + σδ 2 µε3 − 2β 3 µξ3 σδ 2 − βµδ3 σε 2 − β 3 σ 2 µδ3
2 2
Cov[sxy , β̃8 ] =
βµξ3 n
3βµξ3 σε + 3σ µε3 − σδ µε3 − β 3 µξ3 σδ 2
2 2 2
Cov[syy , β̃8 ] =
µξ3 n
We now have each of the components needed to use the shortcut formulae to obtain
the following variance covariance matrix for the parameters µ, α, β, σ 2 , σδ2 and σε2 when
the estimator β̃8 is used.
In order to use the shortcut equations, the quantities Cov[x̄, β̃9 ], Cov[ȳ, β̃9 ], Cov[sxx , β̃9 ],
Cov[sxy , β̃9 ] and Cov[syy , β̃9 ] are needed. Further, to obtain these quantities, the co-
variances between each of the first and second order moments (x̄, ȳ, sxx , sxy , syy ) and
the fourth order moments that occur in β̃9 (sxyyy , sxxxy ) must be obtained. Also, the
variances of these fourth order moments must be obtained, as well as the covariance
between them. The formulae for these quantities are very lengthy and full details will
be given in Gillard [?]. However, there is a potential difficulty.
As can be seen from the shortcut formulae, a key component of the variance covariance
matrices is V ar[β̃]. For the estimator of the slope based on fourth moments, V ar[β̃]
28
depends on the sixth moment of ξ. High order moments may be difficult to estimate
reliably, so the authors believe further work is needed to establish whether this formula
is of practical value. Again, full details will be given by Gillard [8].
29
Bibliography
[15] P. Sprent. Models in Regression and Related Topics. Methuen’s Statistical Mono-
graphs. Matheun & Co Ltd, London, 1969.
31