Professional Documents
Culture Documents
Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at http://www.jstor.org/page/
info/about/policies/terms.jsp
JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content
in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship.
For more information about JSTOR, please contact support@jstor.org.
Institute of Mathematical Statistics is collaborating with JSTOR to digitize, preserve and extend access to The Annals of
Statistics.
http://www.jstor.org
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
TheAnnalsof Statistics
1981,Vol. 9, No. 1,24-44
Purdue University
In a multivariate "errors invariables"regression model,theunknown mean
vectors ul: p x 1,U2t: r x 1 ofthevectorobservations xi,,X2,,ratherthanthe
observations themselves, areassumedtofollowthelinearrelation: U2,= a + Bui,
= 1,2, . . ., n. It is furtherassumedthattherandomerrorset= x,- u, x I =
x'2,),ut= (ul'i,u'2,),arei.i.d.random vectorswithcommon covariance matrix Ie-
Sucha modelis a generalization oftheunivariate (r = 1) "errorsin variables"
regression modelwhichhasbeenofinterest tostatisticiansforovera century.
In thepresent paper,itis shownthatwhen1e = 2Ip+r, a wideclassofleast
squaresapproaches toestimation oftheintercept vectora andslopematrix B all
lead to identical estimators & and B oftheserespective parameters, and that&
andB arealsothemaximum likelihoodestimators (MLE's) ofa andB underthe
assumption of normally distributederrorse,. Formulasfor&, B and also the
MLE's U1 anda2 oftheparameters U1 = (ull,* **,uln)anda2 aregiven.Under
reasonable assumptions concerning theunknown sequence{ul,,i = 1,2, *. . ), &,
B and r1'(r + p)a2 are shownto be strongly (withprobability one) consistent
estimators of a, B and a2, respectively, as n -- oo,regardless of thecommon
distribution of theerrorse,. Whenthiscommonerrordistribution has finite
fourth moments, &, B and r-'(r + p)a2 are also shownto be asymptotically
normally distributed. Finallylarge-sample approximate 100(1- v)% confidence
regions fora, B anda2 areconstructed.
where
(1.2) u21= a + Bu1l, i= 1,29 * , n,
ReceivedAugust1976;revised December1979.
1This research
was supported by theAir ForceOfficeof Scientific Research, AirForceSystems
Command, USAF underGrantsAFOSR-72-2350B andAFOSR-77-3291 at PurdueUniversity.
AMS 1970subjectclassiflcations.
Primary 62H10,62E20;secondary 62F10,62F25,62H25,62P15,
62P20.
Keywordsandphrases.Errorsin variables, linearfunctional
equations,factor multivariate
analysis,
leastsquares,generalized
linearregression, leastsquares,orthogonally
invariant norm,maximum like-
lihood,strongconsistency,
asymptotic distributions,
approximateconfidence regions,
principal
compo-
nents. 24
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
"ERRORS IN VARIABLES" REGRESSION 25
and that
(1.3) theei's are i.i.d. withmean vector0 and covariancematrixle.
The parametersB: r x p and uii: p x 1, i = 1,2, ** *, n are assumedto be unknown,and are
to be estimated.The parametera: r x 1 is eitherknownto be 0 (theno-intercept
model),or else
is unknown(the intercept model) and mustbe estimated.If consistentsequencesof estimators
forB are desired,theparametricformof le cannotbe completelyunspecified(see Section5).
In thepresentpaper,it is assumedthatle = U2Ip+r wherethescalar a2 > 0 is unknownand is
to be estimated.That is, theassumption(1.3) is specializedto thefollowing:
(1.3') The e's are i.i.d. withmean vector0 and covariancematrixle = a2Ip+r.
X (X U= ( :l (P + r) x n,
have columns xi, x2, *.. , xn and ul, U2, *.. , un, In termsof thesematrices,the
respectively.
model (1.1), (1.2) becomes
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
26 LEON JAY GLESER
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
"ERRORS IN VARIABLES" REGRESSION 27
(1.12) B
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
28 LEON JAY GLESER
2. MLE and GLSE estimators. Because the analyses of the no-intercept and intercept
formsof the model (1.4) involvesimilarsteps,an attemptis made in thefollowingsectionsto
treatthesetwo formsof themodel simultaneously. To thisend let
(2.1) Cn = I,, fortheno-intercept
model,
= In - llnln, fortheinterceptmodel.
The estimatorsconsideredin thispaper are based on the followingreductionof the data X:
Let
(2.2) W = XCnX,
and let di : d2 **dp 2 dp i . dp+,' 0 be the (ordered)eigenvaluesof W. Let
satisfy
(2.5) G'G = Ip+r= GG'
(2.6) W= GDG'.
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
"ERRORS IN VARIABLES" REGRESSION 29
I
Therefore,G22#1 0 and G22is nonsingular.A similarargumentshowsthatthenonsingularity
of G22impliesthatGii is nonsingular.0
For any s x t matrixA, s c t,let X,[A] be the ith largestnonzerosingularvalue ofA; that
is, X[A] is the ithlargesteigenvalueofAA', or ofA'A, and
>As[A]? O.
JA] : A2[A]- **:*.-
Recall thatifA is an s x s positivesemidefinite
matrix(A is p.s.d.),thenX1[A],* * X,[A] are
also theeigenvaluesofA. If A1,A2 are s x s symmetric matrices,thenotationA1 > A2 means
thatA1 - A2 is p.s.d. It is well knownthatifA1,A2 are s x s symmetric,
(2.12) : XJ[A2],
A1 A2 impliesthatAX[AI] i = 1,2, *, S.
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
30 LEON JAY GLESER
g(D?1PX) = g(X)
for all X: s x 1, all diagonal matricesD?1 having ?1's on the diagonal, and all s x s
permutation matricesP.
LEMMA2.4. Let g(.) be an s.g.f.on s-dimensional
space.
(i) The functiong*(Xi) definedon si-dimensionalEuclidean space, si S s, by g*(xl) =
g((XI, 0')') is an s.g.f.
(ii) If XA'1= (XM),... , X
X())', j = 1, 2, and IX( iC 1 i = 1, 2, s, theng(X1l))
g(X(2)).
Q(a, B; X) Ijo?
11 11Q(a*(X), B*(X); X)I|o,
whereQ(a, B; X) is defmedby (1.5).
THEOREM 2.1. If G-1 and G-1 exist,thenoi(X), B(X), (1(X) definedby (2.9), (2.7) and
(2.8) are OLSE's ofa, B, U1relativetoanyorthogonally normon (p + r) X n matrices,
invariant
andfurtheroi(X) and B(X) are GLSE's relativeto any orthogonallyinvariantnormon r X n
matrices.
= rT(B)Q(a, B; X),
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
"ERRORS IN VARIABLES" REGRESSION 31
where
(2.18) 17(B) = (IJ + BB')-1/2(-B, Ir)
matrix.In consequence,p singularvaluesofR(a, B, U6((a,B);
is an r x (p + r) row-orthogonal
X) are zero,withtheremainingr singularvalues beingequal to thesingularvalues of Q(a, B;
X). From Lemma 2.3 and Lemma 2.4(i), thereis an orthogonallyinvariantnormjj jodefmed
on r x n matricesQ such that
(2.19) 1jR(a, B, U1(a,B); X) jj = jjQ(a, B; X) Ilo,
forall a, B, X. Notingthat U1 = U01(Q,B), it followsthatifit can be shownthat(xand B are
GLSE'$ of a and B relativeto any orthogonally invariantnorm11jlo on r x n matrices,then
the OLSE-characterof (x,B, U1 realtiveto 11 11will followas an immediatecorollary.
Hence, let lobe any orthogonally invariantnormdefmedon r X n matricesQ. Fix B and
note thatforall a, X,
withequalityholdingforall i in (2.24) if
(2.25) 1r = A(G21, G22)
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
32 LEON JAY GLESER
(2-30) f p + r)
~~~~~~~~n( npr/
(2.30)
{(2re)(tr[Dmb}])
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
"ERRORS IN VARIABLES" REGRESSION 33
whichholds forall X forwhich Gi1and G22are nonsingular,all a, B, U1 and a2. Since the
multivariatenormaldistribution is absolutelycontinuouswithrespectto Lebesgue measure,
it followsfromLemma 2.2 that(2,32) holdswithprobabilityone. Since theclassicalEuclidean
normis strictlymonotonic,it followsfromTheorem2.2 thatd, B, L1}are the unique (with
probabilityone) estimatorsof a, B, U1 thatachieve,the upper bound in (2.32). Finally,it is
well knownthatthe upper bound in (2.32) is uniquelymaximizedas a functionof a2 by 62.
Plugging62 intotheright-hand side of (2.32) yields(2.30). 0
where S = (n - 1)-1 W = ((sj1)) is the sample covariance matrix. That this formula yields the
same resultas (2.7) can be seen by notingthattheeigenvaluesdi and d2of Ware solutionsof
the quadraticequation
0 = d (tr[W])d + I WI = d (n - l)(sll + s22)d + (n - S)2( 22 -12),
so that
(2.34) d (n 1) {sll + S22+ [(s51 + S22)2- 4(s11S22 -12)
gg - (n - I)1 di -
g2lgll S12
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
34 LEON JAY GLESER
is positivedefinite.
Let
A and Assumption
LEMMA3.1. Under(1.4), Assumption B,
(3.5) limn,) n 1W = 0, withprobability1 (w.p. 1).
n1 = n_1UE' -ue
UCnEt
where
(3.9) a = n-1Uln, e = n-'Eln.
By AssumptionA, the SLLN and (3.2),
(3.10) limno ii = IL, limn-+ae = 0, w.p. 1.
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
"ERRORS IN VARIABLES" REGRESSION 35
Let y > y2 > * >- -yp-0 be the (ordered) eigenvalues of A(Ip + B'B), and let D, =
diag (yi, y2, ** , yp). Using the fact that the yi's are also eigenvalues of (Ip + B'B)112A[(IP +
B'B)112]' for any square root (Ip +B'B)112 of Ip + B'B, it can be shown that thereexists
4':p x p nonsingularforwhich
(3.13) A(1p + B'B)41 = D-y, 4'(Ip + B'B)4= Ip.
COROLLARY3.1. Undertheassumptions
ofLemma 3.1,
(3.17) lim,. n lDmmj = U I2 W.p. 1,
and thus
(3.18) limh P1(p + r)62 = a2 W.p.1.
In the proofof Lemma 3.3, W = (xl, x2, ...) is a point in the probabilityspace of all
sequencesof observations, n on a samplequantity(e.g., W(n) indicatesthat
and a superscript
thatquantityis calculatedfromthe firstn elementsof w. This notationis discontinuedonce
Lemma 3.3 is provun.
PROOF. Fix X such that(3.5) and (3.16) hold. Note thatsinceeach G(n)is orthogonal,the
sequence G(n)} lies in a compactsubspaceof (p + r)2- dimensionalEuclidean space. Thus,
each subsequenceof { G (n)) has a convergentsub-subsequencewithlimit,say,
H= (Hi, H12\
kH21 H221J
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
36 LEON JAY GLESER
nW ((n)
iF1
) D)))
(n
D ),
(n)
G21 \ G21
of H that
thiseigensubspaceis unique. Hence, it followsfrom(3.14) and the orthogonality
thereexistsZ :p x p orthogonalsuch that
Note that the limitof (Ginl)(Gil)' over the indices of any convergentsub-subsequenceof
{G (n) } iS
as can be seen from(3.20), the fact that Z is orthogonal,and (3.13). Since this limit is
independentof thesub-subsequence,factsabout limitsof sequencesin Euclidean space imply
that
(3.21) limn,. (GiPl)(Gil)' = (Ip + B'B)1.
Using the factthatthe determinant of the limit(Ip + B'B) 1 is positiveand (3.21), it is now
straightforward to showthatthereexistsn(co),0 c n(co)< 00, such thatG(n)is nonsingularfor
all n - n(co),provingthe firstpartof the lemma.Equation (3.19) is establishedby a similar
argument,noting firstthat for all convergentsub-subsequencesof { G) that (Gil)1
eventuallyexists,and thennotingfrom(3.20) thatthe limitof B(n) = G(nl)(G (n))-1 is H21Htl
= B independentof the sub-subsequence.0
The second equalityin (3.23) is a consequenceof (2.5), (2.6), (2.7) and (2.10).
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
"ERRORS IN VARIABLES" REGRESSION 37
Taking limitsas n -m00 in theequality(3.26), using(3.4), (3.5), (3.19) and (3.25), yields
4. Asymptotic
distributions.From (3.6), themean of W is
(Ip, B')[n S
(W-(W))](B, - I)' = [n-2(B -B)]'(n inG22DmG22)(Ir + BB')
(4.2)
- (Ip + B'B)(n-1GiiDmaxG'i,)[n 12(B - B)]'.
are identical.
PROOF. The assertionof Lemma 4.1 is a directconsequenceof (3.19), (3.25), (3.27), and
(4.2). 0
From (2.6), (2.7), (2.10) and (4.1),
+ r)&2_ a2] = tr{G22G22(B,
rn1/2[r-1(p - Ir)[n-112(WJ_- (J(W))](A, - Ir)')
(4.4)
+ tr{G22G22[n1/2(B -B)][n -1 U C U'][B-B]').
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
38 LEON JAY GLESER
LEMMA 4.3. In theintercept model,assumethatn /2( - ii) and n 1/2(W - &(W)) havean
asymptotic and thatAssumptions
jointdistribution, A, B, C hold.Thentheasymptoticdistributions
ofn 1/2(& - a) and
(4.7) m= (-B, Ir)[n1/2(S - U)]-
Define
(4.10) Zk = eke;k- Ip+r + (Uk t)e + ek(Uk -
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
"ERRORS IN VARIABLES" REGRESSION 39
z\Vll V12I
whileundertheintercept
model (Cn = In- n-llnln)
since,by Corollary4.1, n1/2e-L MVN(O, a2Ip+r) and since,by (3.2), lim .(ii -A) = 0. Since
also n 12(x - ii) = n /2e the following theorem is a direct consequence of (4.13), (4.14) and
Corollary4. 1.
+ ys,6'(ej'eiej) + Hsg'(ee,eje,)
(ii) In the intercept model,the elementsof n-11/2(W- &(W)) on or below the diagonal,
together withtheelementsofn1/2(X - ii), havean asymptotic 1/2(p+ r + 3)(p + r)-variatenormal
distribution with means zero and with covariance matrix V*. Thus,for example,V1* =
((vt j),(i,,j))) gives the asymptoticcovariancesbetweenthe (i, j)th and (i', j')th elementsof
n -1/2(W - &(W)).
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
40 LEON JAY GLESER
where
p = u2{1 + 1,L[Cu2A-1(I + B'B)1A1 + A1]U1},
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
"ERRORS IN VARIABLES" REGRESSION 41
moreconvenient
whichis equivalentto thefollowingcomputationally region:
where
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
42 LEON JAY GLESER
and in the presentmodel the uti'splay the role of the vectorof independentvariables.Using
simulationsin thecase p = r = 3, Gleserand Watson(1973) under-sampledone dimensionin
3-dimensionalspace. The resulting estimatesof B bore littleresemblanceto thetruevalue. On
the otherhand, when all dimensionswere nearlyequally sampled,the estimatesof B were
reasonablyaccurateeven in moderatelysmall samples.
Withrespectto assumption(ii), it should be notedthatif the requirement in Assumption
A that 2e = U2Ip+r is replacedby the moregeneralrequirement
(5.1) le= 2 zo, Io known,
ze= 2QP9O )
Then~fthecommondistribution
wherea2, K are unrestricted. of thecolumnsofE is multivariate
normal,thelikelihoodof thedata has an infinite
supremum.
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
"ERRORS IN VARIABLES" REGRESSION 43
REMARK 1. The argumentsin the proofof Theorem 5.1 easily extendto cover cases in
which1e has theform
e
1
(Ki/21 1/2) l/2O ( )
I
2 Ir
1/? c2X2
wherelo can eitherbe fixedor allowed to varyover any class of positivedefinitematrices,
and whereK1, KC2 > 0 are unrestricted.
REFERENCES
[1] ADCOCK, R. J. (1878).A problem inleastsquares.TheAnalyst5, 53-54.
[2] ANDERSON,T. W. (1948). The asymptotic distribution equations.
oftherootsofcertaindeterminantal
J. Roy. Statist.Soc., Ser. B 10 132-139.
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions
44 LEON JAY GLESER
DEPARTMENT OF STATISTICS
PURDUE UNIVERSITY
WEST LAFAYETTE, IN 47907
This content downloaded from 147.188.128.74 on Wed, 11 Nov 2015 12:37:08 UTC
All use subject to JSTOR Terms and Conditions