Professional Documents
Culture Documents
)
f ¡ ..
.
AN INTRODUCTION TO
LINEAR STATISTICAL MODELS
Volun1e I
FRANKLIN A. GRAYBILL
L-'rojl'.~sor
of 1lfoth e111ulical S tatistics
Colorad o S tutc University
F ort Colli11.$, Colorado
ISBN 07-024331-x
12 13 14 15 16-MAMM-7 6 5 4 3 2
FAC. CIENCIAS
Preface
This book was \vritten with the intention of fulfilling three needs:
( l) for a theory textbook in experimental statistics, for undergraduates
or first-year graduate students; (2) for a reference book in the area of
regression, correlation, least squares, experimental design, etc., for
consulting statisticians with limited mathematical training; and (3)
for a reference book for experimenters with limited mathematical
training who use statistfos in their research. This is not a book on
mathematics, neither is it an advanced book on statistical theory. It
is intended to be an introductory mathematical treatment of topics
which are important to experimenters, statistical consultants, and
those who are training to be statistical consultants. The only mathe-
- matics that is required is generally obtained before the senior year in
college. The mathematical leve] of this book is about the same as
"An Introduction to the Theory of Statistics" by A.M. Mood, except
for the fact that a great <leal of use is made of matrix theory. Chapter l
is a review of the mathematical concepts, especially matrices, that
will be used in the book. For students who have had a course in
matrix algebra, Chapter 1 can be omitted, since in most cases the
theorems are stated without proof and simple problems are given to
illustrate them. Chapter 2 is a review of those statistical conccpts
which are usually obtained in a first course in mathematical statistics.
In these two chapters the important concepts are presented in the
form of theorems so that they can be referred to as they are needed in
later chapters. In the remainder of the book sorne of the proofs are
omitted, but most of them are presented in detail, since this book was
written for those who will have limited mathematical training.
vii
viii PREFACE PREFACE •
IX
Among other things, the statistical consultant is involved in aiding Professor Osear Kempthorne, first for his hel durin
in the interpretation of observations that are collected by experi- work at Iowa State College and Iater for read. pth
'
? my graduate
mg e entire manuscript
menters. This interpretation can often be materially iinproved by the I a lso wan t t o express my appreciation toproJ.essor
r L . Wayne J ohnson ·
use of "statistical models." This use involves three parts. The first t.he hea d of t lie Mathematics Department of Oklahoma St . '
is the specification of the structure of the observations, or, in other s1ty, wh~ made personn~l available for typing and proofre:: Un~v~
words, the selection of a model to represent the real-world situation manuscr1pt, and to David Weeks and Van . ng_o t e
that the observations describe. The second part is the n:iathematical students irr statistics at Oklahoma State a~a~a1 ~eshadr1, graduate
treatment of the model to obtain results about various components of entire manuscript. mversity, who read the
it. This is generally in the nature of point estimation, interval I wish to thank the typists who helped to t h d .
estimation, hypothesis testing, etc. The third part is the use of the into typescript; Miss Nancy Adams M' puC m~ an wr1tten notes
results of the mathematical treatment of the model to make dccisions Carlotta Fine, and Miss Biruta Stakle' iss aro yn Hundley, Mrs.
in the real world. This book is concerned only with the sccond part, I am indebted to Mr F W B jy
namely, the mathematical treatment of statistical models, and no of Standards at Boulde; C~l:r::' fi irector ~f t.he National Bureau
t. f h ' ' or perm1ss1on to reproduce
attempt is made to justify any model for a given real-world situation. por 10n o t e material in NBS report 5069 Gra h
Percentage Points F( ) fi h . '
d
P s an Tables of the
ª
Five important statistical models are discusscd, which are the basis for
many of the applications in statistics, and these ti.ve models are to b bl. h d b L 'V1,~2,P or t e F1sher-Snedecor Variance Ratio
e pu is e y ew1s E. Vogler and Kenneth A No t '
presented from the point of view of infinite-model theory. Through- I am also indebted to Professor E S p J. r on.
out the book emphasis is placed on the power of tests and on the rrang, an d U mvers1ty
. . :J. •
College for permission to
earson . Neyman p C
'. ' . .
width of confidence intervals. The Power Function of the Analy.sis of V . rTeprmt t~e tables from
A set of notes from which this book was taken has served as a two- rn us t rat1ons
· r r ariance ests with Tables and
J.Or Their Use, from Statisti l R .
semester course for seniors and first-year graduate students at Okla- volume l. ca esearch Memoirs,
homa State University; the prerequisites are one year of calculus, one I am also indebted to the Biometrika Truste . .
semester of matrix algebra, one year of introductory statistical theory reproduce a portion of "Tables of p ~s for perm1ss1on to
from a book such as "Introduction to the Theory of Statistics," by Beta (F) Distribution" by M M . ercentage Pomts of the Inverted
F' . · errmgton and C. M. Thom son
A.M. Mood, or "Statistical Theory in Research," by R. L. Anderson . mally' I w1sh to expresa my thanks to . p . .
and T. A. Bancroft, and one year of statistical methods. Man y of the reheved me of most of the household h d ~y w1fe, J eanne, who
students who were not statistics majors had no formal course in that I worked on the manuscript. e ores urmg the many evenings
matrix theory, but sorne had audited a matrix course, and sorne were
self-taught. These students were required to work the pertinent Franklin A. Graybill
problema in Chapter l.
Volume II will contain such topics as sample size, multiple com-
parisons, multivariate analysis of variance, response surfaces, dis-
criminant functions, partially balanced incomplete block designa,
orthogonal latin squares, randomization theory, split-plot models, and
sorne nonlinear models.
This book was written with the direct and indirect help_ of many
people. I owe a great debt to my mother, Mrs. J...ula Graybill, and to
my two sisters and their husbands, Mr. and Mrs. Wayne Carr and
Mr. and Mrs. Homer Good, for their help and encouragement. Had it
not been for them, I probably would not have been able to attend
college.
I want to thank Professor Carl E. Marshall for introducing me to
statistics and for encouraging me to do further work in this field, and
Coutents
Preface . vii
· Chapter 1. Mathematical Concepts . 1
1.1 Matrices . 1
1.2 Quadratio Forms 3
1.3 Determinants 6
1.4 Miscellaneous Theorems on Matrices . 6
1.5 The Derivativas of Matrices and Vectors . 11
1.6 ldempotent Matrices 12
1. 7 Maxima, Minima, and J acobians . 17
Chapter 2. Statistical Concepts . 27
2.1 Sample Space, Random Variable, Frequency Function 27
2.2 Statistical lnference. 32
2.3 Point Estimation 33
2.4 Interval Estimation. 37
2.5 Testing Hypotheses . 39
2.6 Concliision 46
1 Chapter 3. The Multivariate Normal Distribution 48
3.1 Definition 48
3.2 Marginal Distributions . 51
3.3 Moments of the Multivariate Normal. 53
3.4 Linear Functions of Normal Variables 56
3.5 Independence 57
3.6 The Conditional Distribution . 62
3. 7 Additional Theorems 67
1 74
Chapter 4. Distribution of Quadratic Forms .
4.1 Introduction. 74
4.2 Noncentral Chi-square . 75
xi
1
CONTENTS xiii
CONTENTS
xii 12.3 Two-way Cross Classification: No Interaction . 258
77
4.3 N oncentral F 12.4 N-way Cross Classification: No Interaction . 261
82
4.4 Distribution of Quadratic Forros · 12.5 Two-way Classification with Interaction . 265
84
4.5 Independence of Quadra:tic Forros . 86 12.6 N-way Classification with Interaction 272
Independence of Linear an~ Quadrat1c Forros
4
4.6 87 12. 7 2n Factorial l\fodels • 279
12.8 Using Interaction Sum of Squares for Error Sum of Squares 283
4. 7 Expected Value of Quadrat1c Forros · · · 88
4.8 Additional Theorems · · Chapter 13. Two-way Classification with Unequal Nwnbe1·s in Subclasses. 287
93
/':.i Chapter 5. Linear ModeIS 93 13.l Two-way Classificatiori l\fodel with No Interaction. 287
5.1 Introduction. 96 13.2 Computing Instructions 302
5.2 Linear Models 103 Chapter 14. Incomplete Block J.\fodels 306
5.3 Model Classification .
106 14.1 Int.roduction. 306
Chapter 6. Model l: The General Linear Hypothesis of Full Rank
106 14.2 Point Estimation 309
6.1 Introduction . · llO 14.3 Interval Estimation and Tests of Hypotheses 312
6.2 Point Estimation 120 14.4 Computing 313
6.3 Interval Estimation · 128
6.4 Tests of Hypotheses Chapter 15. Sorne Additional Topics about Model 4 318
149
Chapter 7. Coroputing Techniques 15. l Introduction • 318
149 15.2 Assumptions for Model 4 318
7 .1 Introduction · · · · . . 149 15.:J Tests of Hypotheses 320
7 .2 Solving a System of Symmetr1c Equa~1ons ·.. 155 15.4 Test for Additivity . 324
7 .3 Coroputing the Inverse of a Symroetr1c Matr1x
165 15.5 Transformation . 332
Cha.pter 8. Polynomial or Curvilinear Models .
165 Chapter 10. Model 5: Variance Components; Point Estimation 337
8.1 Introduction · · · · · '. · . · · ' · ·~l Mod~l · : : . 165
8.2 Estimating and Testing Coefficie~ts m a Poly~:m1 ffven Set ofData 166 16.l One-way Classification: Equal Subclass Numbers . 338
8.3 Finding the Degree of a Polynom1al that Descr1 es a 1 . . . . 172 16.2 The General Balanced Case of Model 5 . 347
16.3 Two-way Classification . 348
8 .4 Orthogonal Polynomials · · · · 183
16.4 The Balanced Twofold and Fourfold Nested Classification of Model 5 349
8 . 5 Repeated Observations for Each X 186 16.5 One-way Classification with Unequal Numbers in the Groups. 351
Cha.pter 9 . Model 2: Functional Relationships 16.6 Twofold Nested Classification in l\fodel 5 with Unequal Numbers in the
186
9 . 1 Introduction and Definitions · 187 Subclasses 354
9.2 Point Estimation · · · • · · · · 193 16.7 The Unbalanced Two-way Classification in l\fodel 5 359
9.3 Interval Estimation and Tests of Hypotheses 16.8 The Balanced Incomplete Block in Model 5 . 362
195
I Chapter 10. Model 3: Regre~sion Models Chapter 17. Model 5: Var;iance Components; Interval Estimation and
195
10.l Introduction · · · . · · · • 197 Tests of Hypotheses · 368
10. 2 Case 1: The M,_ultivar1ate Normal· 206 17 .1 Distribution of a Linear Combino.tion of Chi-square Varia.tes . 368
10.3 Correlation · 216 17 .2 Tests of Hypotheses 374
l o.4 Case 2, Model 3 217 17.3 Ratio of Variances . 378
10.5 Case 3, Model 3
223 Chapter 18. Mixed Models 383
Chapter 11. Model 4: Experimental Design Models
.· 223 18.1 Covariarice 383
11.1 Introduction · 226 18.2 Two-way. Classification l\fodel with Interaction and with Fixed and
11.2 Point Estima.tion ·. 241 Random Effects . 396
11.3 lnterva.1 Estimation · 241 18.3 Balanced Incomplete Block with Blocks Random: Recovery of
11.4 Testing Hypotheses · · · · . · 245 Interblock Information . 403
11.5 Normal Equat.ions and Computmg 250
11.6 Optimum Properties of the Tests of Hypotheses Appendix. Tables 421
254
Cha.pter 12. The Cross-classification or Factorial Model lndex 461
254
12.l Introduction · · · ·. · · · 255
12.2 The One-way Olassificat1on Model.
1
Mathematical Concepts
1.1 Matrices
The theory of linear statistical models that will be developed in this
book will require sorne of the mathematical techniques of calculus,
matrix algebra, etc. In this chapter we shall give sorne of the impor-
tant mathematical theorems that are necessary to develop this theory.
Most of the theorems in this chapter will be given without proof;
for sorne, however, the proof will be supplied.
A matrix A will have elements denoted by aH, where i refers to the
row andj to the·column. If A denotes a matrix, then A' will denote
the transpose of A, and A - 1 will denote the in verse of A. The symbol
IAI will be used to denote the determinant of-A. The identity matrix
will be denoted by 1, and O will denote the null matrix, i.e., a matrix
whose elements are all zeros. The dimension of a matrix is the number
of its rows by the immber of its columna. For cxample, a matrix A of
dimension n X m, or an n X m matrix A, will be a matrix A with n
rows and m columns. If m, = 1, the matrix will be called an n x l
vector. The rank ofthe matrix A will sometimes be denoted by p(A).
Given the matrices A =(a¡;) and B =(bu), the product AB = C =
n
(cii) is defined as the matrix e with pqth element equal to L ªpsb sq• For
8=1
AB to be defined, the number of columns in A must equal the number
ofrows in B. For A + B to be defined, A and B must ha ve the samc
dimension; A + B = e gives C¡¡ ~-a;¡ + b¡;. If k is a scalar and
A is a matrix, then kA means the matrix such that each element is the
corresponding element of A multiplied by k.
l
2 LINEAR STATISTICAL l\lODELS MATHEMATICAL CONCEPTS 3
A diagonal matrix D is defined as a square matrix whose off-diagonal that AB and BC exist and if A and C are nonsingular then
elements are ~11 zero; that is, if D = (dii), then d¡; = Oif i -=F j. p(AB) = p(BC) = p(B). '
+ Theorem 1.1 The transpose of A' equals A; that is, (A')' =A. + Theorem 1.18 If the product AB of two square matrices is O,
then A = O or B = O or A and B are both singular.
+ Theorem 1.2 The inverse of A-1 is A; that is, (A-1)-1 =A.
+ Theorem 1.19 If A and B ar~ n x n matrices of rank r and 8
+ Theorem 1.3 The transpose and inverse symbols may be per- respectively, then the rank of AB is greater than or equal t~
inuted; that is, (A')-1 = (A-1)'. r + s - n. ·
+ Theorem 1.4 (AB)' = B.' A'. + Theorem 1.20 The rank of AA' equals the rank of A' A equals the
+ Theorem 1.5 (AB)-1 = B-1 A-1 if A and B are each nonsingular. rank of A equals the rank of A'. ·
+ Theorem 1.6 A scalar commutes with every matrix; that is, 1.2 Quadratic Forms
kA = Ak.
+ Theorem 1.7 For any matrix A we have IA =Al= A. .If ~is an n X 1 vector with ith element y¡ and if A is an n x n matrix
w1th iJth elcmcnt equal to a¡;, then the quadraticf orm Y' A Y is defined as
+ Theorem 1.8 All diagonal matrices of the same dimension are n n
commutative. .I .I YiY;ª;;·
t=l J==l
The rank of the quadratic form Y' AY is defined as the
+ Theorem 1.9 IfD 1 and D 2 are diagonal matrices, then the product rank of the matrix A. The quadratic form Y' A Y is said to be positive
is ~iagonal; that is, D 1D 2 = D 2D 1 = D, where D is diagonal. d~finite if and only if Y' A Y> O for ali vectors Y where. Y =I= O. A
The ith diagonal element of D is the product of the ith diagonal quadratic form Y' AY is said to be positive semidefinite if and only if
element of D 1 and the ith diagonal element of D 2 • Y'A1:"·~ O for all Y, a~d Y'AY =O for sorne vector Y -=FO. The
matr!x A .ºf a quadratw form Y' A Y is said to· be positiva definite
+ Theorem 1.10 If X and Y are vectors and if A is a nonsingular
(senu.defimte) '~hen the quadratic form is positive definite (semi-
matrix and if thc cquation Y = AX holds, then X = A-1Y.
defimte). If e IS an n X n matrix such that C'C = 1, then e is said
+ Theorem 1.11 The rank of the product AB of the two matrices to be an ortlwgonal matrix, and C' = c-1.
A and B is less than or equal to the rank of A and is less than or Consider the transformation from the vector z to the vector y by
equal to the rank of B. t~e , matrix P such that Y. = PZ. Then, Y' A Y = (PZ) 'A(PZ) =
+ Theorem 1.12 The rank of the sum of A + B. is less than or equal Z P A~Z. .Thus, by the transformation Y.= PZ, the quadratic
to the rank of A plus the rank of B. form Y AY is transformed into the_quadratic form Z'(P'AP)Z.
+ T.heorem 1.13 If A is an n x n matrix and if IAI =O, then the • Theor~m 1.2~ If .Pis a nonsingular matrix and if A is positive
rank of A is less than n. ( IAI denotes the determinant of the . defimte (sem1defimte), then P' AP is positive definite (semidefinite).
matrix A.) + Theorem 1.22 A necessary and sufficient condition for the
+ Theorem 1.14 If the rank of A is less than n, then the rows of A symmetric matrix A to be positive definite is that there exist a
are not independent; likewise, thc columns of A are not inde- nonsingular matrix P such that A = PP'. ·
¡)endent. (A is n x n.) • Theorem 1.23 A necessary and sufficient condition that the
+ Theorem 1.15 If the rank of A is m ~n, then the number of · ~atrix A be positive definitc, where
linearly independent rows is m; also, the number of linearly
independent columns is m. (A is n x n.)
+ Theorem 1.16 If A'A =O, then A ~O.
+ Theorem 1.17 The rank of a matrix is unaltered by multiplication
A = (:: ::: ... ::¡
by a nonsingular matrix; that is, if A, B, and C are matrices such ~.~ . ~.~ ....... ~.~
4 LINEAR STATISTICAL MODELS MATHE.MATICAL CONCEPTS 5
1i
is that the following inequalities hold: + Theorem 1.31 For every symmetric matrix A tl .
orthogonal matrix C such that C' AC - D. ·h ie:c ex.1sts an
.· . w h ose d'1agonal elements are th- 1 '·'\ ere. D• is a d1auonal
mat 11x 0
e e 1aracter1strn roots of A.
ª11 ª12 I·>0, ... , + Theorem 1.32 Let A A A b ·
>0 • 1, 2, • · • , t e a collect1on of sym metric
1 ª21 ª22 n .?< n matrices. A necessary and sufficient condition that there
ex1sts ~n orthogonal ~ransformatiOñ C such that C' A C C' A C
1 2 '
· · · , C AtC are all diagonal is that A A b
. d . s·
.'
i i e symmetr1c for all
+ Theorem 1.24 If A is an n x m matrix of rank m < n, then i an J.t . ~fnce dall the A; are symmetric, it follows that A A is
A' A is positive definite and AA' is positive semidefinite. symme r1c l an only if A; and A, commute. ¡ i
+ Theorem 1.25 If A is an n x m matrix of rank k < m and + Theorem 1.33 Let an n X n matrix C be written
k < n, then A' A and AA' are each positive semidefinite.
+ Theorem 1.26 If C is an orthogonal matrix, and if the trans-
formation Y = CZ is made on Y'Y, we get Y'Y = Y'IY =
Z'C'ICZ = Z'C'CZ = Z'Z.
In order to develop the theory of quadratic forms, it is necessary to
define a characteristic root of a matrix. A characteristic root of a
p x p matrix A is a scalar A such that AX = .A.X for sorne vector where C;is the ith row of c. Thus e 1-8 th t f ·
X =I= O. The vect.or X is called the characteristic vector of the matrix A. l i e ranspose o an n x 1
coffiu~n vt~ctoCr. The following two conditions are necessary and
It follows that, if A is a characteristic root of A, then AX - .A.X = O· su cien .ior to be orthogonal:
and (A - .A.l)X =O. Thus, A is a scalar such that the above homo-
( l)
geneous set of equations has a nontrivial solution, i.e., a solution other for ali i -=!= .i
than X =O. It is known from elementary matrix theory that this (2)
for ali i
implies IA - .HI =O. Thus the characteristic root of a matrix A
could be defined as a scalar A such that IA - .A.11 =O. It is easily That is t~ s~y, any two rows of any orthogonal matrix are ortho -
seen that IA - .UI is a pth degree polynomial in .A.. This polynomial onal (the1r mner product is zero) and th . g
row with itself is unity. ' e mner product of any
is called the characteristic polynomial, and its roots are the characteristic
roots of the matrix A. We shall noi.v give a few theorems concerning
characteristic roots, characteristic vectors, and characteristic poly-
nomials. In this book the elements of a matrix will be real.
+ Theorem 1.27 The number of nonzero characteristic roots of a
• Theore~ 1.34 Let C - (
• · P
::¡
x! - . . . be f
P rows o an n X n orthogonal
matrix A is equal to the rank of A. Cv
+ Theorem 1.28 The characteristic roots of A are identical with the matrix. That is to sav let e
characteristic roots of CAC-1 • lf C is an orthogonal matrix, it vectors suc'Ii that e¡¿ ~ O (i' ;
2
'. ·~. ' e,, be the transpose~ of p
follows that A and CAC' have identical characteristic roots. (i - 1 ? ) i J - l, 2, · · · , p) and e.e. = 1
- , -, ... , ? . ~hen there exist n - p vectors L su~h ti1 t
+ Theorem 1.29 The characteristic roots of a symmetric matrix C; f J = O for all i andJ and f~f. = 1 (i = 1 i , ª
·f · -1- • 'I,h '· 1 '
2' • • • , n - P) f f - O
are real; i.e., if A =A', the characteristic polynomial of IA - .U¡ I i ,_J. us the theorem states that if we are i ' i i -.
= O has all real roots. such as C1, there exists a matrix C2 of di:Uension (n g ve) .a matr1~
C 1) - p X nsuc
+ Theorem 1.30 The characteristic roots of a positive definite that ( C = C (C1 forms the first p rows of C and C the last n _ p
matrix A are ¡Jositive; the characteristic roots of a positivo 2 2
In this section a few of the important theorems on determinants snm of the diagonal elements of A; that is, tr(A) = 2, aii·
¡.,. ¡
will be given. It will be assumed that the student knows the definition • Theorem 1.45 tr(AB) = tr(BA).
of a determinant and knows how to evaluate small ones. In linear- Proof: By definition, tr(AB) is equal to 2,a¡;bii. By definition,
hypothesis applications it is often necessary to..§.olve systems involving ij .
a great ma~y equations. It might at times be necessary to evaluate tr(BA) is equal to¡ b,kaki· But it is clear that 2, a;;bii = 2, bikaki;
~ ij ~
large determinants. There m·e many methods of doing these tlúngs therefore, tr(AB) = tr(BA).
that are adaptable to automatic and semiautomatic computing
machines. These methods will be discussed in detail later. It will • Theorem 1.46 tr(~BC) = tr(CAB) = tr(BCA); that is, the
be assumed here that the student knows how to evaluate determinants trace of the product of matrices is invariant under any cyclic
by the method of minors or by sorne other simple method. permutation of the matrices.
Proof: By Theorem 1.45, tr[(AB)C] = tr[C(AB)].
+ Theorem 1.35 The determinant of a diagonal rnatrix is equal to
the product of the diagonal elements. • Theorem 1.47 tr(I) = n, where I is an n x n identity matrix.
+ Theorem 1.36 If A and B are n x n matrices, then IABI = • Theorem 1.48 If C is an orthogonal matrix, tr(C' AC) = tr(A).
IBA! = IAI IBI. Proof: ByTheorem 1.46, tr(C'AC) = tr(CC'A) = tr(IA) = tr(A).
+ Theorem 1.37 If A is singular, IAI = O.
It is sometimes advantageous to break a matrix up into submatrices.
• Theorem 1.38 If C is an orthogonal matrix, then ICI = +1 or 'fhis is callcd partitioning a matrix into submatrices, and a matrix can
ICI = -1. be partitioned in many ways. For example, A might be partitioned
+ Th~orem 1.39 If C is an orthogonal matrix, then IC' ACI = l~I into submatrices as follows:
+ Theorem 1.40 The determinant of a positive definite
positive.
~atrix is A= (Au Aj2
A21 ~9
+ Theorem 1.41 The determinant of a triangular matrix is equal
to the product of the diagonal elements. where A is m X n, A 11 is m 1 x n 1 , A 12 is m 1 x n 2 , A 21 is m 2 x n 1 ,
and A 22 is m 2 x n 2 , and where m 1 + m 2 = m and n 1 + n 2 = n.
+ Theorem 1.42 1n-11 = I/IDI, if !DI -:F o. The product AB of two matrices can be made symbolically even
+ Theorem 1.43 If A is a square matrix such that if A and B are broken ~nto SQbmatrices. The multiplication proceeds
as if the submatrices were single elements of the matrix. However,
the. dimensions of the matrices and of the submatrices must be such
that they will multiply. For example, if Bisan~ x p matrix such
where Au and A 22 are square matrices, and if A 12 = Oor A 21 =O, that
th_en IAI = IA11l IA22I· B= (Bn Bj2
+ Theorem 1.44 If A 1 and A 2 are symmetric and A 2 is positive defi- B 21 B
nite and if A 1 - A 2 is positive semidefinite (or positive definite),
wherc Bii is an ni X P; matrix, then the product AB exists; and the
then IA 1 1 ~ IA2I· corresponding submatrices will multiply, since A¡1 is of dimension
m; X n; and B;k is of dimension n 1 x P1:· The. resulting matrix is as
1.4 Miscellaneous Theorems on Matrices follows:
In this section we shall discuss sorne miscellaneous theorems con-
cerning matrices, which we shall use in later chapters.
l\:IATHEMATICAL CONCEPTS 9
8 LINEAR STATISTICAL MODELS
This is clear since, by Theorem 1.43,
+ Theorem 1.49 If A is a positive definitc symmetric matrix such
that IBI = IIA;21I = IA2ll
B = (Bn B12)
B21 B22
~I
A~)I
and if Bu and Au are each of dimension mi X mi, etc.,
A;l = B11 - B12B;iB21
Proof: Since A= B-1 , then AB =l. Thus, The corresponding sub matrices are such that they multiply; so
An Bn - AnB12B2f B21 = 1
Multiplying by Ali 1 gives the desired result. . .
It is known that Bti1 and Añ 1 exist, since A and B are pos1t1ve
definite matrices and since A 11 and B 22 are principal minors of A
and B, respectively. By Theorerr_i 1.23, t.he. dete~"?inant ?f ~he
principal minor of a positive defimte matr1x is pos1tive. S1m1lar
1 For a given set of aii and Y; (that is to say, for a given matrix A and
equations can be derived for At:i.1, Bñ1, and B; •
vector Y), does there exist a set of elements xi (that is, a vector X)
+ Theorem 1.50 Le.t the square matrix A be such that such that the equations AX = Y are satisfied ~ Three cases must be
considered:
A= (Au A12) 1. The equations have no solution. In this case there exists no
~l A.¿2
vector X such that the system of equations is satisfied, and the system
is said to be inconsistent.
lf A 22 is nonsingular, then IAI = IA 22 l IA11 - A12A~ A21I·
1
2. There is just one set of X¡ that satisfies the system. In this case,
Proof: The determinant of A can be written as follows: there is said to exista unique solution.
3. There is more than one vector X that satisfies the system. If
more than one such vector X exists, then an infinite number of vectors
1 o) exist that satisfy ~he system of equations.
where B= ( We shall consider two matrices: the coefficient matrix A and the
-A;;iA21 A;f
10 LINEAR STATISTICAL l\IODELS
r
:1
~
t MATHEl\IATICÁL CONCEPTS 11
~t
augmented matrix B = (A, Y), which is the matrix A with the vector
Y joined to itas the (ni + l)st column; that is to say, (~ t.5 The Derivatives of Matrices and Vectors
;
" We shall now...--discuss sorne theorems relating to the differentiation
of quadratic forros and bilinear forms. lt will sometimes be advan-
tageous in taking derivatives of quadratic and bilinear forms to be
B= able to take derivatives of matrices and vectors.
Let X be a p X l vectOr with elements x,, let A be a p x 1 vector
with elemen~s ai, and let Z = X' A = A'X (Z is a scalar). The deriv-
a.ti ve of Z w1th respect to the vector X, which will be written oZ/oX,
We shall now state sorne important theorems concerning solutions will mean the vector
to the system of equations AX = Y. ·
az
+ Theorem 1.51 . A necessary and sufficient condition that the
system of equations AX = Y be consistent (have at least one
vector X satisfying it) is that the rank of the coefficient matrix
A be equal to the rank of the augmented matrix B = (A, Y).
+ Theorem 1.52 If p(A) = p(B) = p, then m - p of the unknowns
x, can be assigned any desired value and the remaining p of the
x, will be uniquely determined. It is essential that the m - p of
the unknown xi that are assigned given values be chosen such that
the matrix of the coefficients of the remaining p unknowns have + Theorem 1.54 If X, A, and Z are as defined above, then ·
rank p. az¡ax =A.
+ Theorem 1.53 lf p(A) = p(B) = m < n, there is a unique vector X P1·00J: To find the ith element of the vector oZ/oX, we find
that satisfies AX =Y.
Asan example, consider the system of equations _a_z = a(;-::
f 1 a.x.)
, '.
x1 - X2 =6 ax,
2x1 -2x2 =3 which equals a,. Thus the ith element of oZ/oX is a,; so
az¡ax =A.
This can be put into matrix formas
+ Theorem 1.55 Let A be a p x 1 vector, -B be a q x 1 vector, and
,, Let
X be a p X q matrix whose ijth element equals x...
f/ 11
Z = A'XB = k,
"" ~
k, a n'x nmbm
It can easily be verified that the rank of the augmented matrix 111==1 n=l
Then az¡ax = AB'.
(
1 -1 ª) Ptooj: éJZ/oX will be a p X q matrix whose ijth element is
2 -2 3 o Z/ oxiJ· Assuming that X is not symmetric and that the elements
of X are independent,
is 2. Therefore, the system of equations is not consistent, and there
exist no values x 1 and x 2 that satisfy it. This fact is also easily seen
if we multiply the first equation by 2 and subtract it from the second az
equation. We get O = -9, an impossible result.
12 LINEAR STATISTICAL l\IODELS
MATHEl\IATICAL CONCEPTS 13
Thus the ijth element ?f oZ/oX is aib;. Therefore, itfollows that ' use of idempotent matrices in our ensuing work. A square matrix A
¡8 a synunetric idempotem matrix if the following two conditions hold:
oZ = AB'
ax 1~
ij
( 1) A=A'
+ Theorem 1.56 Let X be a p x 1 vector, let A be a p X p sym-
p p
á
¡;
(2) A=A2
metric matrix, and let z = X' AX = 2 L xixp¡¡; then az¡ oA = "
tJ For brevity we shall omit the word "symmetric." That is to say
i=lj=-1
2XX' - D(XX'), where D(XX') is a diagonal matrix whose fl when we say a matrix · is idempotent, we shall mean symmetri~
diagonal elements are the diagonal elements of XX'. ~~ idempotent. We shall make no use whatsoever of idempotent
Proof: By oZ/oA we shall mean a matrix whose ijth element is ~·~ matrices that are not symmetric.
oZ/oa¡;. Thus, ~ ~
+ Theorem 1.58 The characteristÍc roots of an idempotent matrix
{t J, "'m"'nªm•)
~-
. are either zero or unity.
iJZ = Proof: If A is idempotent and if A is a characteristic root of A
ºª'iJ ºªiJ (: there exists a .vector X =F Osuch that AX = .!.X. If we multipl;
.,
both sides by A, we get .
lfi =j, oZ/oaii = x¡. Ifi =Fj, then oZ/oail = 2x¡X; (remember-
ing that a¡¡ =a;¡}. Thus oZ/oA = 2XX' - D(XX'). A2X = Á.AX = ).2X
+ Theorem 1.57 Let X be a p x 1 vector and let A be a p x p But A 2X =AX= .A.X; so we have
symmetric mat.rix such that Z = X' AX; then oZ/ oX = 2AX.
Proof: The derivative of the scalar Z with respect to the vector X : A.X= A. 2X
will mean a p X 1 vector whose ith element is oZ/ox¡. (..\
2
- Á)X =O
az acx'AX) aCt J 1
xmxnamn)
But X =F O; so A. 2
-:- ). must .be zero.
+ Theorem 1.59· If A is idempotent and nonsingular, then A = I.
Thus J. = o or ). = l.
ª" = i=l
n
I a7;
Proof: Suppose (1) and (2) are given. Then B = I A¡ is idem-
1
potent, and there exists an orthogonal matrix p su~h that
But if aii = O, then a¡; = O (for j = 1, 2, .•. , n); that is, the
elements of the ith r~w are all zero. But A = A'; so the
elements of the ith column are also ali zero.
P'BP = (~ :)
~na r~x arn nonnegative; so, 1f their sum is zero, they must all be
+ Theorem 1.65 If A is idempotent and P is orthogonal, P' AP is 1dentrnally zero. Also bv Theorem I • 62 , the Jast p - r rows and
J
idempotent. p - r columns of each P' AiP must be zero. Thus we may write
Proof: (P'AP)(P'AP) = (P'A)(AP) = P'AP.
o
' º)
0
(1) A 1, A~, ... , Amare each idempotent.
1>& Since C'B,C is idempotent, by Theorems 1.61 and 1.62 we haYe
(2) The sum B = LA¡ is idempotent.
i=l
(3) AiA; =O for all i =F j. i = l, 2, ... , m; i =fo t ·
16 LINEAR STATISTICAL MODELS
l\IATHEMATICAL CONCEPTS 17
where K; is an {r - t) X (r - t) matrix. Thus we see that
Proof: If any .two conditions of Theorem 1.68 hold, this implica
C'B;CC'B,C = O that there ex1sts an orthogonal matrix P such that the following
are true:
which implies
P'BP =(~' :)
:~
... (a)
and i = 1, 2, ... , ·m; i =F l where the rank of B is r
2
B = (
m
i=l
)2
L A; = L
i=l
m
A¡ +2
i>Fi
A;A; =
m
2
i=l
A¡ =B
(b)
P'A P=(~· :);PV={:
1
o
o
1,:: :} ... ;
o
We have shown that the sum is idempotent, and condition (2)
is satisfied.
Finally, suppose (2) and (3) are given. By Theorem 1.67,
P'V=(: Ir,,,
o :)
there exists an orthogonal matrix P such that P' A 1 P, P' A 2 P,
where the rank of A, = 1·;· Thus the result follows .
. . . , P' AmP are each diagonal (since A;A; = A;Ai = O). Since
the sum of diagonal matrices is a diagonal matrix, it also follows
that P'BP is diagonal. By condition 3, we know that 1.7 Maxima, Mínima, and Jacobians
P' A;PP' A;P = Ofor ali i =F j. But the product of two diagonal
We shall now state some theorems concerning the maxima and
matrices P' A¡P and P' A;P is zero if and only if the corresponding minima of functions.
nonzero diagonal elements of P' A;P are zero in P' A;P. Thus,
if the tth diagonal element of P' A;P is nonzero, the tth diagonal + Theorem 1.70 If Y = f(x1,X 2, .•• ,xn) is a function of n variables
element of P' A 1P (for allj =F i) must be zero. Since P'BP = Er, ~nd if a~l partial d~ri:atives oy/axi are continuous, then y attains
the tth diagonal elemcnt must be O or 1 for each P' A¡P. For, 1ts max1ma and mnuma only at the points whei:e
if the tth element of P' A;P is equal to k =F O, then the tth diagonal
element of the remaining P' A¡P (i =F j) is zero. But the tth
diagonal element of B is O or l and is the sum of the tth diagonal
elements of P' A;P (i = 1, 2, ... , m). Thus k = O or k = l.
Since P' A;P is diagonal, the characteristic roots of A; are dis-
played dowll' the diagonal, and, since these roots are either
O or 1, A¡ is idempotent, and the proof is complete.
Theorem 1.71 If f(x1,x 2 , ••• ,x11) is such that all the first and
It is of special interest to note that, if second partial derivatives are continuous, then at the point where
m
2A;=1
i=l
the function ·has
then condition (2) of Theorem 1.68 is satisfied. In this situation,
condition ( l) implies condition (3) and vice versa. ( 1) a minimum, if the matrix K, where the ijth element of K is
02.f/ ox;ox1, is positive definite. . ·
+ Theorem 1.69 If any two of the three conditions of Theorem (2) a maximum, if the matrix -K is positive definite.
tn
1.68 hold, the rank of 2 A; equals the sum of the ranks ofthe A¡. In the above two theorems on maxima and minima, it must be
i=l remembered that the xi are independent variables. Many times it is
._
18 LINEAR STATISTICAL l\IODELS
subject to the condition h(x 1 ,x 2 , ••• ,xn) = O. Since the x, are not f~
{~
independent, Theorems l. 70 and l. 71 will not necessarily give the í1 The x 1 , X2, • • • ,xn, l can now be considered n + 1 independent vari-
aesired resu}t. If the equation k(x 1,x 2 , ••• ,xn) = 0 could be so}ved 3:z ables. We now state the theorem:
fJ;=;
for x t such that
~ + Theorem 1.72 Iff(x 1 ,x 2 , ••• ,xn ) and the constraint h(x l> x 2' • • • '
j;i
xn ) = O are such that all first partial derivatives are continuous
then this value of xt could be substituted into /(x 1 ,x 2 , ••• ,xn) and ~ then the m~ximum or mínimum oí /(x 1 ,x 2 , ••• ,xn) subject t~
Theorems 1.70 and 1.71 could be applied. ~ the constramt h(x 1,x 2 , ••• ,xn) = O can Qccur only at a point
Asan example, suppose we want to find the minimum of .f =xi - ~ where. hthe
. derivatives
h of F = f(x ¡, x 2> ... , x)
n - .lh(xl' x 2' • • • ,Xn)
2x 1 + ~ - 6x 2 + 16. Using Theorem 1.70, we get i vams ; 1.e., w ere
~
of = 2x¡ - 2 =o oF oF
-=-=···=----0
oF oF
OX¡ OX¡ OX2 oxn - o.A. -
a¡
-=2x2 -6=Ü
if oh/ox, =P ofor ali i at the point.
OX2 Thus we now have n + 1 equations and n + i unknowns and we
The solutions yield x 1 = 1, x 2 = 3. need not worry about which variables are independent for w~ treat all
n + l as ií they were independent variables. '
The matrix K is given by
This will be generalized in the following.
• Theorem 1.73 To find the maximum or minimum of the function
f~x1,X2, · .. ,xn) subject to the k constraints h¡(x 1 ,x 2 , ••• ,xn) = O
K=
1: = 1, 2, ... , k),
(i form.. the function F =f(x1> x 2> • • • 'x) n
-
.I l,h,(x1,x2, · · · ,xn)· If oF/ox, (i = 1, 2, ... , n) are continuous
then/, subject to the constraints, can have its ma~ima and minima
i= 1
K is positive definite; so f has a minimum at the point x 1 = 1, x 2 = 3.
'
If we want to find the minimum of f subject to the condition only at ~he points where the following equations are satisfied (if
x + x = 1, we proceed as follows. Substitute x 1 = 1 - x 2 into a Jacob1an loh,/ox;I =PO at the point):
1 2
the f function, and proceed as before. This gives
j = (1 - x2) 2 2(1 - ~) +~ - + 16 oF al, oF oP"' aF oF
- 6x2
OX¡ = OX2 = ... = a-x-n = -OA-1 = -º~- = ... = -OA-k = o
oj = -2(1 - x2 ) + 2 + 2xi - 6 = Ü
OX2 Let g(x1,X2, · • · ,xn)? where - oo < x, < oo (i = 1, 2, ... , n), repre-
senta fre~~ency-dens1ty function. This is equivalent to the following
The solution gives x 2 = .¡.. In this case the matrix K consista oí a two cond1t10ns on the function g(x l' x 2' • · • ' x n ) ·•
single term o 2f/ox~. So K = 4 and is positive definite. Thus /,
subject to the constraint x 1 + x 2 = 1, attains a minimum at the (1) g(x11x 2 , ••• ,xn) ~ O
point x 1 = -!, x 2 = !.
or
If the constraint is a complicated function if there are many
constraining equations, this method may become cumbersome. An
alternative is the method of Lagrange m:uUipliers. For example, if
(2)
Í ao f f
00
-oo -oo
• • • 00
__ 00
g(x1,X2, • · • ,xn) dx¡ dx., • • • dx = 1
- n
20 LINEAR STATISTICAL MODELS
In the example abo ve, if we solve the equations for y 1 and y 2 , wc get
21
¡~)
6
r
4 8 6 2
3 10' 11 15 1
J = -6 and IJI = 6. So we have 8
9 8 4 4 13
k(y1 ,y2 ) = ~ cC<4111+11tl +(2111-112l l
11 2
6 2
7T
1.2 Find the invcrsc of thc matrix
Thus it is quite clear that the Jacobian J will play an important part
in the theory of probability distributions. 2
J
and if the equations satisfy mild regularity conditions, and the
solutions for the y, yield
i·= 1, 2, ... , n
A=(: -7
verify that rank A = rank A' = rank AA' = rank A'A.
then the J acobian J (if J =I= O) can also be given by l/IK*I, where 1.4 Prove Thcorcm 1.16.
the ijth element of K* is oy~/ox 1 • 1.5 Let
This theorem might be useful if the equations wcre such that the
oxi/oy; were difficult to obtain but the oy¡fox1 were relatively easy to A = (: _: -:) B = ·(: : :)
obtain. 8 1 7 4 3 5
22 LINEAR STATISTICAL MODELS
r·· l\fATHEMATICAL CONCEPTS 23
Find AB, BA, p(A), p(B), p(AB), and p(A + B), and verify Theorems 1.11,
1.12, and 1.19. Find a set of linearly independent rows of A.
1
~
Show that the ort.hogonal matrix
1 1
1.6 If t
¡,. v'2
o
v'2
and rt. C'= l
l
~
- -v2 o v'2
show that (AB)' = B'A'. Also show that (AB)-1 = s-1A-1 •
1.7 Given o l o
A=(! !) B = ( t
. -!
-!)! is such that C'AC is a diagonal matrix with characteristic roots on the diagonal.
1.16 Prove the following: If C is an orthogonal matrix, then C' is an orthog-
onal matrix and c-1 is an orthogonal matrix.
show that AB =O, and thus verify Theorem 1.18. t.17 If A and B are orthogonal matrices, prove that AB is an orthogonal
1.8 If matrix.
o o
( ( -~2) ~2
1.18 Let
D, = (:
4
o
_:) and D2 = :
-1
o :) and C=
l
show that D D 2 = D 2D 1 is diagonal, and thus verify Theorems 1.8 and 1:9. v'2 v'2
1
t.9 If A is any n x m matrix, prove that AA' and A'A a~e symm.etr1c. Show that C is orthogonal. Find the characteristic roots of A and of C'AC, and
t.10 Given the quadratic form 2Xf +ax¡ + .6X1 X 2 ,.find its matr1x ~- . show they are equal.
1.11 If C is any matrix and A is a symmetr1c matr1x, prove that C AC 18 1.19 Let l 1 1 v'2
symmetric. 3 3 3 v'i
1.12 Given the matrix
l 1 1 1 l 2
o
v'4 v'4 v'4 v4 v6 - v'6
1 1 l l
o o o o
v'2 v'2 v2 - v2
C= 2 1
1 1 Find a vector C 2 such tha.t C = (C ) is orthogonal, thus illustrating Theorem
o
v'6 v'6 - v'6 1.34. C2
/t.20 Prove that, if a matrix A is positive definite, then A' and A-1 are also
1 1 1 3 positive dcfinite. -
v'12 v'12 v'12 v'12 1.21 Prove that, if D is a diagonal ma.trix, where d¡ is the ith diagonal
element, and if di :F O for any i, then the inversa of D exists, and the ith diagonal
prove that e is orthogonal. 2 2 • • element of n-1 is l/di.
1.13 Given the quadratic fonn !XJ + 3X1~2 + 4X2 + 6X3, find its matr1x. .1.22 Given the following systems of equations, find which are consistent and
Test to see whether the matrix is positiva defimte. which are inconsistent. Find solutions where possible.
1.14 Given the matrix X1 + 2X2 + 3X3 = 6
p = (6 2 (a) X1 - X2 2
3 1 X1 - = X3 -1
verify that PP' is positiva definite and that P'P is positive semidefinite.
X1 - X2 + 2X3 = 2
1.15 Given the matrix (b) X X X 3 = -1
4 o 2) 2X1
1 -
- 2X2
2 -
+ X3 = 2
A=
( o O 3
2
O
4
.
X1 + x-;-+ X 3 + X 4 = 8
~ ~-~-~-~=6
find the characteristic polynomial and the characteristic roots. 3X1 + X 2 + X 3 + X 4 = 22
24
1.23
LINEAR STATISTICAL MODELS
l
~
t.27
f X'AX =2 L ,2xix;
Ixr + p2 i=lj=l
find aQ/aX. 'i': p i=l
1.24 Given the matrix i:;Ci
f
A= (12 13 2)1 ¡:-
e
But X'AX = Í
p p
L x,xiaii = i=l
p
Í xiau + ,2 ! x,xiai;
p p
i=lj=l i=lj=l
i:;Cj
ations AA'X = Y has a vector X satisfying the f. The coefficients of the xi term are the diagonal elements of A, which are equal to
show tha.t the system o f equ . Th t · t
Y · al to any column of the matr1x A. a is o say, p-2 in this case. The coefficient of x,xi (i :;é j) is the ijth element of the matrix A,
system when the vector is equ k f h ted
show that the rank of the coefficient matrix AA' equals the ran o t e augmen which in this case equals p- 2 • Thus, in the example, the matrix A has every
. (AA' , A·)
ma t r1x , , where A·i is any column of A. element equal to p- 2 • ( ¡ p )
1.28 Use Prob. 1.27 to find the matrix of the quadratic form px2 x = - X¡ , !
1.25 lf and show that the matrix of this quadratic form is idempotent. pi =-1
p p
(~
1 2 -1 -3
i=l i=l
o 1.30 In Prob. 1.29, find the matrix of the quadratic form for each of the three
A= ( -1
2 1
2
2
_:)· B =
1 -2 2
4
•-1
~.
i-=l
p
I ex,_-:- x>2, px2.
-2 -2 1 2 1 -4 l.31 In Prob. 1.30, show that each matrix is idempotent.
2
1.32 Let B 1 represent the matrix for px2. Let B 2 represent the matrix ofthe
p
! ! quadratic fonn ,2 (xi-x) 2 • Show that B 1 + B 2 = 1 and B 1 B 2 = B 2B 1 =O for
u
p = 5. ical
! -! 1.33 Without using Theorem 1.08, prove the following: If A 1 + Az = 1 and
and C= if A 1Az = O, then A 1 and Az are each idempotent matrices.
-! !
-! -! =l) 1.34 If A is a matrix such that AA = k 2A, where k is a scalar, find a sea.lar m
such that mA is idempotent.
show: 1.35 If
· 'd mpotent and find the characteristic roots of A, the trace of A, show that A 1 + Az = l. Show that A 1 and A 2 are each idempotent. Show that
sho\vth a t A 1s1 e •
A 1Az = A 2A 1 =O, and fi.nd the rank of A 1 and A 2 •
and the rank of A.
,•.<i'······.
l.37 For the matrices in Prob. 1.36, it can be seen that the conditions of
Theorem 1.67 are satisfied. So there must exist an orthogonal transformation
C su ch that C'A 1C = E 2 and C' A 2C = E 1 , where E 2 and E 1 have ranks of 2 and l,
respectively. Verify the fact that the orthogonal matrix
1
v'2
1
C'= - v'2 2
o
X1 = Y 1 cos Y 2 ~!
X2 = Y 1 sin Y 2 ~ In this chapter we shall review sorne of the elementary notions of
evaluate the Jacobian J ==~axi/ay1 1. Also evaluate K* = jüy¡/ax;I by solving r mathernatical statistics, including randorn variables, frequency func-
tions, estirnation of pararneters, and testing of statistical hypotheses.
the transformation equations for Y1 and Y2 • Show that IK*I = ¡J¡-1 •
r
r
.1
.1
. ( 2) Leo
-oo
f f
00
-oo
0
• •
00
-~
f(X1,X2, · · • Xk) dX1 dX2 • • • dX. = 1
k
will consis ~ o~ a t~et o~ s; nu~ber; (i1 ,xi2-'~ 3 ,x4 ,x5 ix6 ). 1:he sample In multivariate distributions we must also consiºder · l d
space cou 1 e ie s1x- 1mens10na ~ uc 1 ean p ane, smce every
~
~
·t· l dº ·b .
con d1 1ona 1str1. ut10ns.
. If the chance
..• , xk h ave the JOmt frequency f(X X
(random)
margma an
· bl
varia es x 1 , x 2 ,
X ) th en th e margma · l
conceivable outcome is a point in this six-dimensional space. ll . t ·b t· f i' 2, • • • , k ,
2.1.2 Random Variables. Another concept important in prob-
1:
~
d1s r1 u ion o a subset of x 1 x x say x ( k
. . b ' 2' • • • ' k' ' ¡, X2, ••• ' XV p < •)
IS g1ven y · '
ability methods is that of a random variable. A randoni variable ~.·
is any function defined on a sample space. For example, in the
experiment on throwing a die in Art. 2.1.1, if x represents the outcome,
~
L
g(X1,X2, · · · ,Xp) = f00f00 · ••f00
-co -a:> -O')
J(X X
l' 2' ••• '
X) dX
1.; P+l
dX
:11+2 ••• dXk
then x can equal 1, 2, 3, 4, 5, or 6. The function u(x) = x 2 is a random
That is to say, the marginal distribution of a subset · bt · d b
variable; in fact, the outcome itself is also a random variable, as is ~ · t t · ti · . is o ame y
log x, ex - 1, cte. l m egra mg . ie Jomt frequency function over the random variables
A statistic is defined as any function of observable random variables ·~ that are not m the ~u?set. For example, suppose the random variables
x 1 , x 2, x 3 have the 3omt frequency function
that does not involve unknown parameters. r
2.1.3 Density Function. A density function on a sample space ¡ /(X1,X2,X3) = c<X1+X1+X3) foro< x. < ex:>,. ,•t --
i
1 •) 3
' .... ,
is a nonnegative function whose integral over the sample space is L
unity. For example, if the sample space is the Euclidean line; i.e., then the marginal distribution of x 1 is obtained by integrat'
if the outcome of an experiment is the number x 0 , where x 0 is any /(X1,X2,Xa) over the variables X 2 , X 3 ; that is, the marginal of x:r~;
number x such that - oo < x < oo, then any functionj(X) is a density Y1{X1) =e-Xi O< X 1 < oo
function on thc sample space if
A marginal distribution is also a frequency functi'on A d·t· l
(1) f(X) ~O for ali X such that -oo <X< oo dº t ·b t• f · con 1 iona
. is ri ~ ~on o a certain subset of the chance variables
IS the JO t d. t ºb t• f 1
(2) i)x¡ dX = 1 .. m
X¡, X2, ••• , xk
1~ r1 u ·Ion o t 1is subset under the condition that the
rem.ammg variables are given certain values. The conditional distri-
Given a random variable y, it is said to have a density functiong(Y) if but10n of X1, X2, ••• 'x.v, given X,,+1• X21+2• ••• , Xk, is
e f
~ x 2 ~ d), is given by ff(X 1 ,X 2) dX 1 dX 2 •
e-<X1+X2+X 3 )
-~~--=e-Xi
e-<Xi+X3 ) O< X 1 < oo
30 LINEAR STATISTICAL l\lODELS STATISTIOAL CONCEPTS 31
+ Definition 2.1 Two random variables x 1 and x 2 are said to be distribution with mean µ and variance a 2 if the frequency function of
statistically independent if and only if the conditional distri- y is given as
bution of x 1 , given x 2 , equals the marginal distribution of x 1 , j(y) = 1 c<u-µ)!/2a!
I
-oo<y<oo
i.e., if y27T0'2
From this basic distribution rnany important distributions are derived.
This can be extended to Tlie Olii-square Distribtttion. If y 1, y 2 , ••• , Yn are normally and
+ Definition 2.2 A set of random variables x 1, x 2 , ••• , xk is said
independently distributed with mean O and variance 1, then
to be jointly statistically independent if and only if the joint n
frequency function equals the product of the marginal frequency U= I y¡
i=l
functions; i.c., if
is distributed as chi-square with n degrees of freedom, designated by
f(X1,X2, · · · ,X.1:) = f1(X1)f2(X2) · • · üXk) z2(n). The functional form of chi-square is
2n12 r(n/2)
cu/ 2 o< u
- < oo), P(O - Á.1 < (J < O + ~) ~ P(O - ii1 < O* < O+ ~) (2.1)
2.2 Statistical Inference
where O* is any other estimator for O. That is to say, sin ce each randóm
In conducting scientific investigations the investigator often knows sample from f(x ;O) will give.rise to a different estimate, it seems that
or is willing to assume that the population from which he takes obser- a ';good" estimating function fJ = h(x 1 ,x2 , •• • ,xn) would be a function
vations is of a certain functional form f (x ;O), where the parameter O such that a large percentage of the samples produce estimates that
is not known but is the quantity that his investigation is attempting to fall very close to 8. In other words, it seems desirable to have an
describe. For example, it might be known from theoretical consider- estimator Ó such that the probability of the estímate falling in the
ations or from previous investigations that the frequency of weights interval (} - .A. 1 to O + .A. 2 is larger than the probability of any other
of the 100,00ú inhabitants of a given city asan approximation has the estimate of O falling in that interval. Although we should like to
functional form · have an estimator that satisfied (2.1), for practica! situations methods
f(x;µ) = 1 cha:-µ>' -oo<x<oo for finding such functions do not exist. Therefore, we shall enumerate
-J2w various other properties that we think are desirable in an estimator.
where the average weight µis unknown. In general, if we say that x On the basis of the nature of each specific scientific investigation, we
is distributed asf(x;µ), we shall mean that µis the unknown param- must select those properties which seem the most reasonable substitutes
eter in the frequency functionf(x). for condition (2.1) from among those properties which are attainable
34 LINEAR STATISTICAL l\IODELS STATISTICAL CONCEPTS 35
in the situation at hand. In many cases there is controversy about 0 ~o. If one hunts for an unbiased estimate Ó, it may be that
which criteria should be preferred. so~e values of Ó will be negative. If one does arrive at a ncgative
Important desirable properties for estimators include: (1) suffi- estima.te of (J under these conditions it would obviously be better to
ciency, (2) unbiasedness, (3) consistency, (4) efficiency, (5) minimmu take (} 0 =O. However, if a gre~t many estimates are to be obtained,
variance, (6) completeness, and (7) invariance under transformation. it is also wise to report the unbiased estima.te of (J even though it is an
These will be discussed briefly. impossible result in_itself. A .
Sujficieney. Let f(x 1 , ••• ,Xn; 01, ••• ,Ok) be a joint frequency Consistency. An estimator O,, (more accurately, a sequence of est1-
function involving k parameters. The statistics 01 = h 1 (x1 , . • • ,xn), mators {On}) is said to be consistent for 8 if the limit of the probability
82 = h 2 (x 11 ••• ,xn), ... , Oni = hm(x1 ,x2 , ••• ,xn) are a set of sufficie1ú that IÓn - 01 < €is 1 as n approaches infinity. Ón is the estimate of O
statislics if g(x 1,x2, ..• ,xn 1 01, ••• ,Ók) is independent of the param- based on ·a sample of size n.
eters O;, where g(x1 , • • • ,xn 1 01 , ••• ,0 is the conditional density
111
) When Jarge-size samples are used to estímate 8, consistency seems
function of the observations, given the statistics. to be an important property for an estimator to possess. H01vever,
We see that in any problem there may be many sets of sufficient if a relatively small number of observations is used, the concept of
statistics. The original observations x 1 , x 2 , ••• , xn always form a consistency is not too important unless the above limit gets clase to
set of sufficient statistics., but we generally want a smaller set since unity while nis relatively small.
this will be easier to work with. We shall then be interested in a Efficiency. An estimator Ón (more accurately, a sequence of esti-
"smallest" set, which is referred to as a minimal su.lficient set. A mators {On}) is said to be e.fficient when the following two conditions are
set of snfficient statistics 01 , 02 , ••• , (Jm is a minimal sufficient set if satisfied:
m is the smallest number of statistics that satisfy the definition of I. Vn(Ón - O) is asymptotically normally distributed witii mean O
sufficiency. and variance, say, a 2 , where nis the sample size.
Next we shall give a criterion to aid us in finding a set of sufficient 1 2. The variance a2 is less than the variance of any other estimator
statistics. O! that satisfies condition 1.
• Theorem 2.2 Assume that the range of the x/s in the frequency We see that the concept of efficiency is a large-sample concept.
function is independent of the parameters (it will always be .Mínimum Jlariance. An estimator fJ is said to be a minimum-
independent of the parameters for the problems in this book). variance estimator of 8 if E[Ó - E(Ó)] 2 ~ E[O* - E(0*)] 2, where
If the frequency function /(x 1 , ••• ,xn; 01, ••• ,01:) factors into O* is any other estimator for O.
the product of two functions such that If an estimator is efficient, then it is consistent and unbiased in the
~
t
-¡
limit but need not be unbiased for finite sample sizes. An unbiased
estimator is not necessarily consistent. The concept of minimum
/1(X¡, · · · ,xn)/2(Ó¡, · · · ,(Jm; 01, • • • ,Ok) !: variance is not a largeAsample concept.
where / 1 (x1 , . : . ,xn) does ?~t conta~n. the parameters, then ~ Oompleteness. An estimator fJ is complete if there exists no unbiased
Ó1 , 02 , ••• , Óm is a set of suflic1ent stat1st1cs. :' estimator of zero in the frequency function/(Ó;O) (except zero itself).
I nvariance. An estimator ó of (} is said to be an invariant estimator
Unbiasedness. An estimator Óis said to be an unbiased estimalor of J01· a certain class of transforrnations g if the estimator is g(O) when the
8 if E(Ó) = 8, where E denotes mathematical expectation. transformation changes the parameter to g(O).
Thus, if we have a great many unbiased estimates of O, then the This principie of an invariance estimator seems very reasonable.
average of these estimates equals the parameter 8. This criterion is However, because ofthe advanced nature of the mathematics in volved,
quite important in our present system of scientific investigation and we shall not do much with this principie of estimation in this book.
reporting, where many observers report estimates of the same param- The following theorem will aid in finding minimum-variance
eter. These can then be combined into one "good" estimate if they unbiased estimators.
are unbiased. In many cases, however, the concept of unbiasedness
leads to very poor estimates. For example, a sample might be selected + Theorem 2.3 If O is a sufficient statistic for O, if fJ is complete
from the population f (x ;O) to estimate (} where it is known that (no function of the sufficient statistic for Óis an unbiased estímate
'
')'
For a full discussion of these criteria the reader is referred to the I e¡= ! [x, - /;(01,82, ••• ,01;)]2
i=l i= 1
references in the bibliography.
Even if one has decided upon "good" criteria for estimators to have, is formed, and the value of the O, that minimizes this sum of squares is
one must still devise methods for arriving at estimating functions the least-squares estimator for the 8i. -
h(x1,X 2 , ••• ,xn) that meet these criteria. There are many important In the method of least squares, the forro of the frequency function
methods, two of which we shall discuss, since they will be used in later nced not be known, but the method can be used if it is known. In
chapters: (1) the method of maximum likelihood, and (2) the method many important cases, even if the form of the frequency function is
of least squares. unknown, the method of least squares gives rise to estimators that
lJ{aximum Likelihood. The likelihood function for a random sample are unbiased and consistent and under certain conditions minimum-
of n observations from the frequency functionf(x;µ) is defined as the variance unbiased. Also, in at least one very important case, the
joint frequency function of the n variables, which is equal to method of least squares and the method of maximum likelihood give
rise to the same estimators. These cases will be discusf:ed in greater
L(x¡,X2, · · · ,xn; µ) = f(x¡;µ)f(x2;µ} · · · f(xn;µ) detail in a later chapter.
"
'J·.
conducts an experiment and takes observations to see how closely ~ -cannot be made with certainty, and probability measures its un-
they conform to the hypothesis, and on the basis of these observations ~: certainty.
he de~ides to accept H 0 as true or to reject H 0 as false. It should be ~ When the hypothesis H 0 is formed and the experimenter takes
1>ointed out that a general hypothesis can never be proved but can [: obsenrations and uses them as a basis for either accepting or rejecting
be disproved. For example, one might try to prove the hypothesis ~· H 0 , he is Hable to two kinds of error, as follows:
that one can always draw an ace on the first draw from an ordinary deck ~ I. The hypothesis H 0 might be true; yet on the basis of his observa-
of cards. No matter how often an investigator performs this feat, fi,: tions the experimenter might reject H 0 and say that it is false. This
he can never prove the hypothesis. One failure, however, will disprove ~ is gene rally referred to as error of the first kind or type 1 error; briefly,
it. On the other hand, if a subject does pull an ace a great many f it is the error of rejecting a true hypothesis.
times without failure, we might be willing to say that, for ali practical 1; 2. The hypothesis H 0 might be false; yet on the basis of his observa-
purposes, our subject can always select an ace on the first draw-but ,~ tions the experimenter might accept H 0 as true. This is generally
the hypothesis as it was set up has not been proved. referred to as error of the second kind or type 11 error; briefly, it is
As another exa~pl~, suppose one has a n~w process of manufacture t the error of accepting a false hypothesis.
that one hopes wdl g1ve added length of life to electron tu bes. An i We should like to minimize the likelihood Qr probability of making
investigator may form the hypothesis that tubes manufactured by i either of these two errors. However, in general, for a fixed number of
this process will have a longer life than tubes manufactured by a f observations, if we decrease the probability of making an error of one
standard process. He proceeds to test this hypothesis by observing •: type, we increase the chances of making the other. P(I) and P(II)
the 1ength of life of a certain number of tubes of each manufacturing ~ will represent the probability of making an error of type I and type II,
1
method. Clearly he can never pro ve his hypothesis, but he can disprove ¡ respectively. We shall generally use the power of the test instead of
it. However, if a great number of tubes manufactured by the new i. the probability of type II error where
process each have a longer life than tu bes manufactured by the standard
process, he may be willing to accept the new process, but he cannot f3 = power of the test = l - P(II)
prove the hypothesis as it was formed.
One may form rules about when a hypothesis will be accepted and Since it is desirable for P(II) to be as small as possible, we shall want
when it will be rejected. In the example above, one might use the p to be as la:rge as possible. The power of a test /3(8*), where 8 is the
following formula: Accept thc hypothcsis if 1,000 electron tubes are parameter under test, is the probal>ility of rejecting the hypothesis
examined from each manufacturing proce~ and if 950 or more of those H 0 : (J = 80 when the parameter (J actually equals 8*.
manufactured by the new process have longer life than the average of In most scientific investigations it seems desirable for the probability
the 1,000 manufactured by the standard method. Here the acc~~ting , ofrejecting the hypothesis O = 00 to increase as the distance between
or rejecting of a hypothesis is identified with the experimenter's f\ Oand 00 increases. That is to say, if the true value (J ofthe parameter
willingness to accept or reject the new manufacturing process and 1'. is close to (J 0 , it may not be too disastrous if the test procedure fails to
not with proof or disproof of the hypothesis. i rcject H 0 • However, if the true value of the parameter is not close but
Testing scientific hypotheses is much more complicated than has ,; quite distant from 00 , we would want the probability of rejecting H 0
been illustrated here. It is not our intention to go into great detail to be relati vely large. The power of the test is a function of the
concerning the philosophical aspects of the problem, but our intention ; parameter as denoted by /3(0), and it is clear that O :s;;; /3(0) :s;;; l. It
is to point out that in testing hypotheses one uses inductive reasoning- ; might be pointed out that, if the hypothesis H 0 is (J = 00 , then
reasoning from the particular to the general. That is to say, from a ! P(0 0 ) = P(I).
small set of observations one draws conclusions about what would {: The procedure for tesliing a hypothesis H 0 : () = 0 0 about a param-
happen ifthe totality of observations were taken, where the number of t eter (J in the frequency function f (x ;O) is as follows:
observations is probably unlimited and many ofthem are in the futurc. 1· l. A random sample of values is taken from f(x;O), and sorne
Thus the scientific hypothesis is concerned with taking a relatively few ,, function of these values is formed, say, O = g(x 1 ,x 2 , ••• ,xn)·
! l'
r;.
~~~¡
Rr
and is the probability that (J falls in the region Rr when O = 8*. 1 where L(Ú) is the maximum of L respect to 01 and 02 subject to
Of course, we should like to minimize P(I) and maximize p(O). f¡ the condition that 01 and 8 2 are in íl. L(w) is the maximum of L with
In general it can be shown for a fixed size of sample that, when p(O) ~ respect to 0 1 and 8 2 subject to the condition that 81 and 02 are in w.
increases, P(I) increases, and vice versa. Therefore, the general pro- f If L(w) is close to L(i"!), we expect H 0 to be true and we have no reason
cedure is to choose P(I) in advañce and then, from all functions (j f¡,· to reject H 0 ; if, however, L{w) is quite distant from L(i"!), we expect
and all regions Rr such that P(I) is fixed, choosc a particular function l H 0 to be false and we reject H 0 • In other words we reject H 0 if L
and region that maximize p(8). ~ is small and not if L is large. Clearly, L is su ch that O ~ L ~ l. So
'I;.
1
In examining the "''goodness" properties of a test function we need · we need to find a constant A and reJ· ect H 0 if our observad Lis less than
: ¡,. to examine only the power function. If the power function of a given or equal to A, where A is so chosen that the probability oftype I error
test is p(O) and if p(B) ~ p 1 (8), where {J 1(0) is the power function of any is «. To do this we need to find ·the frequency distribution of L when
H 0 is true. Let this be g(L ;H 0 ), with O ~ L ~ l. Then Jet A be
other test function, then we say the test function corresponding to
p(O) is uniformly most powerful. · Unfortunatcly, a uniformly most l· such that
powerful test function exists only in very few cases; p(O) may be
greatcr than any other power function ior sorne values of O but not for ~ P(I) = L'
g(L;H.J dL = "'
others.
To examine these ideas further we shall assume that we have a
sample of size n from the distributionf(x; 8 1 ,8 2). The ideas will hold
for a function of s parameters 0 1 , 8 2 , ••• , 811 , but we shall discuss the
r
~/
l'
r
11
¡.
The power of the test is
P<H,) = r g(L;H,) dL
2 .. -2 log Lis approximately distributed as a chi-square variable Now Lis a monotonic decreasing function of u; hence we can use u
if the sample size is large. as our test function. N ow the critica} region of L is O ~ L ~ A.
3. For most of the models we shall discuss, the likelihood ratio will This gives e ~ u < oo as the critica! region for u, where
give a test that has reasonably "good" properties.
4. The likelihood test is always a function of sufficient statistics. c=(-1 -l)(n-1)
A21n
Thc importance of property 4 will be$iven in Theorem 2.6.
We shall illustrate with an example. Suppose a random sample of Also, we see that u is distributed as F(l, n - l) if H 0 is true.
size n is selected from a normal population with mean µ and variance We see that the test function in the above example is based on the
a2. Let the sample value be x 1, x 2, ... , xn. Suppose we desire to sufficient statistics i, :E(x, - x) 2 • We shall now state an important
test the hypothesis H 0 : µ = O and that the alternative is µ =fa O. theorem concerning the power of a test and a set of sufficient statistics.
Now the points in the n space are all points µ, a 2 such that - oo < + Theorem 2.6 Suppose we desire to test a hypothesis H 0 about the
µ < oo, O~ a2 < oo. The points in the w space are all points µ, az parameters in the frequency function /(x; 01 , ••• ,O,). For any
such that µ = O, O ~ a 2 < oo. Since the w space is not just a single test function </l(x), there exists a test function based on the set of
point, the hypothesis is composite. The likelihood is sufficient statistics that has a power function equivalent to the
power function of the original test function </l(x).
1
.': exp [ - __.!:.._ :E(x. - µ)2] The discussion above has been based on the assumption that the
(a221T)"'2 2u2 '
sample size is fixed. That is to say, for a given set of observations,
l [ 1 - 2] e-n/2 an experimenter should use the "best" test function available to test
Clearly, L(!)) = ( 7TÚ2}1112 exp
2
- 2éí2 L(X¡ - x) = (
27ra2)"'2 the hypothesis H 0 •
However, if the experimenter has sorne control over his observations,
where a2 = ..!:. L(x¡ - x) 2 ~ he should take them in such a manner as to sattsfy the following two
conditions:
l. The probability of a type I error should be equal to a preassigned
In the w space, µ = O; so the o:iy quantity we need to maximize with
2
1 number-oc.
respect to is a • Clearly, !.··.
2. The power of the test should be equal to a preassigned number {J,
1112
L(có) - n e-n/2 when the true value of t~e parameter is larger than specified.
- (21T ~x~y112 For example, if a researcher is testing the hypothesis H 0 that two
varieties of wheat yield the same, he may not want to say that they
So
L = [:E(xi - x)2] n/2 1 yield differently unless the difference is at least 5 bushels/acre. On
:E~ the other hand, if the difference in yield is 5 bushels/acre or more,
~¡
:E(xi - x)2 Jn/2 ~;
he may then want the probability of rejectiiig H 0 to be quite high;
i.e., he may want fJ to be close to l.
= [ :E(x¡ - x) 2 + nx-2 We shall illustrate sorne of these ideas with an example. Suppose
a random variable X is distributed normally with mean Oand variance
1 ]fl~ 1. We wish to test H 0 : O = O. A random sample of size nis selected,
- x2
[ l+n and the sample mean X is used as the test function. Therefore, the
_:E(x¡ - i)2 distribution of X when O = O must be found. Under these conditions
X is distributed normally with mean O and variance l/n. We shall
= [l + u/~n - lJ'' take P(I) = .05 and Rr as the two intervals - oo to -oc 0 and oc 0 to
+oo. We ha'\le
- <Xo .j- f.oo .j-n e-!nXa dX
where U=
n(n -
:E(x¡ - x)
l)x2
2
.05 = P(I) =
i-oo .../271'
--=n i -oa
e-:a"n•L dX + -
«o .j2;
STATISTICAL CONCEPTS 47
46 LINEAR STATISTICAL MODELS
which can be written more concisely 2 .5 If a random sample of size nis taken from the distribution e-X, where
X -:;, o, find the moment-generating function of tbe mean of the n observations.
1.9s¡v;; fñ usÍng Theorem 2.1, fmd tbe distribution of the mean of the n obser,•ations.
P(I) = l -
Í-t.96/V; ..j2;
_v_ e-ln..f' dX
.
2.6 lf yl' ... , Yn are independent and each is distributed N(µ,a 2 ), find the
maximum-likelihood estimates of µ and u2.
1.so¡v; fñ 2.1 In Prob. 2.6, show that the estimates are sufficient statistics.
_ ~ ch(.f- 9> dX
2
Also, p(O) = 1 -
Í -1.96/V n .j2TT
2.8 In Prob. 2.6, find the likelihood-ratio test of H 0 : µ =O with the
altornative hypothesis µ :#= O.
2.9 In Prob. 2.8, show that the likelihood-ratio test is equivalent to the
If we let Y = (X - O)Vn, we get ·
l.96-ov'; · 1 conventional t test.
2.10 In Prob. 2.6, it can be shown that x and a2 are sufficient complete
p(O) = l -
Í-1.gs-ov' n
_-= cY /2 dY
.j2TT
1
(a) 2
1
ª
z
n
i=l
(xi - µ) 2 13
14
Table IV, Oliver & Boyd, Ltd., London, 1946.
R. A. Fisher: Applications of "Student's" Distribution, 111etron, vol. 5, 1925.
G. W. Snedecor: "Sta.tistical Methods," 5th ed., Iowa. State College Press,
1 n Ames, Iowa, 1957.
(b) u2 L
(xi - µ)Z + z2 15 "Student": Probable Error of the Mean, Biometrika, vol. 6, 1908.
i=l
n (xi - /t)2 16 R. A. Fisher: On the l\fathema.tical Foundations of Theoretical Statistics,
(e) ¡
1=1
nz2uz -
ta (X· _ µ)2 17
Phil. Trans. Roy. Soc., London, ser. A, vol. 222, 1922.
R. A. Fisher: Theory of Statistical Estimation, Proc. Cambridge Phil. Soc.,
2.4 Find the moment-generating function of ~ ª 2 • vol. 22, 1925.
i=l ª
THE MULTIVARIATE NORMAL DISTRIBUTION 49
We shall now show that the function (3. l) satisfies conditions that
qualify it for a frequency function, i.e.,
(a.) f(y1 ,y2 , ••. ,y p) ~ O for ali values of Yi
f
•
1
i . ao oo ioo ,z
·i = l, 2, ... ,p (3.1)
r¡
50 LINEAR STATISTICAL MODELS THE MULTIVARIATE NORMAL DISTRIBUTION 51
The Iast integral can be written as • corollary 3.1.1 The multiple integral
and does not depend on the value of the elements of the vector p..
- .
By using the frequency function of a univariate normal distribution + Definition 3.2 In the multivariate normal distribution
we see that ' J(e-!CY-¡.i.)'R(Y-µ)
j2;
Thusweget
L oo e-lci,1:
-oo
dt¡ =
1'/-;¡; we shall define (Y - p.)'R(Y - p.) as the quadratic form of the
multivariate normal.
For condition (3.3b) to be satisfied, we must set (3.4) equal to unity. variables in the vector Y. This will be the text of the next few ·
Weget theorems.
21
fidt + Theorem 3.2 The marginal distribution of any random variable
K í=t (3.5) in the p-variate normal is the simple normal (univariate normal)
= (2?T)P/2 distribution.
Thus the function given by (3.1) is a frequency function, for the The proof will consist in showing J:,hat the marginal distribution
conditions (3.3) are satisfied. We point out that ](is nota function of of y 1 is normal (the proof is general if we use y 1 instead of y¡}·
p.; (3.1) is a frequency function regardless of the value of the elements Proof: By definition,
of the vector µ. or of the positive definite matrix R.
1,
The joint distribution of y y 2 , • • • , yP will sometimes be referred
to as the ~istribution of the vector Y, and we shall sometimes write
_ g(y,) = I: I:· ··L: Ke-l<Y-iú'R<Y-•> dy2 dy3 • • • dy, (3.6)
f(Y) in place of/(y1 ,y 2 , ••• ,yP). Partitioning the vectors and the matrix of the exponent gives
+ Theorem 3.1 In the p-variate normal as given in Definition 3.1,
(Y - p.)'R(Y - p.) =[(y¡ - µ¡),(Y2 - u2n(R11 R12)(Y1-:-- µ¡ \
K=~
(2?T)P/2
R21 R22 Y2 - uJ
Proof: By Theorem I.41, where Y 2 and U 2 are the vectors containing the last p - 1 elements
%> of Y and p., respectively. .R 11 is the element in the first row and
IDI = Ild¡ column of R (that is, R 11 = r 11). The dimensiona of the re-
i=¡
maining matrices R 12 , R 21 , and R 22 are, therefore, determined.
Thus Multiplication gives
Since Pis orthogonal, we use Theorem l.39 and get IRI = IP'RPI
tf - (Y - p.)'R(Y - f!.) = (Y1 - /.t1)r11(Y1 - µ1) + (Y1 - µ1)R12(Y2 - U2)
=!DI; the result follows. + (Y2 - U2YR21(Y1 - µ¡) + (Y2 - U2)'R22(Y2 - U2) (3.7)
52 LINEAR STATISTICAL l\IODELS THE MULTIVARIATE NORMAL DISTRIBUTION 53
Sincc R is symmetric and posith-e definite, we know that R 12 :::: which is the frequency function of the univariate normal, and the
R; 1 , that R221 exists and is symmetric, and that R 11 = r 11 > O. proof is complete.
Equation (3. 7) can be written
The result of this theorem gives us a logical reason to call the
(Y - p.)'R(Y - µ.) = (Y1 - µ¡)(Rn - R12R2lR21HY1 - µ1) frcquency function (3.1) a p-variate normal. \Ve shall now generalizc
+ [(Y2 - U2) + R;iR21(Y1 - µ1)J'R22[(Y2 - U2) + R;iR21(Y1 - µ 1)] this theorem.
(3.8) + Theorem 3.3 If the random variables y 1, y 2 , ••• , y 11 are jointly
Note that the first term on the right-hand side of (3.8) does not normal, the joint marginal distribution of any subset of s < p
involve any random variable except y 1 and, hence, that (3.6) of the random v~riables is the s-variate normal.
can be written The theorem in vector terminology would be stated: If the
vector Y is normally distri buted and is such that
g(yi) = KFe-HY1-1'1><Ru-R1 2R;~Ru><v1-111> (3.0)
where F is the multiple integral
«» i<:n • • • ioo e-HY:-(tTz-R221 lti1<111-1'1>lr~:{Yc(Uz-R221 Ru<Y1-111>n dY where Y 1 is a vector containing s of the elements of Y, then thc
~l1 i-oo -oo -oo
2
marginal distribution of the vector Y 1 is the s-variate normal.
1 where Proof: The proof follows the method of Theorem 3.2 almost
¡ y 1 is not a variable of integration; therefore, if y 1 occurs in thc
exactly, the only difference being in the partitioning of the matrix
and vectors in the exponent. That is to say, R 11 is no longer of
integral, it will be considered a constant with respect to the dimension 1 x 1, but is now s x s. Also, the scalar y 1 - µ 1
integration. If we let the vector U 2 - R;21 R 21 (y 1 - µ 1) = ~ is replaced by the s X l vector Y 1 - U 1. The resulting marginal
we can write (3.9) distribution of the vector Y 1 is
1
g(y¡) = l\.Fe-l<u1 -11 1><Ru -R12R22 Ru><Y1 -111) e-HY1 -t11)'(Ru -R12Ri:R:1><Y1 -U1)
't h(Y) - - - -8 - ------ ,--
where l' = J:oo J:'X) ... J:'X). e-HY:i-bYR:i2<Y:i-b> dy2 dy3 ••• dyP 1 - (211) ' 21R11 - R 12R-
22 R 21 ¡-!
1
l
normal distribution, all the parameters are functions of only the first
E(y¡) =
i
-oo
oo 'l/
.1
aJ2w .
:1
e-<111-P1> 'ªª dy1 =
2
µ1
My_"'(T) =
1
Í
e!T'R- T -co •••
00
Í
00
-oo Ke-l<Y-h)R(Y-h) dy1 dy2 ••• dy11
and the proof is complete. which, by Corollary 3.1.1 (if we let the ijth element of R-1 equal
'Ve shall now investigate the variance of the random variables y¡ si;), gives
and the covariance of the random variables y¡ and Y; in the multivariate
normal distribution as given in Definition 3.l.
]f[Y -"'(T) = e~T'R-1T = exp (l f f t,t,.s¡;)
l=l ;=1
(3.12)
In a p-variate normal there will be p v'riances, one for each random In Eq. (3.11), let us evaluate
variable y¡, and !P(P - 1) covariances, oñe for each combination of two
of the p random variables, such as Yi and y 1 (i =/= j). We shall form o2MY-11-(T)
a p x p matrix whose diagonal elements will be the variances of the y, otm otn
and whose ijth element (i :¡!: j) will be the covariance of Yi and y1• at the point T = O. We obtain the equation
The matrix will be symmetric. If we let V represent the variance-
covariance matrix of a vector Y and if the elements of V are v¡;, then, o2My- 14(T)
'.:lt '.:ll =
[ 11
E[(ym - µm>CYn - µn)] exp .~ t¡(Y-; - µ,)
]
by definition of variance and covariance, we have for the frequency 1 u mu n •-1
function given in Definition 3.1 to be evaluated at the point T =O. The final result is
E{[Y - E(Y)][Y - E(Y)]'} = E[(Y - µ)(Y - µ)']=V <P.ilfY-11-(T) 1 . {cov (Ym·Y") if m # n
atm uln'.:l = E[(ym - µm>CYn - µn)J =
+ Theorem 3.5 V= R-1 • That is to say, the matrix in the T=O var (Ym) if 71l = n
exponent of a p-variate normal distribution is the inverse of the · The above is valid only if the function under the integral is
variance-covariance matrix of the vector Y. ,~;: such that the variances and covariances exist, such that differ-
Proof: We must show that the covariance of y¡ and y 1 (i =i= j) is the entiation can be perforrned under the integral, and such that
ijth element of R-1 and that the variance of y¡ is the ith diagonal integration and taking oflirnit of the vector~ are interchangeable.
element of R-1 • The proof will consist in finding the joint !' The conditions hold for the p-variate normal. From (3.12) we
moment-generating function of see that
(Y1 - µ1), (Y2 - µ2), · · · '(Y11 - µ11)
(J2Jfy-11-(T) 1 =s
otm Oln T-0 mn
If T is a p x 1 vector with elements t¡, then, by definition, · and the proof is complete.
56 LINEAR STATISTICAL l\IODELS THE l\IULTIV ARIATE NORMAL DISTRIBUTION 57
The vector of means p. and the variance-covariance matrix V
From (3.11) and (3.12) we have
completely specify the multivariate normal distribution. If we speak
of a vector Y as being distributed normally with mean p. and variance E(e<Y-11)'At] = e!l(A'VA>t = elt (A'VA>
2
But, since y, and y 1 are independent, we can use (3.14), and the
integral becomes
(-~~)
where E(y¡m) = µ¡m, and equals
Q = {Y1
\y2
-6)'( 3-l)f\y2 -6)
-2 -1 2
Y1
-2
Q =(Y - fL)'R(Y - fL} = Y'RY - fL'RY - Y'RfL + fL'RfL
we sce that Y'RY is the only term that involves only second-degree
6 terms in y 1 and y 2• Thus, from Q we select only the second-degree
we see that E(Y) = (. ). But if the multiplication is effected so terms and evaluate the matrix R. We have
that Q is given as 2
= (~ j = ~)
V=(-~ -~ D
subvectors such that Y and ff Y*
V= (Vu V12)
and
V21 V22
Suppose it is desired to write the joint marginal distribution of y 1 and
are the corresponding partit.ions of Y*, 11, and V, then the con-
Ya· From Theorem 3.3 it follows that y 1 and Ya are jointly normal with ditional distribution of the q x 1 vector Y 1 given the vector
means 1 and 2, respectively, and with covariance matrix
Y2 = Y: is the multivariate normal distribution with mean U 1 +
V12V;l(Y: - U 2 ) and covariance matrix (V 11 - V 12V221 V 21 ).
Proof: By the definition of a conditional distribution we get
1
y y* _f(Y1,Y:) - [l/(217)Jl/21Vl!]e-~<Y*-11rv- (Y*-11) •
VT3 is obtained by striking out the second row and second column from g( 1 1 2) - t 2(Y*)
2
- , * . _1 * >
[1/(217) 1(J>-q>1v221•1e-~<Y: -U:> Vu<Y:-U:
(3.16)
V (since the joint density of y 1 , y 2 , and y 3 is to be integrated over the y 2
space). Since Vt3 is diagonal, we see that y 1 and Ya are independent. where /(Y 1 , Y:) is the joint distribution of the vectors Y 1 and
If it is desired to obtain the marginal distribution of y 1 , we should Y 2 with Y: substituted for Y2 , and f 2 (Yi) is the joint marginal
strike out the rows and columns pertaining to y 2 and Ya (the second distribution of the elemtmts of Y 2 evaluated at the point Y 2 = Yi.
and third) from V and obtain the variance of y 1, which is equal to 2. Ifin (3.16) we replace Y*, p., and V with the corresponding sub-
Using Theorem 3.8, we see that Ya is independent of both y 1 and y 2 matrices and simplify, we get (using Theorem 1.49)
but tl1at y 1 and y 2 are not independent of eac~ other. g(Y1 1Y * tri
2)
Suppose we wanted to find the distribution of the linear combination
of Y given by A'Y, where A' = (1, -2, 1). By Theorem 3.6, we know
that A'Y has the univariate normal distribution with mean equal to
A' 11, which is 5, and variance A'VA, which is 13. We can also find K (3.17)
by examining Q. From Theorem 3.1 we have where K* = (217)q 12 IV 11 - V 12V221 v 21 ¡1 12, and, since V 11 -
V 12V221 V 21 is positive defiñite, the theorem is proved.
K=~= - 1.
(217),,,2 1v11 (211"),,,2 It should be pointed out that g(Y1 1 Y:) is a frequency function of
Y1 for any val ue of Yi. If the matrix in the exponent of J(Y) is
Hence, K= 1 R = V-1 and if, corresponding to the subdivision of V in Theorem
(217)
3
'2.Jª 3.10, we get
R = (Ru R12\
3.6 The Conditional Distribution
R21 R2j
Suppose we have two random variables y 1 and y 2 that have a tben Rn = [Vn - V12V2°21V21r1
bivariate normal distribution with means µ 1 and µ 2 , respectively,
So, if we know the expone1~t matrix ofj(Y) and want the exponent
and covariance matrix V. Suppose we want to estimate the expected
matrix of the conditional distribution of Y1 given Y 2 = Yi, we just
value of y 1 for a given value of·y 2 • This suggests the use of the con-
cross out the rows and columns of R pertaining to the variables in Y 2 •
ditional distribution of the multivariate normal.
This leaves R 11 • The covariance matrix of the conditional clistri-
In this section we shall prove sorne important theorems about con- bution is Rii 1 •
ditional distributions.
64 LINEAR STATISTICAL MODELS THE MULTIV ARIATE NORMAL DISTRIBUTION 65
+ Theorem 3.11 E(Y 1 1Y:) =· U 1 + V12V22 (Y: - U 2), where 1
wc shall not explicitly indicate the value of Y 2 • If we consider
E(Y 1 1Y;) means the expected value of the random variables in the ••• , yP as given, we get for the mean of y 1
y , y3,
conditional distribution of Y 1 given Y 2 = v:. 2
p p
Proof: The proof is immediately obvious from inspection of Eq.
(3.17).
E('fii) = µ¡ - ::¿ piµi +¡::2
i=?.
L f3iYi (3.20)
+ Th'!orem 3.12 The covariance matrix of the conditional distribu- In (3.20), we see that the rate of change of E(y 1} per unit change in
tion of Y 1 given Y 2 = v:
does not depend on v:. y is f3J· Thus /3; is called the partial reg1·ession coefficient; it indicates
p,..oof: The proof of this theorem is obvious when we examine Eq. tlie change in E (Y 1) per uní t change of y i w hen the other y¡ (i =I= j) remain
{3.17) and see that the covariance matrix of the conditional distri- unchanged.
bution is Rií1 = V 11 - V12Vi;/ V21 and does not in vol ve Y:. In the conditional distribution of Y 1 given Y2 , we shall define the
partial correlation coejficient Pii·rsl···p to be the correlation between
We shall adopt the notation vii = ai1 to indicate the covariance of
'!J; and y, (which are in Y 1) in the conditional distribution of Y1 given
= j) in the joint distribution of y 1 , y 2 , ••• ,
y¡ and Y; {variance of Yi if i
Y 2 , where Yr, Ys, y,, ... , and y,, are in Y 2 • For example, p 13.567 will
y,,. We have shown that the covariance of y, and Y; equals <I¡¡ in the
represent the correlation between y 1 and y 3 in the conditional distri-
joint marginal distribution of any subset of Y1> y 2 , • • • , y,, that contains
bution of y 1 , y 2 , y 3 , and Y.a given y 5 , y6 , and y7 •
Yi and Y;· However, as is stated in Theorem 3.10, the covariance of
r·.
Yi and y 1 in the conditional distribution of Y1 given Y 2 = v: is the
From this it follows that
ijth element of Ril.1 • The following notation will be adopted: ai;·rst··-p
will mean the covariance of Yi and Y; in the conditional distribution of
Y 1 given Y 2 = Yi, where the subscripts on the right-hand side of the
dot all appear in Y 2 and the subscripts i andj must appear in Y 1 • For The partial correlation coefficient Pii·ral···p indicates the correlation
example, o-12•456 means the covariance between y 1 and y 2 in the joint bctween Yi and Y; when the variables Yr, Ys, y,, ••• and yP are held
conditional distribution of y 1 , y 2 , and y 3 given y 4 , y 5 , and y 6 • constan t.
The multiple-correlation coefficient p 1 will be defined as the correlation
+ Theorem 3.13 a 11 •23 •..P ~· ª11· between the two random variables y 1 and z, where
P1'Dof: The proof is as follows: ªn·2-'3···p = R-¡1 = r¡~ = v11 -
V 12Vi21 V 21 ~ v 11 = aw since V12V221 V21 is positive semidefinite
and, hence, cannot be negative. Thus,
In Eq. (3.17), if Y 1 is-1 x l and we make the substitution _ E{(y1 - µ 1 )[z - E(z)]}
P1 - --:::============:::::::::=
.J E(y 1 -µ E[z - E(z)]
1)
2 2
e= (y1 - µ 1) - V12V;~(Y2 - U 2) (3.18)
Now E{í_y1 - E(y1 )][z - E(z)]} = E[(y 1 µ 1 )V12V;i(Y2 - U2 )]
-
it f ollows that e is normally distributed with mean O and variance
= E[(y1 - µ1HY2 - U2)'V2;1V21] = V12V'2clV21
v 11 - V12V;}V21 = a2 (say)
Also
In Eq. (3.17), if we let the 1 x (p - l) vector V 12V221 = ~ = {{3¡),
we get (writing y1 for the conditional random variable) and E[z - E(z)] 2 = E{[z - E(z)][z - E(z)]'}
(3.19) = E(V12V;l(Y2 - U2HY2 - U2)'V;fVn]
= V12V;lV21
which is called the multiple regression equation of y 1 on y 2 , y 3 , ••• , y,,.
To indicate the conditional distribution of Y 1 given Y 2 = v:, we So (3.21)
shall simply write itas the conditional distribution of Y1 given Y2 and
66 LINEAR STATISTIOAL l\lODELS THE MULTIVARIATE NORl\fAL DISTRIBUTION 67
+ Theorem 3.14 The multiple-correlation coefficient satisfies the a.s follows: Assume that y1 is normally distributed with mean µ1 +
inequality O~ p'f ~ l.
Proof: py = 1 -(u11 .23 •..p/<1 11 ). ByTheorem3.13,0~(<1 11 . 23 ...p/<111) ~ 1
f p;(Y; - µ;) and variance a 2 • It is then sometimes desired to find
and the result follows. t.'1~ increase in the variance of y1 if regression is not used; that is to
say, it is desired to find the marginal distri bution of y 1• It is impossi ble
From Eq. (3.21) we obtain to obtain this from the conditional distribution of y 1 given Y 2 unless
C111 - ªii°·23···p 2 either (1) the joint distribution of y 2, Ya, ... , y.,, is known, or (2)
=Pi
<Tn y 2, '!Ja, .•. , y'P are jointly independent of y 1 • This can be seen from
u 11 is the variance of y 1 in the marginal distribution of y 1 ; u11 •23 ...P is the the fact that by the variance of y 1 when regression is not used we mean
variance of y1 in the conditional distribution of y 1 given y 2 , y3 , ••• , yi>. the variance of the marginal distribution of y 1 • It is clear that the
The quantity cr 11 - <T1i.23 ...P is the amount that the variance of joint distribution of y 1 , y 2 , ••• , Yi> (from which the marginal distri-
y 1 (when y 2 , ••• , yP are ignored) can be reduced by sampling from the bution of y 1 must be obtained) is equal to the product of the conditional
distribution of y1 when y 2 , y 3 , ••• , Yi> are not ignored but are known. distribution of y 1 given y 2 , Ya, ••. , y P and the joint marginal distribu-
For example, let us assume that the heights and weights of indi- tion of y2, y 3, ... ,y.,,. Ofcourse, ify 1 isindependentofy 2,y3, ... ,yri,
viduals in a large city follow the bivariate normal distribution. If then the conditional of y 1 cquals the marginal of y 1•
we ignore the weights of people and compute the variance of the heights Let us consider the case in which p = 2, and let y1 =y and y 2 = x.
of all the population of the city, we get a number which we shall call The regression equation is
u 11 • If, now, we compute the variance of the heights of only those y = µ 11 + {J(x - µx) + e (3.22)
,'I
individuals who weigh 150 lb, it is quite clear that we shall get a number, where x is known and where e is distributed normally with mean O and
say, u 11 .2 , su ch that <1 11 ;?: u 11. 2 • That is, it is evident that the variance variance a 2 • This indicates that we know that, when x = x 0 , x = x 1 ,
of the heights of all people will be equal to or larger than the variancc or x = x 2, then y is normally distributed with mean µ 11 + p(.T0 - µx),
of the heights of people in the class restricted to persons who weigh Pv + p(x 1 - µz), or µu + P(x 2 - µz), respectively, and with variance
150 lb. If <T 11 = u 11• 2 , then u 11 - u 11 . 2 = O, and the variance of a2. The probability that y is between a and b (this requires the mar-
heights is not reduced by sampling from the restricted class. If ginal distribution of y) is not known unless we know the relative
<T 11 . 2 = O, then u 11 - u 11 . 2 = <T 11 , and thc variance is reduced by an probabilities that we are sampling from the distributions correspond-
amount u 11 by sampling from the restricted class. In many investiga- ing to x 0 , x 1 , and x 2 • In other words, we need to know the joint
tions the important thing is not the actual amount that the variance distribution of x 0 , x 1 , and a: 2 •
is reduced but the fractional reduction. Therefore, the square of Sometimes we are not primarily interested in how regression reduces
the multiple-correlation coefficicnt is an important quantity; it gives the variance of y¡ in Eq. (3.19). We postulate that y 1 is distributed
usa measure of the fraction of reduction in the variance of a random according to Eq. (3.19) for ccrtain known Y; values, and we want to
variable when we use related variables. estimate µ 1 , µ.i, the p¿, and u2 and to test certain hypotheses about these
If we examine the multiple-regression equation (3.19) we can also parameters. This will be discnsscd in dctail la ter.
seo thc utility of the multiple-correlation coefficient p 1 • In (3.19),
if y 2 , y 3 , ••• , Yv are given (fixed), then the variance of y1 equals the
3.7 Additional Theorems
variance of e, which is v11 - V 12Vz-21 V 21 • On the other hand, if the
y¡ are not fixed but all that is lmown is that y 1 , y 2 , ••• , Yi> are jointly Sorne additional theorems will be stated in this section; sorne proofs
normal, then the variance of y 1 is v11 • As before, the fractional will be asked for in the problems (y 1 , y 2 , ••• , yP are assumed to be
reduction in the variance of y 1 that is dueto y 2 , y 3 , ••• , Yv is p¡. jointly normal).
Sometimes rnultiple regression is considered apart from the multi- + Theorem 3.15 In the multivariate normal distribution, y 2 , y 3 , ••• ,
variate normal. That is to say, let us assume that the random variable
y P ar.e jointly independent of y 1 - µ 1 - V 12V2z1(Y 2 - U 2).
e is normally distributed with mean O and variance u2 and that y¡ (i =
2, ... , p) are known. Also assume that the relationship of the random + Theorem 3.16 If cr 12 = cr 13 = · · · = <l¡p =O, the partial regres-
variable y 1 to these quantities is given by (3.19). This could be stated sion coefficients P; in (3.lD) are ali zero.
.'fi'f'':•"I''
j ••
1.
Q is the quadratic form (Y - µ.)'R(Y - µ.) of the joint distri- + 2y3y 4 - Oy 1 - 2y 2 - 6y3 - 2y 4 +8
bution of y 1 , y 2 , ••• , Yp· We shall find the following: (1) f(y 1 1y 2 , y 3 , y 4 ), (2) f 1 (y 1 ), (3) Pu>
It was shown in Theorem 3.6 that, if Y is distributed as the multi- (4) p 1n, and (5) the multiple-correlation coefficient p 1 of y 1 on y 2 ,
variate normal, then any linear combination of Y is distributed norm- Ya' and Y4·
ally. For example, if A = (a.,) is a p x 1 vector of constants, then First we shall find R and V. These can easily be found by examining
. p
only the second-degree terms of Q, that is, ay¡ + 2yi + 2yi + y¡ +
the scalar quantity A'Y = I <X;Y; is normally distributed. This will
'2Y1Y2 + 2y3Y4 - This gives
i==l
be generalized to the case where A is not a vector but a q x p matrix.
(: ~ : :¡ (-!
-l
+ Theorem 3.22 If the p x l vector Y is distributed normally with
mean p. and covariance V and if B is a q x p matrix (q ~ p) of R= V=
rank q, the vector Z = BY is distributed as the q-variate normal o o 2 1 o o : _ :21)
distribution with mean B µ. and covariance BVB'.
o o 1 1 o o -1
The proof of this theorem will be left for the reader.
This theorem in effect says tliat the distribution of q linearly Next we shall find p.. This can be found by solving iJQ/iJY =O. We
independent linear functions of variables that have a multivariate get
normal distribution is a q-variate normal. <JQ = 6y¡ + 2y2 - 6= o
Since a multivariate normal distribution is completely specified by ()yl
its mean and covariance, we shall use the notation "Y is distributed
N(p.,V)" to indicate that the p x 1 vector Y is distributed normally oQ = 2y1 + 4y2 - 2 =o
with mean p. and covariance matrix V. ºY2
If the covariance matrix of a normal distribution is diagonal, then,
as shown in Theorem 3. 7, the variables are mutually independent. In OQ = 4y3 + 2y4 - 6 =o
ºYa
particular, if the covariance matrix of a multivariate normal is a 2 1,
where CT2 is a scalar, it follows that the variables are independent and oQ = 2y3 + 2y.. - 2 = o
that each has a variance a 2 • If a vector Y is distributed N(p.,a21), oy4
~-.~·1'-~·
~ rr•' '
R 134 = (: :
o 1
~)
1
R!a\ =
-
(! ~ -
o -1
0
1)
2
So
o
l. We must exhibit the mean and varíance of the distribution of P13·2 = .J(!)(l) =O
f(Yi 1y 2 , Ya, y4). We shall do this two ways.
a. We shall use Theorem 3.10. We get 5. The multiple correlation coefficient of y 1 on y 2 , y 3 , y 4 is
Jv12 V;:lV21
Vi2 = (-l O O) U2 =(O 2 -1) Pi=
Jvu
v22 = oi
(
o1 -1
º) v;:f = (! : ~) where V¡¡ is defined in part 1 of this example. Thus
o -1 2 o l l
Pi= .J11_! = fI_
Ji ,J6
u1 + vi2 v;f(Y2 - U2> = 1 - !Y2 Another way to find p 1 is to find a 11. 234 first. This we do by striking
out rows and columns 2, 3, 4 from R and inverting the result. We
V11 - Viz v- 1
22 V21 -- ~
o -
_i._ -
15 -
1
3 get a 11 . 234 = r
Also, V 11 = a 11 = i; so
Therefore, f(Yi 1y 2 , y3 , y4) is a normal distribution with mean l - !Y2
and variance .¡.
_j~n -
Pi-
ª1i
O'u234 _
- Ji - A
-2--
o
l _ -
6
b. We shall use Theorems 3.20 and 3.21. In that case we get
R 11 = 3; Riii = l = variance. For the mean we get Problems
3.1 If y 1 , y 2 , and y 3 are jointly normal with quadratic form
oQ = 6yi + 2yz - 6 =o
ºY1 Q = 2y~ + 3yi + 4yi + 2Y1Y2 - 2Y1Y3 - 4Y2lJ3 - 6yi - &y2 + 10y3 + 8
(a) Find R.
Solving for Yi gives y 1 = --ly 2 + 1, which is the mean of the distri- (IJ) Find µ.
bution/(Y11 Yz,y3,y4)· (e) Findf(Y1 1 y 2,y3)·
2. To find the mean and variance of / 1{y 1), we strike out the rows (When a normal distribution is rcquired it will be sufficient to state the vector
mean ami covariance matrix of the distribution.)
and columns of p. and V that correspond to y'2, y3 , y 4• This shows us
3.2 · If a vector Y has a multivariate normal distribution with mean µ. =O
that Yi is distributed N(l, i). and covariance matrix ·
3. To find p 12 we must first find the covariance of the joint distri- '
bution of y 1 , y 2 • We get (let V* be the covariance matrix of y1 , y 2 )
Furtber Reading
1 C. R. Rao: "Advanced Statistical Methods in Biometrfo Research,'' John
and covariance matrix Wiley & Sons, Inc., New York, 1952.
2 S. S. Wilks: "Mathematical Statistics," Princeton University Press,
-~)
Princeton, N.J., 1943.
V= ( : : 3 T. W. Anderson: "An Introduction to Multivariate Statistical Analysis,"
John 'Wiley & Sons, Inc., New York, 1958.
-1 o 3 4 A.M. Mood: "Introduction to the Theory of Statistics,,, McGraw-Hill Book
find (a), (b), (e), (d), (e), (f), (g), and (h) of Prob. 3.2. Company, Inc., New York, 1950.
3.4 If a vector Y has a multivariate normal distribution with mean O and 5 M. G. Kendall: "The Advanced Theory of Statistics," vols. I, II, Charles
covariance matrix Griffin & Co., Ltd., London, 1946.
6 H. Cra.mér: "Mathematical Methods of Ste.tistics,'' Princeton University
V~(: _: -:)
Press, Princeton, N.J., 1945.
t¡ S. N. Roy: "Sorne Aspects ofMultivariate Analysis,'' J ohn 'Viley & Sons, Inc.,
.New York, 1957.
(~
2 o
V=
o 3 _¡)
o -4
(a) Find P12• P1a• P23• PH• P24.• and P3.t·
(b) Find p12,31 , illustrating Theorem 3.19.
3.6 If y 1, y 2 , and !Is are jointly normal and if
(a) p12 ::fo O and p13 ::fo O, can p23 = O?
(b) p12 = O and p13 = O, can p23 =fa O? .
3.7 Prove Theorems 3.16, 3.17, 3.18, and 3.19, and state the hypothcs1s of
each theorem as a condition on the covariancc matrix V.
3.8 Prove Theorem 3.21.
3.9 The bivariatc normal can be written
f(x,y) = J
2
exp { - 1 [(X---µz) 2
UzO'v V 1 - p (27r) 2( 1 - p 2 ) '1z
Find R and V, and show that var(x) =u;, var(y) =u;, and cov(x,y) = '1zO'vP·
DISTRIBUTION OF QUADRATIC FORMS 75
In vector notation this theorem could be restated: If Y is distrib-
uted N(µ.,I), then Y'Y is distributed as a noncentral chi-square
with p degrees of freedom and parameter A =!µ.'p..
Proof: The proof will consist in finding the moment-generating
'P
function of I y¡, finding the moment-generating function of the
4 i -1
distribution given in (4.1), and then using Theorem 2.1. The
p
moment-generating function of W = ¡ y'f is, by definition,
i=l
Distrihution of Quadi;atic Forms oo ioo · · · ioo ~
1
mw(l) = E(ew') =
i -CXl -CXl -oo (27T)
4.1 Introduction We shall work with the exponent, to get it into a more suitable
form. We have
If the reader is acquainted with the arithmetical techniques of the
analysis of variance, he knows that they involve the summing and -!L(y; - P,;)2 + tLy~
squaring of ohscnTations and solead to quadratic forms in the observa-
tions. The process known as the analysis of variance can be described = -!(L!fi- 2tL!fi - 2Ly¡µ¡ + L¡}¡)
ns the partitioning of a sum of squares into a set of quadratic forms.
:Eµ~
2
It is, therefore, olear that quadratic forms play an important part in - 1 - 2t[:E( µ¡ \ :Eµ¡ ]
many segments of statistical applications. In this chapter we shall
- - - 2- Y; - l - 2eJ - (1- 2t)2 +l- 2t
examine the distribution of quadratic forms when the variables are
Lµ~ ( 1 - -
=--• 1 -) - -
1 --2t µ. )2
L ( y.---•-
normal. 2 1 - 2t 2 l 1 - 2t
Thus we get
4.2 Noncentral Chi-square
Since Y'Y is a sum of squares, we know that, if Y is distributed
p
N(O,I), then Y'Y = I y¡ is distributed as x (p), that is, as a chi-square
2
)2] dy
i=l xexp [ - -l -- :2t
E ( y.---'-
µ.
1 ···dy
variate with p degrees of freedom. We shal1 also investigate the dis- 2 ' l-2t 1J
f00f(W) dW = i e-?i =
+ Theorem 4.3 If a vector Y is distributed N(¡i.,u21), then Y'Y/á2
1 has the noncentral chi-square distribution with A = fL' ¡i./2a2 •
Jo i=O i!
This proof also is left to the reader. ·
since we can interchange the operations of integration and infinite
summation. The moment-generating function of the noncentral + Theorem 4.4 If a vector Y is distributed N(¡i.,D) where D is
chi-square is diagonal, tlien Y'D-1 Y has the noncentral chi-square distribution
oo oo e-A,ti(w)l(P+2i)-lcw/2 with p degrees of freedom and parameter A = i¡i.'D-1µ.
mw(t) = E(ew 1) =
J: ewt 2 . dw Proof: Sin ce D is positive definite, we know that there exista a
O i=O i!2l(P+2i)r(.P ~ 2i) nonsingular p x p matrix B (not orthogonal) such that B'DB = l.
Let Z = B'Y; by Theorem 3.22, Z is distributed N(B'¡i.,I).
co e-AA,i Loo (w)!(J>+2i)-le-w/2 So Z'Z is distributed as the noncentral chi-square with p degrees
=}; - - e"'t dw of freedom and parameter A = !¡i.'BB'¡i.. But from B'DB = I,
_ i==O i! o 2 l(P+2i)r(P ~ 2i) we get BB' = D-1 ; so Z'Z = Y'BB'Y = Y'D-1 Y and ).. =
!11-'BB' IL = i11-'D-1 ¡i., and the proof is complete.
The integral is, by definition, the moment-generating function
of the central chi-square distribution, and equals (1 - 2t)-H11+ 2 0. 4.3 N oncent.ral F
Thus
Since the noncentral chi-square has properties similar to those of the
(4.3) central chi-square, we consider the possibility that the ratio of two
noncentral chi-squares might have sorne properties similar to
Comparing Eqs. (4.2) and (4.3), we find that Snedecor's F distribution. In fact, the ratio of a noncentral chi-
square to a central chi-square will play an extremely important role
mw(t) = mw(t)
in the theory of the power of tests in regression and analysis of variance.
and, since they exist in so me neigh borhood of zero, we use Theorem Therefore, we shall derive what is termed the noncentral F distri-
2.1. This completes the proof. bution.
If A. = O, that is, if µ 1 = µ 2 = · · · = µ'P = O, the noncentral chi- + Theorem 4.5 If a random variable w is distributed as x' 2(p, íl),
square distribution degenerates into the central chi-squaredistribution. that is, as t.he non~entral chi-square with p degrees of freedom and
\~ ~~-
~ ....
parameter A., and if another random 'rariable z is distributed as If we let (v/2)(1 + I/x) = r, then dv = 2x/(1 + x) dr, and
x' 2 (q, O), that is, as the central chi-square with q degrees off~~eedom,
and if w and z are independent, then the quantity h(x) = L
DO
O¡x-l<11+2)
leo ( 2x --
)l(2i+11+0)r6(2i+P+o-2>e-r dr
i=O O 1 +X
U=~~
_pz co
0 .xl<2i+i>-2>r
'
(2i +2P + q) 21c2H11+0>
is distributed as the noncentral F distribution with p,q degrees = i~O (1 + x)l(2i+11+0)
of freedom and with noncentrality parameter A.. The frequency
function of u is If we make another substitution, u ·= (q/p)x, we get.
DO
r (2i + 2p + q) (l?_)l(2i+JJ)
q
A_ie-A
'U!(2i+J>-2) .
r(2i +2.P + q) (l!_q)l(2i+i>)
00
Aie-A
ul<2HP-2)
g(E2; p,q,l) = i
w
Let - =X W =V . ~-cl(_E2)l<2i+1>-2>ci _ E2)l<11-2>
z
'=º r(¡) r(2i: P) i!
The ~acobian of the transformation is v/x2 • Substituting, we get . o ~ E2 ~ 1 (4.5)
HA =O, g(E2 ; p,q,A. =O) is the beta distribution. Similarly, when
l =I= O, we term g(E 2 ; p,q,A.) the noncentral beta distribittion.
We are interested in evaluating the integral
where
o,= it2•1•rW21121+»r(2i ~ P)
E!
L 0
g(E2; p,q,IL) dE2 (4.6)
The marginal distribution of x is where the quantity ~ is obtained from the integral ..~01 • A ¿'%.
l ~~,.~~l~ •• "'..,.<~~.
.
J,
Ea.
1 g(E2; p, q, A = O) d.E2 = IX b .... )(
)
1( E>
-:::-- 0
¡:¡ • 7)
...~
~~ ~·
..,,v:,,..._....
~º
fi'AC. ClENnIA fil
80 LINEAR STATISTICAL l\fODELS DISTRIBUTION OF QUADRATIC FOR1\1S 81
where IX is given. Before cxpla.ining how the tables prepared by P. C. here J is the degrees of freedom in the numerator of the F statistic.
Tangare used to evaluate (4.6) and (4.7), we shall discuss the impli. ;he pro~edure for computing the power {J(A.) for a given l is as follows
cation of these integrals in the F' distribution. If we transform from (f and/2 will be given):
E 2 to F' by the relationship \. Choose the probability of type 1 error IX; that is, set IX = .01 or
E2 = pF' ex = .05.
q+pF' 2. Find E! from the Tang tables.
we get J. 1
g(E2 ; p, q, Á =O) dE 2 J. 00
f(ll'; p, q, ¡ =O) dF'
3. Compute <fo = V2l/(/1 +
1). _
IX= 2
=
4. Find P(II) for the appropriate values off1 , f 2 , and q,.
E« F«
5. Then {3(l) = 1 - P(Il).
where F« = qE';_/(p - p.E;_). If a. = .05 (say), we can read F« from Sorne examples will be given in thc use of the tables.
Snedecor's tables for the F distribution with p and q degrees of freedom. 4.3.1 Example. If « = .05, p = 6, and q = 10, we see that
For example, suppose we wish to test thc hypothesis l = O in the ,E20 ,, = .659. Continuing the example, we see that, if A = 14 (say),
F' distribution and use a type 1 error size IX, where the interval tl~;n </> = 2, and {J(l = 14) = 1 - .142 = .858. That is, when
F« ~ F' < oo is the critica! region. In other words, if the observed .a = 14, the probability of rejecting the hypothesis that A. = O is equal
F' falls in this interval, we reject the hypothesis A. = O. The power to .858 when a type I error probability of .05 is used.
of the test {J(J..) is the probability that the observed ll' falls in the critical
region F« ~ F' < oo when A =fa O, and is equal to the integral 4.3.2 Example. Supposep = 2,q = 4;find{J(l = 2l)when1X = .01.
Now E~01 = .900, </> = V 13'fL = 5; so_ l - {J(A. = 1/ ) = .095 and {1(.A. = 12.!i)
{3(Á) = f.
F«
00
f(F'; p,q,l) dF' (4.8) = .905. Thus, if we are usingF' as a test criterion, we compute E 2 and
reject H 0 : l =O if E2 > .900. The probability of rejecting H 0 when
If in (4.8) we transform from F' to E 2 , we get l = Il. is equal to .905.
4.3.3 Example. Suppose that a random variable y is distributed
{3(J..) = f,1, g(E2; p,q,l) dE2 N(µ,l) and that we want to test the hypothesis H 0 : µ =O. Suppose
1'Ja. further that we decide to take a random sample of size n (denoted by
So the integral in Eq. (4.6) equals l - {"J(J..), or unity minus the power y 1 , y 2 , ••• , Yn) from the distribution and use
of the test, which is the probability of a type II error.
U=
y2n(n - 1)
Since most of the test functions in regression analysis and in the :E(y. - 'fi) 2
analysis of variance are based on the F distribution, it is apparent that as the test function. In order to find the critical region for a fixed
the integral in (4.6) will be very important. The tables calculated type I error, it is necessary to determine the distribution ofthe random
by P. C. Tang allow us to obtain {J(l) for various values of p, q, A. and variable u when H 0 is true, i.e., \vhen µ =O.
for IX = .05 and .01. -
Since the basic variables y 1 , y 2 , ••• , Yn are independent and normal
Tang evaluated the integral
with mean O (assuming H 0 is true) and variance 1, it follows that
is known that l:(yi - 'fi) 2 has the central chi-square distribution even we equate the moment-generating function of v to the moment-
if µ =fa O, and the two are independent. Thus, when µ -:f: O, the test generating function of Y' A Y, we ~et
function u is distributed as F'(l, n - 1, nµ 2 /2). If n = 7, we reject p
H 0 if the calculated u is greater than 5. 99. The power of the test II (1 - 2tdii)-1 = (1 - 2w1c12
i=t
can then be evaluated for different · values ·of .A. For example, if
It is clear that there exists a neighborhood of zero such that, for
A. = 8 (i.e., if µ 2 = 1/!), then cf> = 2.8 and p (A. = 8) = .899. If n = 21
all t in that neighborhood, the quantities on the left and right
insteasJ. of 7, then E~05 = .179 and p (A. = 8) is about .955. This gives
sirles of the identity exist. It can be shown that, for this to be
us sorne idea of the value of increasing the number of observations
true; k of the d¡¡ must be equal to unity and the remaining dii
that are taken.
must be equal to zero. But, since du are the characteristic roots
of A, this implies that A is idempotent of rank k.
4.4 Distribution of Quadratic Forms We shall extend this theorem to the case in which the mean is not
zero. The proof foll~ws along the lines of the proof of Theorem 4.6.
Since the <:Iistribution of quadratic forros will be very important to
us in succeeding chapters, we shall prove sorne theorems that will + Theorem 4.7 If Y is distributed N(µ.,I), then Y'AY is distributed.' ( )
enable us to ascertain the distribution by examining only the matrix as x' 2 (k,A.), where A. =-!µ.'Aµ., if and only if A is an idempotent ·
of the quadratic form and the covariance matrix of the random vari- matrix of rank k.
ables. • Corollary 4.7:i If Y is distributed N(µ.,a 2 1), then Y'AY/a2 is-
+ Theorem 4.6 If the random vector Y is distributed N(O,I), a distributed as x' 2 (k,A.), where A = µ.' Aµ./2a 2 , if and only if A is an
necessary and sufficient condition that the quadratic form Y' AY idempotent matrix of rank k.
be distributed as z2 (k) is that A be an idempotent matrix ofrank k. We shall now prove sorne theorems that will help us to determine
Proof: We shall prove sufficiency, but shall only sketch the proof the distribution of a quadratic form Y' AY when the vector Y has as
of necessity. If A is idempotent of rank k, there exists an or- its covariance the matrix V instead of the identity matrix l.
thogonal matrix P such that P'AP = (~ :). where 1 is the + Theorem 4.8 If a vector Y is distributed N(O,V), then Y.'BY
is distributed as x2 (k) if and only if BV is idempotent of rank k.
k X k identity matrix. Define a p x 1 vector Z by the relation- Proof: By Theorem 1.22, there exists a matrix C (not necessarily
ship Z = P'Y. Then Y'AY = Z'P'APZ = z~z 1 ; where Z = orthogonal) such that C'VC =l. If we let Z = C'Y, Z is dis-
(~j- By Th:orem 3.22, Z is -~&tributJiN(O,I), and it follows tributed N(O,I). Also, Y'BY = Z'C-1BC'-1Z, and, by T-heorem
4.6, it follows that Z'(C-1 BC'-1)Z is distributed as z2 (k) if and
only if c-1 BC'-1 is idempotent. That is, we must show that
that Z~Z 1 = ! z;, which is a sum of squares of independent
C-lBC'-1 is idempotent ifandonlyifBVisidempotent. HBVis
i=l
84 LINEAR STATISTICAL l\IODELS DISTRIBUTION OF QUADRATIC FORMS 85
and C-1BC'-1 = C-1BC'-1C-1BC'-1. Also, if c-1Bc'-1 is idem: . = Z'DTZ = L duzJ, where du is the jth diagonal element of
j:a 1
potent, then C-1BC'-1 = C-1BC'-1C-1BC'-1 = c-1BVBC'-1 and D 1 ; if we apply the same transformation to Y'BY, we get
BV = BVBV. This completes the proof. Pi +Pi
Y'BY = Z'P'BPZ = Z'DiZ = L d 2 /j, where d 21 is the jth
+ T~eorem 4.9 If Y is distributed N{µ.,V), then Y'BY is distributed i=P1+1
as x~(k,A.), where A. = iµ.'Bµ. and le is the rank of B, if and only diagonal element of D 2 • Since Y' A Y depends only on the first
if BV is idempotent. p 1 element of Z and since Y'BY depends only on the elements of
Proof: As in Theorem 4.8, if we Jet Z = C'Y, then Z is distributed z from p 1 + l to p 2 and since all the elements of Z are inde-
N(C'µ.,I), and, byTheorem 4.7, Y'BY = Z'(C-1BC'-1)Zisdistrib- pendent, Y' A Y and Y'BY are also independent. This completes
uted as x' 2 (k,1.) if and only if C-1BC'-l is idempotent (where the sufficiency part of the proof. The proof of necessity will be
A = tµ.'CC- 1BC'-1C'µ. = iµ.'Bµ.). It is necessary, then, to show left for the reader.
that c-1BC'- 1 is idempotent if and onJy if BV is idempotent; + Theorem 4.11 If Y is distributed N{µ.,I), the set of positive semi-
but this has been proved in Theorem 4.8. definite quadratic forms Y'B 1 Y, Y'B 2 Y, ... , Y'BkY are jointly
independent if and only if BiB; =O for ali i =fo j.
4.5 Independence of Quadratic Forms Proof: The proof is an extension of Theorem 4.10.
+ Theorem 4.12 If Y is distributed N(µ.,I), the set of quadratic forms
The preceding four theorems give us methods of determining when· Y' A 1Y, Y' A 2 Y, ... , Y' AkY, where the rank of Ai is n¡, -are
quadratic forms are distributed as the central or noncentral chi-square. jointly indepen9,ent and Y' A¡Y is distributed as x' 2 (n¡,.íli), where
It is also very important to have methods available to aid us in deter- A¡ = lp.' A¡µ., if any two of the following three conditions are
mining when quadratic forras are independent. This will be the text satisfied:
of the following seven theorems. ( l) Each A¡ is idempotent.
k
+ Theorem 4.10 If Y is distributed N(µ.,I), the two positiva semi- (2) ,L A¡ is idempotent.
definite quadratic forms Y; A Y and Y'BY are independent if and i=l
only if AB =O. {3} AiA; =O for all'i :¡!= j.
Proof: To pro ve sufficiency, assume AB = O. Taking the trans- Proof: By Theorem 1.68, any two of the conditions imply the
pose of each side gives B' A' = O, but, since A and B are sym- third. By Theorem 4.7, condition (1) implies that Y'A¡Y is
metric, we get AB = BA = O; tha~ is to say, A and B commute. distributed as z' 2 (n¡,li) where ni is the rank of A¡ and íl.¡ =
Hence, by Theorem 1.32, there exists an orthogonal transforma- !p.'Aj(.1.. By Theorem 4.11, condition (3) implies that the set is
tion P su ch that P' AP = Dt and P'BP = n:, where Dt and jointly independent. Thus the theorem is establish~d.
n: are each diagonal. If AB = O, it follows that P' APP'BP = An important case of Theorem 4.12 that occurs very frequently in
O, which gives DfDi = O. Thus, if the ith diagonal element the analysis of variance and in the theory of regression is when :EA¡ = l.
of Dt is nonzero, the ith diagonal element of n:
must be equal to In this case, condition (1) is necessary and sufficient for condition (3)
zero. We can then write, without loss of generality, (and, of course, vice versa). So we shall state the following very
useful theorem. k
:. :.)
D1 O + Theorem 4.13 If Y is distributed N{µ.,I) and if Y'Y = L Y' A¡Y,
i=l
Di= O O either of the following two conditions is necessary and sufficient
(
o o for Y' A;Y to be distributed as x' 2 (n;,A;), where n; is the rank of
A; and A;= ip.'A;µ., and for the set Y'A 1Y, Y'A 2Y, ... , Y'AkY
where Di is a Pi X Pi diagonal matrix with rank Pi and O* is the to be jointly independent:
p 3 X p 3 null matrix where p 1 + p 2 + p 3 =p. If we Jet Z = {l) A 1 , A 2 , ••• , Ak are each idempotent matrices.
P'Y, then Z is distributed N(P'µ.,I). AJso, Y'AY = Z'P'APZ {2) A¡A; = Ofor all i =I= j.
86 LINEAR STA'.l'IS TICAL J\IODELS
DIS '.l'RIBU '.l'ION OF QUADRATIC FOR!\IS 87
The proof is immediate from Theorem 4. 12. Another important + Theorem 4.17 If Bis a q X n m atri x, A is an n x n matrix, a nd Y
form of t his theorem is
is distril>uted N(µ ,cr2 1), t he n t hc linear forrns BY a re independent
k
• Theorem 4.14 If Y is distributed N(µ ,I) and íf Y' Y =L Y' A .y of t he quadrat io form Y' A Y if BA = O.
i~ l ' ' Proof: Suppose Pisan orthogon al matrix such that P '.AP. = .D ,
a n ecessary and sufficient cond ition that Y' A iY be distributed as
x' 2(n;,A;), where n¡ is the rank of A ; and A¡ = i µ ' A ;µ , and for the
Y ' A 1 Y, Y' A 2 Y, ... , Y'AkY to be j oint ly indepe ndent, is that the
1
~::·;::~:.~;~go::tm:tcix~p ~:dP~Y ~ (~;' th:). :~:,:~:::
r a nk of the sum of the A ¡ be equal to the sum of the ranks of
the separate A ¡; t hat i~ to say, t hat Lp(A;) = p(LA;) = p(I). a diagonal m atrix with nonzero elem e nts on the d iagonal. Now
P roof: It was shown in Theorem 1.69 t h at for t he sum ofthe ra nks
óf t he A; to be equal to the ra.nk ofth e sum ofthe A ; is a necessary 1
O= BA = BAP = BPP'AP = CD = ( Cn C
1-j (º1 º) = (º)
a nd sufficient condition for condi tions 1 and 2 of Theorem 4.13. ,C 21 C2 O O O
Hence the result.
This impli es that C 11 D 1 = O and C 21D 1 = O, which imply
The fa mo us Cochran-Fis her theorem is a special case of Theorem C 11 = O a nd C 21 = O. So C can be \\Titten C = (O,C 2 ), where
4. 14, n a mely, for the specia l case when µ = O. The proof of Theore 111 C12\
4. 14 was first g ive n by lVIadow. (
C2 = C 22J" Now
• Theorem 4.15 If Y is distributed N(µ ,cr I), the positive semi-
2
1
d e finite quadratic forms Y' A Y and Y' BY a re independent if BY =
.
BPP'Y = cz = (O, C.)(z )
- Z2
= c.z2
-
tr(A B ) = O, or , in other wonls, if the covariance of Y' A Y and
Y ' BY equals zero. and
k D1
• Theorem 4.16 Let Y be distributed N(µ ,cr2 I) , a nd let 2: Y'A;Y = Y' A y = Y'PP' APP 'Y = Z 'DZ = ( Z~, Z~) ( O
i= I
Y ' Y, where the rank of A ; is n;. Any one of the three conditions
listed below is a necessary and s ufficient condition that the folJow- So BY d epend s on t h o ele me nts ofZ in Z 2 , a nd Y'AY depends on
ing two statements be true : the elements of Z in Z 1 , but a ll the elements in Z are ind ependent ;
(1) Y' A ,Y/cr 2 is distribu ted as x' 2 (n;)•;), where A;= µ 'A ;µ /2cr2. hence, the elernents of Z 1 are independent of t he elements in Z 2 •
(2) Y ' A ;Y and Y' A;Y are independent if i =F .i· Therefore, t he result follows.
The con di ti o ns are:
(1) A ; is idempotentfor a ll i = 1, 2, .. . , k. 4.7 Expected Value of Quadratic Forms
(2) A ;A; = O for all i =F j.
k
\Ve shall now state two theorems con cerning the expected value of a
(3) .L ni
•- 1
= n; that is, the rank of tho sum of the A ¡ is equal to the quadratic form.
sum of the ranks of the A ¡.
+ Theorem 4.18 lf Y is distributed with mean O and covariance
The proofs of Theorcms 4.15 a nd 4.16 wi l l be left to the reader. matrix cr2 I , the expected value of the quadrati c form Y' AY is
equal to cr2 tr(A).
4.6 lndepende nce of Linear and Quadratic Forms Proof: E(Y' AY) = E(l,Y;Y;ll·;;) = E(2,a;¡y¡) + E(2,2:y¡yp;;).
ii ,,. t~~
By Theorem 3.22 we know that, if Y is distributed N(µ ,V) and if But, if i =F j , then E(Y;Y;) = E(y;)E(y;) = O; h e nce, E(Y' A Y)
B is a kno wn matrix of constants, BY is d istributed N(Bµ ,BVB'). L a;;E(y¡} = a 2 2: a;¡ = cr2 tr(A).
It will be important to be able to d etermine when the linear forms BY i i
and a quadratic fo rro Y' AY a re statistioally independent. This is the + Theorem 4.19 In Th eore m 4. 18, if A is a n idempotent m atrix of
substance of the next theorem.
rank k, the expect ed value of Y' AY is equal to ka 2 •
88 LINEAR STATISTICAL l\IODELS DISTRIBUTION OF QUADRATIC FORl\lS 89
The technique of analysis of variance can be viewed as the process • Theorem 4.23 Suppose u is distributed as F'(p,q,Jt.); then the
of partitioning a sum of squares Y'Y into component parts such as random variable v = u/k is distributed approximately as F(r,q),
8 where k = (p + 2.A.)/p and r = (p + 2A.) 2 /(p + 4.A.).
Y'Y = 2 Y' A,Y. If Y is distributed N(µ,u 2 1) and if (say) µ' Ak = O
i=l Using this theorem, we obtain for Eq. (4.8)
and if we desire to test the hypothesis µ'A t = O (say), then we can
immediately see the utility of being able to determine the distribution {J(l) = f,eo f(u; p,q,l) du
and independence of the Y' AiY. For, if we choose as a test function Fa.
u= Y'A,Y/Y'AkY, we can determine the distribution of u; if, for
which is approximately equal to
example, A, and Ak are idempotent and A,Ak = O, then (nk/n,)u is
distributed as the noncentral F with noncentrality l = EL' A,µ/2a2.
Also, we see that l = O if and only if µ'A t = O, and in that
case (nk/nt)u is distributed as Snedecor's F. Thus we can choose a where Fa. is such that
f ClO
(l/k)Fa.
g(v; r,q) dv
j(u; p,q,A.) is the noncentral F distribution, and g(v; r,q) and f(u;
4.8 Additional Theorems p, q, A. = O) are central F distributions. .
4.8.1 Example. We shall illustrate Theorem 4.23 by referrmg
+ Theorem 4.20 Suppose that the n x 1 vector Y is distributed to the example of Art. 4.3.1. We have a= .05, p = 6, q = 10, and
N(µ.,D) where D is diagonal. Then Y' A Y is distributed as we want to find {3 (l = 14). The quantities in Theorem 4.23 are
x' 2 (n - 1, A.), where Ji.= !µ'Aµ, if A = D- 1 - (D-lJD-1/l'D-11).
Also, A. = O if µ = 1µ. (1 is a vector with each element equal to k = p +
2.A. = 6 28 = 5.67 +
1, and J = 11'.) . p 6
+ Theorem 4.21 Suppose that the n x 1 vector Y is distributed r = (p + 21)2 = (6 + 28)2 = 18.06
N(µ,V); then Y'AY and Y'BY are independent if and only if p + 4A. 6 + 56
AVB =O.
Also, ~05 = .659, and from the F table we get F.05 (6,1.0) = 3.22. So
+ Theorem 4.22 Suppose that the n x 1 vector Y is distributed
3 22
N(µ,V). .!_F = · = .568
k+l k IZ 5.67
Let Y'AY = 2 Y'A¡Y, and suppose that Y'AY, Y'A 1Y, ... ,
To find {J (A. = 14) we must find the probability of exceeding an F
i=l
Y' AkY are each noncentral chi-squares with parameters n,l; value of .568 with 18 and 10 degrees offreedom; i.e., we mustevaluate
n 1 ,Jt. 1 ; • • • ; nk,A.k; respectively. If Y'Ak+ly is nonnegative (i.e.,
semidefinite), it is distributed as z' 2(nk+ 1, Ak+i), where nk+l =
k k
fQO
.568
g(v; 18,10) dv
n - 2 n¡ and lk+I =J. - 2 A.i, and Y'A 1Y, ... , Y'Ak+ly are
i=l i=l where g(v; 18, 10) is the central F distribution with 18 and 10 degrees
independent. of freedom. The integral is approximately equal to .86, which agrees
The extent of Tang's ·tables is limited, and it might be desirable to well with the value .858 obtained by using Tang's tables.
find the value of the integral in Eq. (4.8) for values of p, q, and J. not
included in the tables. Patnaik gave an approximation that will allow Problems
the noncentral F to be used to evaluate Eq. (4.8). This is given in the 4.1 Find the mean and variance of a random variable u distributed as z12(p,i.).
following theorem. 4.2 Find the mean of a random variable u distributed as F'(p,q,l).
DISTRIBUTION OF QUADRATIC FORMS 91
90 LINEAR STATISTICAL :MODELS
4.3 }i'ind the mean of a random variable u distributed as E2(p,q,J..). 4.15 In Prob. 4.14, find E(u).
4.4 Show, by using the transformation 4 .16 lf the 3 x l vector Y is distributed N(O,I), find the following:
(a} E(Y'AY).
E2 = pF (b) E(Y'BY).
q+pF (e) The joint distribution of Y'AY and Y'BY.
(d) The distribution of u = Y'AY/Y'BY, if
that 1(205 in Tang·s tables cot'r<.•sponds t.o the F. 05 rn.h1P of thc ccntt·al P for thc
followin~ vah1<'s of p and q:
p 6 7 2 8
·
=!)i B = (! : !)
} ! !
q 3 10 20 60
u 4.17 lf the p X l vector Y is distributecl N(O,I),
4.5 Evaluate the following integral (and find E~05 ):
(a) Find the matrix of Q1 = p'fi2.
p
(b) Find the matrix ofQ2 = :L (Yi - 'fi) 2 •
p(J..) = f12 g(E2; p,q,.t) dE2
(e) Find the distribution of Q1 •
i=l
JE.o¡¡ (d) Find the distribution of Q2 •
for the following values of the parameters -(use Tang's -tables): (e) Show that Q1 and Q2 are independent.
(j) Find the distribution of Q1 /Q 2 •
. p
2 4 5 6 3 7 7
(g) Show that l 'Y and ! (Yi - 'fi) 2 are_jndependent.
i-1
q 6 2 18 30 5 2 4 (h} Find the expected value of Q1 .
(i) Find the expected value of Q2 •
6 10 18.75 15 4.5 36 64 4.18 Prove Theorem 4.19.
4.19 If the n x 1 vector Y is distributed N(µ,a 2 1), find the expected value of
4.6 Prove Theorem 4.2 by the use of moment-generating f unctions. Y'AY, where A is idempotent of rank p.
4. 7 Prove Theorem 4.3. 4.20 Using Prob. 4.1, find the noncentrality paramoter A of the distribution
4.8 If Y is distributed N(µ.,V), prove that the quadratic forro of the p-variate Y'AY/a2 ·¡f Y is distributed N(µ,o- 2 1) and if A is idempotent.
normal 4.21 Prove Theorem 4.20. Hint: Show that AD is idempotent.
Q = (Y - µ)'V-l(Y - µ.) 4.22 Prove Theorem 4.21.
is ·distributed as x2 (p). 4.23 If the n x l vector Y is dist.ributcd N(O,A), whm·e A is positive semi-
4.9 If Y is distributed N(µ,V), prove that Y'V-lY is distributcd as x'2(p,.i.), dcfinite of rank k, use Theorem 4.9 to provc that. Y'Y is distributed as z'2 (k,0) if
where .t = !11:v-1µ. - · ~ ·-: · A is idcmpotent.
4.10 lf X is an n x p matrix (n > p) of rank p, show that A = X(X'X)-lX' i:;: 4.24 Suppose that y¡ (i = l, 2, ... , n) are independent and y¡ is distributed
is idempotent. N(µ¡,af). Show that
1
4.11 If the n x 1 vector Y is distributed N(O,I), find
!n 2 (y¡ - y•)2
E{Y'fl - X(X'X)-lX']Y} i=l <1¿
4.12 lf the n x 1 vector Y is distributed N(O,I), find the distribution of is distributed as z'2(n - 1, l), where J.. = O, if µ 1 = 11 2 = ··· = Pn· The quantity
(a) Y'X(X'X)-lX'Y. y* is defined by
(b) Y'[I - X(X'X)-lX']Y. * _ :E ( l/af)y ¡
4.13 Ifthen X lvectorYisdistributedN(X~,l),where~isap X lunknown - :El/af y
vector and X is as in Prob. 4.10, find the distribution of Hint: use Theorem 4.20.
(a) Y'[X(X'X)-lX']Y = Q1 4.25 Let Y' = (y,x) be distributed as a bivariate normal with mean µ.' =
(b) Y'[I - X(X'X>-1 X']Y = Q2 (µ , Pz), variances 1 and correlation p. Let Y 1,Y2 , ••• , Yn be independent
1
(e) and show that the quantities Q 1 and Q 2 areindependent. samples from this distribution. Show that for p2 =/= l
4.14 In Prob. 4.13, find the distribution of
n -pQ¡
:E (xi - 2pX;Y; + Yi)
U=--- (a) q = l - p2
p Q2
92
LINEAR STATISTICAL M:ODELS
is distributed as x'2(2n,A.), where
,l
1
= n(µf - 2Pµ1µ2 + µ~)
2(1 - p2)
(e) Use the resulta of (a) and (b) and Theorem 4.22 to find the distribution of
5
q
2
= L [(x, - x)ª - 2p(x, _ x)(y, _ 'fi) + (y¡ 'fi) 2]
1 - p2
Notice that q = q1 + q2 • Linear Models
4.26 Use Theorem 4.23 to evaluate the i t 1.
p, q, and l in Prob. 4.5. negra in Eq. (4.8) for the values of
Further Reading
5.1 Introduction
1
A. C. Aitken: On the Statistical lnde end .
Normal Variates Biometrika vól 37 p93 9ence of Quadrat1c Forms in
2 A T e · N ' ' · • PP· - 6, 1950. One of the ·aims of science is to describe and predict events in the
. . ra1g: ote on the Independence of Certain Quad t. F world in which we live. One way this is accomplished is by finding a
Math. Statist., vol. 14, pp. 195-197, 1943. ra ic orms, Ann.
3 P. C. Tang: The Power Funct" f th A · formula or equation that relates quantities in the real world. We
Tables and Illustrations for ~~:irº Useª sZz~~~•s;f Variance Tests with may be interested, for example, in the relation between temperature
pp. 126-146, 1938. ' is· esearch Mem., vol. 2, and pressure in a chemical process or in the relation between the number
4 H. Hotelling: On A Matrix Theorem of A T e ·
15, pp. 427-429, 1944. · · raig, Ann. Math. Statist., vol. of apples on various trees in an orclíard and the amount of fertilizer
5 that each tree receives, etc.
W. G. Cochran: The Distribution of Qu d t" F .
l'roc. Cambridge Phil. Soc., vol. 30, p. l ;8,r~9~4; orms m a Normal System, Suppose1 two variables U* and V* are functionally related by
6 W. Madow: The Distribution of Quad t" F . j( U*, V*) = O. If U* and ·V* can take on every value in a given
Random Variables Ann Math Stati:st ra
7 F A G b'll ' . .
llorms
"vo · • PP· 100-101 1940
;e
Non-central Normal m intcrval, then they cannot be measured exactly, and, instead of observing
i~ th.e Gray 1 1ªL~d G. l\Hfarsaglia: ldempotent Matrices and Q~adratic Forros U* and. V*, we actually observe u and ·v, where u = U* + d and
ti = V* + e, where e and d are errors of measurement.
enera mear ypothesis An Mi th o · However, if
1957. ' n. a . otatist., vol. 28, pp. 678-686,
the error of measurement is extremely small, we may be willing to
8 !~:· !:~~ ~~ ~t· T.voCl.ra2i9g:' pp.
• • UJ .,
On the Decomposition of Certain X2 Variables
608-610, 1959. , neglect it. So me may argue that f (U*, V*) = O is simply a mathe-
9 0. Carpenter: Note on the Exte · f C · ,8 matical abstraction and that no such functional relationship can exist
Variates Ann !lfath Sta/," t
10 R A h : " F" 1
r;;on
. . . is ., VO • ' p. 455, 1950.
°
raig Theorem to Non-central
in the real world. Be that as it may, it is nevertheless true that the
. . IS er. Stat1st1cal l\lethods for Research '"orkers " or d concept of a functional relationship between sets of events in the real
B oyd, Ltd., London, ¡946 . , 1ver an
world is important. And, although the relationship may not be exact,
it may be so clase that the approximation is invaluable in prediction.
Therefore, when we say that a functional relationship exists among
1 Throughout this chapter we denote mathematical variables by starred
capital letters U*, Vt, v:, ... if they are not observableand by U, V1 , V2 , ••• if
they can be observed. We denote random variables by lower-case starred
letters yf, uf, vf, ... if they are not observable and by u, Xp x 2 , ••• if they can be
observed. The letters a, b, d, e will be unobservable random variables. Greok
letters will denote unknown pa.rameters.
93
1
94 LINEAR STATISTICAL MODELS LINEAR MODELS 95
a set of variables, we shall mean that in the real world the function n If a functional relationship can be found that relates observable
describes the relationship among the variables to a very close approxi- -· variables, then a knowledge of X 1 , X 2 , ••• , Xn and· oí the function f
mation. can be used to predict Y accurately. For instance, the law of a body
For example, consider the relation between time and the distance falling in a vácuúm under the influence of gravity is Y = {JX 2 , where
that a particle falls under the influence of gravity; that is, S = {JT2. y denotes the distance the body falls and X denotes the time the body
For many practica! purposes we consider this a-functional relationship, fa.lis. lf an experiment is run where a body is dropped and the time
and we can predict the distance quite accurately in a real-world and distance observed, then fJ can be evaluated. However, if another
experience. Suppose 8 and T 2 cannot be observed. We should then reading of time and distance is taken and fJ is evaluated again, the
write the equation as 8* = {J(T*) 2 • If we can observes and t 2 , where second {J will be quite likely not to agree with the first. The reason
s = 8* + e and t 2 = (T*) 2 + d, we can substitute and obtain for this disagreement may be either ( l) that the relationship is not
s = {Jt 2 - {Jd + e or s = {Jt 2 + b, where b is a random error. The given by Y = {JX 2 ; or (2) that, even though the relationship is given
problem here may be to estimate fJ by observing values of s and t 2 • by y = {JX 2 , the time and distance cannot be measured accurately,
There may be events in the real world thatare related notfunctionally and so an error is introduced; this second type of error will be called
but in a more obscure manner. For example, consider the weight w measurement error. If reason 1 is operative, then the real functional
and the height h of individuals. Therc scems .to be no formula that relationship may be Y = f(X,Z), where Z is the initial distance the
will enable us to predict the height of an individual from his weight. body is from the center of the earth, or the distance may actually
Although there does not seem to be any functional relation between depend on many other factors, such as the mass of the body, the position
height and weight, there does seem to be sorne kind of relationship. of the moon, etc. If the distance dependa on many factors X 1 , X 2 ,
For example, let us assume that w and Ti have a bivariate normal distri- . .. , Xn, besides the time X, the relation could be written Y =
bution. In the conditional distribution of k given w, there is a linear /(X,X 1 ,X 2 , ••• ,Xn) or, more explicitly,
functional relationship between w and the expected value of h. In
'j.. other words, if we divide the individuals into weight classes and
Y= {JX 2 + g(X,X1 , ••• ,Xn)
evaluate the average height of the individuals in each weight class, lf Y = {JX 2 + g(X,X 1 , . • • ,Xn) is a relation_s,hip that can be used to
we find that there is a functional (linear) relationship between weight predict Y, then observations will no~ Ü.gr~e exactly when the equation
and the average height of all individuals who have that weight. We y = px2 is used instead. This °diságréeirierit of observation with
see, therefore, that, even though we may not be able to predict w theoretical relationship will be called equation error, or, in other words,
exactly by lmowing only h, the quantity h may still be valuable for error due to the use of the wrong equation. It may be realistic to
predicting w. assume that, for repeated_ observations of distan ce Y and time X, the
In scientific investigations the concept of cause and effect is often variables X 1 , X 2 , ••• , Xn assume values such that g(X,X 1 , ••• ,Xn)
quite obscure, and we shall not enter into a discussion of it here. We · acts more or less as a random variable, and its distribution may be
shall not use cause-and-effect terminology to describe an equation such inferred. If this is the case, the relationship-can be written Y =
as Y = <X + {JX; we shall say .that X can be used to predict Y rather px2 + e, and Y cannot be exactly predicted from a knowledge of X only.
than that X causes Y. Although functional relationships are assumed to hold in many
In many fields of scientific endeavor, for example, in physics, rela- fields of science, such as pl)ysics, there are many scientifi.c areas, such
tionships can often be expressed as functional equations Y = F(X 1 ,X2 , as biology, economics, meteorology, etc., where relationships are much
... ,Xn)· For instance, Ohm's law states that the electromotive force more obscure. . .
Y is equal to the resistance X 1 of the conductor times the current X 2 ; For example, the yield of ;~heat in a given plot of ground cannot be
that is Y = X 1 X 2 • The law of gravitation states that the force of predicted accurately. Many of the factors affecting this yield are
attraction Y between two unit-mass points is inversely proportiona.I known, but the equation relating the quantities is obscure. It is
to the square of the distan ce X 1 between the points; that is, y = fJIx¡. known that tem¡mrature X 1 , rainfall X 2 , amount of ~unshine, X 3 ,
There are many other such relationships: Boyle's gas law, Kirchhoff's fertility X 4 , and many other factors influence the yield Y. Although
law in electricity, Newton's laws of force and acceleration, Newton's all the factora that affect this yield are not known and although the
law of cooling, and so forth. relationship is not known, it is useful nevertheless to assume that
96 LINEAR STATISTICAL MODELS LINEAR MODELS 97
there exist a fi?ite number of factors X 1 , • • • , Xn and a function g If the distribution of the random variables is given, this will be
such that the y1eld Y can be cxactly determined by considered part of thc model, and there may be unknown parameters
in the distribution. For instance, an equation S* = p(T*) 2 for the
y = g(X1,X2, ... ,Xn)
distance that a body falls in a vacuum is a model that does not contain
Thus, to find relationships that aid in predicting events in the real random variables. If s = S* + e, where e is a random variable,
world, wc shall assume that the following proposition holds: substitution in the first equation givcs s = p(T*) 2 + e as a model.
+ Proposition Suppose Y is a variable quantity in the real world If it is further assumed that e is a normal variable with mean O and
which we want to prcdict. Then there exist a finite numbe; variance a 2 , then this distributional property will be considered a part
of variables X 1 , X 2 , ••• , X n and a function g such that Y = of the model.
g(X 1 , X 2 , ••• , Xn)· . In general there is no re~son to restrict the equation in our definition
to any particular type, such as quadratic, exponential, etc. However,
We are ~ot saying that X 1 , X 2 , ••• , Xn are capable ofbeing observed, there may be certain instances when we want to do so. The major
~mt that, if they could be observed, we could predict Y exactly; that part of our discussion will concern equations that are restricted to be
1s, Y, X 1 , X 2 , ••• , X n are said to be functionally related. linear in a certain fashion. This does not imply that other equations
~.,or example, consider again the problem of predicting an individual's are not important but only that linear equations have received most
he1ght. We know that we cannot predict his ·height exactly from a attention from mathematical statisticians and that methods of treat-
k~owl~dge of his ~eight only. However, we may be able to predict ment are available for them.
lus height exactly 1f we use a number of other quantities, such as his Therefore, in succeeding chapters we shall be mainly concerned with
father's weight X 1 , his father's height X 2 , his mother's weight X 3 , linear models, according to the following definition.
etc. \Ve assume that the equation
+ Definition 5.2 By a linear model we shall mean an equation that
h = g(w,X1 ,X2 ,X3 , .•• ,Xn) involves random variables, mathematical variables, and param-
holds. \Ve might also want to assume that we can write this eters and that is linear in the parameters and in the random
variables.
h =o:+ pw + f(X1,X2 , ••• ,Xn)
and assume that, for a given value of w' the quantities X. 1> X 2' • • • ' X n For example, if y 0 , y 1 , y 2 are unknown parameters, then y 0 + y 1 X
change and f (X 1 ,X 2 , ••• ,X n) acts as a random error. The function g + y 2 Y =O is a linear model, and so is y 0 + y~ex + y 1 Y cos X =O.
in the proposition above may be·. -··.known and the elements X l> X 2t However, m_odels such as y~ + X sin y 1 + y 2 =O are not linear in
... ' xn ~ay be known,. but it still may be impossible to predict y the parameters y 0 , y 1 , y 2 •
In what follows we shall often use only two variables; however, the
exactly oWíng to our inability to measure sorne or ali ofthe.Xi exactly.
To sum up, we shall be interested in two type_s of error: resulta will hold for any finite number of variables.
1. Equation error: the error of not knowing the function g or of Let us consider the general linear model y 0 + y 1 X + y 2 Y = O.
not knowing all the variables X 1 , X 2 , ••• , Xn. . We could also have the model Yo + y 1 X* + y 2 Y* =O or Yo + y 1x
2. Measurement error: the error of not being able to measure all + y 2 y = O or even Yo + y 1 X + y 2 Y* = O, depending on whether
the xi exactly. the variables were random or fixed, observable or unobservable.
However, we see that it is impossible to have models such as y 0 +
y x + y Y = O, since this would imply that a random variable x was
1 2
5.2 Linear Models a function of parameters and of a mathematical variable Y.
5.2.1 Definitions and Classification. We shall give the defini- We can write the linear model
tion of a model that will be used in this book. (5.1)
+ D,efinition 5.1 By a model we shall mean a mathematical equation
hlvolving random variables, mathematical variables, and param- y=_ Yo_ Y1x º(5.2)
as
eters. Y2 Y2
!)8 LINEAR STA'.l'ISTICAL ll'fODELS 99
L I N EAR l\IODELS
aboru t y or E(y) or tl~e })~rametc~s in t he m,odel y = o: + o: X + . ely . The ob1·ect in this model is to observe values of y
espec t iv
0 1 1
ª2X2 +e whcn xi Yaned m sorne m terval. 'I he experimental design :nd estimate ai, ~' an.d a 2 (a2 is the_ variance of e). .
model is s_o mewhat different. In this model Xi takes only the valuea
F or. ox-ample , in measuri ng the rntrogcn .
content of t he fo h age o n a
. . · t' f
O and ~ . For c~ample, s uJ?pose :ve wan_t to examine the durabili·.,y of ccrtam · t l ·ee ' there a re two m a1· or sources of vanat10n: t 11e vana 1011 o
two d1ffe re nt kmds of pamt. .rhe pam ts are each subjectcd to a on t he tree and the variation due to the measuremcnt or
t he leav es ' h 1
friction machine, and the t ime it takes for t his machine to wear thc Jabo ratory error. S uppose we take n leaves from t h.e tree, w ere t le
paint_is r~corded. We shoul~ like to have a formula for predicting . trogen content of t he ith leaf is ai . In t h1s case t h e a 1 are
ac t na 1 111 . · ., s
the t im e 1t takes for each pam t to wear away. Paint 1 and paint 2 . d m variables from a distribution w1th vanance a¡. uppose m
bear no numerical relationship to each other. The prediction equation ian o . 1 't t t
measu rements are made on each leaf to determ1~e t le 111 rogen con en
can, however , be written y = o: 1 X 1 + a 2 X 2 , where X and X takc of that particular leaf. If we !et Y;; denote th_e Jt~1 m~asur~ment ~f the
1 2
t h e values O and 1 a n d where o:; is the t ime it takes paint i to wear away. ith Jcaf, the e;; a re random variables from a d1stnbut10n w1t h vanance
Thus, y= a 1 when X 1 = 1, X 2 = O; y = oc 2 when X 2 = 1, X = o. a2. 'l'he model can be written
1
Thercforc, _if we !et Y; be the time it takes for paint i to wear away,
~ve can wnte t he modcl y = ix 1 X 1 + o: 2 X 2 as Y; = <X;- Actually, y Y;;=µ + a;+ e ;; j = 1, 2, . .. , m; i = 1, 2, ... , n
~s a fun ction of X;, b u t, since X; = O o r 1, X; may not seem to appear
where y¡ 1 is the o bserved ni trogen con ten t of ~he j_th mcasurement on the
rn t he formu la. In addition, there may be an error in measuring y,
ith lcaf. The q uantity µ is a constant wh1ch 1s the average value of
or there may be other factors besides the effect of thc paint that cause
One obJ. ect of this model is to es ti mate ai and a2 from t he observed
values of Yu· Vio note t 1rnt t 11e van·a n ce o f· Y;; equa1s 0'1 + a 2 ,• that
y to take on a particular value. We may, t he refore, wish to write t he - y¡¡- 2
equation Y; = o:; + e, where e is a ra ndom error.
is, it, is a li near function of t..he variances.
For another example, suppose a researcher has developed a new
variety of wheat, which h e wants to compa re with a standard variety.
He wan ts to examine the yield of t he two varieties; so he p lants both 5.3 Model Classification
under uniform conditions. If wc Jet <7. 1 be t he average y ield of the In the previous section we discussed the models from t h e point of
n ew variety and <X 2 the average y ielcl of t he standard variety, we can view of a researcher_ In this section we shall define models_ as a
write the model for the observed y ield as
staListicia n Jooks at them . In subsequent chapters we shall d1scuss
y = ix1 X1 + ix 2 X2 +e in detail the five models defincd below. _
_ 5.3.1 Model 1. Suppose Xv X 2 , . • . , Xi.: are known math e matical
wh ere y is t he observed yield, e is an error, a ucl X 1 and X take on the ya riables, y is an observable random variable, {3 0 , f31, f32, · : · , f31.: ~l'e
2
values O a n d l. Alt h ough the experi menter may attempt to plan t unknown parameters, and e is an unobser vable ra ndom vanable w1th
t he two varieties under very similar condit ions, there a re a great mean O. Und er t hese conditions th e model
many factors, such as soil fertili ty, humidity, moisture, a nd so forth,
that rn a ke the observed yield diffe r from t he tr ue average y ield of t he Y = f3o + f31X1 + fJ2X2 + · · · + f31.:X1.: + e
variety undcr observation. These effects constitute t he random error
t erm e. For example, if X 1 = 1 and X 2 = O, t he n t he corresponcling will be clcfined as model l.
y value is y 1 = o: 1 + e. vVe see t hat t he observed value of y is not Man y of t he models discussed in the last section w~ll fit int~ the
t he true average effect o: 1 but is equal to ix 1 p lus the error due to t he framcwork of this model. For example, suppose t here is a funct1onal
uncontrolJed factors. W'e may want to estimate cx 1 , which is E(y ) . relationship between Y* and X 1 , given by Y* =~ o +f31X1. \~e
A similar s ituation holds for y wh e n X 1 = Oand X = l.
1 cannot observe J'*, but we observe y, where y = Y"' + e, whe1:e e ~s
2
5.2.6 C ompo n e nts - of-variance Model. This model is sim ila r a meas urcment error. vVe t he n h::we y = f3 o + f3 1X1 + e, wh1ch is
tv t he experimental desig n model in t hat the X variables again take model l.
t h e valucs O and 1 o nl y. The model can be writte n y = a X + S uppose the functional relationsh ip Y = g(_X 1,X 2 , . . . ,X~) can be
ª2X 2 + e. However, a 1 and a 2 are not parametcrs, but a re unobserv-
1 1
wn'tte n )r = ,.,R 0 ,_¡- f'R 1 X' i + J¿(X 2• · · · ,.;.\,"''
)' ) wluc h we a1)prox11nate by
1
able random variables from distri butions with varia n ces e?¡ and a~, " 11 = /30 + /31X1 + e
104 LINEAR STATISTICAL MODELS LINEAlt MODELS 105
T~1~s
is a mean-related model with equation error, but it fits the defi- 5.3.5 Model 5: Components-of-variance Models. Let a 1 , a 2 ,
mt10n of model 1. If y is not observable, we write it a be unobsenrable random observations from a distribution
••• , p • 2
with mean O and variance ªª; let b 11 , b 12 , ••• , bpq be unobservable
y* =Po + f11X 1 + e random observations from a distribution with mean O and variance a¡.
If Y = y*: + d, we get y =Po + p1 X 1 + a, where a = e + d. This Lct y¡; be an observable random variable such that
also satisfies the requirements for model 1. Y;; = µ + a¡ + bii
5.3.2 Model 2: Functionally Related Models with Vari-
ables Subject to Measurement Error. Suppose the mathematical whcre µ is an unknown parameter. Models such as these will be
variables Y*, Xi, ... , XZ are not observable but that y, x 1 , ••• , xk ca11 called nwdel 5. One object in this type of model is to estimate a!
be observed, where y = Y* +e and x. =X~ +e.1 (i = 1 2 k) and ai.
Suppose further that there is a functional 'relati onship 'a~~~~' th¿
mathematical variables, gi ven by Further Reading
1 l\f. G. Kendall: Regression,StructureandFunctionalRelationship, Biometrika,
Y* = Po + f11X: + · · · + {3kx: parts I, II, vol. 39, pp. 96-108, 1952.
2 H. Scheffé: Alternative Models for Analysis of Variance, Ann. llfath. StatÜJt.,
'Ve can substitute and write vol. 27, pp. 251-271, 1956.
3 C. P. Winsor: Which Hegression? Biometri.cs, vol. 2, pp. 101-109, 1946.
Y = Po + f11x1 + · · · + Pkxk + e - P1e1 - f32e2 - • • • - f3kek 4 H. F. Smith: "Variance Components, Finite Populations and Statistical
When the above specifications are met, we shall define the model as Inference," N. Carolina Inst. Statist. Mimeo. Ser. 135, 1955.
5 :M. H. Wilk and O. Kempthorne: Fixed, Mixed and Random Models, J. Am.
1nodel 2. Statist. Assoc., vol. 50, pp. 1144-1178, 1955.
5.3.3 Model 3: Regression Models. Suppose y, x1 , • • • , xk are
a set of jointly distributed random variables such that the expected
value in the conditional distrib. ution of y given •x.J. = X.i (i = I ' ...,,
?
... , k) is given by Po+ p1 X 1 + · · · + pkX1.:. We can then write
Yx = Po + /31X1 + · · · + P,.~Xk + e
When the above conditions are satisfied, we shall define the model as
model 3.
The only difference between, model 1 and model 3 is that the X.
in model 1 are mathematical variables, whereas in model 3 they ar~
particular values of random values.
5.3.4 Model 4: Experimental Design Models. Let y be a
random variable, and let X 1 , X 2 , ••• , X k each be equal to either O
or 1. If /3 1 , {3 2 , ••• , /3 k are unknown parameters, then
Ytr·k = µ + Pi + «¡ + · · · + e;r·k
where ¡3¡, «s, ... are unknown parameters and eii···k is a random ¡
variable. >
JllODEL 1 : GENERAL Ll NEAR HYPOTHESIS OF FULL RANK 107
Equation (6.1) is sometí mes referred to as a prediction equation.
Suppose we wish to predict, for example, what t he value ofy will befar
givcn set ofthe X; . Sioce e, and co nscq uently y, is a random variab le,
~'C cannot in genera l predict t he exact value of y for a given set of t he
X· but we may be able t o p redict an inte rval a nd establis h t hc prob-
~ bili ty t hat t he interval will con tain y . On t he other hand, we may
6 not be particularly interested in p rcdicting t he value of y, bu t we may
desire instead to estim ate E(y), t he expectecl value of y, p ertaini ng to
11 given set of the X; .
Suppose a physicist, in study ing t he motion of a certain type of
Model 1: The General Linear Hypothesis particle, concludes t hat there is a fun ction al rela t ionship between t he
distance d t h at the particle m oves in a certain t ime interval t. That
of Full Rank is to say, h e is willing to assume that t he fun ctional equation d = vt
rcl<Ltes t he d istan ce and the ti me a particle moves. Suppose further
thaL t he vclocity v is not kno"·n , so t he experimenter decides to
measure corresponcling values of d a nd t a ncl t hus ascertain v. H ow-
6.1 Introduc tion / cver, suppose he cannot measurc d acc urately . That is to say,
instcad of measuri ng d he can o nly measure y, where y = d + e,
J n t his chap.ter "'.º shall derive thc d istribution of pertinent statisLics where e is a measurement error t hat is nor ma lly and indepcndently
n ccded for est1mat1on of certain parameters in model 1 and for testing distributed a bou t a mean of O fo r each selected value of t. Therefore,
hypothescs abou t them. the model fi ts t he de finition ofj(y;x;{J) .
6.1.1 Definitions and N otation. Consider t hefrequency fun ction lf we can obtai n an estím ate of V (say, v), t hen the estimated distance
f(y; XvXz, · • • ,x,,; fJ ~· : .. ,{J,,) of a rand om variable y, which d epcnds el that th e par t icle trnvels duri11g t ime t is g iven by J, = vt. Jn t his
on JJ known quant1ties x 1 , . . . , .;P a nd on p unknown p arameters sit uation an experi menter may be interesteu in setting confide nce
fJy f32, .. · , fJ~ . .Tn ~his chap t er 1 we s ha ll in vestiga.te various proper- Jimits o n t he d istan ce t he part iclc t ravels in time t; Lh is wo uld
ties of the d1stnbu t10n of t he random variable y . This freq ucn cy necessitate setting confidence limi ts on E(y) = d .
function will be denoted by f(y ;x;{J), f(y;{J), or f(y). We shall assume For another examp le, Jet ns consider a large ind ustria l fir m t hat
1' has many thousands of employecs . L et us assu me that a n ind ivid ua l
thro ug ho ut the discussio n that E(y) = I {J;x;, where the {J. - are un- is conte mplating accepting employment wit h this co mpany and t hat
i- t '
known para meters, and t h at t he varia nce of y equals a2, where a2 he wants to get sorne idea of its salary system . Suppose he knows
d ocs not depend o n t he {J; or on t he x.. The vector P '. = (y . x x that t he following prediction equatio n is satisfied:
•11 ' J " il ,. ;2,
: · · ,x1,, ) w 1 r epresent a n observatio n fro m this distribution. That y = rt.1 -!- CX2X +e (G.2)
1s t o say, when a n observation y is taken from t his distrib ution t he
corrcspo_nd ing X; v~lues must be specified. Throug hou t t his cha~ter, where y is thc a nnua l sala ry, x is t he number of years cmployed wit h
th e X; w il l be cons1dered known constants, no t random variables. the cornpany, a nd e is a random varia.b le that is, say, normally distri b-
. If in f(y;x ;f3) ."'e make t he transformation e = y - 'L{J;X;, thcn e uted wit h mean O a nd varia nce a 2 • 'l 'hat is to say, t he average salary
1s ~ random variable such t hat E(e) = O a nd E(e2) = a2. This can be is linearly related to the number of years employed. Suppose he
written wanLs to p redict what his sala ry will be 10 years hence. From (6.2)
.,, he knows t hat t he average salary of a li em ployees who have becn wit h
y= I fJ;x; +e
i=l
(G. l) the company 10 years is a 1 + 10:x 2 • HO\rnver, he a lso knows that t he
1 salaries of ali p eoplc w ho ha ve becn employed exactly 1O ycars with
\Ve s ha ll n ot nccessarily a dhere Lo t he notation t hat lower-case letters
the company varíes around t he mean cx 1 + 10o: 2 wit h a normal fre-
rep1·csc n t random varia bles, etc., which we introduced in Ch a p. 5.
106
r1 quency function whose variance is a'. T hus, t he<e may be people
108 LINEAR S'l'A'.l'IS1'1CAL J\CODELS
J\fODEL 1: OEN JmAL LJ;NJ;:AR llYPOTHESIS OF l'ULL RA;-,rK 100
wh o _h a\·e. b een e m p loyed 1 0 . years wi t h t hís company and whose
salar~es m 1g ht ~e extremeJy h 1g h or cxtrcmcJy low, bu t most of th nd ti¡e11 ra.,ndom ly . se lect a n obser vationr Y1 fro m t he d ístributio n
sala n es (approx1mately 05 per cent) a re with in 2a uni ts of a + l O 0
ª .x 1
= x 11 , x. = x 1 2 , • . • , x,, = xh>). Then we select anothe r set
.
I f a is . JI o:, f(!J . ,· ay x -x x - andan observatio n 11 • at ra.ndo rn fro rn
0f X s S
i , • 2 l • • 22 • • · • ' . '"" .~ -
q ui te sma , a n e mp loyee can p redfot wi t h a fairly hígh prec · · ·
. . . · 1s1on · d 1'.stri bu t ion f(11; x 1 = x 2 1 , x. = x 22 , . . . , x,, = x 2 "). \Ve rcpeat
wh at h 1s sala ry will be 111 10 years. In genera l a; will not be k nown· thc -~ - • ] • J
. . cess until n values of y are d ra wn . No st1p u at1on las yet been
so !le rn ust takc so me obscr vations of sala ries wi t hin t hc campan,' Lh1s pto .1 · · • f t t i1cy m1g · l
esti m ~te rx; , a nd t hen s u bstitute t his estimated value of a; into E), put on t) ie ·X·· ,,. They . neccl n ot a.U be u1sL1nct; . · J11_ ac,
· b 1t
be t he samc. H owever , sorne restn ct1on w1ll e p u t on t hc
(6.2) rn o~·der to pr~dict. If o-_2 is so la rge t hat t he prediction is i:dt 11
ªmatllX ••• of the :i; . in s¡)ccia l instan ces. Therefore, whenever the model
very precise, one m 1g ht Lry a d1ffercnt p rediction equation, say, rJ • ed b 1
y = X {3 + e appears , il will be ass1l11ied thc!l i f was consfruct y l1~e
. g process de'ined above. On t h e has1s of t h e observed matn x
(6.3¡ sampl in :¡• · d ~ •
X a nd obser ved ra.ndo m vari able Y , est1mators fo r t he /3; a n or a-
-~ere x 1 is t h e 1:um ber _of ycars employed, x 2 is t he number of ycars "·ill be deri ,·ed. . . . .
of: fm•mal ed uca~10n, and t he rem a ini ng sym bols are as defined in (l>.2) · 1'he prcceding will be for m ulated m t h e fo llowrng de fi111 t10 n .
w1t h t he exccpt1on Llrnt Lhe r:tndo m n orrn'al varia ble f h as va,ria nce cr.
+ De finition 6.1 T h e modcl Y = X{3 + e [w hc re t he q u an t,i ~i es
T hus, if a¡ < a , onc can do a " better" jo b ofpredícting wit h Eg. ((i.~j
2
t ha n wit h Eq. (G.2). o-iven in (G.G) a r e such t hat Y is a random o bscn ·ed Yector, e _ i ~ a
Now Jet us consider thc general problem, whcre thc m od el is a linear
~andom vecto r, X is an n x p matri x of lrno w11 fixed quant 1t res
fu nction of p variab les as given i n (G. l ) . . a nd {3 is a p ·x 1 Yector of u nknown par::w1etcrs] will be ~alled
rnode l l , t he genc ra l-linear -h ypothesis moclel of foil ra nk, 1f t he
S ínce, in ge neral, the /J; in Eq. (6. 1) are not kn own , we s ha ll havc to
estímate t hem in ordcr to utilize the prcdictíon e quatíon. To estímate r a nk of X is cq ua l top where p ~ n.
t he /3;, a random sam ple of size n wíJI be takcn fro m t he distribu Lion Two cases concerning t he d istribution of t he Yector e \\·ill bo
f (y;x ;{J) . T he sample will be denoted by P ;, P~, . . . , P;., an d thc cxam ined:
relations wí t hi n t hc systcm of obscr vations can be wri tten . Case A: e is d ístri bn ted .N(O,a2 1), w hcrn a 2 is unk nown.
Case B: e is a rando m vector s uch t ha,L E(e) = O a nd cov(e) =
Y; = L" /3;X1; + e; j = J, 2, . . . , n (6.4)
E(ee') = a 21, whe rc a 2 is unknown . . .
i=l Thc defi ni t ion of case A is eqnivalent to saymg t hat ea.ch e, is
or, in Yector fo r m, norma lly distri bu ted wi t h mean O and vari an ~e a 2 a nd t ha._t t he e,
Y = X{3 +e (G.5)
n,re joint ly independcnt. That of case B is equ1 valcnt to saymg t hat ·
the cxpected valuc of each e; is zero, the e; a re un~orrelated, a nd t hc
e. have a comm on un k now n varia n ce a 2 • Case A w1ll be referred to as
'
the nonnal-theory case. . . .
where Y= In usíng Eq . (G. 1) as u. m odel we s ha ll be i~1 tcrcstccl 111 ~scert~1mng
its ma ny p ro pe r t ics. Below is a li st of t ite t h111gs w~ sh all rnves_t1gatc:
J. T he point Cl;t im ation of a 2 , t he {3¡, a 11 d t he l111ear fun ct10ns of
the /3 1
2 . T he point es t ima.tion of E(y) . . .
3. The inter va.I esti mation of a 2 , t hc f3 ¡ , a ncl t hc )mear funct1ons of
(G.G) t he{3¡ . .
4. T hc intcrval cs t imation of E(y) andol a fu t ure obser vat1on y
·5. T be intcr val estim a.Lion of x fa r a n obser ved y when p = 2, t h at
is, in simple li near models
Examining Y = X{3 + e in more detail, we see t hat \ve must first i
Se)ect (eit he r at random 01' by d esign) a set of x's say x X ~·
' ' º 11' " 1 2 , · · · 1 ·'-'l pJ
r G. T h e test of t he hypothesis t h at f3 1 = /3f, f3:? = /Ji, · · · , (31> = f3;,
where t he f3't are g iven constants
JlO
LINEAR S'.l'ATI S'.l'ICAL lllODELS JllODEL 1: GENERAL LINEAR HYPOTHESIS OF F U LL RAN K 111
. 7·, The t~st of t he hypothesis that the linear fun cti ' . '1'1.Lking t he deri"rntive of logf( e ; (3 ,a2 } with rcspect to t he yector (3,
?o, \\ hc re ris a kno wn vector and r . k on r (3 IS equaJ to
8 TI < o IS a nown sea lar Lhc above can be written
. ie test of t he hy p othesis t hat a . •
a set of known constants w he re tl1f~en s_u~set of the /3i is equaI to
. · ' 1e ma111111
t~1 at is, ; test of the hypothesis t hat = {3* /3g ~· :;re unspec1fied;
"herc f3 • , {3*
. . .
º
- [logf(e ; f3,a 2)J = O
0(3
J_ Llog f (e; f3 ,a2 )] =
ªª2
O
2' . . . '
/3*k are 1{nown constants a nd 1'
le 2 <- /32 , . . . , f3k -- I';
11•
,
~~ @ ~
(21T1Ti )11f 2 -a
º
of31 Llogf(e ; (3,a2)J = o E( ~) = E(s- 1x ' Y) = s- 1 x 'E(Y) = s- 1 x 'E(Xf3 + e) = s- 1x'Xf3 = f3
º [.,
~/3 ogf(e ; (3 ,a2)] = O
JJ
E(ii2 ) = ~ E[Y'(I -
n
xs-1X ')'(I - xs-1X ')Y I
º
oa2 [logf(e; (3,a2)] =o 1 In t his a1·ticlc we s hall use Lhe noLn.Lion Lhat the symbol ,.._, refe1-s to a m axi-
function g(a 2 ,{J 1 ,{J 2 , . • • ,{J,,). S ince '12 , {J 1 , {J 2 , • . • , {Jf) for m a set of f3 = s-1X 'Y, a2 = ---....:'--- --'--
n - p
114 LINJ•:AR S'.l' A'l'IS '.l'ICAL MODEL S i\ JODl~L
,
l: GF.N F.RAL LINEAR HYJ>O'VH B SJS OF J<'ULL JtANK 115
The value of f3 t hat m inimizes e'e is given by t he solu t ion to (2) E(ee') = a2 1
,.
a
- (e'c) = O .' Lhc bcst (m inimum-va riance) linca.r (linear fun cLions of_ the_ Y~
0(3 unbiascd cstim ate o f f3 is g ivcn by lcast squarcs; Lhat is, f3 -
a ~ S - JX ' Y is t hc bcst linear u nbiascd estima.Le • ·
of (3 . .
el let A '1' - AY·
We get o (3 (e'e) = 2X ' Y - 2X 'Xf3 = O. T he least-squa rcs estima.te P roof: Let A be a ny p X n co nstant irn~~nx a n t-' - - ,
(3 * is a, acneral linear fun ction of Y , wh1ch "·e shall tal..e as ~~
of f3 is, t hcrefore, - . . 'tcºor A \Ve must specify t he elcmcnts of A so th a t (3 * ."'J
--- ~ = s- 1x 'Y . es L11n a t-' · . .
bethe besL u11b1ased est 1maLcof (3. L ct A -
_ 5 - 1X ' + B. Srnce
·r A
w hich, of eourse, is t he same as t he maximum -likelihood estímate 1 5 - 1X ' is kno wn , we m ust find B in order to be ab le to spem Y ·
u nder nor mal theory . l\Iinimizing t he sum of squares e'e <loes not F or unbiasedne:s, we h avc
p rovide an estí mate of a 2 . However, t he unbi ased estím ate of a 2
b ased o n t h e lcast-squ ares estim a.te of f3 is g iven by E((3 ':') = B(AY ) = E!(S- 1X ' + B )Y l = (S- 'X' + B)Xf3 = f3 + BX(3
á2 = (Y - X~ )'(Y - X~) = _Y_'(_I_-_ 1
x _s_- _x_n)_Y But. Lo he unbiasecl , E( f3 *) m ust ecp1a l (3 , a,n~ t his implies that
?J, - p n-p BX~ = o for ali (3 . T hns, unbiascdness s pec1fles Lhat_ BX - O.
. ti roperLy of " bcst" "·e rn ust fi nd t hc mat n x B t h_at
F or le P b · t t t he restr1 c-
N ext we sh a ll investigate t he p ropert ies of t he least-sg uares esti- m i ni mizes vR,r({Jf) , where i = 1, 2, .. : , p , su JC C .º
mators. Sinee t he frequency form of t he ra nclom vector e is unsp ecified, ti on B X = o. To exa mine t.h is, cons1dcr t he coYana nce
in gene ral it will not be possible to exami ne t he "good ness" of t he
estimato r ~ rclati,-e to a ll fun ctiom;. Instead, we sh all ha,·e to lim it cov(f3 *) = EL( f3 * - f3 )( f3 ~' - (j)'J
1
o urselves to a s ubset of fu nctions: for example (since ~ is a li near
fu n ctio n of t he y;), t o t h e set o.f ali linear fu nctions of t he Y; · W e havc r = E{[(S- 1X ' + B )Y - f3 l[(S- X'
1
+ B)Y - (3]'}
,,u'" ·" ·
11 6
LINEAR S'l'A'l'IS'l 'ICA.L l\IODELS '' l\IODEL l: GENERAL LINEAR H YJ:'O'l'JIESIS OF FULL IlAN K 117
S ul>stituting X~ + e for y a n<l using BX = O, we get This is not true for funetions i n ~1cra1. For example, if 'rn wish to
cov( ~ *) = E(S- 1X'ee'xs- 1 + Bee' B ' + s - ix'ee'B' + Bee'Xs- 1
estimateii = (/3 1 + /3 2 )/2{33 , t he value ofthe best linear u nbiased esti mate
= a 2 (S - 1 + BB' ) . ) of 1¿ is not gi ven by
3+ G 9
= - -
L.et BB ' = G = (g;;)· T hen cov( ~ *) a2( S - 1 + G) (2)( - 4) 8
diagonal ele mcn ts of co"\j ~ *) a re t he . t· . · The
{Ji* · 1,o m1111111
. . . respec ive vi.Ll'iances of t i
1zc each var({Jt}, we must t h . ~ . . . . ie 6.2.3 Point Estimation of E(y). To estímate t hc m ean of y for
diago nal ele me nt of cov(~ *). Since ª2 ~nde~-~1e, m1111m1ze each a given set of x 1 , x 2 , ••. , x 11 , we can u se t hc fact t hat E(y) is a linear func-
must find a m atri x G s uch that e 1 a· a re constants, we tion of thc {J;, that is, E(y) =
P
2: {J;X;- Hence, invoking the t hcorem
minimum. B u t G - BB '. . ~c l ia~onal element of Gis a i= l
. - is pos1tive sem1definite·henceg >-
Tl rns t 1rn d1agona.l elemcnts of cov( ~ *) ,·11 tt· . ' . . ;~ ,_... O. concerning estimators of linear functio ns of t he {3;, we h ave
when g;¡ = O for 1: = l 2 "i. a am t heir mm1murn
, ' ... , p . But, If B = (b ..) t he n g .. - ./"-. :V
. F01· exam
· l 1·r t h e 1east-squa res estimators of {J fJ
Pe, d (3 . of X ' Y is 2: X;k'!Jk · T hcse quant it;ics cn.n 1·cadily be fo und by us ing a
tively are 3 6 el v 2, an 3 , rcspec-
esti m~tor of 'say' ~{J1 - iR th+cn5/Jth~ val ue of the best linear ~mbiasccl desk calculator."
, < ' l - 1' 2 3 l S (3)(3) - (2)(6) + (5) ( - 4) = -23. 6.2.5 Choosing the X matrix . The X;; values rnust be kn own
IIS LIN EAR ST A'.l.'I S'l'TCAL l\IODELS i\IODEL l: GENERAL LIKEAlt llYPO'l'JlESI S OF FlJLL RANK 119
b efore the Y; valnes are sclected at ra nclom. In some cases X;J Yalues 'J' m in irni ~c var (/j 2 }, we choose ou r
s uch t hat L(x; - x) 2 is as la rge as
X;
rnay h e p ickccl or chosen by t he expcrimc nter in a ny way he wishcs· ºssible. To minimizo var (fj 1 ), we choose t he X; such t hat L.xlf
in other cases t hey ca.n not be so con trnlled. lf t hc experim enter ca1~ po 9 •
)'(°¡; . _ x}- 1s as sma
,~( .
11 as poss1. ble. • ,.incc ,¿,,, ,~ 2
x, - x-)2 ~ kJX¡, t hc
ch oose t hc x i; valu cs, t he question of h ow to select t he m arises. In .., · ·('p· ) is mi n imum if t he X; a.re chosen s uch t hat x = O. Th is ~Liso
va1 1 • • .
general, it wo nld secm t hat the best way to pick th em is so t hat thc niakcs t,hc cov(f3 1 ,f3 2 } = O. Note that we ha ve assumed t hat n is
vari ance of certain estim ators will be as s mall as possible . fi xc<l. .
F or exarnple, the varia nce of the estimator of A.' 13 is a 2 A.' S- 1 A., a nd wc To estímate a 2 , we not e t hat a ny of t hc fo llowrng for mulas can be
rnigh t wa nt to ehoose t he X;; t hat minimizo A.' S- 1 A.. \ Ve cannot do t his 11sed:
in general for a ll vectors A., but we mig ht be a ble to for so me. \~Te shall 1
discuss t his fur t her when we discuss sp eci fie examples. / (¡2 = ~ (Y'Y - Y' xs- 1 X'Y) = - - (Y' Y - ~'X'Y)
r:.. •. 6.2.6 Exa mples . A simple linear mod el is n - 2 n - 2
i = 1, 2, ... , n r: l . .
1 •
= - - (Y - X(3 ) 1(Y - Xl3 )
w here {3 1 and {3 2 are un k nown scalar constants a nd t he x, are k nown n - 2
scala r cons tants. T herefore, referring to Eq. (G.6}, we fi nd t hat
(p = 2) 1
= -- (
n - 2
Y'Y
-
A' SA)
t' t' = n--J.-2 {'~"-' (Y;. - -)2
Y -
LL (x; -
~
x)(y; -
-2
L(x; - :i:)
ü)J2}
1
l
X1)
X=
X
z As ~n oth er example, Jet us s upposc t hat we wa nt to predict the
distance s t ha.t a pa rt icle will t ravel in t ime l if t he velocity is constant
(
and cq ual to v a nd if the ini t ial clista.n ce fro m a certa in refcrence poin t is
l x,. ([ . T he relationship is
I t is easy to see t hat 0
s = d0 + vl
n
s-1= l ( LX~ - L X;)
S = X 'X = - - - - -2 Ho\\·ever, Jet ns assume t hat we can not mcasure s accurately but t hat
( L X; nL(x; - .¡;¡ - LX; n
t hcrc is a. random error atta.ched ; i.e., we measure n ot s but d, where
el = s + ~' whcre e is a rando1n errnr . T hcn ou r rclationship is
and
d = d0 + vt + e
Thus,
. (Pi) = s- x 'Y = - -1- - (I,x7I,Y; -
13 = 1 Lx;Lx;Y;)
Since d 0 an d vare not lrnown , we sha ll takc a series of ob servations of d
µ2 n L (x; - :f:)
2
- Lx;LY; + n'I,y;x; nnd t and estim at e o ur nnknown constants d 0 a nd v. If d ; reprcsents
J _ L(x; - x)(y¡ - ?i) • • the ith distancc measurcd at time l 1 a ncl e; is t he correspond ing error,
0 1' f'9 -
- L(x; - x)2
f31 = 'fi - f32x we ha.ve
i = l , 2, ... , n
(3 2 is t he slope of t he line, or, in other worcls, it is t he cha nge in E(y)
per unit cha nge in x, a nd /3 1 is the value of E (y) when x = O. ·
· mo del (p = '>)
a genera l -J•mear -h ypo t h es1s ~ . TI1e obser·,,at1·011s taken
Since t he cov( ~ ) = S - 1 a 2 , we see t hat
are gi,·en below.
· =- a2Lx -
COY P'
- ( 11P2) J'.( - x-:)2
n....,x;
1
19 20 45 55 78
Dis tnnce el 9 15
- <.12 10 12 18
and var ({3 ) = - - - -2 Ti me t 1 2 3 4
2 "f,( x; - x)
120 L LNEAH. S1'A'.l'IS'.l'TCAL MO D }:r,s 1110DEL 1 : Gl.:NlmA 1~ L INEAR H YP01'HESIS OF lWLL RANK 121
T he form ul as abo ni g i ve is distribu ted as t(n - v) . T h us,
.354G9
s s--1- - .020G5G )
B··~·
_ f3· ~ t,,12)
= ( 7
50
50)
59 (
- .029G5G .0041518 f'ª"
-la.to
t(u) du = P ( - tª
12 ~
, f <JC¡;
= 1- a.
á 2(n - p) - á2(n - p) -- •2
a (n - p) _ _ _
( 1 l) to t he p oint (xvx 2 , ••• ,xv)· Since t he estímate of E(y) is a random
varia ble, subj cct to ra,ndom Auctuations, we might d esire a conficle nce-
et.o Cl.1 et.o et.,¡
intcr val estímate of E(y). To o btain this, let
. To. set. a 1 - o: con fidence inte rval on fJ ;, wc use t hc fact t hat p· ·
d 1stn buteu N({J i• e¡;a 2) ' \\·JlCte
. C¡; .IS t he iJth
.. clc mc nt of s - 1 = ;C.
is 7J
It is a lso t ru e that (n - p )á 2
/u =
2
w is distribu ted as z2 (n - p) inde- . a random sam p le of k v alues of y is drawn from
J,et us S uppose t hat
p e nd cntly of z. Therefore,
thc f req uency func twn /
z f (y ; Xo1•Xo2' · · · ,xop)
U= - -===
áJ x·s-1x
whme . each of the . k values of y is selected from t he same frequency
is dis tributed as t(n - p), anda 1 - IX confidcnce interval is obtaincd . ¡ e
funct 1011 , . . , w 1th the samc X ; values. d
from I et these k values of y be denoted by y 01 , Y 0 2 , ••• , Yo"' a n let Lhc
( to:12 _, of these sample values b e de not ed by
1- IX = )_ t(u ) di¿ = P(-ta¡2 ~ i¿ ~ ta¡2)
!llean
-let.t~ 1 k
Yo= - L Yo;
Substitutin g for n, w e get k ;=¡
S upp Ose a confiden cc inter val is desi red on the mean y0 of t hese k
· 1 1
values of y. T hat is to say, suppose :ve const1:uc~ a 1: i ~tcrval s uc l t lat
. babili ty that the m ean y 0 w11l fall w1thm it is e qual t o sorne
the p 10
reassigned value 1 - IX . · The inte~v:al w1ll e called
. b
pre ictio.n _ª a· .
After so rn e ma11ip11lations, we a rrive at
~interva l 0'J.r size l - CJ.· ' since the
· probab1 h ty t hat the mean• Yo of a fu t m e
sample of y values_se_l e~te~ at ran dom from t he dens1ty f(y; x 01 ,x 02 ,
P(Li;J;x; - tª 12 ) á
2
x·s-1 x ~ "L{3;X; ~ "L¡];x; + tª12 ) á 2x'S- 1 x) = 1- IX
... ,x01,) will fall w1th m 1t is egual to 1 - IX. •
(6. 11) Since f3 and cr2 are not known, the p rocedure is to select n sample
a n d the quantity in p a re ntheses defines the desired confidence inter- va1ues P ·' -_ (y;,·x 11'. • • • ,.x 111) ' w here J. = 1, 2, . .. , n, use t hese values
val o n E(y) . Th is is a specia l case of the sit u ation described in the t~ esti m~te f3 a nd cr2 , a nd construct the predict ion ~nter~al usi n~ the
estima t es ~ and a2 . T o do t his we observe that, s111ce Yo and ¡3 are
prcvious section. T he width of the inter val is 2to:12 Vá2 x 'S- 1 x.
independent, the quantity
6.3.3 Interval Estímate of the Mean of F uture Observations.
L c t us s upposc that thc eguation w = {3 0 + fJ 1t describes t he weight of z = Yo - ~'xo
chic kens th at h a ve bce n fed a new ration fo r a givcn t ime t. 'l'he weight
w is in g ra m s, and t is in \rneks. S uppose t hat t can be mea s ured wit h no
error but t ha t thei.:e is eith er a n equation error and/or a n error in
is distributed N[O, cr
2
G+ x óS - 1 x 0) J, where X0 is a JJ x 1 vecto r
m easuring the weig ht w . \iVe obser ve y, whcre y = w +e. So thc whose i t h element is Xo; · It follows t h at
mode l w ill be writtcn
Y = Po + f31t +e
where fJ 0 is the initial weig ht of a chicken and {J 1 is t he growth rate. If
we wan t to estím a t e the {J; by observing the weight of a few chickens, is distributed N(O,l ) . Also, v = (n - p )cr~2/ cr-? is disLri buted as
we can u se t he theore ms developed in t h e previous articles. Suppose x2(n - p) a nd is independent of w . Hence,
t h at a farmer p la ns to s hip le chickens to marke t after feeding them for
12 weeks a11d t hat he wan ts to lmow what the average weight of t hese k U=
wJn=-p
chick cns w ill be. The average weight of all chicke ns tha t have b een fcd Jv
for 12 weeks is {3 0 + J 2(3 1 . However, the far mer may not be interested is distributed as t(n - p) . Therefore, 1 - IX = P( - laft ~ u :::;; ta¡2) .
in estimating this quantity; h e may want instead a confidence interval Substitut ing for u, we obtain
on the m ean of k \' a lues of y selected from the p opulation with !Uean
~
{3 0 + J 2{1 1 . Thi s typ e of problem is the s ubject of t his article. We
s h a ll de ri ve formu las more genera l t han those illustrated in the above
example .
l - (/. = p
[
-
t
a/2 ~ J G+ ~'xo
¡¡2
Yo - ~
t '"]
xóS- lxo) "' •. -
LINEAJt S'l 'A'l'I S'l 'ICAL J\TO OJ.:T.S i\!ODEL I: GENERAL LINEAR HYJ'O'l'HE S IS OF FULL l tANK 125
After some m an ip ulations " ·e ani,·e at
63
.s Inte rva l Estímate on x for a n Observed y in a S imple
P[{3'xo - ta12)á2G + XóS- lxo) ~ Yo . ·ear Model. Let us conside r t hc relations hip betwce n growth of
Ltn. kens anda g1ven · ·
tune · 1Ta1 t 1rnt we e1·1s cusse d a t t· ilC b cgmnmg
mte . .
o º
111
of Art. 6.3.3. 1 f . . t d .
uppose t hat ~ rn a rm er is now rn~eres e 111a so mew1a 1 t
P( l.1 ~ {J1 ~ 11.9) = .95 Taking thc dcrivative with r espect to a 2 , f3v {3 2 , :i· 0 g ivcs, aJte r sorne
T l!c width of the co nfidcn ce in terval is 4.9 _ l. l = 3.8. simplificaLion (a 2 is conected for bias),
To seta95per ccntconfid enceinterval on{J 1 _ 2(3 we use Eq ( 6 l O)
2
where r ' = (1 , -2) . " cgct ' · · ' - Yo - P1
Xo = P2
r 'f3-= ( 3.0) = 4.0
(l , - 2) - .5 fi _ ..,(y ; - g)(x; - x)
2- 2:(:1:¡ _ x)2 (6. 13)
We substitu te into Eq. (6. IO) a nd get
a.,2 = [l +
a·, -
k
~ Lx r. .
nL(X¡ - i) 2
2
+ ·x 2O
L(x,. - i) 2
_ -'>XoX-
l:(x, _ x)2
J (6. 14) Xo = X + /J2(i¡ -
J.
Yo) ± ta¡2Ó"
J.
J1.(~ + ~) +
n l.: L(x; - i)
2
(Yo - 9) 2
~i ta12Ó"JA(~ + ~)
2
- f32W - Yo) + + Wo - '[/) if A. > 0 (G. 17)
~ }, A n k L(x, - i)2
is dis tributed as x2 (n - 2) a ncl is inclcpendent of it. Also, the quantity
If }. < O, it can be seen by exarninin g (6.16) t h at confidence lirnits on
x rnay not exist. Even if co nfidc n ce limits do exist when J. < O, t hcy
V2 = 11+"'"'k (
Y; - Yo -
- )" 0
will in general b e too wide to be of a.ny p ractica! use.
2 L.,; ?
G i =n+l v· 6.3.6 Width of Confidence Intervals. When an experimenter
decides to u se a con fidence in terval, h e s h ould decide wh at widt h is
is distribu ted a,s 7., 2 (k - 1) and is ind ependent of u and v 1 . Hence,
necessary in order to make a decision . Y.,Te must, t herefore, investigate
the width to see wh a t is in volved. W'e shall not d iscuss this h ere, but
shall de vote a ch apter in Vol. II to this i m portant s ubject.
6.3.7 Simultaneous Confidence Intervals. The frequency
is distributed as x2 (n + le - 3) ancl is indepcnden t of u Thcrefore, interpretation of the foregoing resu lts on confide nce inter vals is that,
the quantity . if man y samples are taken a ndan inter val su ch as that give n b y (6.9) is
constructed , then, on the average, 100( 1 - a) per cent of these intervals
u.Jn + k - 3 will cover t he t rue (unkn own ) value {3¡. This is t rue for one par t icular
ªu·./vja 2 value of i only; i.e ., a sarnple can be selected anda confiden ce interval
set on {3 1 , and the above frequency interpretation will h old. B u t if t he
is distributcd as t(n + k - 3).
same data are used to set confiden ce intervals on {3 1 and {3 2 , the peoba-
If we dcsire a 1 - C1. confidence interval abo ut x 0 , we can write
bility is not equal to 1 - C1. t hat t h e result ing confidence intervals will
include both {3 1 and {3 2 • For each set of ob.servations only one confidence
statem.enl can be made using the above form:ulas. T h e same is true for
Eqs. (6. 8), (6.1 0), (6.11), a nd (6.12) .
This procedure does n ot seem to be realistic-setting a confidence
128 LINEAR STA'l'ISTlCAL J\lODELS )JOPEL ) : C:E1\E.RAL LINEAR HYPOTJIESTS 01'' Jo' ULL RANK 129
intcrval on only one {l¡. It seems t hat a n experimente r will want to set . ilar set of values of t he paramcters a 2 , {J 1 , (J 2 , • • • , {J'P) in Q t h at
nn.r t JCt • . ) .
a confidence in terva l on cach (Ji and to .know the p robability that al! r .· 1 ·zest he likclihood fun ction , and let /_,(L) be t he max1mum value.
111nx1111 • . • •
these interYals contain t hcir respective {l¡. If the inteiTals were J,(cíi) has a simila1·.defi111t1on 111 the para.meter space r cstncted by 1:fo·
independent, t his p roba bility would be the product of t he individua! 'fhat is to say, w 1s t he space of a 2 , {3 1 , {J 2 , ••• , (J,, defined by the 111-
probabilities; i.e., in simp le linear models we should calculate two ~quafüy o < a 2 < oo a.ncl by /3 1 = f3i, {3 2 = {3i, ... , {J,, = p;. Actually,
intervals similar to (G.ü), one corrcsponding to {3 1 and one for {3 2 , and · one-dim ensiona. l spacc. To fincl L .(w) and L(ñ) we shall work
(o) IS a . .
t he probabili ty t hat both intervals wcre correct, would be (1 - 0:)2. with the logarithm of t he hkcl1hood funct1on.
Unfortunately , the intervals are not independent, and (1 - cx) 2 is not To find L(w} we procccd as follows:
the correct probabi lity. The problem is more complicated than this ~ n ., (Y - Xf3*)'(Y - X(3*)
and will be dealt with in sorne detail in Vol. II. ¡ f(e· a2 (3*) = - - log 27T - - log a- - .;._---'------
og ' ' 2 2 2a2
\
Since the {l[ are fi xed, this eqnation is a function of a 2 only, and the
6.4 T ests of Hypotheses
value of a2 t hat maximizes it is the solntion to
6.4.1 Testing the Hypothesis (3 = (3*. To test the hypothcsis
d (l f ( . 2 A*)) = (Y - X(3 *)'(Y - X(3 *) - _!:..._=O
(3 = (3 * ( (3* is a known vector) in the general-linear-hypothesis moclei - og e,a , l'"' •4 -2
of foil rank, we s ha ll assume that the vector e is distributed N( O,a2I) · dcr 2
'a 2a
un less it is specifically stated otherwise. Thc solut ion is
Testing (3 = (3* ( (3 * known) is equivalent to testing simultaneously á2 = (Y - .X(3*)'(Y - X(3 *)
that each cocffi cicnt {3 1 equals a given co11stant fJi. It is quite important n
fo r an experimentcr to have · such a test available. In t he model
Y = X (3 + e, for examplc, if (3 = O, a knowledge of t he factors that n"'2e- 11f2
and L( ') - - - - - - - - -- ---:::
correspond to t.he X; does not aid in the prediction of E(y). However, w - (27T)"' 2 ((Y - X (3 *)'(Y - X (3 *)]" 12
if (3 -:p. O, t he x 1 factors wi ll be val uable in predicting E(y).
To test the hy pothcsis H 0 : (3 = (3 * wc must devise a test fun ction, ProceeclinCY
0
in simila r fashion to obtain L(~)), we see t hat t he point
say, u = g(y 11y 2 , •• • ,y,.,x 11 ,x12 , ••• ,x,.v), t hat is a function of the where Jogf(e ; a 2 , (3) attains its maximum is t he solu Lion to t he equations
observations Y; a nd x 1; such t hat t he distribution of u is known when o X'Y X'X(3
(3 = (3 * . In orcler to cvaluate the powcr of the test, the distribution - [lo<Tj(e· a 2 (3)J = - 2 - -2- = O
0(3 t:> • ' ' á - á
must also be known whcn the a ltcrnative h ypothcsis H 1 : (3 -:/=- (3 * is
true. () ~ (Y - X{3)'(Y - X(3) n
- (logf(e· a- (3)) = - - = O
\Ne shall use the likelihoocl ratio Las t he test function. The likeli- ºª2 ' ' 2á4 2á2
hood equation is
n"'2e-,,¡2
1
f (e; (3,a2) e- e'e/20• Wc get
= j(e1 ,e9, ... ,e,.; (3,a-) =
_-2
L(Ó.) = (277 )"' 2 [(Y - X{3)'(Y - X {3 )]"12
- (27T) n/2an
i\! OD'EL ] : Cl~l' 'ERAL LINE Alt HYl'O'l'HBSI S OF FULL RANK 133
132 L1NEA R s·r A'l'lS'r ICAL i\lODELS
thc rninimum \\'it h respcct to í3 whcn t hcre is no rcstriction on f3 ,
The power of t he LcsL is gi \'en by
and
fJ(J,) = J:,,/ 2
(v; p , n - p, },) dv
.?. = . .:.í3(_ - -=-í3_*"'-)'_
2a2
S (;..:._
f3_- ~í3_*)
whe re ). =- (f3 - f3 *)'S(f3 - f3 *) Onc of thc most imporLant quantit ies in the AOV tab le is the
2u2 ror mean square Q 1/n - p. (It should be notcd that t hc
and m ay be eYaluated for differcnt Yalucs of p, n - p , and ). by usin ~:iantity á2 that appears in t he confidence-inte r val equa,tions of Sec.
Tang's tables. g (i.3 is equal to Q1/n - p.) T o calcul at e Q1 we can Jet f3 * = O; thcn
t,hc total sum of s quares Q is Y~Y. l3ecause _o [ its impor tancc we
TAUL"E Ü. J ANA l.Y:-iTS OF VAlUANCE FOB. '!'ES'J'l NG {3 = {3 * rcpcat the fac~ that Q 1 = Y 'Y - . f3'X' Y , where í3 ~s t he sol11 t i_o n of t.he
normal equat10ns . We s ha ll d1scuss mcthods fo r computing t hese
sv Dl<' s~ MS quanti t ics in a later chaptcr.
6.4.2 A Test of the Hypothesis That a Give n Linear Function
Total n Q = (Y - X{3 *)'(Y - X {3 *)
r '{3 (r Is a Known Vector) Is Equal to r0 . In d e ri ,·ing Eq. (G.10) ,
Q2
Duo to !3 JJ Q2 = (Y - X{3 *)'XS- 1 X '(Y - X{3 *) Art. 6.:l. I, we no ticcd Lhat n was distribu ted as Stude nt 's l with n - JJ
p
<logrees of freedom , whe re
E LTO I' n - 1J Q¡ = Y '( I - xs- lX ' )Y ~
n - JJ (6.26)
The evaluatio n of 'IL is gencra lly put in the for m shown in T a Lle G. l · To test H 0 : r ' f3 = r 0 , we simply replace r ' f3 b y t he kno wn s~alar r 0
t his is a n a_nal ysis-of-variance (AO \ ) table, in which thc d egrees of in (6.26) a nd co mpare t hi s value of ii wit h t he tabulated value of t he
free<lom (DF) , s um of squares (SS), and mean square (l\'l ) corre- Student 's t d istributio n with n - p d egrees of freedom . In deri ving
sponcli ng to t he , ·a rio us so urces of variat.io n (SV) a re lisLcd. Thc pro- Eg. (6. 1O) we notcd t haL
ced u re fo r calc ulating n is to obtain Q a nd Q 2 by dircet comp 11 tation
r' ~ - r 0
a nd to get Q 1 by using t he id en t ity Q 1 = Q - Q2 • V = -;::='====
By obser v ing t he likeli hood equation it is casily sho,,·n that L can be .J
a 2 r's- 1r
obtained by evaln nting t he fol lowing [w he rc min (Y - Xf3) '(Y - Xf3) ·
is distributcd .N (µ , l ), wh e re µ = (r ' f3 - r 0)(a2 r ' S - 1 r) - l; t hcreforc,
m eans the m ini m11 m n ilue of (Y - Xf3)' (Y - Xf3) with respect to f3 ]:
v2 is distributecl a s x' 2 ( 1, ?.), w here
Q = (Y - Xf3 *)'(Y - Xí3*)
= min 1(Y - Xf3)' (Y - Xf3) when H 0 : f3 = f3 * is t rue]
Q 1 = (Y - X~ )'( Y - X~ )
= min 1(Y - Xf3)' (Y - Xf3) wi t h no restri ction on f3]
L = (Q 1/Q) "' 2 'l'hus,
'\Ve now state Lhe fo llowing impo rtnnt thcorc m.
+ Theorem 6.5 [n t he general-linear-hy pothesis model g iven in isdist ribu ted aslt''( J , n - p , .?.) and red uces to Sned ecor 's ft' ifn,ncl o nly .
Defini t io n 6. 1 \\'h cre e is distributed .N( O,u 2 I ), t he q ua nt ity ifH 0 : (r ' f3 = r 0 ) is t rne.
6.4.3 Testing a Subhypothesis, i.e., the H y poth es is fli - fn ,
n - JJ Qz
1l = - - - P2 = {3;, .. . , {3.,. = p¡ with the R ema i n ing '(J, U n specifi ed . In
p ()¡ this art icle we s hall examine t hc problc m of Le8Li 11g t hc h y poLhesis
is distrilrnted as F ' (p , n - p, .?.), where Q1 + Q 2 is the m ínimum P1 = {3':', {3 2 = f3i, . .. , /Jr = fJ: (r < p) in thc model Y = X f3 + e in
with resp ect Lo f3 of (Y - Xf3)'(Y - Xf3) wh en H 0 is true, Q 1 is Definition 6. l. This is a vcry useful a nd very impo rtant test .
134- LINEAR S TATlS TICAL lllODELS Jl[ODEL l : GE ·1m.AL LINEAR HYPOTHESIS OF FULL RANK 135
For t he case p = 2 t he equation is . test is equi valent Lo testing t he hypothesis {3 1 = {3t, {3n = {3:, ... ,
t 111s - -
R ={Ji (r < p) in thc mod cl
(6.27) /' r V
Now suppose a n cxperimen ter wa nts to know if he can rcpl ace Eq. Y1 = 2 x;;{3i + e;
i= l
(6. 27) wi t h t he model
'].'he test of the hypothcsis H 0 : y 1 = Yi and its construction are stated
(6.28)
in t he followi ng.
In other words, he wants to test t he hypot hesis {3 2 = O in t he model
• Theor e m 6.6 Let the model Y = Xf3 +e given in Definition 6.1
given in (6.27) wit h no stipulation abo ut t he value of {3 1 • This t est is
be partit io ned so that
clcarly d iffere nt from t he test discussed in the preceding a rt icle; t here
a procedure was devised to test whether {3 1 and {3 2 , for example, were Y = X1 Yi + X2Y2 + e
simu ltaneously equn.l to zero. H ere, if t he conclusion is drawn that
{3 2 =I= Oin mode l (6.27), it follows t hat t he two factors x 1 and x 2 togcthcr where y 1 is of dimcnsion r x I , a nd let e be distribu ted N{O,a2 I ).
are bett er for p redi cti ng y than t he factor x 1 a lone. On t he other h an<l Then to test thc h ypothesis H 0 : y 1 = yf by the likelihood ratio,
if {3 2 = O, x 1 and x 2 together are no better t han x 1 a lo ne. ' the p roccdure is :
A t est of t h.is typc is extremely valua ble in simple linear m odel ~, c.g., (1 ) Obtain the minimum value of e'e wit h respect to thc un-
when x 1 = 1 m model (6.27) . In this case {3 1 is t he in tcr cept ofthe line known pa.rameters in f3 in t he model Y = Xf3 + e , a.nd d enote t his
a nd {3 2 t he s lope. In ma ny cases it is extremcly desirable to be a ble to mín imu m value by Q0 .
t est whether t he slope {3 2 is zero regardless of t he valuc of t he in tcrcept (2) Obta in t he m ínimum value of e'e with respect to t he
{3 1 , or to test w hethcr t he intercept {3 1 is zero regard less of t he s lope. unknown para meters in y 2 in t he model Y = X 1y t + X 2Y2 + e
In the model (this w ill be called thc model reslricted by the hypothesis H 0) .
Y = Xf3 + e Denote this mínimum value by Q0 + Q1 .
(3) L et Q = Q0 + Q 1 + Q2 , where Q = (Y - X 1 y t)'(Y -
partit ion t hc matrix X and the vector f3 so t hat X 1y j"). Then the qua ntity it g ivc n by
n-pQ¡
i¿= - - -
r Qo
and where
Ol'
Since Bis posit ivo dc fini te, ti is d istributed as F(r, n - p) if and
\iVe desire a test of the hy pothesis y 1 = y i (y i k nown) wit h no on ly if H 0 is t rnc, i.c., if a nd only if y 1 = Yi-
stip ulations o n y 2 • Since P1'0oj: The likclihood fun ction is
f( Y ; Y1·Y2,a2 )
- f31
f32 ) = l
(2..a2)"' 2
cxp [- ~(Y
2a-
- X 1y 1 - Xzy2)'(Y - X1Y1 - X2Y2)]
Y1 = ...
and
( (6.29)
whe re ~ = S - LX' Y is t he value of {3 t haL maxim izes {6.29). 1f To find thc dis tribuLio n of L, let u s examine the quantities Q0 ,
wccxaminee'e = (Y - X{3)'(Y - X {3),weseethat t;hc m inimuin Q aud Q~. In (6.32) we havc stated in effect that X~X 2 h as an
valuc of thc e'e with res p ect to {3 (denoted by Q0 ) is in~ersc. This foll ows from the fact that X~X 2 is a principal mina r
of thc posit ive tlefinite matrix X ' X. Hence, X~X 2 is a lso posi t ivo
Q0 = (Y - X~)' (Y - X~)
definite. From the fact that
n"12e- nf2
Therefore, L(Ú) - Q - 1112
(6.30)
X'{ I - xs- X ' ) = o
1
- - (27T) n/2 O 1
f(T; y 2 ,cr2 ) = l~
(27Tcr-)" 12
exp [ - ~
2cr
(T - X 2 y 2 )'(T - X 2y 2 )] (6.3 1) Ji'rom (6.3-!), Q0 ca n be wTitten
Q0 = (Y - Xtyi)' (I - xs-1X')(Y - X 1 yi_')
If wc lct :rz(y t) fLnd á 2 (yt) de note r espectively Lhe valucs of y 2
Using t he n otatio n
and cr2 t hat maximizc Eq. (G.3 1), we get
T = Y - X1Yi
Y 2 Y1*) --
A { (X 2'X 2 )- 1x 2·T -- {v·x
~ 2 )- x '(Y
1
2 - x 1y i*)
A = 1- xs- 1x'
(6.32) A2 = I - X 2 {X2X 2 )- 1
X2
we s hall study the idc ntity
S ubsLitu t ing t hc n1lucs in (6.32) into t he likc lihood fun ctio n (6.3 1),
we have T 'T = T'AT + T'(A2 - A )T + T'(I - A 2 )T = Q0 + Q 1 + Q2
(6.35)
It foll ows t h aL T is distributcd N[X( {3 - {3 *), cr21], where {3 * =
If we examine e'e in t he model restrictcd by y 1 = yi , we see that (y:). since
(6.33)
E(T) = E(Y - X 1 yt) = Xf3 - X 1 yt = Xf3 - x(y:)
138 T., INEAR STA'I'IS1'ICAL l\IODELS
l\fODEL 1: GENERAL LINE AR HYPO'l'HESI S OF F ULL RANK 139
In U:q. (G.35) t he idc ntity ondin 0rr e le ments of the ri ght-ha nd side X ' Y. Now Q1 =
corresp -,X ' Y - 'X'Y J - · ti
I = A + (A 2 - A) + (I - ~)
, , ,
Y'X s - IX' Y - y X 2(X :!X2)- 1 X2y = ~ -: Y:_ 2 ) " : 1ere y 2 I S le
1 1
fon to the reduced normal equat1ons X 2 X 2y 2 = X 2Y.
is obtained. Using (6.34) it is easily verificd that A, (A 2 - A), so ~!h e qua.ntity y~X~Y will be called t hc red'uci'ion d1ie lo y 2 ignoring y 1
a nd (1 - A 2 ) a re each idempotent matrices. H e nce, using
Theo re m 4.1 G, we sce t hat : and will be writtcn R (y 2). . . . r
(x~xo 1 x;x2
o ) (Y•)
y2 =
(X~Y)
x ;Y
We need the quantities ~ ' X'Y a nd Y' Y, which remain t he same for all
fi vc tests. And we need the follow ing for the respective tests:
J. R({32 ,{3 3 , •.• ,{3,,) that is, the reduction duc to {32 , ••• , {3,, ignoring {31
From this we obtain thc equations
2. R(f31,f33, · · · ,(3,,)
(Yi\
1
3. R(íJ1,f32, · · · ,fJ,,_1)
yJ = ( (X' XO )- (X;XO )(X'X~YY)
1 1
4. R(f32,f34,{JG, ... ,{J,,)
)- 1
2 5. R ({J1 ,{3 4 , . • • ,{J,,__1), etc.
l\1ost of these tests require that the appropriate f3.i be set equal to zero
~ = (Y1) = ((X~X1 )- X~Y)
1
and in t hc normal equations and that the remaining equations be solvcd.
y2 (x;x2 ) - 1x;v 'l'his is a t edious and time-cons uming job. However , if {3 1, (3 2 , • • • , {3,,,
Also, R(l3) = Y'X1(X;X1)- 1 x;y + Y'X2(X;X2)- Lx;y are ali orthogonal, then a li Lcst,s similar t,o !,hose listed a bove aro easily
made. If {3 1, ... , {3,, are o rthogo na l, t hen X ' X is a diagonal matrix,
and R(y 2 ) = Y'X2 (X;~)- 1 x;Y and the redu ction duc to a ny set of t he f3; adjusted for any other set is
so - R(y¡ 1 Y2) = R (l3) - R(y2) = Y'X1(x;xl)- 1 X~Y = R (y¡) simply th e reduct ion due to that particular set of the (3; ig noring the
lf we wanted to test t he hypothesis y 2 = O, t hen , by a method oLhcr set.
similar to t hat used to obtain T able G.2, we would need R(y 2 j y 1 ) = For example, let us suppose in t he model Y = X 1/3 1 + X 2 /3 2 +
Xi13 + e t hat thc normal equa.t io ns a.re
R(l3) - R(y 1) = R( l3) - y ;x;Y. rrx;x 2 = O, wcseethatR(y 1 jy 2) =
R(y 1) a nd R(y 2 I y 1 ) = R(y 2), and we can use T able 6.2 to t est any onc 2pl =4
of t he three h y p othcses 13 = O, y 1 = O, a nd y 2 = O.
The noncentrali ty p arameter in t hc test y 1 = O g iven in Art. 6.4.3 is 3p2 = 3
X{X2(X~X2)- X~X1lY1
1
). = y ax;xl - p3 = :3
2a2 / Since X ' Xis diagonal, the{J,areorthogon al. W egetp 1 = 2,p 2 = 1,
This can be written {J3 = 5, and R(l3) = WX'Y = 36. Snppose Y'Y = 48; then the e rror
1 1 sum of squarcs is 12.
,
A = -2? Y1'X'X ' X' 2 (X'X
l Y1' X 1
1 iY1 - ::;-;; 2 2 )- x :!·x iY1
Suppose we wish to test t he following
· cr ~a-
Since the power fiinction {3(J,) is an increasing junction of J., we wa nt). to l. f31 = o
be as large as p ossibl e. rrx;x 1 is fi xed , J. is a maximum ifX;x 2 = O. 2. /32 = o
Thereforc, if possible, we choose t he X matri x such that x ;x 2 = O. 3. f31= f32 = o
For 1 we need R({3 1 1 {3 2 ,(33 ), which equals R(l3) - R(/3 2 ,/33 ) . But
W e n ow formulate the defini tion of ort hogonali ty.
1-12 LINEAR S'J'A'l'LS'l'J CAL J\fOD ELS
J\fODEL l : GENERAL L TN EAR HYPO'l'lrnSJS Ol? FlJLL RANK 143
R(/3 2 ,{1 3) is obtained from t he no rmal equations by setting /JJ = o and SoJdng the e<)IHLLions g ivcs
ig n ori1~g the first equation. V1Te get /f 2 = 1, /J3 = fi; so R({Jdl 3 ) = 2
8
T herefore, R( f3 ) - R((J 2 ,{J3 ) = 8. B u t t his is R(p 1 ). •
- ( ::!i.7!) )
I
Fortest2 weneedR(P 2 p 1 ,p3 ), which equalsR( f3) - R(p 1 ,{J3 ). Bya f3 = 8 .79 D* = 22.79 + 8.7!)'1'¡ - 2.69T2
p rocedure sim ila r to tha.t give n abo,·c we can s how t hat R(P 2 p ,{J) _ I
R(f12) = 3. i a - -2.G9
Now, for_tes~ 3 we need R(p 1 ,p 2_ I P3 ), which equals R (f3) - R(f3 ). 'fo estima.te a 2 ancl to test the hypothesis fJ 1 = {3 2 = O, we sha ll con-
B u t , by settmg fJ 1 and P2 zero a nd us111g t he last of t he normal eg uations
3
struct an AOV ta~le. The quantit ies needcd are given below.
I
we get R(/33 ) = 2 5. So R(/3 1 ,/1 2 {33 ) = 11 . B ut ~his eguals R({Jt>[l .): First, R( f3) = f3'X' Y = (22. 79)(290.0) + (8.7 9)(3,264.9) - (2.69) x
whic h is R(P 1 ) + R(p 2 ) . -
(5,!J67.2) = 19,24-0.5.
6.4.6 Example. T he distance a part icle Lra vels fro m a givcn
refere nce p oint is g iven t heoretically by t he curve If we set y 1 = G:) a nd y 2 = {3 0 , the rcduced normal eq:.1ation is
D* =Po + f31 T1 + /32'1'2 obtainedfrom t he model y =Po +e ; thatis, weset{J 1 = {3 2 =O . \Ve
where D * is t he d istan ce, T 1 is t he t ime t he par ticle moves, and T ~ is thc rrct t he normal eq u<Ltior1 8/j0 = 290.0 a.nd y2 = /Jo = 2!)0.0/8 = 36.3
tempera.tnre of t he medium t h rough \Yhi ch t he part icle m oves .- 1'hc ~nd R( y 2 ) = y~X~Y = (200.0)2 /8 = 10,5 12.5: R(y 1 1 y 2) = R( f3) -
t ime a.ncl te mperature can be m eas ured "·it hou t error, b u t, instead of R(y 2 ) = 8,74-3.3.
observing D*, we observe y = D* +e, where e is a normal ra.ndoin The AO\T is g ive n in Table G.3.
er ror wit h mean zero. Th e fo llowing set of measurements were taken:
'J.'ADJ.E 6.3 ANAT,YSI S OF VAR I ANCE
y 6.0 13.0 13.0 29.2 33. l 32.0 46.2 l l 7.5 SV DF SS MS F
2 3 4 5 6 8 20
Total 8 19,286.!)
10 12 11 14 15 D110 to j3 3 19,240.5
18 30
Duo to y 2 (unadj) l 10,5 12.5
l . Find ~ . 8 2• D uo to y 1 (adj) 2 8, 728.0 4,364.0 470.3
E r ror 5 46.4 9.28
2. F ind D* .
3. Test t he hypothesis p1 = {3 2 = O.
We get From this we geL a2 = 0.28. Using Eqs. (G.8) and (G.n) we can set
confide nce inte r nils on the {3; and on a 2 •
+~
4!)
120) 6.4.7 T est of t h e H y pothes is A.;13 = A.~ f3 = · · · = A.;(3 = O. In
X'X = ( 555 1,014 this test we assumc t hn.t t he r x p m atri x
~ (::i
120 1,014 2,110
T he norma l equat io ns a re
120/J 0 + l ,014/j1 + 2,110¡32 = 5,967.2 has rank r . L et G be a p X p matrix of rank p such that G = (~:).
LetG- 1 = /1 = ( /1 1 , .ó. 2 ),where .ó. 1 hasdimension p X r: Now G isa
n.nd
X'Y (3.:::::)
= k no\nl matrix, hence G - 1 = 11 is also k nown . Let
5,9G7.2
144 J.I ' fü\lt S'l 'A'r lSTICA L l\IODJn.s ;\IODEL 1: CBNERAL LINEAR llYrOTJIESIS O"E' FULL RANK 145
Let Hcro,. = 2 ancl 1..; = (O , 1, - 2); A.~ = (l , - 1, O); t hc on ly ne\\" quantity
, ,c nced to compute is R( cx 2 ) . I n the model wc put {3 1 = (3 0 = 2{3 2 •
1
'l'his gives us
We can write t he_ rn odel Y = X 13 + e as Y = XG- 1 G13 + e oras Y = 2/32 + 2íJ2T1 + /32'1'2 + e
Y = Za. + e. T h1s bccomes Y = Z 1 a. 1 + Z 2 cx 2 + e. Now cx 1 :::: 0 or y = {32 (2 + 21.'¡ + 'l.'.J +e
is equintlent to G 1 13 = O, which is the test we desire. \Ve now follow
which wo "-rite as y= (32x 2 + e
the procedure given in Art. 6.4.4 to find R( cx) and R( a. 2 ). R( cx) is the
same as R( l3), and wc set a. 1 = O and compute R(cx 2 ) . From thes wlicrc x 2 = 2 + 2T 1 + T 2 • Tho reduced normal equation is fJ ,290iJ2
quantities we gct a.n AOV t.able . e = 13,077, which g ives p2 = 1.41. R(a. 2) = if~X;Y = 18,407.7, and
To sum up, we have the fo llowing instructions: To test :J.;13 "" H(¡3) - R( a. 2 ) = R( a. 1 1 a. 2 ) = 848.1. The F value is 848.1/9.28 =
A.~ 13 = · · · = :t..; 13 = O we ha.Ye two models : !11 .:HJ. This is larger than the tabu latcd F valuo fo r 1 and 5 <log rees of
frcc<lom at t ho 1 per cent level.
l. The nnrest rict.ed _:i10del Y = X13 + e, from wh ich we obtain thc
normal eq uations X'X13 = X'Y nncl get R( l3) = j3'X'Y Problems
2. T he model rcstricted by the hypo thesis :t..; 13 = :t..; 13 = · . . ""
1..;13 = O. This wc sha ll write as Y = Z 2 a. 2 + e ; from it we obta,iu the 6.1 Prove LhaL , for s imp le lin ear mo<le ls s u ch ns t ho1;0 o í A r L. ü.2.G,
normal equu.tions Z~ Z 2 éi 2 = z;y a nd get R(a. 2 ) = a;z ; Y. ~[~(y¡ -
2
y)2 - [.E (xi _- x){y¡_-=- 'fj) ] ] = _ l _[(Y - X~)'(Y - X~)]
vVe compute Y' Y n,nd obtain t hc AOV of Table 6.4. For example, n - 2 l.(x; - x) - n. - 2
suppose wc have t he model 6.2 An e xperirnonLer has theore tieal reasons to b cl ie \·o that, in s toring ice
crcum at low Le mporatu1·es, tho a.vernge we ig ht loss E(y) = .& oí t he ice c ren.rn (in
Y1 = f3o + f3tx11 + f32:C21 + f33x31 + {34x4; +e; pin L contn.ine rs ) is linea rly re lated Lo the storage t,ime . 1 10, t here fo re, ossumes
tl mL Lho linear m odel
TAilLI~ G.4 ANALY S IS or.· VARIANCE FOlt T.ESTJNG L = E(y) = {J 1t
t..;f3 = t..; f3 = "~13 =o
rol1llc;; a\·era go we ig h t loss t o storage timo. H e lm om; t hc ro are m an y othcr
.,- DF SS l\'IS F fllct ors bcsiclcs Lim e t hat affect w e ight loss, but it seerns feas ib lo to assumo that
thcsc factors togothe r aet a s a random e rro r c. llenco Lhc mod c l can be wriLton
Total n Y'Y y = {J 1 t +e
Dueto f3 p ~ 'X'Y
Due t,o a 2 (unadj) JJ - T a:;z;v To csLimaLe {J., h o conduets an ox pe rime nt in whie h h e m emH1rcs tho we ig hL loss
Due to a 1 (adj) ,. ~ 'X'Y - a 2 Z~Y of pint.s of ice c rea m O\·er an ex t e nde d p erio d . The resulls (t i8 in weeks; y is in
Error n - 71 Y ' Y - ¡3' X ' Y grnrw;) a re
I¡ 2 3 4 7 8
and we want to test {3 1 - {3 2 = O, /3 1 - 2{13 = O. We can s u bstitute
/3 2 = (3 1 a nd {3 3 = ~{3 1 and gct . 15 .2 L .30 .41 .'19 .!i!l .72 .83
Y; = f3o + f3 1(xli + X2; + ~X3;) + f34X4; +e; EHtinmte {31 and a2.
6.3 Find P.. í32, ÍÍ:¡o á 2, and P1 - P2 + 2/ia fo r
for thc model Y = Z 2 a 2 + e rcstricted by thc hypothesis. ~h e lineal' mode l Yi = {3 1 + {J 2 x 1 i + {J3 x2i + e;
Lhe uaLa lit>low, a s:;u111 i11g Lho.L
flts Definit.i on G. I.
As another examplc, supposc we wanted to test thc hypothesis
/33 = (3 4 , (3 1 = (3 2 • • 111.>stituting, wc obtain y 1 1 5 1 o 1 4 4 - 1
?J; - {3 0 + {J1 (x11 + x 21 ) + {33 (x31 -t= x 41) + e1 ~-l--2-·-1-,-3- 3 3
for the modcl Y = Z 2 a. 2 + e rcstricted by t he hypothesis.
6.4.8 Example. Suppose that, in the example of Art. 6.4.6, we X2 -1- -1--2- 1_1_ 2 3
~
6
recorded:
6.4 Ir X ( : : ) , find X ' X , (X ' X ¡->, X ( X ' X¡- ', X(X'X)- 'X', and 1 _
T ADI.E 6.5
X(X'X)- IX' . -y X3 X4 X5 X6 X1
--
Xo X1 }(2
6.5 J n s implo linea r m od ols, as illustrnted in A1·L. G.2.5, provc Lhat, if the :t
can b e solect~d a nywhe ro in the intcrval (a,b) and ifn is un oven in tcger, t hen th' 4 1 1 1 1 1 1 1 1
variance of P2 is minimized if we selected n/2 valucs of X; equal to a a nd n/~ 5 l 1 - 1 l - 1 1 - 1 -1
values cqual to b.
3 1 1 l - 1 l - 1 - 1 - 1
6 .6 Provo thc id enLit y in Eq. (6.7). 2 l l - 1 - 1 - 1 - 1 l 1
ó.7 U se tho X m a Lt-ix in Prob. 6.4 a nd find t r(XS- 1X ') o.nd Lr(I - XS- l X'). 6 l - 1 - 1 1 1 - 1 l - :i
ó.8 Lct kr and k~ be pos itive constants. Provc Lhat 3 1 - 1 l l - 1 - 1 - 1 l
4 l - 1 - 1 - 1 1 l - 1 1
8 l - 1 l - 1 - 1 1 1 - 1
1
is pos iLivc d efin iLc, wh cre X is an n x p m aLrix of rank JJ < n.
ó.9 In Prob. G. 8 , s h ow t h at (a) Find X 'X.
(b) Find R( f3) , R({J1 1 Po.P2• · · · ,fJ1), R({J3)·
v- 1 l
= -;; xs- 1x· + -q1 (I - xs- 1 x·¡ I
(e) Show t hat R(f12 ,f13 Po,P11fJ4, · · · ,{J7) = R(fJ2) + R({J3)·
kr (d) Find t he reductio n due to H 0 (adj) , wh ere H 0 is P1 = P2 = {J3 •
(e) S how t,hat t he re is no unbiased estimate of a 2 •
6.10 In Theore m 6.4 , if V = kiXS- 1 X ' + k~( I - XS- 1 X '), sh ow that ~ is 6.17 In Lhe proof of Theore m 6.6, s how Lhat Lhe matrices A , (A 2 - A) , a nd
Lhe so.me os it is if V = l. I - A. are each ide m potent.
6.11 Provo Lh e fo llowing statem en t, made in A1·t . G.2.6: vn r (P1 ) is o. minimum 6.18- In P rob. 6. 17, show t hat the prod ucLs A (A 2 - A), A(I - ~) and
if tho X; aro c hoson s o LhaL x = O. (A2 - A )( I - A 2 ) a re each e qual t.o t hc null m atrix.
6 .12 r •rovo 'rhoorc m 6.3. 6.19 In P rob. 6. 15, s uppose a value of y equa l (,o 5.3 is selected from an
6 .13 P rovo Thcore m 6.4. unknown X value. Use Eq. (6. 17) to set a 95 por ce nt confid once inte rval on the
6.14 In Lhc m o d c l Y = X f3 + e , whe re n = 10 a nd JJ = 3, t ho foll owing nor. unknown X.
mal c qu u.Lio ns wc re co mputod (Y' Y = 58): 6.20 In t he examplc of Al"t. 6.4.6, set a 90 pc1· cen t con1idence inter val on
f11 - · 2P2·
3/i1 + P2 - ·2Pa = 1 6.21 If n = 20, p = 6, :< = .05 in Eq. (6.8), find t he oxpccLcd width of t he
confidcncc intcr val.
P1 + 2P2 + P3 = 1 6.22 R opea.t Prob. G. 2 1 for n = 40 , p = 6, r1. = .05 in Eq. (6.9 ).
6.23 In Prob. 6.21, s how th at lim E (w) = O, wherc E(w) is t h e expected
wid t h . " - 00
(ci) F ind ~. 6.24 In Eq. (6.9), find Lhe d istribu t ion of tho width w.
(b) Find R(¡3 ) = ~ 'X'Y. 6.25 In P rob. 6.24, find E(w) a nd var (w).
(e) Set, confidenco limit s o n a 2 , p1 , P2 , p3 , and P1 - {J2 • 6.26 Provo that Lhe cocfficient of a 2 in Eq. (6.1 5 ) is equa l to
(d) F ind R(y 1 ) , who rc y 1 = P1 •
.E(x; - x 0 ) 2
I
(e) Fi nd R(y2 y 1), w he ro y; = ({J2,{J3). 1
6.15 ln !he m od c l Y; = /1 0 + /11:i·; + e; (j = 1, 2 .. .. , 20), tl H' fo llowin¡? 11o r· k + n~(.c; - f-)2
m a l cquations wc rc compu tcd p:.yJ
= 90):
6.27 Jn Eq. (6. 12), fincl x~s- 1 x 0 for t he s imple linear m odcl
wherc
10/j0 + 6/j1 = 22
1
(a ) • et a 95 por cen t confid cnco in ter val o n E(y), whon X = 3. Xo = ( )
:r
(b) If wc sclcc Lu. su.mplo of 30 values of y fro m the above m od cl wh cn X = 10,
find a 95 p or cent confid cnce interval on the. m ean of t hese y valucs. U se Eq.
(6.1 2).
148 LINEAR STA'rIS'.l'J CAL l\IODELS
F u r th cr Rea d in g
1 O. K empt horne : " D es ig n a nd Analysis of Expe rirnen ts," John " "ilcy &
Sons, Inc., New York , rn.::;2.
2 R. T.... Andci·son and T. A. Bancroft,: "Stat,ist,ical Theory in R esearch ,,
1\IcGraw-Hi ll Book Compa ny, Inc., Ne,,· York , 1952. '
3 c.. .R. Rao: " Advan ced Statistica l l\fct,hods in B iometric R esearch ' " J om
1
Wiley & Sons, Inc. , Ncw York, 1952.
4 H. B. 1\1Iann : " Analysis ancl D csign of Experimen ts," Dover P ub lications
5
New York, 1949.
F. N. D avid a ncl J. Neymn n: Exterrnion of t h e 1\farkoff Theorem on Lea ·t
"
7
Squa res, S tatist. R esearch .111em., vol. 2, pp. 105- 116, 1938. s
6 C. R. Rao: A T h eorem in L east Squa res, Sankhyü, vol. 11, pp. 9- 12, 1951
7 C. R. Rao: On Tra ns formations U seful in the Dis t ribution .Prnble ms of .
L east, Squares, Sankhyü, vol. 12, pp. 339- 346, 1952-1953. Co1nputing T echniques
8 C. R. Rao: Genern.Jizat,ion of 1\forkoff's Theorem and Tests of Linear
Hypotheses, Sankhyü, vol. 7, pp. 9- 19 , 1945- 1946.
9 C. R. Rao: On the L inear Combin aLion of Obpervations ancl the General
Theory of L east Squarcs, Sankhyéi, vol. 7 , pp. 237- 256, 1945-1946.
10 R. L. Plackett: Some Thcore ms in Lcast, Squares, Biometrika, ...-ol. 37
pp. 1-1-9- 157, 1950. , 7. 1 I ntroduction
11 P. C. T a ng : The Powcr Funct,ion of A OV Tcst,s, Statist. R esearch .M cm
vol. 2, J 938 . ., It was sh own in t hc p reccd ing ch apter t hat; to estima.te f3 or an y
12 A. \ Vald: On t,he Powe r l<unction of i,he Ana lysis of Variance T est, Ann." component {3 1 from the modcl Y = Xf3 + e, we must find thesolutio n ~
Jlfath. S tatist., vol. 13, pp. 434--439, 1942. of thc set of equations X' X~ = X ' Y. W e can sol ve for ~ by finding t he
13 A. Wald: On the Effic ient, D esign of Statis t ica l Investigations, Ann . .llfath. in verse of X'X, and get ~ ·= (X ' X)- 1 X'Y. If we want on ly ~ ' we can
S tatist., vol. 14, pp. 1:34--140, 1943.
solvc t hc system of equations X'X~ = X 'Y wit hout actually fi nding
14 S . \ V. Nasl 1: N ote on .Power of 1,he Ji' Test, Ann. 111ath. Statist. , vol. 19
p. 43'1, 1948. ' thc inversc of X 'X . B ut , if we want a lso the covariance matrix of ~'
15 J. \ Volfowitz: The Powcr of the Classical T ests Associat,ecl with the Normal weniust find t he inverse ofX 'X, sin ce cov( ~ ) = cr2(X 'X )- 1 • Both ways
Dis t1·ibution, Ann. :N!ath. S tatist ., vol. 20 , pp. 540- 55 1, 1949. will be discussed. Section 7.2 will be dcvoted to a method of solving a
16 J. E. l\'faxfield and R. S. Ga rclne r : Note on Linear H y potheses wit,h ºPro- system of symmetric equations wit hont actually fi nding t he in verse of a
scribed Mat,rix of Norruul Equ ations , Ann . .lliath. Statist. , vol. 2Ci, pp.
149- 150, 1955.
matrix, a nd Sec. 7.3 will be d evoted to finding t he inverse of a
17 R. A. Fishe r : The Goodness of Fit a ncl R egression Formulae, a nd the symmetric m atrix.
Distrib ution of Regression Coefficie nts , J. R oy. Statist. Soc. , vol. 85, part There are many ways to invert a matri x, and t he method to use in a
I V, 1922; roprinted in "Contribu t ions to 1\fothematical Statis t ics," John particular ins tan ce depends upon what type of calculator or computer is
\Vi ley & Sons, Inc., New York , 1950. a.vailable. \ i\Te sha ll present only one method , the Dooli t t le method,
18 S. Koloclziejczyk: On a n Import,nnt Class of Statis t,ica l H y p ot.hcsos,
Biometrika, vol. 2 7, 1935.
which is especia lly adaptable to ordina ry d esk calculators.
no uzero c lc me n t. Th c s paccs whe re a n a s teris k ( *) a ppears need not b Jfowever, t,hc red nction d uc to f3 can b e obta i ned ve ry easily from
filled, sincc Lhe elcmc nt s in t hcm would be zero ifthcy " ·ere co mplete~ Table 7.2. It is
via thc instru ction. After each r ow is complcted, it sh o uld be totaled R((3 ) = ~ 'X'Y = R1or10 + R2or20 + R3or30
to sec whcthcr it e quaJ s Lhe elemen t in t hc check column. which is t h e s nm of the prod uct of t he t hree pairs of Yalues in t hc C 0
col um n. .
T An1, e 7.;!, Auorn:nATED D ooLI'rl'LE l\fET H OD F OJt .·01.v1 NG EQ. (7. l ) Supposc wc wa.11 ted to tes tthe hy pothcs1s (33 :== O. \Ve lrnow from
Ins Lr uc Lion Row Theore m G.5 a nd T a ble 6.2 tha.t we need to_ ob~am:
C2
-
C1 C3 Co Check
l. Thered uctiondueto{3 1 ,{J 2 ,an d{3 3 , wluch1sR( f3) = ~ 'X'Y,whcre
R'J 2 4 2 G M ~ is the solu t ion ~o Eq. (7. 1) .
H.; 10 2 18 34 2. The reduct1on duc to {3 1 , {3 2 (when {3 3 = O), w hich is R(y 1 ) =
R¡ o
~'X'Y (~
3 12 - lG 1
whc rc y- 1 = \ is Lhc sol u tion to t h e norma l equation s in (7 .1)
@ Y1 1 • ~J
R~ R 1 2 ® (i 1-J.
tR1 r 1 l 2 l 3 7 exccpt w ith P:i = O a nd the {3 3 equation (7.1.3) omitted. That is, we
sol ve
R;¡ - Ji 12 r¡; H2 * 2 @ (i G
(7.2)
tR2 r2 * l - L ;{ :1
1
1
R~; - R 13 r 1¡ - R~rv - R3 * * 8
l
- LG
- 2
- 8 and get R(y1 ) = y~X~Y = íJ1(6) + iJ2(18)
tR3 1"3 * * - 1
1 Then the s u m of sq u a res fo r {33 a d j usted fo r {3 1 a n d {3 2 is
R (y2 1 Y1 ) = "
Í: R ;or;o
Total 1i - 1 Í:" (Y; - y) 2
j - 1
i = k -J l
¡J - 1
\Ve s ha ll uot prove t his t heorem , b u t it is cxtrc mcJy important . The P - l Í: rfoRfo
i- 1
rcader s ho uld become well acqu ain ted wit h it if he is going to use t he
p - k-1
abb reviatcd Doolittle technigue t o o btai11 thc qua n t it ics nced ed t o test P - k - l Í: rfoRfo
h y p ot,h eses in model l. i- 1
7.2.3 Alte rnative Method of Presenting the Analysis-of- p- l
v a riance Table. lVfany t imes t h e X matri x i n mod el l h as uni ty for /,; Í: rfoRfo
i - 1>-k
evcry clc me n t in t he first column, and th e cguatio n takes t he form
11 p-l
1)-1 E rror n - 11 Í: (Y; - 'fi) 2 - Í: rfoRfo
YJ = ¡¿ + iI= l xdl; + e; j = 1, 2, . . . , n (7.3) j - 1 i - l
1 1 - 9
=
2 4 2) ~/l3
-
T3
- -
*
-
*
·-
1 - 2 -
;¡
8
- - - - - - - - - - - - - --
l.
ll t - ·a9
X 'X
( 4
2
10
2
2
12
R15r 1i
Rz6r2i
+ Rzsr2i + R3 51"3;
+ R3Gr3i
B1
B2
11
_l.i .. -¡j3
~Q
5
1i l.
8
vVe can find B 1 by solving U37T37 B3 J.
8
~)
4 2
Using columns 0 1, 0 2 , 0 3 , a nd E 2 ofrows rv r 2 , andr3 we get
(: 10
2 12
2
_.1..l.)
by t he a bbreviated D oolittle mcthod tind B 2 by solving
B, ~( ":
(
2
4
4
10
2
2 1
º) nnd, by a similar process, we get
2 2 ¡ :,¿ o
and simila rly for B 3 . \i\fe ctLn save wo rk by p utting Lhc matrix in thc B, ~ (-:ª)
form
2 4
2 11 o º)
21 o l o
-:)
2!) - 11
( : l~ 12 1 o o l Thcrefore, B = (B 1 , B;¡. B 3 ) =~ - 11 5
(
and reducing to tria ng ular for m. This is il l11stratccl in Tab le 7..J, -a I
where all t he operations excep t t hose of t hc last t hrcc rows are exactly
t he sa me as in Table 7. 2. There is a short-cut method for obtaini ng B by using t hc ins tructions
for the last t hree rows ofTable 7.4 and nsing only colu rnns E 1 , E 2 , and
o
Using columns l > 02, 03, a nd E 1 of rows 1" ¡ , r2. anc.l T3, wc find thc
E3 . This gi•es t he clements of B immcd iatcly. In co mput ing t he
element s of B 1 to be
20) rows B 1, B 2 , B 3 notice t he pivota! ele menLs R 15 , R 25 , etc.
B, ~ (- •:;
\Ve rnight poi nt out that t h e dclcrmin ant of A is equal to Lhe prod uct
of t he first nonzcro elements in rows R 1 , R 2 , and R 3 . ln this part icula r
problem it is eqna l to (2)(2)(8) = 32.
COJ\lPUTl NG TECH N I QUES 159
158 LINEAR STATIS'l 'ICAL llIODELS
TAUJ,E 7.5 p P
.3.2 The General Case . W e shall look at t he solut ion fo r a
AaaitEVIATED DooLrCTLE FonMAT Fon. A x 7
SYMllffiTRIC MATRI X eneral p x p symmetric m atri x. E verythi ng proceeds as for the
~ X 3 case except t hat we shall have more rows. T bc instruction and
InstrucLio n Row CP2 · · ·CJJC 0 E1 ···E.;,¡11 row columns fo r a p X p symmetric matrix are give n in Table 7 .5.
'J'he inst ruction A t means : Divide t he elcmcnts in t he row by the first
110
nzero element of row R t. _
7.3.3 E xample . W e s ha ll demo nstrate t he abbreviated D oolittle
technique by a n example . In measuring t he various constit uents ~f
R',,
cow's milk it would be d esirablc to find a fu nction rclat ing t he q uant1-
- ties: solids nonfat, fat, a nd protein. Supposc we wish to estimate
protein by using a knowledge of factors fat Xi and solids nonfat X2 by
---··------
t he equation
Y; = µ + f3 ix;i + f32X¡2 + e,
By taking samples from l 6 cows t he values shown in Table 7 .6 were
obtained.
TABLE 7. 6 DATA FOR ExAMP1.i:: OF A nT. 7.3.3
8.86
--·- - - - - - - - -1- -- 1-- - - - - 3.75 4.74
3.66
9.50
8.56
3.16
:l. 6fi
3.36
:l.64 9.2 1
3. 19
8.54 3.36 3.92 8.93
2.99 4.27
8.62 3.60 2.99 9.16
:l.46 4.0:l
9. 35 3.87 :1.28 $ .'1.5
p t :1. 27 3.5 1
:3. 14 3.23 9. 09
R' . - ~ R 1·
R,, 3.27 3.97 S. 39
"' _,¿_, <IV Q}
3.23 7.87 :1 .00 :3.Gfi 8.:l6
q- l 2.78
3.79 !.Ua 3. 18 4. 23 9. 28
rV a. 59
IG.00 59.50
5!).50 224 .47
-; - - ------------t- - ---l- - - -- - - - - - (
141.50 527.0 1
LR;_ v+l!-JriJ
i- t
Thc equations are red uccd by t he abb rc,·iated Doolittle method and
thc desired quant ities computcd. This is illustra.ted in T a ble 7 . 7. The
AOV is g iven in T able 7.8. Soh·ing fo r t hc rcgressio n coefficients givcs
/j 2 = .34 2566, /j 1 = - .OL77!HJ, /t = . 365;{-1 0 . Usi ng Theorem 7. 1, we
get
R( f3) = ~ 'X'Y = 177.65 and R (µ ) = 177.29
COMPUTING '1'.IW H N IQUES lGl
Thc F valne found in Table 7.8 is less t ha n t he tabulated Ji' val ue ¡1,t t hc
5 per cent lcvel. A test of t he bypothesis µ eq~ al to zern (or to so me
·- -·- --
•O(!:) •n o other constant) can be made by t be t test, us111g the elements from
'º C> o
<X> <X> M O'l
o e oo o in
M
- L-
M'-'l
<O <O
(X'X)-1giyei1in lines B 1 , B 2 , and B 3 ofTable 7. 7. Similarl y for tcsting
<N C'I l.":I "d'i -o
.,., .....
-i
·"''--
"'' O'l
-1 o o - <O
'J'ADLE 7.8 ANALYSJS 01•' VARTANCE FOR DATA OF TABLE 7.6
·-
C-.1
-""
o O>
~ 1 1
sv DF 8S M:S
,_ Total 16 178.66
000 o o o o 00 "" CN<NM
O'l <O o R(¡L,{11 ,pz) :3 177.65
000 o o o o o o <X> ..... o 1 177.29
o o o o o R(p)
~ ~"' º
o º
0º0
o o o
o
o
o
o
o o O
<O
M
<O '<!< <O
Ir.> <X> M
R(PpP2 11) l 2 . 36 .18 2.30
rll
< ..... ºº O M <O O M
13 1.0 l .077
E-l "¡' 1 Error
...
o
< p1 or {J 2 . Also , t he in verse eleme nt s can be used to test a li near com-
o·-
M L') 0
"'
A 000
o 00
00
00
ºo
M M t-
M _,
<X> <O
bination of µ , {J 1 , (3 2 equal to so me constan t and for setting confide nce
o 00 0 0 ..... '<!<
o 00 00 oo "" L') CXl Jimits on µ , {3 I> or {3 2 and on a linear cornbination of t he parameters .
000 00 O M ""o
,..... We shall refer to t his example again later in t he chapter.
1 1
Now we sha ll test the hypothesis {3 1 = {3 2 = O using-the abbreviated
M M C--l Doolittle method on the correctecl s um s of squares and cross products.
o
o o o
o 00
00 o
ºo""
'º .,., "''
1- Ir,)
1-0 O')
M <X>
The corrected sums of squares and cross products are en tered in a table
CX>O .... <O
o <O
º
o ºº 00 o o
·- -
-
M
1 1
<O
..-<
O
O'l <O
L')
.,.: cÑ
1 1
Jike Table 7. 9 and reduced as d emonstrated t here. The correspo nding
TABLE 7.9 A n nn.Ev rA'rEo DooLl.'l'.L'.LE Fo1i M A T J.' OR DATA O.I!' TABLE 7.6,
L') --< CoR1mCTEO fiuM OF SQUARES
'º
L-
·-
M '<!<
<O
CX> O'l <X> Check
<O C\I
<N
~
"'
L.-,
oP""'"1
00 r.-i
t-
e.o
C\I
MM
C\l
~ -
"''º
<O
01 02
ºº E1 E2
C)
.-< '<!<
IJ?
R{ 3. 2018 .80 158 .22 198 1 o -
ri.22ñ:rn
R'2 o 6.06!)36
·-o- ·--oo
3. 1822 1.08558 1
lQ ~ 00 -
,_
L')
<X> <X> R1 3.20 18 .80158 .22 1!)8 1 o 5.22536
o .....
IQ
~e-: ~
0
t-
LI')
º """
l.~~
M
..... <X>
<O -
o'º
00 t:'-1 ·- o
<O
~~
CN .-< --
'"1 1.0000
-·--
.25035 .069330 .31234 .00000 l. 63200
'<!< <N .,.,
,_..¡ lQ ~) .....
""" 2.98 152 l.0300 1 - .25035 1.00000 4. 76 118
..... Rz *
'"2 * 1.00000 .:i4546 - .08397 .33540 1. 59689
- ----- - - - - -- - -
.. ·-oo
lf:)
MO
'<110 B1 .:~3334 - .08397
C) 00
B2 .33540
""~ * *
M-
AOV is given in Table 7 .10. li'rom Table 7 .9 we see t hat thc valu es of
o
o
o o
o o P an d p are the sam e a s whe n t hey are computed in Table 7.7 , except
- -
1 2
<O ¿ .....; * * * * for ro un<ling errors. T h e a nalysis of vn.riance in Table 7 . l O gi ves the
samc mean s qua re for error a nd for t he reduction due to {3 1 and (3 2
~01 ¡.."' 1 adjusted for µas T able 7 .8 a nd, hence, the sam e J? value . The work is
160
CO~J PUT!NG 'J'ECHNIQUES 163
162 LINEAR STA'rIS'l 'lCAL i\fODELS
T ABLE 7.10 ANALYSlS OF VARTA NCE Q}' CoRRECTED Su~IS OF nnd find t,he inve rse of t h e m atrix.
SQUARES AND Cuoss Pnoo uC'r s 1.2 R epeat Prob. 7. 1 for t ho system
• Theore m 7.2 If Bis a n approximate in verse of t he matri x A , t hen use Theorem 7.2 to improve t h e approx imation. .
a n improvement on Bis given by B *, whe re 7.4 In Al"I;. 7.2.3, prove that t h c maximurn-likelihood estim ate of {3 0 is t he
samc whother we use the model in t he z; or x;.
B* = B (21 - AB ) 7.5 In Art. 7 .2.3, prove tho.t R({J 0 ) = (I:y) 2 /n. .
7.6 \Vorlc t ho problom in the example of Arl. 6.4.6 using t he Dooh ttle
, upposc we w:mt t he inverse ofth e matrix
( 2 -3)
method. . . 2 . . .
7.7 Provo that t ho total corrcotccl sum of s quares cl1v 1dod by a is dist nbuted
2 2 2
l\S 7.2(n - l ); i.e., prove t hat I:(y; - y) /a is distributod as :r. (n - 1) if y is
A=
- 3 5 distr ibutecl N(O,a 2 1).
7.8 T'rove t,hat, if t h e p X p matrix
The invei·se is A - 1 = (: :)
o l o o
Sup pose, h owever , t hat we have on ly an approxima te inverse
B O o o
B = ( 4.!)89 3.011)
3.012 1.001
o o o
1.058 - .0±9)
Then 21 - AB = (
- .093 1.078
o 1 o o
4.908 3.001)
and , t hcrefore, B * = B (21 - AB) =
( 3.002 then B- 1 = O o o
l.!)!)!)
If still further acc uracy is desired , the opcration can be repeated on
B*. o o o l
164 LINEAR STA'J'ISTICAL lllODELS
7.9 If X'X~ = X 'Y is a sys tem of normal equations, whcrc ali t h o e lements
in t,h o first column of X are equal to unity ancl where ~ · = (fi,P1, ... ,p»-tl,
p rove t hc fo llowing:
j = 1, 2, ... , n
...
If we ]et· Xi; -- .x ;, ·x 2i -- ·x21 ' •x ª'. = ·x~1, we ~get t he ]Jarticular modcl
Y; = {3 0 + {3 x; + {J x; + {J 3:c~ + e;
1 2
165
166 LINEAR S'l'ATISTICAL J\[ÜDl':LS
POLYNOl\llA L OR CURVILTNEAR l\lODELS 167
This model is exactly model l. The p ert incnL matricf':s are We s hall assu mc that f(x) can be expand ed in to a Taylor series, so
that we obtain
1 X1 x¡? x3
1
+ 0 x + 0 x + · · · + O~ + · · ·
L X¡ :Ex7
fa) = J(x) = 2
y 00 1 2
X =
1
1
X2
X3
x2
2
?
x3
X~
X~ X 'X = ("
L X· :Ex;
LX; LX~
:Ext LX~
L x1 LX 5
'fhis can be writtcn y = 0 <X + a x + e wherc e is t he rcmainder in a
1 1, 1
Taylor series, which--we assume fo1· our purposcs is random error (see
LX~
the equa.Lion-crro r model in Chap. 5). If the fit to a. linear equation is
Lxt L.x1 LX~j 110 t satisfactory, we assu me a new model
l x,. x;, •x3
n
1
Error n - 2 Y ' Y - R(cx 0 ,CL 1) El R(y 0,y 1,Y21Y3} = YoLY; + Y 1LY;Xi + JÍ2LY;X [ + y3L.y¡xf, a nd R(yo,Y 1·Y2)
=== y0 Ly¡ + i\Ly;xi + y2 L.yix'f, where y0 , y1 , y2 are obtained from (8.4}
by striking out t he Jast row and colurnn of t he 4 x 4 matrix a nd
T o get t he qua nt ities in Table 8.2, we find the norma l equations,
w hich are TADLE 8.3 ANALYS1s OF VA!UA NCE Fo 1i Cuurc PoLYNOlltrAL
sv DF SS MS F
(8.2)
Total n
Duo Lo Yo• Yt • Y2• Y3 4 R (yo,J't>Y21J13)
Duo Lo y 0 , y 1, y 2 (unad j) 3 R(Yo•Yt>Y2) = R(f30,fl1,fl2)
and R(f30,f3 1.f3 2) = PoLY; + P1LY;X; + P2LY;XT. R(f30,f31) = /JoLY; + 03
/J 1 LY;X;, where Po a n d /J 1 a re solut ions to t he equations D uo to y 3 (udj ) 1 R(y3 1 Yo•Y 1•Y2) 03
E3
Error n - 4 Y 'Y - R(yo,Y1·Y2•Y3) E3
---
p o nen t of x 2 in it . T his may int rodu ce a bia s in t he e rror surn of
squa res for linear. Sim ila r remarks hold whe n we decide t hat a pth. sv DF SS MS Pr
degree p oly no mia l fits thc data. T hc rema indcr sum of s quares for
10 1,452.00
the lower degrees may be biased. If a n indcpendent est imat e of the •rota l
v uc t o cxo, ex¡ 2 1, 3!) l.l G
v a ria nce of t he rnndom varia b le e is available, a bett er met hod is the 1 846.40
onc deseri bed in Scc. 8.5. Duc to cxo .
l)uc to CX¡ (adJ ) l 544.7G 544.76 7 1.58
Some sta t is t ieia ns reco m mend t hat, two consecut ivo n onsignifieant J~rro r
8 60.84 7. Gl
res ults appear befo re a d ceision is made o n t he dcgree of poly nornial.
Suppose , for exa mple, t ha t t hc first t wo consecut ive nonsig ni ficant Since we conclude tha t IX ¡ =I= O, we nex t fit a quadratic. vVe get
resul ts a re fo und when t es ting {l 2 = O a nd whcn t est ing Ya = O; t hen
we conclud e that a lin ear poly nomial fits t ir e data. 10.00 17 .00
32.20 ) 02.0 )
2 . The com puting involved is not cliffi cult for IO\\·-cl eg ree poly- X 'X = 17 .00 32.20 65.!)6 =
nomia ls. Table 8 .1 req uires no d ifficult corn pu tations; Table 8.2 (
32.20 65.!)G 142.G8
X'Y
( 114.0
15G.0
reg uires t ha t we so l ve a syst e m of th rcc eq ua Lio ns wit h t hrec un lmowns,
t o obtain t he s u m of squares due Lo {3 0 , fJ 1 , {3 2 • Ta ble 8.3 req uircs t hat \Ve get Po = 4 2.9 6, P
1 = -28.GS, P
2 = 4.6G, a nd R ((J 0,(J 1,/32)
we sol ve a sys te m of fo ur cqua t io ns and fou r un knowns, to obtain t he 1 409.35. R(/3 0 ,/3 1 ) = R{et. 0 ,et. 1 ) = 1,39 1. lG.
sum of sg uares d ue t o y 0 , y 1 , y 2 , y 3 . H , ho wever, t he data fi t a pqly- ' The AOV is given in T a ble 8.5.
no rnia l of degree 3 or m ore, t he co mputation ca n bccome quit e t rouble-
::;ome. 'fABf,E 8.5 A NALYS IS OF VARI A NCE FOit Q UADl tA'l'IC J:'OLYNOlllIAL l\fODEL S
(ªº)
&1
= ( 31.0424)
- 12.8480
Y' Y = 1,452.0
Tota l
Mean
JO J , 452.0 0
846.40
2
O r thogonal polynomials can be used \\·he t h er t he x val u es are c qually
or un e qua.!ly spaccd. Howe ver , t hey s imp lify computation g reatly l Co + .(3:. ___- ___:n)e¡
_. . :
" ·h en t hc x va,Iues a,re in e qua] stcps; so we sh a ll discuss t hat case only. X= 2 X' X =
8.4.2 Orthogon a l Polynomials for Linear Models . W e shall
d iscuss t hc ideas fo r a linear model first a.nd t h c n point out sonic
gencralizatio ns . l Co + .(n - l )c
:. ____ ~
1
2
S uppose we havo sorne data (y and x values) to which we wis h to fit Since th e constants c 0 ande 1 are at our disposal , we sh all ch oose c 0 = O
a polynomial. ' nppose t he x values are in equa l s teps ; i .e., s upposc such t,Jrn.t X'X is diagonal and c 1 s uch that fractions are climinated in
x 1 = a + h; x 2 = a + 2h; ... ; X; = ci + ih; ... ; x,. = a+ rih. t he model. If n is an odd number, c 1 = l; if n is even, c 1 = 2. The
For cxamplc, if t he X; a r e 4 .3, 4 .5, 4 .7 , 4.9, 5. 1, t h en a = 4. 1, h = .2, modcl becomcs
and n = 5. F irst we s hall fit the linear m odel Y; = Po + P1 x; + e¡;
bu t, instcad ofwrit ing i t this way, we s h a ll write it Y; = rt.o + ª1P1 i (. n+l)
-2- +
- ei = <Zo + Ct.1C1 (2i - 2n- l) + ei
V; = ª o+ a 1P 1 (i - ·i) +e;
where P 1 (i - i) is a fi rs t-d egr ce poly nom ia l in i - i, w hc rc 'i is thc whcrc P(i - ~)
l 2
2i - n -
2
l
if n is a n odd integer
mean of t he i values . T hat is to sa.y, s ince i takes on th e val urs (8.6)
1, .. . , n, ·i = (l/n) í:i = n(n + 1)/2n = (n + 1)/2. Thcrcforc, P1 ( i - n ;
1
) = 2i - n - 1 i f n is a n cven in tcgcr
P 1 (i. - .-)
t -_ 1,, (.
1
i- +
11 -
-
2
1) = C
o + e1 (.
i - n
-- +2 1) 1'his gives
is a firs t-d eg ree poly nomia l in i - í , where c 0 a.nd c 1 are cons t a nts . lf X'X= c2 n(n +Ol )(n - 1)
)
=
(n
0
(8.7)
we set t he polynomial i n X; equ a l to th e polynomia1 in i - i, wc get 1
12
Since th is is a n identi ty in t he .x., " ·e can equate cocfficic nts (re m e mber- , 2
ing x, = a + ih) a.nd get So t he normal equations X ' Xéi = X'Y are easily solved. ·w e get
Y; = ªo + ª 1[co + C1 (. i -
n+l)] + e;
~
Substituting into Eq. (8. 5), we can easily obtain the f] 1.
174 L INEAR S T ATIS'l'ICAL J\IODEL S POLYNOJ\IIAL OR CURVILINEAR l\TODBLS 175
If wc wi sh to test t hc h y pothesis cx 1 = O wc need the following From Eq. (8.8) we get
quantitics : red1i,ction clue to a 1 adjii,sted for a 0 , dcnoted by R(a 1 j CY.o),
and reditction dite to a 0 and cx v d enoted by R(cx 0 ,a 1). Because of the
orthogonality of a 0 and o: 1 , these qua nti t ies are ~ = e~º)
_ R o: _ & X y _ {Ly¡P 1 (i - (n + 1)/2]} 2 From Eq. (8.7),
+ l)/2]}2
(~
R (CJ. 1 CJ. 0 )
1 - ( i) - i i - L{P¡[i _ (n
2
X 'X = 1
:)
a nd R(CJ.0 ,CJ.1 ) = R(cx 0 ) + R (a1 ) = &0X~Y + &1X~Y = (:Ey¡) From Eq. (8.5),
n
{:Ey;P 1[i - (n + l)j2]}2
&o - 3&1 = Po + 4.8¡1 1
Pi (i _ ~- l) (2i - ;• - l)
n = Ci
~= ( - 14.1)
3.0
can be cvalu ated for various values of n. The values of this poly- Also, R(a 1 j a 0 ) = 8.1 ; R(cx 0 ,a 1 ) = 53.1 ; Y' Y = 55.0. So the AOV is
nomial have b een tabu latcd [1] up to n = 104. The computations as shown in Table 8.8.
for a fow values of n a re given in T a ble 8.7 [Eq. (8.6) is used].
TAill..E 8.8 A N ALYS IS OF VARL\NCE FOR LINEAR POLY NOJ\UAL
TAilL E 8.7
2 3 5 G ¿p¡
sv DF SS M:S F Pr
n 4
Total 5 55.0
3 - 1 o 1 2 R(cx 0 ) l 45.0
4 - ~! - 1 l 3 20 R(cx1 1 cx 0 ) 1 8. 1 S. L 13.5 < 5%
5 - 2 - 1 o 1 2 10 E rror 3 1.9 .G
G - 5 - :3 - 1 l 3 5 70
The cen tral feat ures to notice in using orthogonal poly n omials to
8.4.3 Example. Give11 t he d a ta
fit a first-d egree poly no mial when t he xi are equally spaced a re as
'!J LO 2.0 4.0 3.0 5.0 follows:
l. Since, by Eq. (8.5), fJ 1 js zero ii a nd only if cx 1 is zero, we can test
X¡ 5. L 5.4 5.7 6.0 G.:3 {J 1 = O by using the test fo r cx 1 = O.
2. \Ve can obtai n point estimates and interval estimates of {3 0 a nd {3 1
we scc that the X¡ a re equa lly spaced , anda = 4. S, h = .3, n = 5. By by using Eq . (8.5) .
u sing Table 8.7 for n =- 5 we get t he table bclow. 3. The AOV on t he o:i is very easy to comp ute, since t he qua ntities
in Eqs. (8.7) a nd (8.8) are easily obtained.
Y; 1.0 2.0 1 4.0 3.0 5.0 15 = L.y¡ If onl y a linear polynomia l is to be fitted, orthogonal polynomials are
~J-o- - =_1'-~y-,-
.P-_(_i___ n_;_
1
P i (i - 'i) - 2 _2_ , __9___ not wort,h while . But if we want to fit first a linear, t hen a quad ratic,
1- 1_)_
cte., until we fi nd a " best " fit a nd if the X; are eqnally spaced, t hen
1
ort hogo na l po.ly nomia ls are a great h elp. In the next article we s hall
L.y'f = 55 l:[P 1 (i - i)] 2 = L.Pr = 10
use them to fi t a second-d egree poly nomial.
176 L ! N" i': Al t S'l'A'l' IS'l' ICAL l\IODELS P OL YNOl\lIAL OR CUR\' f LT KBAR MO DELS ] 77
8.4.4 Fitting a Qu a dratic b y O rthogona l Poly nomia ls. The and X 'X =
m odel wc sha ll fi t is
Y; = Yo + Y1X; + Y2x7 + e; (8.!l)
,. ¡P+ -n~ l)
'
wh ich we shall \\Tite as
fr.(i- n; l) f [P,(i- ~ J)J •1
?11 = Oo + c51P1( i - n ~- 1
) + 02P 2( i - n ~ 1
) +e; (8. l O)
í.P:( . n + 1) ,L. 1, (· n + J)P (·
i - -?- 1 t --
9 - : n
1 - - ).- + 1) L-'. [l'--:_.r\ i _ 11 +
._,. L)J:
where P1 (i- n +2 ) = +
1
d0 d 1 (1 - n + 1)
2
d 1 -::/= O
¡. - i -
an d P 2 ( i . - + -1) = fo + (.
n- 2 f1 +
n---¡--1) + f2 (. +
n~ 1) 2 'l'hisgi,-es usd 0 = O; f 0 = - f,:_[(n + l )(n - l )]/ 12; and f 1 = O,since
i - i -
di =f. o.
With t hese values t hc p oly nomia ls can b e wri tten
T he gu a nt it ies d0 , d 1 , f 0 , f 1 , f 2 a re a t our d isp osal to determine as we
please. The b \·o poly no mials (8.!J) a nd (8 .1 0) can be rewritten (since
X; = a + ih) P(i - ~)
l 2 = d(i - ~)
2 l
Y; = Yo + }'1(a + i h ) + y 2 (a + 2
ih) +e; (8.14)
P.(i _ +2 1) l)(n + l ) + (i _ + 1)2]
(8. ll )
! 2[
Y; = Óo . (·
+ O¡[do + tl1 i - n+l)]·+ ~2 [f o + f1 (· n+
z- ~l) i - -
n = - \n -
u 2
n
we h ave X =
1P ( n--
i n- + l) ( n·(- 1)
P2 n - - -
2 2
178 LINEAR S'l 'A'J'ISTICAL l\IODELS POLYN Ol\UAL OR CURVILINEAR ll10DELS 179
An important p oin t to notice here is t hat t he coefficicnts of c5 0 and d The AOV is given in T able 8.10.
are the samc as thosc of t he poly no mia l when only a linear is fitted'.
that is, 80 = &0 ancl 81 = & 1 (where &0 and &1 are as in Art. 8.4.2)'. TABLE 8. 10 ANALY SJS 0 .F VAH I ANCE FOR QUADRATlC Por,YNOM I AL
T he reason that the computations are so easy is t hat P 1 (i _n; 1) All the entries in Table 8. 10 are t he same as in Table 8.8 except the
quadratic and the new error term, which is obtained by subtraction.
and P 2 ( i - -n +
2
- 2)
h ave been tabulatcd [1].
The estimates of 0 0 and 01 a re the same as thosc of a 0 a nd (/.¡, respec-
tively, in Art. 8.4. 2. If we want to estímate t he y 0 , y 1 , and y 2 in Eq.
Table 8.!) g ives t he poly nomials for a few valucs of n. They are (8.9), we use Eq. (8.13). We get
computed from thc guantities in Eq. (8.14). , <>2 .07
Y2 =- = - - = - .7!)
.00 .0!)
TADLE 8.!) 91 = 12.0
·Po = - 3!).6
n
: =~ ::! ~ ; 1 2 ! ::11::: 1 _: 1 2 1:
discuss pth-degrce or t hogonal polynom ials.
Let a pth-degree poly nomia l
+ f31xi + {32x¡ + · · · + {3 Px:' + e;
Y; = f3o (8. 15)
The values for P 1 are the same as those given in T a bl e 8 .7 . where t he xi a re equally spaced (x; = a + ih), be rcpresented by t he
8.4.5 Example. Suppose we want to fit a quadratic poly nomial polyno mial
to the data in Art. 8.4 .3. We get Y; = a:0 + a: P 1 1 (i - i) + r1..i.P2 (i - i) + · · · + a:J)P ,(i - i ) + e¡
1
P2 2 - 1 - 2 - 1 2 ¿p~ = 14
J:.>' = P t (.i - i
_) = p ' ( i· - n2
- +- = 1) aº' + <t11
(·
i -
n2
- + -1).
+ ª_21(i - n ~- )2 + · · · + ªu (i - n ~ 1)'
1
(8.17)
Coefficient of i 0 :
The equ ations a re
1 p (l _n + 1)
1 2
p
p
(i _ +2 1)
11
(8. 18)
..... ... . .... .. .. ... . ... . ... . . ... ... .. .. . . . . . ... ....
Coe.fficient of i,,:
Ifwc use the notation L.P ,P. to mean ;~
1
Pr ( 1,.
1
1 -
n~ 1)
+ 1::>• (i. -
+
n~ 1) ,
Y Ci\lº 1944 1945 ] 94-6 1!)4 7 1948 1949 J!l50 1 195 J 1 1952 I 1953 , 1!)54 , l 95,¡ •rotal 12 12,8 15.99
Rain- - - - -- - -¡ - ¡- -,- ,-
Hc<luctio n for mean 1 12,538.87
fa ll :rn.2 32.2 :15.1 34-.2 39. l 41.3 36. l 30. l
1
30.5 20. 1 24.8 ! J 2s. 2 Hcmninder fro m m ean l L 277. 12
Fro1\1 [1] we record t he linear, quadratic, cu bic, a nd qua rtic poly nomial Lincll r 1 7 1.55 7 1.5:3 3.48 > 2%
for n = 12. We s ha ll use the 2 pcr cent F point as t he index for adc. E rror fo1· linea r 10 205.57 20. 56
guate fit. The AOV is g iven in Tabl e 8.1 2.
Quadratic 1 103.96 103.96 9.21 < 2%
T An u ,; 8. 11 · PoLYN OMIALS F O lt R A r NJ.'ALr, DATA
Error for quad ratic 9 101.GL 11.29
- 11 - 9 - 7 - 5 - 3 - 1 ~ 5 !l 11 572
1
Problcms
8.1 It is known LhaL t ho follo wing data fit a c ubic modol. Set a 95 p er cent
conficlonce inton ·a l o n each of Lhe coofficion ts y 0 , y 1 , y 2 , ami y 3 .
8.2 Jn Lhe exam ple o f ArL. 8.:t2, íi t, n, ou uic n.nd qua r Lic poly nom ia.l by using
t he abbrcv iaLed Doolit,Lle 1110Lhocl, ancl inLerprnL Lho rnsul ts .
8.3 In Lho moclc l Y; = (J 0 + {J 1x , + {lzXi + e¡, fine! t he maximum -likolihood
esLimato o f t ho m nx imum o :· minimum value of E(y) (assumo i = l , 2, ... , n).
8.4 In a n ex porimen L on co1Tos io n of m e tal Lho fo llowi11g cla t,a were collected:
Yolt.ago applie d 1.5 2.0 2.5 1 3.0 3.i'i 4.0 4.5 5.0
bccau se of mcasurement errors; Y; a.nd xi are ob served , w h cre Y;"" Jf wc s um t ho n equatio 11s in (9.2), wo gct
Y; +e; a 11d X¡ = X; + d;. The e; are independont norm al variables
with mean O and varia nce a~. T he d; are simila rly distrib u ted with {J- 'L(y, - & - /iX,) L(x, - X,) =
va ria n ce a~. \1\fe shall attempt to find maximum-likelihood estimates •'.!
a,
+ a,
-2 0 (9.D)
1
of (J., {3, aJ, aml a~. Th c lik elihood func t ion is
JL.we s ubstit u to Eq . (!l.4) in to the first term of Eq. (9.9) , then (9.9)
beco mes
(!U)
'L(x, - X,) = o (9.10)
S ubstitu ting fo r e¡ a nd d;, we gct If we t ra nsfcr the socond term of (9.2) to t he right-hand side of the
equa lity sign, we got
(9.7)
- ~[L.(y¡ - (J. - {3.l()2 _!.... E( x; - X,)2] (0. 15)
2 a; ' Áa~
(9.8)
Wc now have n + 3 unk now n parameters X 1 , X 2 , .. . , Xn, (J., {3, a nd
190 LlNEAit S 'l'A'l'IS'l'TCAL J\lO DELS J\fO DEL 2: FU NC'l'IO J\ AT~ RELA'l'IONSHIPS 191
a~. T a king t he p a r t ia l d c ri rn.t ives \\" Ít h respcct to each of t hese n -L and whe re t he sig n u sed for t he first 1,crm is t lrn,t which will maximize
parameters, \\"C obta in ' 3
the Jikelihood function. If wc substi t ut,e in to (9. L7), we get
alog L (.111 - & - {JX,)/j X¡ - X, 1
/j(x¡ -
:;--K
<J.L t
= ·2
ar + ) ... .a~
= O l = l ,2, ... ,n (!l.lG)
a;, = 2 l:(y; - !í)[(y, - y ) -
·n
i )J (9.23)
á;
1
(!l. 19)
•2
(Jo = - _1-2 ["'( - ¡¡"'(X¡ -
I''-' -
.1.:}(!J¡ - '!i)J
n-
From (9. l fl) we gct \Ve have the following theorcm:
+ Theorem 9.1 If Y; , X¡, Y¡ , X ; satis fy the co ndi t ions in Scc. 0.1 a nd
if ;. = aj/a; is known , then t hc maximum-likelihood estimates of
a nd from (!) .16) a~, u., and f3 (t he divisor of a; is changed ) are
fi = ± ( u2 + 1Y + u
w hich girn I:(x; - X;) = O a nd, finally, a.= fj - Px
& -= fj - /j:f; (9.20) a~= n-::.
~[L(Y; - f/) 2 - ~'2:.(?J; - fi)(x ; - :i)]
Sol ving (0. 16) for X 1 g ives
These cst im ates a re consistent cstinrnLes of t heir rcspectirn
x = x, + J. flyt - ;,{J& parameters .
t 1 + ;.p2 (9.21)
From (9.18) " ·e get 9.2.3 Example. Suppose the following n d ucs ·of y ¡, X; are
recorded under t he assu mp t ions of t he prcccding article (assumc
f1 = L:(y¡ - &)X;
). = 1) :
I:(X ;)2
?J¡ 6.3 7.9 8.2 9.4 10.8 l.l .i3 12.9
Substituting (9.20) a nd (9.21 ) into this last equation, after some
simplification, g ives
X¡ 11.l 9.9 S.2 7.3 0. 1 G.O tJ-.2
f1 2
(},2:(y¡ - y)(x; - x)] + {J[L:(x; - x) 2 - J.L:(y¡ - Y) 2 ]
Weget l:,(x; - x) 2 = 34.337 ; L,(y; - [¡}2 = 3 1. 7 1 -~ ; !:,(!/ ; - fj)(:¡;¡ - x) =
- L:(y; - jj)(x; - i) = O - 32.311 ; n = .7; x = 7.543 ; '[j = 9.5 71 ; U = .O-l0;í8!l ; ¡i = - .060;
This is a qua dratic in f1 a nd, whe n solved, gives & = 16. s 1 ; a; = . 13 o.
9.2.4 Case 2 : Controlled-independent-va riable Model. Let
f1 = ± [ uz + 1] + ul (9.22)
thefunctional relations hip be V = u. + {JU, wh erc ca pi tal lc t,tcrs st a nd
for m athe matical quantities a nd lower-casc le ttcrs f'o r ra.nc.l om v a ri-
ables. \'Te assu me that V a nd U cannot be o bscrnx l b11t, tha t, \\"e can
where U = L(Y; - '[¡)2 - (l /í.)l:,(x; - x)2 observe x and y, whe re
22:(x; - x)(y; - y) y = V +e (9.24)
]!)2 T, l N 1.;A lt S'l'A'l'JS'l'ICA L J\fODELS i\lODEL 2: l !'U "CTION.-\.L RELA'l'IONSHIPS 193
where d and e are errors of measurcment. In case 1 wc assumed that which becomes
e and d were norma l a ncl indcpend cnt; s incc U is a mathematical vari-
a ble, t his would makc rl a nd x functionally rclatecl. For case 2, h ow.
cvcr , wc s ha ll assume that x is set cqna l to predetermined quantities 13ut t his does n ot fit the de fln it ion of m odel 1, sincc the error· term
xi a nd , h ence, t ha t X takes the role of a ma thematical variable. This e_ f3 1d + {3 2d 2 - 2{3 2 Xd d ep ends on X.
forces U to be a ra ndom varia ble; hcnce V a lso is a random variable. 9.2.5 Exampl e. Suppose an experimenter has a 1·csistor that
Vle sha ll rewri te (9 .2-l·) he wants to meas ure. H e assumes that t he la w E = RI holds, where
X=u-1-d (9.25) E is thc voltagc ofthe circuit a nd I is t he current rneasurcd in a mper es.
a nd assume tha.t e is independent of u, v, ancl d but that v, ii, a nd da.re lle cannot measure E and I exactly, but he ca n observe y = E + e a nd
fonct ionaUy relatcd. The p hysical meanin g of this assumption is this: x = J + ll. To use t he controlled-in<lepenclent-variable method, he
V-le must decide bcforehand wh at values x will take (we denote these decides he will bring his a mmeter to t h e following values : X = 10, 15,
20, 25, a nd 30, and rcad t he corresponding vo ltages y. H e observes a
val ues by X ) and thcn bring t he rccordi ng i nstxument t o those values
a ncl meas ure y. Accordingly, X is termcd a conlrolled variable. voltage for each of the a mpere read ings a nd get ::; a set of val ues:
\ Ve can now rewrit e v = r1. + {3ii
y- e = r1. + {3(X - d) O l' y = r1. + {3X + (e - {Jcl) (9.2G) --=---~ 15.0 ~ ~1~
Jf we let f = e- {3d, we gct y 3. l 4.8 7.2 8.4 1 9.7
y = rJ. + {JX + f
From t hese h e computes R = "Ly;XJL.Xl = .33; this cst.irnate of R
a n<l t his fits thc definiti on of model 1 in Chap. G. H ence t he theory in has ali the propcrt ies listed in t he consequcnt of Theorc m ü. L.
Chap. G app lies.
If we ha ve t he more general function al relationship
9.3 I nterval Esti matio n a n d Tests of Hypotheses
V = r1. + {3 U + {3 U + · · · -1- (3kU,,
1 1 2 2
(9.27) Thc o nly case we sha ll cl iscuss for in terval estirn atio n an ti Lcsts of
and if y = V+ e x; = U;+ di i = l , 2, ... , l.: hypotheses is the controlled -variable model, case 2. The rcsu lts of
we can control X; (i.e., set :i;; = X;) and gct Chap. G a re a pplicaLlc. It s hou ld be noti ccd that the varia ncc of y in
this case is· equal to a~ + (J2a~ . This appears in t he no nccntl'ality
v = r1. + f31Ui + · · · + fJkuk parametcr in tests of hy p othcses, and the po wer fun ction is affoctcd.
(9.28)
i = l , 2, ... 'le
Substituting, we get Problems
9.1 The fune t iona l re lat ions hip Y = a + {JX is known lo hold. Tlw rnat he ·
Y= r1. + f31X 1 + f32X2 + · · · + iJkXk + a (9.29) maLical yar iables Y n nd X ca nno L b e obscn ·ed , buL the follow ing value,; o f
where g is a r andom e rror snch that !/¡ a nd X¡ a re obscr ved , whe rc '!}; = Y, + e; und X; = X ; + d ;:
9.4 In an c xporimon t, d esig ncd Lo m ens ure t,hc d ens it y of gol<l, i,he following
v a lues w ere obtaincd ( w i::; wc ig ht, v is volu m c) :
-
w 19.0 44.0 G0.7 74.0 82.0 90.3 109.0 156.0
Th e fun c tional re la tio nship is lV = óV, whc ro 11' is weig ht in grnms and r is
v olnm e in cubic ccnt,im ct.ers . Howe vor , s in ce t.h cro are errors of m easurement
w a nd v were reco1·ded. Assuming ;, = 1, find t,he maximum -likelihood estimat~
of t he dens it y ó.
10
9.5 In P 1·ob. 9.4, a ssume t h at the re ·is n o erro r in m eas ur ing V, and find thc
m a ximum-lik e lihood cstima te of ó. Not,ico Lho difforence betwoen t,he 1,wo esti-
1natcs .
9 .6 Prove that a;
is g r:cat.er t han 0 1' e q ual to Z Cl'O in Theorom 9.1.
l\1odel 3: Regression Models
11
9.7 For the quanti t,ics in Art. 9. 2.2, p1·o vo t,ha.t L (x ,. - X;)X; = O.
i= I
9.8 For t.hc quo nt.it.ies in Ar t . 9.2.2, provc t,hn.t, L.(y,. - ci - ífx;)X ; = O.
9 .9 F o r t he qua n t,it,ic::; in A r t . 9.2. 2, pi·o vo t hnt Eq. (9.2:3) ca n be \\Tittc n 10.1 lntroduction
l l:((y¡ - ?/) - {j(X; - x )]2
Q-2 In this chapter we s ha ll d is cuss wh at is often called nmltiple regres-
• = 2n 1 + ;., p
2
sion. If y, x 1 , x 2 , • • • , X1.: a re ra ndo m variables whose joi11 L<lis Lri bu t ion
Furthe r Reading
is gi\·e n by f(y ,x 1 ,x 2 , . . • ,x,,) and if t hc expcctcd value of y in t he condi-
t ional dist ribution f(y 1 x 1 = X 1 , x 2 = X 2 , . . . , X1.: = X 1.: ) is equal to
1 O . L. D avies: " D csig n and A n a lysis of Indus t r ial E x pcrime nts ," Olive r and {J 0 + {J 1 X 1 + /3 2 X 2 + · · · + f31.: X 1.:, wc shall say tha t therc is a linear
Boye! , Ltd., L o nd o n , 1954.
regression of y on x 1 , x 2 , •• • , X1.: a nd ca ll t his mod el 3. \\'e can write
2 J. B erkson : Aro Thc rc Two R cgrnssio ns ?, J. Am. S talist. A ssoc., vol. 45, pp.
164- 180, 1950.
3 E. S . K eeping : Not,e on Wald 's l\lo thocl of Fit,ting a S tra ig ht, Lino when
Bot h Variables A ro . ºubjcct t o Error, B iom etrics, ,-ol. 12, pp. 445-4 48, 1956.
where the {3¡ are unknow n parameters, wbere X ; are par ticu lar (known)
4 M. S . Bart.lett.: F it,t,i ng n. Straig h t L ino whc n B oth Vmiables Are S u bjcct to
Errnr, B iomclrics, vol. 5, pp. 207- 212, 1!l4!l. values of thc random variables :i·;, a nd wherc e is a ra ndom Ya riable with
5 E . L. Scot t: Not o on Cons is te n (, E s timates of t,he L inear Strnctu ra l R ela t ion mean zero.
bctween Two Varia bles, A nn. Math. S tatist., vol. 2 1, p p . 284-288, 195 0. T his model is simila r to model J , t he d ifference being that in model 1
6 T. C. K oopma ns a nd O. R eie rs o l: Tho I d en t ificaLion of Strnct.ura l Cha rac- the X ,. are mathematical variables and not pa rticula r Yal ucs of random
teristics, A nn. Jlfalh. S tatist., , ·o l. 2 7, pp. 105- 18 1, 195 6.
variables. T h is difference may seem s light at first , but it is quite
7 l\L G . K e nda ll : R cg1·essio n, SLru ct,uro a nd Functio na l R e la Lionship, B io-
m etrika, pa rts I a nd II, ,-o l. 39, pp. 96- 108, 1952. important when we consider its consequences. For exa mple, in model
8 H.. C. Geary : Non L incm· F unc Liono l R e lnLions hip be Lween 'l'wo Var iables 1 there are np Ya lues of X ; involvecl. The probabili t y of t he confiden ce
whe n Onc Variab le Is Cont.ro lle d , J. Am. 8 /alisl. A ssoc., Y O L 48, pp. 94-103, statements a nd tes ts of hypotheses was ha.sed on the fact that repeated
195 3. values of y¡ were to be drawn from these same np v alues of X ;. ' i\'e
9 A . \ \'a ld: The F it,t,ing o f St1·aig h L Lines if BoLh Varia bles Are S ubj ect t o
shall n ow extend Lhis i11ference to oLhcr X ; by using model 3. Also,
Erro r , A nn. 11/ath. S tatisl. , ,-o l. 11 , p p. 284.- :300, 1940- 194 1.
10 D. \ -. Lindlcy : E s t,imaLion of a. Funct,io na l R e laLio nship, Biom etrika, YOL correlation was not defin ed for model 1 (correlation is not dcfin ed for
40, p p. 47-49, 195:3. mathematical variables), but it will be d efi ned a nd usecl for t he random
variables in model 3.
In t he functio11al form s = vt , fo r exa.mple , it is difficu lt to i magine
the t ime t p laying t he part of a random variable . T his wo uld be
possible, but it secms as if we s hou Id wa n t to calcula te t hc d is tance s for
195
l 9G LIN l~ A Ji S 'l'A'J'íS'l 'lCAL l\IODELS
lllODE L 3: RECrliESSJON J\IODEL S 197
prcselcctcd Yal ncs of l. Also, it is clear t ha t , if t is a ra.nd o m variable u 2•
t hen s m ust also be a random varia ble, a nd t he correlation between th~ pa.i·arnetcrs {3 0 , {3 1 , a nd
L
'J'his m ea.ns t ha t t hc j oint distribution of
two guantiLies is equal to uni ty. On t he othc r h and , consider the !/ 3.11d •x can be w ritte11
following examp les, which seem to be d ifferent in charactcr from the
) _ h(x) _ l_ e- Ct/2a'JCv - /1 0 - P 1 :i:}' -oo < y<oo; -oo < x <oo
a bove : f(y, x - . u~
which we sha ll \\Ti te as r = {3 0 + {3 1X. So we shall let =L: [L: h(x,y)Á·(y 1 x) dy Jg(x ) clx .
a nd (10.Z)
Thc term in t hc brackets is Eu 1.,[h(:i:,y)], which is a fun ction of x
o nl y . S ubstitu t ing gives
sin ce
R _ ª 12 E [h(x,y)J = L : {Eu¡.,[h(x,y)J}g(3:) ch = E.,{Eui.Jh(x,y)]}
F1 - - and
a ll
and the lemm a is p ro,·ed.
Sin ce µ 1 , ¡.í 2 , c1 11, c1 22 , ancl a1 2 a re 1rntxim11m -likelihood estimators, t hey
a re co nsis ten t a nd efficicnt . It can also be show:n t hat t hey forma set
No w, E({J.) =E [E ({J. )J = E
1 ., t.<lo: l z
{E [í:(y¡ }:(x¡
t.<lx
- y)(x;_ .i)2- x)]}
of s ufficient estimators a nd they a re compl ete a nd unbiased. \ Ve shall
sum up t his inforniat ion in the foll o \\·ing t heorem. ·
• Theorem 10.1 lf (x;, y ;) fo r 1· = 1, 2, . .. , n reprcsents a random
sample fro m t he bi variate n orma l díst ri bution, th e maxi mum - Bnt
likelihood cstim atcs (correctcd fo r bias) of µ 1 , µ 2 , cr w a 22 , a 12 giYcn
in (10.l ) haYc t..hc fo llowing proper t ies :
and
( L) Cons ísLcncy
(2) Efficiency
(3) Completencss so
(4) i\Iinim um n1 rian ce unbia scd
(5) S ufficiency
Also,
(6) That (t 1 be indepc nden t of é1 11 , c1 22 , c1 12
200 LTNEA.l't S'l'A'J 'IS'I'ICAL MOD"ELS l\IODEL 3: mmRESSlON MODELS 201
So {j 0 a nd p 1 are un biascd csiimato rs of (3 0 a nd (3 1 , respccti,·c ly, but Lhey
a re not norm a ll y <listributcd as when t hc X; a re fi xed. \ Ve ha.,·e !f wo Jet
establish ed the following theorem.
+ T h eorem 10.2 Thc ma.ximum -likelihood cstimates of /3 0 a.nd {3 1 then
givcn i n (10.2) ha ve t.he following prope rt.ies:
( .1 ) Consistency
(2) Efficien cy
(3) l\Iinirnum varia n ce unbiased
10.2.3 T h e Wis h art Distribution. We s h all now find t he joint T hc integral ex ists, sin ce we can choose t ; small e noug h so that
distribution oí t he quan t ities (n - l )all, (n - 1)& 12, (n - l )a 22 . v- 1 - 2Tis positived cfinite. Thcrefore, ifwele t V- 1 - 2T = C,
F irst " ·e s ha ll pro,·e t he fo llowing t heorem. we get
f(bwb12 •bd = ~ 2n¡v¡1112,.112r(~) r(n; 1) ( 10.3) Now, from t he distribuiion of U; in t h e theorem, we s ha ll find t he
moment-gc11erating fun ctio n of L,x~ L.x¡y¡, L,y[. L et us desig nate
lo if Bis not positirn dcfinitc iL by
b 11 b 12)
where 611 ;:;;::: O (
B = b21 b22
Proof: It is clear that 'Lu,u; = B. We s h a ll find t hc joint mom ent-
genera ting function of L-:i?;, "i.x¡y¡, Lyf, s ho" · t h at it is e~ n a~ Lo ~he
= f · · · f ""
-ao
a:l
-oo
e
lr(T!:U·U')
1
• g(x1 ,x2 , ... ,y,.) dx1 dx2
·
• • • dy ..
mome nt-gencrating function of b11 , b 1 2 , b 22 , \\·here t he ch stn but10n By definition of a ra ndom sample of vectors,
of b 11 , b 12 , b 22 is given by (10.3), and u se Theore m 2. 1. It can be
s hown by direct integration t hat
f (b 1 1 ' b12 1 . . . J b ,,,, ) = 7Th>C11- Il;¿ n11/21 Vl "' 2 fr f'(n+ iJ - i) \Ve ha ve s h0\\"11 t hat, if X; and Y; are o bscrvatio ns fro rn N( µ ,V), thcn,
I
l o
i= 1
if B is p ositivc d cfi11ilc
if n is not pos itive <lcfinilc
lctting
v = (~11 ~12)
ª21 ª22
• Theo rem 10.6 If Y ; is a rand om vector from the distribution --:;- ---;.;-¡~ ~--;;-¡-;;-¡--;;-¡-;;¡--;;-¡--;;-1~
N( µ ,V), t hen
The following qua.ntities " ·ere calculated:
A =I
n
(Y¡ - Y)( Y¡ - Y)' .
ª
i=J [¡ = 1í2 = 12.72 fi =
/JO • - -;;-"
/12 /11 = - :.'>() · íi1O
1? '
O"¡ ¡
is d istributed as IV(p, n - 1, V), wherc Y= ( l/n) "LY¡ for n :;;:::
p + l. x= {t 1 = 2.48 Ó-12 = 1.10 P1= ~12 = 13.4-o
Proof: It is cle.?-r that ali
11 •2
I (Y¡ - Y) = o p•2 = - 0" 12- = .90 Ó"n = .082
i=l Ó"n Ó-22
ancl, hence, that t he Y; - Y ar e not linear ly inclepcndent. Let The regression cquati_o n is Y = - 20.GO + 13.40X.
1
204: LINEAR S'rA'fISTlCA L i\lODELS MODEL 3: :i<.EGRESSION J\lODELS 205
10.2.5 Interval Estimation. ln t he biYariate normal distrj. wheref{l,x 1 , • . • ,x 11 ) is thc joint density of t a nd t be X ;. Hence, thc
bution it mig ht be desirable to be ab le to set confidencc li mits on the robabili ty statement in Eq. (G.10) is t rue when the X¡ a re ra ndom
parametcrs µ 1 , µ 2 , (3 0 , and {3 1 • Let z; = ax¡ + by;; thcn , by Theorcn¡ ~a.ria.bles, i .e., when \\·e sample from a mu ltivariate inst ead of from a
3.22 , the z; a re dis tributed indep cndently N(aµ 1 + bµ 2 , a 2 ) . Therc. conditional distri_bution. T he joint distribution , howcver, must be
fore, Studcnt's distribution can be usecl to set confidence li mits 011 such t hat the conditional is normal and satisfies Definition 6.1.
aµ 1 + bµ 2 • If we Jet a = l and b = O, we get confidence limits on µ We sclect ed _o nly onc ty pe of confidence-interval statcment to
and similarly for µ 2 • ¡, dcmo11strate the result ; however, the same result holds for any
The situation looks more complex if wc desire to set confidence limits confidence statement in Art. 6.3. l.
on {3 0 or {3 1 . However, we sha ll s how t ha t the theory of A rt. 6.3 .1 can be The expected width of the confidence interval is different in the two
used. To demonstrate this, s upposc we considcr thc general-linear. situat ions: (1) when the X ; are fixed and (2) when the X; are random.
hypothesis modcl of Definition 6.1 , ancl in particular let us considcr 10.2.6 Example. vVe s hall set a 95 per cent confidence interval
setting a confidence interval on r'{3, as in Eq. (6.10}. This probability 011 p1 for the example of Art. 10.2.5. R egarding these data a s com iug
statcment as we derived it is val id for repeated random samples of Y; [rom model 1, t he X¡ fixed , we can use Eq. (6.9} and obtain the 95 per
from the samc set of X ;. However, if wc change th e X; values and cent confidcnce interval
draw a ra ndom sample ofy; from the new X;, the probabili ty stat ement 9.8 ~ f31~ 17 .l
is still true. H cnce it is true for any set of X¡, sincc the confidence
Accor<ling to the theo ry of Art. 10.2.4 , t his statement mea ns t hat, ifwe
coefficicnt <loes not depend on the X;. Thcrefo re, wc can also select the
tn.ke repeated sets of y mcas ure ments from these xi and calcul aLe 95
X ¡ at_rando m from any distribution, so long as thc conditional d istri-
pcr cent confidence intervals for each set, then , on t he average, 95 per
bution of y given x is normal and satisfics Defi11ition 6.1.
cent of all intervals wi ll conta in the true unknown parameter {3 1 . Th is
To demonstrate further, the quantity
type of infercnce does not scem to be too useful. Howcver, by using
r'~ - r '{3 the theory in the preccding a rticle, we can select repeated sets of leaves
u = -:===== (thc X¡ vn.lues may changc from set to set) and set a confidence interval
J á 2 r ' S- 1r
on fi 1 for each set. \Ve then know that U5 per cent of all intervals
is distributed as Student's t whe n thc X¡ are fixed. Let this t distri- contain t he true unkno\\'n para meter {3 1 .
bu t ion be m·itien g(t 1 x 1 , . . . ,xn)· Tite prol>a.bility statemcnt of Eq. 10.2. 7 Tests of Hypotheses. Thcre are various hypothcses
(6. 10} can be written that are intei:esting to test in thc bivariate normal distribution . Those
- i - (/. = P[ - e,.
- -
~
r'~ - r '{3 ~ t. •
(á2 r ' S- 1 r )' -1-
, J wc shal 1 discuss are:
l. That the vector of mca ns µ.i s equal to a known vector µ. *
la10 2. That thc r egression coeffici ent fJ 1 is equal to a. known const ant f3i ;
= P(-t1112 ~ iL ~ l 1112 ) =
f- 11110 g(t 1 x 1, . . . ,Xn ) dl and similarly for (3 0
There are other t ests that a re important (on the ele ments of V for
examplc), but we shall not discuss them.
If h(x 1 , ... ,x 11 ) is t he density fun ction of thc X ;, then
Tite hypothesis H 0 : µ. = µ. *. To test H 0 : µ. = µ. *, where µ. * is a
J>( - la/2 ~ iL ~ la¡2) = 1 - (/. kn own vector, we shall use t hc followi ng t heorerns, which we sha ll give
= 1{ "° · · · f"°
without proof.
(l - (J.)h(x1 , ... ,x,,) dx 1 · • · dxn
.. -o:> -c:o + Theorem 10.7 If the p x 1 vector Z is distributed N( µ.,V), if t he
= f eo · · ·f eo [f la1o g(t 1x 1 , .. . ,x 11 )/z{x 1, .. . ,x 11 ) dt Jdx 1
p x p matrix A is distributed Jl'(p,n,V), and if the elcments of Z
are independent of the ele ments of A , the quantity
- co -<X> - 12 1~
· · · dx 0 n = Z' A- 1 z 11 - p 1- l
co f eo f la1o p
= f ·· · f (t ,x1 , ••• ,x11 ) dt dx 1 • • • dx ,.
is d istributed F ' (p , n - p + 1, .A), where ;, = t µ. 'V- 1 µ..
-oo - oo - la¡ ~
206 LINEAR STATISTICAL MODELS MODEL 3: REGRESSION MOD:~LS 207
The proof of this theorem was given first by Hotelling. The·:! between y and X is not defined. In the model of Sec. 10.2, however,
quantity [np/(n - p + l}]u is generally referred to as Hotelling' 8 p2. :} correlation does have meaning, as it did in the multivariate normal of
Cbap. 3.
+ Theorem 10.8 If the n vectors Y¡ (i = 1, 2, ... , n) are random ~~ As we saw in Theorem 3.14, the square of the correlation coefficient
samples from the p-variate N{µ,V) and if .~
gives the percentage decrease in the variance of one random variable
- S = -1- k~ (Y¡ - -Y){Y¡ - Y)'
- due to use of information available on a related variable. Also, if the
n - 1 i=l correlation between two random normal variables is zero, the two
variables are statistically independent. However, if the random
then S and Y are independent. variables are not jointly normal, then zero correlation does not
This theorem is an extension of the independence of fi and s 2 in the ' necessarily imply independence. In fact, the correlation coefficient
univariate case. . between two random variables can be zero and they can still be func-
We shall now state the theorem that will be used to test the hypothesis tionally related. For example, suppose we ha ve four points (x, y), each
fL = µ*. with probability !, that is, (1, 1), (2, 4), (-1, 1), and (-2, 4). Now,
+ Theorem 10.9 If Yi (i = 1, 2, ... , n) is a set of random vectors ; E(x) = (I)(!) + (2)(!) + + (-2)(!) =O
(-1)(!)
from the p-variate N(µ,V), the quantity and E(xy) = (l)(!) + (8)(!) + (-1)(!) + (-8)(!) = O
w h ere ex. = µ
<J,, µ:i:
10.3.1 Simple Correlation. A simple correlation coeffi.cient is 11
__ and
defined as a certain operation on two random variables. Hence, in the ; <1z
general-linear-hypothesis model given in Definition 6.1, the correlation · A similar proof holds if P:x:11 = -1.
208 LlNEAR STATISTICAL l\IODELS MODEL 3: RF.GRESS LON i\IO DELS :W9
For example, if x ancl y are ra.nd om variables a nd tbe co rrela tion + Theorem 10. 12 If p = O in t hc bi variate norma l, t hen
bet\\·ee n y and f(x) is equal to uniLy, t he n y = <J. + (Jf( x). In this
respect co rrelation coefficients can be useful in obtaining the relation- r(n') 1)(1 - p2)l(•H)
observed , and it might be extrem ely da ngerous to draw in ferences as if Proof: The proof is immediatc from Theorem 10. l l.
both varia bles were random. Howe,·er, when t he two variables are
+ Theorem 10.13 If p =O, tite n
indeed rando m, t he square of t he correlation coe fficicnt is an estimatc
of the perccntage reduction in t he variancc o[ ono of t he variables by
using the othor.
r (n- 1) (l - p:!) l(n-4 )
~ O ~ p' ~ 1
2
w ·e shall statc so me theorems (somo with out proof) on the corrclation f<P' ;O) ( ) ( )
r n ; 2 r ~ (p2)!
coeffici ent . In T heore ms 10. 11to 10.21 \rn shall assume that U¡= (X¡)
Y· + Theorem 10.14 If p = O, E(¡)2) = I /(n - 1).
for i = 1, 2, ... , n is a ra ndom sample from thc bi,·ariate norm; l
d is trib ution with correlation cocfficient cqual top a ncl t hat • Theorem 10.15 lf p = O, t he qua nti ty
l =
pjn,_ _
- 2
.,,) 1 - p2
is distl'ibuted as Stude n t's l with n - 2 <log rees o( freedom.
• Theor em 10.11 The dens ity function of pis
To t est t he hy pothesis H 0 : p = O wc can use S t u<lcnL's distribu t io n,
f (p; p) = (n - 2)(1
- p
2)~(n-1}
(1 - p2)~(n-4)
f"' dw but to test Lhc hypothesis H 0 : p = p* (p* =!= O) wc m11st know thc
distri bution o ( p whcn p =/= O. lf n ~ 25 , David's La.ble:; can be usccl.
7T o (cosh w - pp)n-I
lf n ~ ~ 5, inLerpolation in David 's tables can be uscd, but wit h so rne
-1 ~ p ~ 1 s:icrifi ce in accurncy.
Thc proof of t his theo rem will be omitted. The theorem can be It may be desirnble to test t he hy pot hesis t hat two or lll O l'C corrclation
obtaincd b y using Theore m 10.3 wit h p = 2, substituting p fo1· á 12/ cocffi cients a re equal. T hi s t;y pe of prob lcm can be ha nd lecl quite
casily by using ~L transformation dnc to R. A. Fis hcr. This is t hc
·•/ a11á 22 , a nd in teg rating o ut á 11 and &22 . substance of t he next few theorems .
The density fon ction f(p) is difficult to work with. T he parameters
in f(p ) are p , which is unlmown, and n , which is known. To test Fisher h as sho wn that, if we m akc t he transformation
h y potheses a bo ut p a nd set confidcnce intervals on p, it is d esirable to 1 +
Z = ~ log. - ~
A
h av e = a rcta nh p
1 -p
Po
'"
m =
Lf(n; - 3)/(n; - 1)1
-'-;......;;.~_;..'-'-"'-~~
which is not signi íiean t at the 5 per cent level. Hence \\"C eannot reject
Z(n; - 3) H 0 • The power functio n for p = p* is
T he improvcd es t i mate givc n iu Theo re rn l O. l Gis used to obt,ain t hc
pooled estimatc. {J(p*) = fP:z.
f(p ;p*) dp
2 12 L I NE AR STA'rlST lCA L i\IODELS MODBL 3: REGRESSTOX MODELS
a nd
- 1
r 111
f (p;.5) dp = .025
Notice t hat a n er¡u a l two-tail t est has been uscd. Using this, wc get
which g ivcs
"
.4 O1 ~ -
l .!)(j
fT77 ~ arctan
vl7
- .0745
.
~ ar clia 1~h
1
1 p ~ .401:2
p ~ .87G!J
l.!l(i
+~
,! L7
n; - 3 21 25 p = .205.
214 LINEAR S1'ATISTICA L i\CODE LS l\I O D~L 3: REGRESSlON '.'TODELS 215
10.3.3 Pnrtial Co rrel a tion. To gafo a better jnsig ht into hypot hes.es a bo11 t t he pa~·t_i al correlatio n coe~ci en ts . Lct us pa rt ition
r clations hips a.mo ng rando m variables in a mu ltivn.riate normal popula- the 11 rnt nx B as ,,.e part1t10ned V. We obta m
tion , it may be desirable to stratify the popubtion into s ubpopulations
in whi ch o ne o r more of the variables is fixed and then determine the B =(Bu B12\
correlat.ions n. mong the other variables. B21 B2j
This was t he case in observations about how average yea rly tempera- Define Bll. 2 to be (Bu - B 12B 2l B 21), and !et b;n denote t he ~jth
ture x 1 a nd average yearl y ra infall x 2 affected the y ield y of a certain clcmen t of B u.2·
agricu ltura! crop. The simple correlations recorded were
+ T h eor em 10.22 Under the conditions expla ined above, the maxi -
Pvi = .90 Pu2 = - .45 PJ2 = - .55 mnm-likeJihood estímate of t he part ia l corrclation coeffi cient
It is su rp risi ng that t here is a negative correlatio n between yield, .and .. 2 ¡,1b11.-
.. + 1 , ... ,P = b lJ.
Pij•q+ J... .,'J) is p" 1)''1 .. • b'1.2·
..
rainfall ( - .45), since rainfall is expected to incrcase the yield. F<tiftli~r'J Proof : Th is t heorem foll ows immediately fro m Lhe invaria nt prop-
investign.tion reveals t ha t temperature and rainfall are negatively erLy of t hc m aximum-likelihood estimates.
correlated ( - .55), as they sh ould be. Thc con clation b etwecn rainfall + Theorem 10.23 Under the conditions abovc, Lhe eleme nts of the
and y ield in t he subpopu lation derived by holding tempc rn.t ure con- 11mt rix B 11 .2 a re dist ri b ut ed as W[p , (n - 1) -- (p - r¡), V i 1.2J.
stant clarificd the p icture greatly. T his implied t hat, for constant We sha ll not provc t his theorem , bu t a proofcan be fo und in [U]. \Ve
t e mperature, rainfall tends to increase y icld , as cxpectecl. are no w in a p osit io n to state a very important Lheorcm.
In Cha p . 3 th e partial eorrelation coefficient was defined as t hc simple
correlation between two ra ndom varia bles '~·l10n certa in other rando111 + Theorem 10.24 If t he p X 1 vector Y is p a,rtitioncd into (y(I >) ·
y 121
varia bles a re held . constan t. Let t he p X 1 vector Y be distributed
.i.V(µ ,V). ] nrther, let us partition Y into a q X l vector Y<ll a nd a wh ere Y(I J is a q X 1 vector, and i f Y is distrib11ted N( µ ,V), then
(p - q) x 1 vector Y<2 1 such t hat a li th e t heorems 10.11 to J0. 2 1 are va.lid for t hc d isLribut io n of
P ii·N l ,.... v if n is replacecl by n - p + q a nd p is re placccl by
y=(y<O)
y c2¡
P;;·q+ 1. .. .,11·
10.3.4 Example . It is assumed that t hc 4 x 1 vector Y is
From Theore m 3. 10 we know t hat t he condit ional distribution of y w distributed .i.V( µ ,V ). On the basis of 24 observations t.hc following
given Y< 2> is n orma l wit h m ean U 1 + V 12V221(Y 121 - U 2) a nd with co- nw.trix was co mpuLed :
vai·iance ma Lrix V 11 - V 12V2z1 V 2 1 . \iVc sl111ll üse thc symbol V 11 . 2 to
-lü)
10 !) - 1
(-!
denote V 11 - V 12V2z1V 21 . The part ia.I corrclation between Y; and !) 20 - 3 - Hi
Y;[Y; and Y; are in Y!l>] hold ing Yo+i • Yo+ 2 , ••• , y,, constant is equal to S=
- 3 5 3
<1¡¡.q+J ... . v
- Hi - rn 3 27
lt is desired
where a iJ·o+ • •.. .,,, is the ijth ele mcnt in the covt~rian ce matrix V 11 . 2 • l. To find p 12 . 3 4
Suppose a random sample of n vectors Y 1 , Y 2 , • •. , Y ,, is taken :!. To test. t he hypothesis p 12 . 34 = O at t he 5 pe r cent· leve! agn.inst
fro m the norma l distributio n dcscribed a.hove. J t was p roved in pre- t hc a lternati,·e p 1 ~. 3 ~ =I= O
ceding sections t hat the ele ments of the matrix :t To set !)5 pcr cent co nfidence limits on Pu·H
n
l. We fi ne! S JJ ·Z Lo be equal to
(n - l )S = 2: (Y; - Y}lY¡ - Y)' = B
i=l
1 ( 49 - 35)
are d ist ri butcd as W( p, n - l , V). On thc basis of t his fact t hco rems
12G -35 1,285
on simple conelation coefficients were deri ved. \iVe sha ll now state
w it hou t proof a simila r theorem t hat wil l enablc us to es t i mate a nd test and fro m this we get Pi2 · 3 .1 - .139.
216 LINEAR S'l'ATISTICAL l\IODELS l\lODEL 3: REGRESSION MODELS 217
2. Using David's tables with paramet.ern - (p - q) = 24 - (4 - :?) parameters "='{:J., or <J. ~rom this it ~ollows that the condition~l distri-
= 22, we find that the two-tailed critical region is -1.00 ~ x ~ - .42a bution of y g1ven x = X 1s normal w1th mean a + {JX and var1ance a2 ;
and .426 ~ x ~ 1.00. Since p12 • 34 = -.139 is not in this critica[ that is,
region, we do not reject H 0 •
3. The 95 per cent confidence is, by Fisher's transformations, - .53 ~ f (y 1 x = X) = _I_ e-<If2o2 )(11-a:-PX>'I.
P12·34 ~ .30. -
a.¡:x;.
10.3.5 Multiple Correlation. The conditional distribution of t0.4.2 Point Estimation. If n pairs of sample values are selected
y 1 given y 2 , y 3 , ••• , yP is defined in Chap. 3. If the p x 1 vect01· Y is at random from the distribution defined in the previous article, the
distributed N(µ,V), the conditional distribution of y 1 given y 2 , ••• , :Y,, }ikelihood function is
. e-0/2a11)'E(1"-a:-p:z:,)2
isdistributed normally with meanµ 1 +V 12V2l(Y 2 - U 2 )and varianco L = J(y1 ,x¡,y2 ,x2 , ••• ,yn,xn) = li(x1 )h(x2 ) • • • li(xn) - - - - - - -
V 11 - V 12VilV 21 • If we let z = µ 1 + V 12V2;1(Y 2 - U 2 ), t.hen the ( 27T0'2) n/2
multiple correlation coefficient is defined as the simple correlation To find the maximum-likelihood estimates of 0t, fi, and u2, the deriva-
between y 1 and z. The importance of this coefficient is explained in tives of log L with respect to the three parameters are set equal to zero.
Chap. 3. It is defined as This gives (Ó'2 is corrected for bias)
... _ ,. . )1
0'11.23 · · · P -
Pi- (1 - A
t hese measuremen t errors are large enough to be of importance or can and gct
:E(u; - ii:)(v; - v)
b e ig nored. and
A
Y1 = IX¡ =
A
ai(l - ui~
U~2) = u2 Pi .04 .05 .81 .72
So it follows that, if Po + p1X is used to represent the per cent protein 100 120 90 95
of a quantity of milk when X is the per cent butterfat of the same
quantity, then, about 95 per cent of the time, the true value y will
(a) Use Fisher's transformation to test p 1 =O against the alternative p >o.
be within 2cr units of {3 0 + {J 1 X. Then, since a 2 ~ ~' the regression (b) Use Studcnt's t to test p 1 = O against the alternative p > o.
1
estímate is generally better than using µ 2 • 10.9 . In Prob. 10.8, use :F'isher's transformation to test p~ = Pz against tho
Since x is never known, the quantity {3 0 + {J 1 X cannot be used even alternattve p1 p2 •*
if {3 0 and {3 1 are given. What is needed is a regression function based 10.10 (a) In Prob. 10.8, use Fisher's transformation to test the hypothesis
p4 = .80 against the alternative p 4 =fa .80.
on ii, the obsm·ved butterfat content. The regression equation Y" =
(b) Compute the power of the test for p = .70, p = . 75, p = .85.
cx 0 + <Y. 1 U, bai:;edon thedistributionofuandy, can beused. However, 10.ll In Prob. 10.8, use Fisher's transformation to test Pa = p4 = .75.
cx 0 and ix 1 are unknowns. So, as explained in the preceding article, the 10.12 In Prob. 10.8, use Fisher's transfo1mation to set a 95 per cent confidence
regression of u on v can be used to obtain an unbiased estímate of o: 0 interval on p3 •
and (X 1 , and then f" = &0 + oc 1 U can be used as the prediction equ&tion. 10.13 In Prob. 10.8, use Fisher's transformation to set a 95 per cent confidence
interval on p 4 •
Problems
10.14 n:
p 3 = p 4 , use the data in Prob. 10.8 to find a combinc~d cstimatc>.
10.15 Thc 3 X 1 vector Y is dist.ributcd .V(p.,V). On thc basis of lOO
10.l If (xi, Y;), wherc i = l, 2, ... , n, is an n-sample from the bivariate observations the following matrix was ·computA.>d:
normal distribution defined in Art. 10.2.2, show:
(a) ¡1 1 = x is an unbia.C)ed estimate of µ 1 .
(b) ¡1 2 = 'fj is an unbiased estimate of µ 2 •
96 67
-10)
-~:
1
48 -8
(e) - -
n - 1
:I:(x; - x) 2 is an unbiased estímate of 0'11 • s = -8
(
20
~(Y;
1 Find Pt2·3 and P13·2·
(d) - - - y) 2 is an unbiased estímate of a 22 •
n-l 10.16 In Prob. 10.15, test thc hypothesis p 12•3 = O against the alternativa
l P12. 3 -=;é O at the 5 per cent level.
(e) -- ~(xi - x)(yi - 'fi) is an unbiased estímate of a 12• 10.17 In Prob. 10.15, set a 95 per cent confidence interval on p13•2'
n - 1 .
222 Lt N"EAH. S'l 'A'l'I S'l 'I CAL MODE LS
Furthe r R eading
1 A . :!\ [. l\lood : ··rn troduutio n Lo Lh e Thcory o f St.atis tic::<," i\IcGraw-lii U
J3ook Compan>·· Tnc .. ~ew York, Hl50.
2 P. C.:. Hoel : ' ·In t roduct.ion Lo í\IaLhem at.ical S t at is t ics," J ol m \r iley & Sons
Inc., X cw Yod ;:. l !:l5ü. '
3 ~ [. G. K endall : " The Ad,·anced-Tl'heo ry oí. 'tatisLics," ,·ols . I, II, Cha rles
C:riffin & Co. L td. , L ondon, 1946.
4
5
0. \ 'V. i::inedccor: '"Statistical l\letho d s, .. Iowa Stat.e Collcge P ress, Ames,
lowa, 1958.
H. L. Ande rson a nd T. A. Banc roft: "Stat is Lica l Theory in Research,"
11
McGt·a\\·-Hill Book Company, In c., New Yol'I<, 1952.
ó T . \ V. A nderson : " A n In Lroductio n to Mu lt in tr iate StfitisLical An a lysis,"
7
.Jolin \Viley & Sons, I nc., New York, 195í.
~- :-;, \\'ilks : " Mat.liem atical SLatis t ics," Princeton Univei·s ity P ress,
Model 4,: Experiinental Design JVIodels
P 1·i111:cl o n, N .J., 1 94:~ .
8 C. H . Rao: "Ad v a nccd fltaListical M ct,h o d s in. B iom et 1·ic Research ," J ohn
\\'il<>y & S ons, In c., New York, U:l52 .
9 F . :": . Dav id : "Tables o f t he Correlation Coef'fic ie nL," Biome t rika O ffice,
Uni vet'Rity College, L o ndon , 1938 . 11.1 Introduc tion
10 H.. A. F ish er: F requcncy Distribut ion of Lh e Values o f the Correhition
(\¡cffl c ien t in Sam ples from an I n d c fin il e ly Largc Popula tio n, Biometrika, In Ar t. G.3.4 it was stated t h at model 4 is defined as
\ 'O I. 10, pp. 5Üí-52 l , 19 15.
11 H. A. Fis het·: On t he Probab le E n ·o1· of a. Coeffic icnt o f Correlation Deducccl Y = f31X1 + f32X2 + · · · + /31;Xk +e
fro m a ~ma l! i::ia.mplc, 1llctron, vol. 1, p a rt, 4, pp. 3- 32, 192 1.
12 H. A. F isher : T he Distrib11 t io n o í t,hc Par t.ia l Corrclatio n Coefficient,
where y and e a re random variables, /3; a re unknown paramet ers, a nd
J fi.ll'un, \"O I. 3, pp. a29- 332, HJ2 l. the X; take o nly t he values O a nd l. T o expla.in t his in more detail ,
13 H.. A. F ishe r : T lw Ucnera l :-iampling Dist1·ibut.io n of t hc Mult iple Cor rclation Jet us consider two examp les.
Coefficient, P roi;. Hoy . Soc. Lo1ulrm, ser. A, vol. 121, pp. 654-673 , 1928. Suppose a manufactu re r of light bu lbs, looking for \\·ays to increase
t be life of his p roduct, has developed two che mical coatings to p ut on
t he fila ments of t he bu lbs. If no coating is put on t he fil a mcnt, t he
bulbs will Jast a n average of µ (unk nown) hr. T he manufactu rer
assumes t hat, if he p uts coating 1 o n the fil ame nt, it will increase t he
fila ment life by T 1 (unknown) hr and t hat , if he uses coating 2, it will
increasc the life by T 2 hr. Now, he would like to fin d T¡ a nd Tz, t he
·number of ho urs t hat coatings 1 a nd 2, r espectively, leng t he n t he life of
the bulbs ( T 1 or T 2 may be zero or negati ve), and T 1 - T 2 , t he difference
between t he effects of t he two coat ings . T o evaluate t hese constants
the manufacturer could m ake a b ulb wit h coating l ét nd t est it to see
how many h ours it will last, Lhe n m ake a bu lb wit h coa.t ing 2 a nd see how
many ho urs it will last. The model can be \\Ti tten
Y1 = µ + +e T1 1
Y2 = µ + 7 2 + e2
where Y; is the n umber of h ours t hat the bulb \\·ith coating 1· lasts, and
where ei is a r an dom error d ue to.all t he uncontroll ed factors, s uch as
nonconstant voltage, imperfections in t he manu fact ure of t he va rio11s
223
2 24 L l NEAR S'l'A'l'l STI CAL l\fODELS J\IODEL 4: EXPERil\lEN'l'AL DESIGN J\lODELS 225
corn p on ents, t emp cra.t ure, an<l h umidit y . No w, t..hc mirn ufacturct or in matrix notation as Y = U ex + e, or
will p ro ba bly not b e contcnt to d raw conclusio ns based on o bscrval,ion Yn 1 O
of o nly one lig ht b ulb for each coating. L ct us sa.y t ha,t he decides to
m akesix bu lbs a ncl coat t hrec wit h chemical 1 a nd i hree wi t h chcmi cal 2. Yl'!. 1 O
T he moclcl can be writtc n o
+ C13
1
Y 11 = µ + Ti + e11 (C'/·1 ) ( 11.4)
o 1
Y 12 = µ + Ti + e 12
Cl.2 e?.1
Y 13 = µ + T i + e 13 ( ) l.J )
o 1 e22
CJ
..L
1
Y2 1 o l C2 1 Y 1:1. + r1 + fi2 +
= fl e 12
(11.5)
Y22 1 o J e~:! ?/21 = µ + T2 + Pt + €21
Y22 = µ + T2 + f32 +
Y23 l o C23
e 22
I n t his case X is <L 6 x 3 matrix, a nd its rank is 2. H encc, t.. his modcl 0 1' Y;; = µ + T¡ + {Jj + €¡; i = 1, 2 ; j = 1, 2
is no t of ful! rank, a nd t he t heore ms of Chap. G a re not i rn mcd iatcly where YiJ is Lhc obser vcd y icld of cor n on a certain plot o f gro uncl. T he
a p pli ca ble for estimating a nd t cst,ing hy potli cscs abo ut t hc comp oncnts rcscarcher is ass uming t hat.. t hc obser vation Y;; is cq11a l t o a constant µ
of [3, t ha.t is, µ , ..1 , T 2 . (Lite average yielcl whcn no chc mi cal is applie<l) p lus t he cffect d ueto the
Suppose t hc ma nufacturer is intcrcs tcd o nl y in µ + T ¡ <Lntl ¡t + -r 2 ith che mical, p lus t he e ffcct d ueto t hc.fth mcthod of app li cation, p lus a
r ather t ha n in T i a nd T 2 . l f wc writ..e µ + T 1 = (/. 1 a ncl ¡t 1- T 2 = !:1. 2 . random erro r d uc to a.JI t he uncont rollcd facto rs, s uch as differences of
we can write t he moclel as fcrt ility amo11g t he p lots. T hc cxpcrimentcr m ;Ly dcsire to test or
Yn =+ e ll
C'I.¡ estimatc t hc pMa metcrs ¡.¿, T;,.fl;· Jn mat rix notat..io n t his m oclel is
Yi2 = C'I. ¡ + C 12 Y1t l l o 1 o µ eu
?Ji 3 = C'I.¡ + <' ¡ 3 Ti
( 11.3) Y12 l o o l e.12
Y2 1 = C'l.z + ez¡
T2
-
o 1 o Pi +
Y22 = + C'l.2 e 22
Y2 1 C:!. l
E( b' Y) = b 'X(3 = A.' (3 and r 2 both sa,tisfy Sr = :A, t hc n r ;X' Y = r~X' Y = :A' (3 = :t..'~-
Proof: L et ~ be a ny solu t ion to t he syste m X'X~ = X ' Y. Using r 1
which implies that b 'X = A.', or X ' b = A.. This implies t hat the
and r 2 , wc have r; X' X~ = r ;X ' Y and r~X'X~ = r~X' Y. But
mnk of the coefficicn t matri x X ' equa ls t he ran k of t he a ug mentcd /'.
matrix (X' 1 A.). If Lhis is trne, the n t he ra nk of X 'X equals t hc r; X'X = r~X' X = :A' by hy pot hcsis; hc nce, :t..' (3 = r;X' Y = :A'~
ra nk of (X ' X 1 A.) a nd , he n ce, S r = :A has a solu tion for r. On thc ,,./"'- ~
and :t.. ' (3 = r~X' Y = :t..' (3 .
othe r hand, if Sr = :A has a solution for r , t iten " ·e can \\Ti te
~
X '(Xr) = :A a nd !et b = Xr. That :t..' (3 = :A ' ~ mcans t ha.t, if :t..' (3 is estim ab l ~, :A ' ~ is t he estímate,
\Ye s ha.ll now proYc a Lheo rcm for mode l 4 a na logous to t hc Causs- wherc ~ iR a ny soluLion to t he n or mal cquations .
Ma rkoff t h eorem fo r model 1. I t wi ll be i m por ta.n t to lrno w how ma n y linear fun ctio ns a re estimable,
but first we sha ll formu late t he fo llowing .
+ Theorem 11.2 In model 4 , case B , t he best linea r unbiased estima te
for any est inrnblc fun ction :t..' (3 is r 'X ' Y, whc rc r satisfics t hc • De finition 11.4 Thecstima b le fun ctions :t..;(3 , ~ (3 , . .. , A.~ (3 aresaid
eg uation S r = :A. to be linearly indepenrl<'nl P~sfimablefimctions if t hcrc e xist vcctors
Proof: L et u s ass ume t ha.t the best linea r 1111 biased estima to r of :A' f3 r 1 , r 2 , . • • , r ,,such t hat Sr 1 = :A. 1, Sr 2 = :t.. 2 , • . . , Srq = A.Qand if
is b ' Y , whcre b ' = r ' X ' + a ' . T hus b is completely ge ne r:tl, t hc vecto rs :t.. 1, :t.. 2 , • • • , Aq are linea rly indepc ndent. If t he A; a re
s ince a is genera.l. \ i\'c must d et ermin e t he vector a s uc h t ha,t not linearl y indepc ndc nL b u t if t he matri x f\. = (:t.. 1 , :t.. 2 , • . . , :t..0 ) has
(1) E(b' Y) = :t..' (3 n1nk t, then t he set conta ins t Jinearly indcp_endent estima ble
(2) var( b ' Y) is lcss t ha n an y other linear fun ct ion of Y t hat sa.tis- fu nctio ns.
fies (1) • Theor e m. 11.4 Therc are cxactly k linearly inclep e ndent estima ble
F or unbiased ness, we get E( b ' Y) = b 'X(3 = (r 'X' X + a 'X ) (3 = functi ons, whe rc k is t he rank of X .
:t..' (3 . Sin cc on r hy pot hesis states t h at Sr = :A, we must have Proof: ·w c must s how t hat t he re a re exacLly k linearly independent
a ' X = O. Also, vecLors :A t hn,L lcad to a solu t ion fo r Sr = :A, whe re k is t he ra nk of
var(b 'Y) = E(b 'Y - :t..' (3)2 X . Tf therc a re q vecLors sat;isfy ing Sr = :A, !et them be A¡, :t.. 2 ,
= E[( b' Y - :A' (3)(b' Y - :t..' (3)'] ... , :t.. 0 (q > k). T hc n , by Thcorcm l l.l , t herc ex ist q vectors
r 1. r 2 , . •• , rq s uch that Sr 1 = :t.. 1 , Sr 2 = :t.. 2 , . • . , Sr"= :t..,,, or
= E((b' X(3 + b'e - :A' (3 )(b'X(3 + b' e - :A'(3)'J S (r 1 , r 2 , • • • , r ,,) = (:t.. 1 , :t.. 2 , • • • , :A"), or SR = 1\., whe re R is Lhe
= E L( r 'X'X(3 + a ' X(3 + b 'c - :t..'(3) p X q 11mt ri x (r 1 , r 2 , . • . , r ") a.ncl f\. = (:A¡ , :A.2 , ••• , :A,,) .
X (r 'X'X(3 + a 'X(3 + b'e - :t..' (3)'] B u t S is of ra nk k; so f\. m ust be at most of ra n k k; hencc t here
can be at m ost k lincarl y inclep cnd ent estimable fon ctions.
= E( b 'ee ' b ) = a b 'b = rr (r 'X'
2 2
+ a'Y(Xr + a)
Lct Xf be the ith row o fX. \Ve sh a ll s how t haL x; (3 is estima.ble
= a~r 'X'Xr + a 2 a ' a for every i. To sh ow t hat x; (3 is es t i ma ule we 111 ust show th at
(using the fa.et that a'X = O). X 'Xr = X ;adm itsasolut ion r . T heran k ofX'X cg ua ls therank
230 Ll ' EAR STA'l 'lS'l'lCA L l\lO UBLS
l\lODBL 4: E XPBRU.IEN 'l' AL DESlGN l\IODELS 231
o f (X ' X 1 X ;). H en ce, x; (3 is es t i mu.ble for i = 1, 2, ... , n. But co111bina tion ofthe bcst linear unbiasetl est.imates ofthe estimable
these X ; fo rm a. set of k linearly indepcndcnt vecto rs, and the fun ctions; i .c., if (X'Y (i = 1, 2, ... , q) al'e the best linear unbiasecl
t hcore m is provccl. est imates of the estimable functions J..; {3 (i = 1, 2, . .. , q), then
Ncxt we sh all extcnd Defi nition 11.3. l.a ;r¡X 'Y is t he bei:; t linear unbiased estímate of La;A; {3.
+ Definition 11.5 L et A bca matrix s uch t hat A = (A 1 , A 2 , . • . , A) Thc proof will be Jcft for the reader.
- 111
where A ¡ is p x J . \ Vhe n we say that t hc matrix fun ction A'(3 is tt.2.2 Examples. To illustrate the preceding theorems, Jet us
estimable, we sha ll m ean that each A; (3 is estima ble (i = 1, 2, considcr the model give n in Eqs. (11. l ) a nd (11.2) .
. . . ,m) .
• Theorem 11.5 X (3 a nd X'X (3 are estimable. Y;; = µ + T ; + e,.; i = 1, 2; j = 1, 2, 3
The proof " ·ill be left for thc reade r. As a consequcnce of t his Wc hiw c a iread y s hown tha t (3 is n ot estimable, since X is oflcss than
thcore m , E(y;) is estima ble fo r e very i , sin ce E( Y) = X(3. full rnnk. Thc first question we shall a nswe r is : I s T 1 - • 2 estimab le~
It is clcar that
+ Theorem 11.6 lf J..~ (3 , J..~ (3 , .. . , J..~ (3 are estimable, a ny linear
combin ation o f t hese is estimable.
Proof: L et a 1 , a 2 , • •• , ª"be constan t scala rs . \Ye must show that
J..' (3 is a n estima ble func tion if A = L:a ;A,. L et r ; be s ueh that
Sr; = A; and r = L:a;r .;. Then r satisfi es Sr = A, u.nd t he proof
is comple te. We rnust see " ·hethe l' t here ex ists a vector r that is a soJu t ion to the
equations X ' Xr = A. Now
• Corollary 11.6.1 If J..' (3 is estim able, t he n J..' = a ' X , whe re a' is a
l x n Yeetor ; t hat is to say, if J..' (3 is estimable, J..' 111ust be a linear
co mbination of thc rows of the matri x X. =
G3 3)
Proof: X(3 con tains /.;, linearl y independe n t estima bl e fon ctions;
X 'X
(3 o 3
3 :l O
sin cc t here a re exactly k such fun ctio11s , J..' mus t be a linea r corn-
bination of t he l'O\\' S of X . and Sr = A. g ives t hrce equa tions and tluee unknowns.
This corolla ry is valua bl e in d et ermining wha tfonct ions a.re estimable
(~ o~ ~)
by inspection of the m od el Y = X(3 + e . This will be illustrntcd in
detail later. : )(::) = (
+ Corollary 11.6.2 If A.' (3 is estimable, there exists a Yector b s uch 3 3 13 - 1
that J..' = b 'X'X, a nd the best linear unbiased est ím ate of J..' (3 is
To in vcstigate the existen ce of a solu t ion , we see that the rank of
b 'X'Y.
X' X is 2. The a ug mented ma trix is
Thus we can d etermine frorn the n o rm a l cquations what is estima ble
(~ :: ~)
ancl what t he best linear nnbiasetl es limate is .
+ Corollary 11.6.3 Let C be a n y 111 x n matrix ; t hen E (CY) is
est imable. 3 o 3 - 1
Proof: Since E(CY) = CX(3, the resul t follows from Thcorcms 11.G
and this matrix a lso has rank 2. \V1·i tin~ out the system a nd soh·ing in
and 11.G.
ter nis of r1
+ Theorem 11.7 The b cst linear unbiascd es timate of a linear r2 = ! - r1
combination of estima ble fonctio11s is g i ven by the same linear
r3 = -k - r1
232 L 1N F.AR S1' A'I' l.S'l'JCA L l\lO D ELS i\LODEL 4: EXPERii\IE N'l'AL DES CGK lllODEL S 233
By fix ing r 1 at any val ue we p icase, we gct a soln t io n to t he cqnation~ .By investigating t hc r a nks " ·e sec Lhat the ~·a1~k. of t he coefficicnt
X 'X r = 'A. For cxa rnp le, if \\'e lct r 1 = O, wc get niatrixis 2 a nd t he rank ofth caug m en Lecl matnx is 3. H cncc T 1 + Tz
is not estima.ble.
This is quite a clifficu lt proccdure (and beco mes increasingly labori o us
ifthe dim e nsio n ofX'X is large) forexami ning the ex istenee of estinu 1.ble
fu nctions. B y using the model a nd T heo re m 11.5 (instead ofTheorem
JI. l ) we can so me times te ll immedi ately whethe r a eerta in fu netion of
= (O,~. -!)(~, )
the parameters is estimable . For exa.mplc, X (3 is estima b le a nd it
and equals
r 'X ' Y
J 2·
+µ Tl
x·v ~ C:)
/L + Tl
s in ce X(3 = ( 11.6)
/L + Tz
2. µ -j- Tz
whe rc Y;. = L Y;;; a nd r ..
j
= L Y;;·
ij
So T1 - Tz is estima.b le, a nd the
µ + Tz
best linear unb iased esLim ate is
or the general t erm of t his yector can be written
E(Y;;) = µ + T¡ i = ]. 2; j = 1, 2, 3
By T heo rc m l 1.6, a ny linear combina t io n of µ + T; is estim a ble.
ln fu r the r demonsLnüion o f Thcorcm 1 1. 1 , we see t hat Also, since X (3 contains k = 2 linear ly inde pe nde nt estima ble
fun ctions, e,·ery estima.bl e fu nctio n m ust be a linear co m bination of the
E (}. _ Y,) = ( +
l.
3
-, ]!} ll
r
T1 + _J_
;{
Í: e¡;
- µ - T .. -
-
2 e2 ; )
- 1-
3
= Ti - Tq
-
rows of X (3 . S in ce µ occurs in every r o\\' of X (3, a ny csti 111ab le fun ctio n
that does not contai n ¡i m ust be s u ch t hat t hc coe ffi cicnts of t he T, acld to
zero. For example, 3µ -l 2T 1 + T 2 is estimable (fi rs t row p lus seeond
i.e., there d ocs exis tan unbiascd estilllate of T 1 - T 2 . row plus fourth row), bn t there is no linear co mbin ation of t he rows of
T o demonstrntc Theorcm 1 1.3, s uppose we take a nothcr solu t io n of x13 that gives T¡ + Tz; hc 11 cc T I + Tz is noL estim able. Therefore, by
X ' Xr = 'A ; t h at i:;, lct r 1 = 1- W e get r ' = (:},O, -·}), a nd displaying X (3 = E(Y;;) we can genernll y see what is a,nd what is not
estimable .
After we d ecide t hat a certain fu nctio n is estim a ble, t he re still
r ' X'Y = C, O. - 5)
(
)> =
y ) Y
··
- ·)Y.
:~ - -· =
Y
1.
- l"2 .
3
remains thc proble m of fi11din g th e best estimatc. Th is can be aeco m-
plished by using thc no rm a l cquations a nd Corolla ry 11 .G.2. F o r
) 2.
cxample, hc re the no rm a l equations are X'X~ = X ' Y, " ·hi ch are eq ua l
th e sam c esLirnatc ns befo re. to
No" · we s hal l in vesLigate wh cLher T i + T2 is estim a ble. In t his case (a) G¡í +
37\ +
3f2 = .r
'A = (O, 1. 1), and X 'Xr = 'A becomcs (b) 3,ii +
3f1 r1. ( 11.7)
(e) a,ú + 3f~ = r 2 .
By Thcorems 11 .5 ancl 1 1. 6, a ny 1inea.r combina Lion of t he rows 0 11 t he
left sid e (wit h h ats r e moved) is estim ab le , a nd , by Corolla ry 11.G.2, t he
best linear u nbiased estim at e is give n by tite sam e li near combination of
234 LINEAR STATJS'l'ICAL l'\IODELS 1\IODEL 4: EXPERii\IEXTAL DESrGX :\L O DT~LS
235
rows o n t hc rig ht -ha nd side of the equations. Suppose we take Eq. ,\s in t he prcvio us ill ustration a ncl by Theorem l l. 5, each i·ow of X'X:
(11.7b) minus Eg. (11.7c); weget lcads toan estima bl e fun ction, ancl the estím ate is cqua l to t hc ricrht-
1Ja n
d side of t he equation; e.g., fro m row (e) " ·e know t hat 2µ + -T
'> ~ +
,,
3-í\ - 3f2 = Yi. - Y2 • p + p2 is estimable a nd t ha t t he best linear unbiascd cstin;ate
O t' f - f - Yi. - Y2. (Corollary l 1.6.~) of this functi on is equa l to }" 2 ,. \Ve see that the
1 2 - 3 foJlowing a re estima ble:
But, by Theorem 11.3,
~
/31- /32
Tl - Tz = A
7"1 -
A
7"2 = Y l. - Y 2. 71 - 72 + /32 - f31
3
71 - 'Tz + f31 - f3z
so t hc bcst linea r unbiased estímate of Ti - T
2
is
etc., a nd that t hc best unbiased estim at es are irnmcdiatcly available
Yi. - Y2 • from Eq. (11.8). For example, t he best linear unbiascd cstimat e of
3 .,. _ 72 is o bta ined by taking ·} times Eq . (11. Sú) min11s ~- tim es Eq.
1
the samc estima.te t hat wc obtained by solving X 'Xr = ).. and using (ll. Sc). This gives
r 'X ' Y. By exam ining the normal equations we see th at 3µ + 3T1 and
3µ + 3T 2 are cach estima ble (since they are rows ofX'X) a nd that they
f1 - f2 = HYi. - Y 2.l = Yi. - Y2.
are li_ncad y i_ndep endent. T~rnt is to say, )..~ = (3, 3, O);)..~ = (3, o, 3) Notice that the only linear com binat ion of t hc T 1 t ha t is estima ble is
are l111earl y rndependent. Smce the rank of X 'X is 2, we know fro 111 l:c¡ T ¡, whcre Le¡ = O. ·This prompts t hc following dc finition .
Theorc m 11 .4 t hat every estimable function ).'(3 must be a linear
combination of )..; (3 and ).~ (3 . + Definition 11.6 In t he model Y = X(3 + e t hc linear combination
of p ara meters "l,c¡{3¡ is ca ll ed a contrast if ~e; = O.
F or a nothcr e xa mplc, Jet us examine Eq. {11. 5). From
11.2.3 R e p a rametrization. In order to ut il izo lihc theorems in
E(y¡;) = µ + T ¡ + /31 Chap. 6 on model 1 wltcre X ' X is of fo il ra nk, we s l1<i ll rcpant111etrizc t he
wc seo that a ny linear co mbination of µ + Ti + {31 is estima ble. If we model of l ess th a n foil ra nk to a model of ju ll rnnk.
examine t he norma l equations, we get for X'X~ a nd X'Y
+ Definition 11. 7 B y a rcpa rametrizatio n of thc rnodcl Y = X (3 + e
4 2 2 2 2 y we s ha ll mean a t ransformation fro m thc vcclor (3 to t he vector
a by a = U (3 , " ·here each element of a = U(3 is an estima ble
2 2 o l 1 yl.
fon ct io n.
X'X = 2 o 2 1 1 X'Y = r 2. Since X 'X is posit ive semidefinite of ra nk k , we know that t here exis ts
2 1 1 2 o y .1 a nonsing ula r matrix W * (p x p ) su ch that
2 1 1 o 2 r .2
A dot in t he s ubscript of Y ;; means that t hey;; a re to be summed over (W*)'X'XW* = (: :)
t he subscript th at is replaced by the dot . Therefore, wc get
where Bis k x k of ra nk l.:. If we parti t ion W * in Lo W * = (W, W 1 ),
(a) 4¡1 + 2-í\ + 2f2 + 2/j1 -!"" 2p2 = Y ..
where W is p x k , we get
(ú) 2µ + 2-r1 + P1 + P2 = r ..
(e) 2/i + 2f2 + P1 + P2 = Y2. (ll.S)
(:;)x'X(W, W1 ) = (: :)
(el) 2¡1 + f 1 + f 2 + 2~ 1 = Y.1
(e) 2¡í, + f 1 + f 2 + 2/j2 = Y .2 which gives W 'X'XW = B a nd W~X'XW 1 = O. T his implies that
236 LINEAR STATISTICAL MODELS l\IODEL 4: EXPERIMENTAL DESIGN l\IODELS 237
W'X' is of rank k and that W~X' = O. vVe can wri te the modet • J
full-rank ~eparametrized model Y = TS + e is used to estimate A' (3,
Y = X(3 +e as Y = XW*(W*)-1 (3 +e, and, if we let (W *)-1 :::: tben the two estimates of A' (3 are identical. Therefore, we shall state
(~J we get
an important corollary to Theorem 11.8 on the uniqueness of estimates.
U* =
+ Corollary 11.8.1 Any full-rank reparametrization giYes the same
Y= X(W, W1 )(~) 13 +e
estimate of the estimable fun:_tion A.'(3.
In the full-rank model in Chap. 6 we found that, if the matrix X'X is
or Y = (XW)(U(3) + (XW1 )(U1(3) + e diagonal, this leads to sorne simplifying results. Wc shall now show
that a model can always be so reparametrized that the resulting matrix
which reduces to Y= (XW)(U(3) +e, since XW 1 -;- O. Letting XW:::: of normal equations is diagonal. W e shall first. state the following
Z and U(3 =a., we can write Y= X(3 +e as Y= Za.+ e, where Z is definition.
n x k and has rank k; hence we ha ve a full-rank reparametrization, i.c.,
• Definition 11.8 By an orthogonal reparamctrization of the model
from a less-than-full-rank model to a fitll-rank model. Thus we ha ve
the following useful theorem. Y = X(3 +e we shall mean a full-rank reparametrization to the
model Y = Za. + e, where Z'Z is a diagonal matrix.
+ Theorem 11.8 Let a. be a k x 1 vector of k linearly independent
estimable functions of the parameters (3 in the model Y = X(3 + e, We shall now show tha.t there always exists an orthogonal repara-
where X is an n x p matrix of rank k. Then thcre exists a repara- metrization.
metrization to the fulI-rank model Y = Za. + e. • Theorem 11. 9 If W * is an orthogonal matrix su ch that
To estimate a. or any linear combination of a. we use the normal (W*)'X'XW* = D
equations (Z'Z)a = Z'Y, and all the theorems of Chap. 6 apply.
Another method of proving that a. = U(3 forms a set of k linearly (D is a diagonal matrix), there exists a partition of W *. say, W* =
independent estimable functions is to show that a., = u,(3 is estimable (W, W 1 ) (where W is J> X /~), such that XW *W *' = XWW' and
far cach i = 1, 2, ... , k, where U¡ is the ith row of U. By Theorem 11. l, a. == W' (3 is estimable and W'X'XW is diagonal of full rank.
this means that there must exist vectors r 1 , r 2 , ••• , rksuch that X'Xr1 Proof: This follows from the fact that there exists an orthogonal
= ui, ... , X'Xrk =u~. Putting these into a single matrix equation, · matl'ix W *su ch that (W *) 'X'XW * = D, wherc t.he characteristic
we get X'XR = U', where R = (r 1 , r 2 , ••• , rk). So we must finda roots of X'X are displayed down the diagonal of D. Exactly k of
these roots are nonzero. \\7e can writ.e W* = (W, W 1), nnd
matrix R satisfying X'XR =-U'. Using (W*)'X'X(W*) = ( : :).
we can obtain X'X(W, W 1) = (U'B, O) by multiplying on the left by (W*)'X'XW* = (W'
W'
)x'X(W, W)= (DO Oº)
1
1
so a. = U (3 is estimable. where D 1 is k x kof rank l.: (nonzero characteristic roots ofX'X are
There may be different matrices W * that will diagonalize X'X and, on the diagonal of D 1 ). Thus W'X'XW = D 1 and W~X'XW 1 =
hence, many different full-rank reparametrizations. For example, in O, which implies that W~X' = O. Now we can write Y = X(3 + e
the model Y = X(3 +e, suppose two full-rank reparametrizations are as Y = XW*(W*)'(3 +e,
represented by Y = Za. + e and Y = T8 + e. By the definition of
reparametrization, this means that there exist matrices V and U of
rank k ancl of dimension k x p such that a. = U(3 and S = V(3. But or Y= X(W, W,)(;) 13 +e= XWW'13 + XW1w;13 +e
there exists a nonsingular k x k matrix A such that U = A V. This
shows that thcrc is a linear relationship between full-rank reparametri- orY = (XW)(W'(3) + e,sinceXW 1 =O. lfweletXW = Zand
zations. That is to say, if the full-rank reparametrized model Y = W' (3 = a., we get Y = Za. + e, and, since Z'Z = W'X'XW = D 1
Za. + e is used to estima te the estimable function A' (3 and if another (diagonal), the theorem is proved.
238 LIN EAR S'l'ATISTICAL l\ TODEL S llWDEL 4 : EXPERIMENTA L DESIGN MODELS 239
+ Corollary 11.9.1 If W is as clefin ed in Theorem 11.0 a nd if wc de- ~ +~
note XW by Z a nd W ' (3 by a , t hen the model Y = Za + e is an
orthogonal repara metrization of t he model Y = X(3 + e.
• Corollary 11.9.2 If W ' (3 is a.n ort hogona l repara metri zation of t he
and (U *)-1
= W* =
(
~
-t + ! +t
- ~·
+!)
J.)
m oclel Y = X(3 + e, t hen X = XWW'.
This corollary wi ll be qui te useful in reparametrizing .
11.2.4 Examples. \~'e sha ll ret urn to Eq . (1 L. 2) to illustrate
reparametrization . Fro m Eq. (11.2) we see t hat µ + -r 1 a nd µ + 70
so w ~ ( _:
t
-!
are two linearl y independent estimab le fun ctio ns. T hey are esti mable-
1 o
since E(y 11 ) = µ + -r 1 a nd E(y 21 ) = ¡.t + T 2 ; t hey a re linearl y inde~ l o
p endent, since
1 o
and Z = XW =
o 1
o 1
o 1
The normal cquatio ns a re Z'Za = Z' Y; so
and
and since A. 1 a nd A. 2 are Jinearl y indep ende nt . To illustrate Thcorem whichimpliesth at& 1 = !iYu and &2 = !Y2 _. Now<X 1 - <X 2 is estim-
11.8, let 0'. 1 = µ + -r 11 <X 2 = ¡.t + T 2 ; we get able, and <X¡ - <Y. 2 = -r 1 - -r 2 ; so f 1 - f 2 = ·} Yi. - }Yv just as i n
the model before it was repa ra metrizcd.
11.2.5 Variances and Covariances of Estimable Functions.
In Theorem 11. 2 we proved t hat, if A.' (3 is estimable, r ' X'Y is the b est
(minimum-varia nce) linear unbiased estímate~ of A.' (3, where r satis fies
Sr = A.. \\Te also found the variance of A.'(3 , a nd n ow we stat e the
To find Z in the n ew model Y = Za + e , \1·e need t.o fi nd W * = (W, W 1) following:
= (U *) - 1 , where + Theorem 11.10 If A.~(3 a nd A.~ (3 are two estima ble functions, the
U* = (~J respective variances of t he best linear unbiased estimat es are
a2r;X'Xr 1 a nd u 2 r~X'Xr 2 , where r 1 a nd r 2 , respectively, satisfy
X'Xr 1 = A. 1 and X'Xr 2 = A. 2 . The covaria nce of the estimates of
Sin ce
u=
1
( .L
l
o
º)
1
A.;(3 a nd A.; (3 is equal to a 2 r ;A. 2 = u 2 A.;r 2 •
P roof: \Ne get
we need to find U 1 s u ch t hat U * is n onsing ula r . It is easy to sec t hat cov(A.~~' A.;~ ) = E[(A.i ~ - A.i(3)(A.;~ - A.;(3)']
U 1 = (O, 1, 1) will satisfy t his condit ion. So = E [(ri_X'Y - A.i_(3)(r;X' Y - A.;(3)']
= E[( r~X'e)(e'Xr2)]
U* =
(
1
1
1
O 1
º) = u2 r i.X'Xr2 = a2A.~r2 = u2 r~A. 2
To find the varia n ces, A. 1 can be set equal to A. 2 , a nd t hen r 1 beco mes
o 1 1 r z.
240 L f).'"EAR S'l 'A'J'IS'J'lCAL i\I ODELS llrODEL 4: EXPERl i\IEN 'l'AL DESIGN l\lODELS 241
To show why an o rthogonal repara.mcLrization is sometí mes uscfu¡ is invariant for any ~ that is a solution to the normal equations
w o sh all prnvc Lhe fo llowing Lhcorc m . ' X'X~ = X ' Y.
+ Theore m 11.11 If Y = Za. + e is a n orthogona l rcp;u a metriza. \Vhen we say that á 2 is úivaricmt we mean that á 2 is the same rega rd-
t ion o f Lhc modcl Y = X~ + e , Lhc cle men ts of a re unco rrelatc<l a lcss of wh ich soluti on to t he normal equations is used. The proof will
Proof: ex = (Z 'Z )- 1 Z 'Y = D ~Z'Y , whe re D 1 is diagonal. . bo lefL for t he rcader.
co,·(a ) = E[(ii - a.)(a - a.)'J = E{[(Z' z)- 1Z'Y - a.][(Z 'Z)- 1z·y - al'} • Theorem 11.14 In. Theorem 11.13 Lhe q uantity
1
= E{((Z' Z)- Z'(Za. + e) - a.]
X [(Z' Z)- 1 Z'(Za. + e) - a.]'} (n - k)á2
a "-
= E{[(Z 'Z)-1 Z 'e][(Z'Z)- 1 Z'e]'}
= u2 (Z 'Z)- 1Z' Z(Z 'Z)- 1 = a 2D! 1 is clistributecl as x2 (n - le), a nd á2 is an unb iased estím ate of a 2 •
So cov(&;,&;) = O if i ::F j, a nd t he t hcorem is proved. The proof of t his t ltcorem will be left fo r t hc reader.
11.2.6 Point Estima tes of a Linear Combin atio n of ~ ; un der
Case A . If e is d istributed N(O,a 2 1), ma ximizing the likelihoo<l 11.3 Interval Estimation
equ ation leads Lo thc samc n o rm a l cquations as t he Jeast-squarcs
m cthod under case B. 11.3.1 Inte r va l Estímate of Estim a ble F unctions . To set; a
L et us reparamc t r:ize t he mocle l Y = X~ + e to a modcl offull rank coníidcnce intcn ral on an estimable function A. ' {3, the proceclure is t o
s nc h as Y = Z a. + e. The n, by t hc inv<uiance property o f 11rnxim111n rep:uametrize t he model Y = X{3 + e, wh crc e is distribu ted N(O,a 21),
likelihood, t he max imum-likclihoocl estímate of a. is = (Z 'Z )- 1Z 'Y,a in to a fu ll -rank moclcl Y = Za. + e, w here, say, 0'. 1 = A. ' ~- Thcn ,
a nd a li t h e propcrtics listed in Theorcm G. 1 apply to a (re me mbe r et¡~ from Lhe normal equations (Z'Z ) a = Z ' Y , t he estima.te &1 can be ob-
a linearly i ndepende nt set of estima b le fon ctions) . This res ult is gi,·en taincd. By Theorems 1 J .12 and 11.14 it follows that & 1 is di strib nted
in t ite fo llowing t heorem. N(a 1 ,c 11 a 2 ), wh ere c 1 t is thc first d iagonal elemen t of (Z ' Z )- 1 , t h at
+ Theorefl] 11.13 In model +, Y = X ~ + e, where X is of rank k, 11.4.1 T e s t of t h e H ypot hesis t h a t a n Estimabl e Function
thc qua n t ity of {3 Is Equ a l to a Kn o wn Con stant. To Lcst thc hypoLh esis
H 0 : A. ' ~ = íl.(l' (w hcrc c<J is a lrnown constant), whcre A.' ~ is an esLi rnablc
á2 =-
1
- (Y - X~)'(Y - X~) fu ncLion, we can repara rnetrize Y = X~ + e to a full -ra nk m od el
n- k
242 LTKEAR STATIS TICAL i\rüDELS l\LODEL 4: EXPERIMENTAL DESIGN l\TODELS 243
Y = ZCI. + e, whcrc o:1 = :A.' (3, a nd use the test funct ion in Art. l l.3.l J,ct t he s X 1 \ecto r C1. be
(J. = (~:)
with o: 1 replaccd by o:J; hencc, if H 0 is t rue, t he quantity
2 (&¡ - rt.~)2
V = ---='-----"'-'--
<f?
C11
is distribu ted as Ji"(l, n - k, }.), where c 11 is the first d iago nal elernent A.~13
1 Furt her, Jet A.: +1 13, .. . , :A.~. 13 be k - s estimable functions s uch t h at
of (Z'Z)- 1 a nd where the noncentrali ty }, = - - 2 (A.'(3 - o:ti)2.
t..;{3, A.; [3, ... , J..~. 13 form a set of k linearl y inclep endent estimable func-
2c11a
11.4.2 Test of th e H y pothes is H 0 : {31 {3 2 = =···=
PQ (q :s;;; k). tions. Also, !et
(
A.~+113)
This is one of the most useful tests in the genera l linear hypothesis of
less Lhan fuU rank . For example, in Prob . 11. 2, it may be clesirablc to
test 7 1 = 7 2 = 7 3 . J..:+213
y =
\iVe shall discuss t he testing of estimable h ypotheses only. That is
we s ha ll considcr the test ing of t he h y pothesis {3 1 = {3 2 = · · · = {Jq ¡f
J..~.13
a nd on \y if t hcrc cxisLs a set of 1inearly independent csti mable functions
:A.; (3, A.; 13, . .. , :A.; 13 such t hat H 0 is t rue if and only if :A.; (3 = J..; ~ ::: Jf we lct S = ( ~) , th~n, by Theorem 11.8, it follows t hat wc can repara-
= A.;13 = O. For example, consicler t he model g ive n by
rnctri:w fro m t he less-tha n-full-rank moclel y = x 13 + e to thc full-
Y;1 = µ + 7; + O;+ e;; i = 1, 2, 3; j = 1, 2, 3 rank mod cl Y = ZS + e . This can be writtcn
Suppose we wish to Lcst H 0 : 7 1 = 7 2 = 7 3 • By exam ining t he modct Y = Z 1 a + Z 2y + e
we see Lhat 7 1 - 7 2 a nd 7 1 + 7 2 - 27 3 are two linearly independcnt Now H 0 is t rue if and only if C1. = O; t herefore, to t est H 0 we u se
est imable fun ct ions. It is also evident t hat H 0 is t rue if a nd only if Theorem G.G a nd test a = O. This is de monstratccl in T able 11.1.
7 1 - 7 2 a nd 7 1 + 7 2 - :.h 3 are sim ultaneously equal to zcro. Thus, The impo rtant qua ntities are :
H 0 is anesLimable hypothcsis,and we s hall test7 1 = 7 2 = 7 3 bytesting l. Sum of squarcs due to e rror , which equa.ls (Y - ZS)'(Y - ZS),
whcre 5 is t hc solution to t he normal eq uat ions Z ' ZS = Z 'Y.
2. Sum of s q uares dueto C1. adjus ted for y , which is cq ua l Lo R (a 1 y )
= S'Z' Y - y'Z~Y, where y is the solution to t he no rma l eq uatio ns
• Definiti o n 11. 9 A hy pothcsis H 0 wi ll be called est i mablc if t here z;zá = z ; Y .
exists a set of linearly indepen dent estima ble functions )..; [3, /..~ ~. These nor ma l eq uations are deri ved from the model Y = Z 2 y + e .
. .. , J..;13 s uch t hat H 0 is t rue if and only if )..;13 = :A.; 13 =- · · · = Jf we set
)..;13 = O. 5Z'Y - y 'Z;Y n - k
V = Y ' Y - S'Z 'Y - s -
In genera,!, suppose " ·e wish to tes t t he hypothesis H 0 : {J 1 = {J 2 =
v is dist ribu tcd asF'(s, n - k, },) .
T he value ofA wi ll be discussed la ter.
··· = {Jq in moclcl 4, the less-t ha.n-fu ll-ra.nk model Y = X13 + e . , 'up-
lt wou ld be quite a tedious job to find t he matrix Z if t his were neces-
pose there cxists a set of linearly independent es t imab le fun ctions A; ~,
sary in order to find v, but fortu nately it is n ot necessary. vVe can use
:A.~ [3 , . . . , )..: (3 such t hat H 0 is true if a nd only if
Theorcm 11.1 3 (,o show that (Y - z5)'(Y - ZS) = (Y - X~ )' (Y - X~).
So t hc de nominator of v can be obtained fro m t he normal equations
J..;13) X'X~ = X ' Y. By t he same token t he t erm (Y - Z 2 y) ' (Y - Zá) can _
:A.213 be obta.i ned from t he normal equations t h a t are dcrived from t hc model
(
= 0
Y = X[3 + e with the condition /3 1
forc provcd t he following t heorem .
{3 2 = =···=
{J 0 • \Ve have there-
J..:13
244 LINEAR STA'l'IS'l'ICAJ, lllODELS llIODEL 4: EXPERIMENTAL DESIGN llIODELS 245
• Theorem 11.1 S In model 4, Y = X f3 + e , to test Lhc hypothesis • Theorem 11.17 Let {3 1 , {3 2 , ••• , {3Q be a subset of t he elements of
{3 1 = {3 2 ---= · · • = {30 (q ~ k), wh ich we assume is cqui valent Q
to testing t he li nearly estimable fon cLions A.; f3 = A.; [3 = · · · : : : f3 in m odel 4. Suppose that .L cJ3; is estimable for every set of
q i- 1
A.~ [3 = o, constants C; sueh that I C; = O. Then t he hypothesis H 0 : {3 1 =
(1) Obtain any solution to t he normal equat ions X'X~ = X'Y i- 1
L ct.us writc X = (X 1 , X 2 , . . . , X v), where X ; is t he ith column of X. ¡11 thosc cquations in t hc model w hich in vo h·c Y]j· So the s um of Yv
The model Y = Xí3 + e can be written as Y = "'ZX J3; + e , where {Ji ovcr those which contain -r 1 is
is thc ith element of í3· Let us exami ne X 1 {3 1 . T he elcments of the
n x 1 vector X 1 are only O's or l 's. T herefore, t he parameter {3 1 occurs
in t he nth observation ofthe model if and only if t he nth elcmcnt of X Next we e\·a luate
is equa l to l. Now exami ne X~Y. T his is the sum of the elements of
Y fo r those elements of X 1 t hat are equal to l. T herefore, thc term E(Y1 .) = E[ t
+ T 1 + rt.; + ev)]
(µ
X~Y can be found by s umming t he elements of Y over those clements in = 2 (µ + T¡ + CI.;)
the m odel that contain {3 1 . This samc procedure applies to all t he {3¡. j
Now t he normal equations can b e obtained by findi ng x ;v for cach {J;, = 2µ + 2-r¡ + CI.¡ + Cl.2
t ak ing t he expcctation of ea ch x ;v to obtain x ;xí3, t hen putting thc
'l'his gives the equa.tion correspondi ng to T 1, wh ich is
s ig n . . . . over each p a rameter. Th.is gives
X~Y = X{X~
The cquations fo r T~ a nd T 3 are obtained in a simi la r fashion. The
x~Y = x~x~ cquation corresponding to o: 2 , for e xamplc, g ives
The set of norm al equations consists of six equations, one corresponding Thc complete set of norm al egua.tio ns can nO\\º be writ.te n, as follows :
to each of the para meters µ , T 1 , T 2 , T 3 , o: 1 , rt. 2 . First find t hc normal ¡t: G(i + 27\ + 272 + 2f3 + :3&l + 3&~ = r ..
cq uation corresponding toµ. 'rhe ele ment of X 'Y corresponding toµ
is cqual to x;y and is t he sum of t he elemcnts Y i i over all t he equations
'T¡: 2µ + 27l + &¡ + &2 = y l.
in the model in w hich µ occurs. By examini ng the model we see that ' 2: 2¡1 + 272 + ª1+ &2 = J r2. (11.9)
µ occurs in every one of t he equations of the model ; hence, 'T3: 2¡1 + 2T3 + &¡ + ª2 = y 3.
CI.¡ : 3,ú + 71 + 72 + f3 + 3&1 - y .1
X~Y = 2,yii = Y ..
;; «2: 3{t + -í\ + 72 + f3 + 3&2 = Y.2
T he next q ua nt ity to find is t he expected value of Y... T his is The work of evaluating t he normal eq uations can be s hortened ma ny
times by finding on ly a represcntat ive equation in each set of theparam-
E(Y.J = E[~(µ+ T; + o:1 +e;;)] ctcrs. For instance, in t he exa mpl e above, Yii = µ + Ti + et.; + e;;.
thcre a.re t hree sets of parametcrs, µ, T, a ncl et.. In st ea.d of fi nding a ll
= 2 (µ
ij
+ Ti + CI.¡ ) thc normal equ atio ns, s u ppose we find t,hc norma l equation co rrespond-
ing to ¡i, t lrn,t corresponding to T 0 (the qt,h T), a nd that fo r e1., (the sth e1.).
= 6µ + 2 T ¡ + 2T2 + 2T3 + 30:1 + 3 0'-2 This givcs ro. for t he rig ht- ha nd side of T(I and Y .• for t he rig ht-h a nd
Putting the symbol . . . . on t he pa1·am et ers, we get for the fi rst cquation sidc o( e1. ,. H cnce we obtain
in t he set of normal equations
ft: G¡1 + 27. -! 3&. = Y ..
µ: 2µ + 270 + &. = r Q. q = l , 2, 3
Next we shall look for t he equ ation corresponding to T 1. Ti occurs a, : 3fi, + 7. + 3&. = Y .• s = 1, 2
248 LCNEAR S 'l'A'l'IS'.L'CCAL MODELS l\[ODEL 4: EXPERillIEN'l'AL Ol •:SIGN i\rODELS 249
where &. = L.&1 and f. = L.f;.· This is <L very eompaet representation The normal equations are g iven in (l 1.0). From these we see t hat
of t he nonnal cq uations and easy to obtain from the model. ,. _ T 2 and T 1 - T 3 are esti mable. H ence H 0 : T 1 = T 2 = T 3 is an
11.5.2 Solv in g the Norma l Equations. Sin ce th~ X ' X matrix c;timable hypothesis. Using the nonestima ble condit ions f 1 + f 2 +
is p X p of rank /..; < p, the~e are in fi ni t,ely man y vectors !3 tltat satisfy f = O a nd & 1 + & 2 = O makes thc n orma l eq uations easy to solve.
3 . •
t he n orma l equ aLio ns X ' X!3 = X ' Y. '~'e ncecl to fincl only one such The solut1on is
vecto r. Now, by Theorem 11.5, X' X~ represen ts a set of k lincarly.. fi=Y ..
independent estimab le functions, a nd every estimable fun ction is a f1 = Y1. - Y ..
linear combination of the rows of X'X~. L et V be a (p - k) x 11
matríx of ra nk p - le su eh t hat V!3 is a set of JJ - k non estimable func- f2 = Y2. -Y ..
t ions and such that no linear eombination of V!3 is an estimable func- f;¡ =Ya. -Y ..
tion. It can be show ri that th ere ex ists a vector ~ that satis fies the &1 = Y.1 - Y..
2p - le eguations X'X~ = X 'Y and V~ = O. This follows from the
&2 = Y.2 - Y ..
X ' X) .
fact t hat the matrix V has ra nk p, s ince V is of rank p - le (by the Therefore, t he quantity ~ 'X' Y in Theore m 11.15 is
(
way ít is chosen) and sin ce the rows ofV are linearly independen t ofthe R(µ,T,a) = f1Y .. + f 1 Y1 . +f 2 Y2 . +f 3 Y3 . + &1 Y. 1 + & Y.
2 2
r ows of X ' X or else so rne linear combina,tio n of V!3 wou ld be a set of The error sum of squares is
estima.ble fun ctions. No w V !3 can often be so chosen as to make the
solution ofthe norma l equations quite easy . Also, n onestimable func- y2 3 J'~ y 2) ( 2~ y 2 y2 )
t ions Vf3 can often be fou nd by examining the modcl.
~
,(,J
2
Y¡; fi 'X ' Y -_ "'
- 1-' y2
..(.J • ii - G.. - ( ;:¿ - '·
~1 2
- -·· -
G
:¿ -" - -··
1- 1 3 6
11.5.3 Exampl e. L et us find t he solution to t he n ormal cquations
g iven in Eq. ( L1.8), ,,·hi ch come from t he model To get ·r z;v, we use t he model Y;; = µ + +a; +e;;; wo have p u t T
Ti = T 2 = 3 T Now µ + will be replaced with µ* . The nor-
= T. T
y, 1 = /L -l T; + <1-; + e;1 i = 1, 2 ; j = 1, 2 mal equations are
Exam ination oftli ese normal eq 11ations re ,·ea.ls t hefact that Eq. (11.Sd) µ*: G,ü* + 3ii + 3éi"2 =
1 Y ..
plus Eg. (1 l. 8e) minus E g. ( ll. 8a) is cqual to zero; also, Eq. (ll.8b) }Jlus
Eq. {ll. 8c) minus Eq. ( ll.8a) is equa l to zcro. H e nce, t here are at
3P,* + 3ii 1 = Y. 1
most t hree lincarly indcpendent equat ions. 'l'he refore, we must find 3P,*
two linearly índependent equa.tions t hat are not estimable. By exam- These can be obtained from (11.9) by letting T 1 = T 2 = T 3 = O and
ining t he model it CéLn be seen t hat we can use using only t he equations for µ, <1-v and a 2 . The solution is easy to
obtain if we use the nonestimable condition ii 1 + ii 2 =O. The solutio n
IX1+ IX2) is
V!3 = (T ¡ + T2
p.*= Y..
\ i\fe could h avc fo und many different fnn ctio ns that co uld be taken fol' ii1 = Y.1 - Y ..
V!3, but we chose t hese because t hey make t he solntion of the normal
equations qui Le casy. Setting V~ = O g iYes &1 + &2 = O a nd f 1 + Also,
ª2 =-= Y.2 - Y ..
f 2 = O. Then, from Eq. (1 1.8ci) wc get µ = tY.. = Y.. ; from Eq. * -, , _* r J'2 ( 2 y2)
(Ll.Sb), 7- 1 = p·i. - Y.. = Y1. - !J .. · Continu ing, we get f 2 = Y2 . - R(,.'' ,<1.) = y Z9- Y = µ 1 . . + <1-_ l r . + _ - Y ·- = -·
1 1 0:9
G· + ;~1
9 ' -·
3 - -·
G·
y21
.,. DF SS F EMS
~
(b} Find t,wo lincarly indcpe nde nt est imable func tio ns .
(e) Find t,hc varianecs and eovariances of the estimates of thc fun ct.ions in (b).
1 11 .2 Fo1· t,hc model
-
Totnl 6 2
L Y;; Y;; = /1 + T¡ + Ó; + C;; i = 1, 2, 3; j = 1, 2, 3
;j
Due to ¡t, 'T, et. 4 Si = R(µ,T,et.) (a ) Find t he normal cquations .
Due to,,, et. 2 S 2 = R(µ.,a.) (b) Find two l ineurly independen t estimable fun et,ions .
(e) Fi nd Lhe variances and covaria nces of t,he estima tes of t he fu n etions in (b).
Duo to 'T (adj) 2 S i - S2 = 'Tss
Trns
-- 0'2 + .L ('Ti - i'.}2
11 .3 For Lhe mod el
E rns i
Error 2 L Y'f; - S1 = E 55
ij
r~ = l. 2
J = l, 2
l 2 .25
1 2 2 .25
l 2 2 . LO
F urthe r R eading
1 G. ' " · S ncdecor: "Stat.is t ical McLh ods ," Co wa StaLe Uollegc Press, Ames,
Iowa, 1!)56.
2 ' "· G. Cochran and G . l\L Co x: " Experi me n tal D cs ig ns," John Wiley &
Sons , Inc., New Yo rk, 1!)57.
3 O. K ompt,lio rne: " Desig n a nu Ana lysis of E xpcrime nLs," John " ' iley &
Sons, I nc., Now Yo rk, 1952.
4 C. H . Goulden: " Me thods of S t a t is Lical Analysis ," John Wiloy & Sons , Inc.,
N ew York, 1!)39.
THE CROSS-CLASSIFICATION OR F AC'rORI.AL MODEL 255
discuss t he two-way classi fi cation m odel a nd later extend ou r discussion
to n,n N-way classification model.
T h roughou t th is eha pter we sh a ll ass ume that t he re exis ts a set of ni.¿fi- + mé2 (12.2)
Obse n ·able ra nclo m Ya.ria bles y .. 1 \\·hose structure is s u ch t hat
l}Ull}q. ••
-
Yr iJm1,q ... t - µ dmvq ... t +e.
11mpq ... t1
wherc /l 11mJlQ
·· . .. t
a re a set of constants rx,:
and e tJmJ>Q
.. ... t a re a set of uncorrelated rn.ndo m Yariables with means O
and constan t Yariance a 2 • F o r confidencc internils a ncl tests of or, more co mpactly,
hypotheses we s ha ll make t he furth er assump t ion that t he e;;mpo ... 1 are
n orma lly distribu tecl. (12.3)
\i\' e s hall start with t he simple. t moclc l in terms of thc numbcr of
a0 : m 0 ¡1 +m & =
0 0 Y.,. q = 1, 2, ... , r
elemen ts present- the o ne-way classifi cat io n mod el, wh ich is some· Since the sum of t he last r rows of Eq. (12.2} is equal to t he füst row,
t im es referred to as the conipletely random?°zed model. W e s hall then t here is at least one linear depende nce. It can be seen also that
254
256 L CNEAlt S'l'ATIS1'1CAL MODELS 'l'HE CROSS- CLASSIF ICA'l'ION OR FACTORIAL i\IODEL 257
thc last r l'O\\"S are linearl y indcpendent. Thus, t he rank of X'X , + C"J., whcre <Y. 1 = C"J. 2 = · · · = C"J., = r1. . There is o nl y one normal
11
is r, a.nd ihere are r Jinearly independent estimab le fun ctions (by eqnat,ion, n n<l, by th e p rocecl u re of See. 11.5, this is
T heorcm 11.4). A lso, by T heorem 11.6, the quantities E(y;;) and any Mji* = Y ..
fL*:
linear fun ction of E(y;;) a re estimable. Thus, f' + r1., a nd a ny linear
fon et io n of t he µ + <:1.; are es timable. vVe can use Lhe (i + l )st cqua,. J{enec, R(1t) = Y .. fi* = Y~/J.ll. The AOV is presented in Table 12.1.
tion of t hc normal equations (1 2.2) a nd Corollary 11.G.2, \\"hieh states
t hat t he best linear unbiascd estima.t e is t he rig ht-httnd side of Lhc cor- TADLE 12. l ANALYSl S OF VATIIANCE FOR TESTING Gl1 = Glz = ... = rx,
/'-.-. -
r espond ing equation; t hat is, µ + r1., = Y ;./nt; = Y;.- If L:c; = O, S \' DF SS MS F EMS
then :Ec;C"J.; is estimable and t hc estím ate is 'f.c,y;.· Suppose 'f.d,C"J.; and -
Totnl 1\ (
Í: Yi1
L.f;C"I.; a,re t wo estimable fon ctions of the C"I.¡. Sinee they are both estim- ij
able, sorne linear eombination ofthe left-h and side of t he normal equa- 1
Y~.
Duo to mcnn -¡¡
tions ( 12.2) must yield L.d ¡&;. Thcn t he same Jinear eombination of the
r•
Ii - '· Y.'.
right-hand side of E q. ( 12.2) will g ive the best li near un biased estímate. Cl.ss CX.ms 1
Due to rx (ndj) r- l ass = a•+ - - ,2 m.(rx1 - ii.)•
The linear com bination is obtained by mul.t iply ing Lhe (i + l )st cqua- Jll 111¡ r- 1 11' DlS r - l .
'
t ion by dJ11i, and adding. F o r Zd;r1.; to be estima. ble it must be true J;rror ilJ - T ll"ss =
L Y,} - _L- yo _.!;
ir••
a•
ij i M-r
t hatL.cl; = O. Theestimateis ,2Y,.dJm,.. Simila rl y,theestimateof '11l¡
i
'Lf,rJ.; is .2: Y ;J¡/m;.· The covariancc of t he esiim at es is
i To find t hc cxpeeted mean square of treatments, eonsider t hat
/""'-... /'"--.. -2 d 1.¡.1
cov("2:d¡r1.,, 'f.f;C"I.;) = E[("i.cl,y,_ - "i.d¡r;.¡)("i.f ;Y;. - "i.f1r1.;)] = <rL - - y'~ Jr2 .,
m, rl.s.~ = I -'"
m¡i
- -··
JI
= I (Y; . -
;;
yJ-
T he two esti ma.tes a re uneorrelated if and o nly if
1 1 l
But Y;. = - Í: Y;;= - Í: (µ + C"I.¡ +e;;)=µ + <f.¡ + - Í: e;;
l.: dJ, = o ?n¡ J ?n; j ?n; :1
?n¡
1
If m; = ni, t he two estima.tes are uneorrelatcd if a nd only if "i.dJ; =O. and Y.. = -;;--[ Í: Y;;
.1• 1J
The Yaria nee of t he esti mare of "i.d;C"I.; is
If ni; = ni, t his variance of a eontrast is a 2 L.dUm. = l;f[t m;(µ + <f.; ) + fr e;;]
12.2.3 Tests of Hypotheses . Suppose it is desired to test t he
hypothesis <Y. 1 = <Y. 2 = · · · = C"J.,. By Theorem l l . l 7 , t his is a n csti m- l 1
able hy pothesis, since L:c;ix; is estimable if "i.c¡ = O. T he qua.ntitics
= '' + -M I; ?n¡C"I.¡ + -11! I;; e;;
l
needed a re R (µ,<Y.), R(¡t) , and fina lly R(C"J. ft) = ll(µ ,C/..) - R(¡L). To Substit uiing gives
obtain a solution t o t he Jtorma l equa.tions (1 2.2), we use t he nonesiim-
able eonditio n 'f.m;&; = O. This gives ¡1 = Y .. ; &a = Ya. - Y .. fo r q = E[fr (V;. - v.f]
1, 2, . .. , r. The reforc ,
y2 ( ) '2
R ( µ ,(/.) = ªt-' 'X'Y = Y . . ¡í + 'f.Y •.. &.' = -Jlf_·· -1- ¿....:i:. _ -··
y2) · [
= E .2:;; ( µ + a.; + m;
-l .2:; e;; - !' - -1 .2: m,,r:1. 1,
JII ,,
- -
l .2: eM
JII
)2]
?n · JI[ 1J(/
•
To find R(µ), we can write the model Y;; = µ* + e¡;, where µ * = = Í: m;(<Y.; - <i.)2 + (r - l)cr2
i
258 LINEAR S'l'A.'l'IS'l'I CAL i\IODELS 'l'HE C.dOSS-CLASSIFI CA'l'ION OR FACT ORIAL l\IODEL 259
?]
1
a set of machines or a set of chem ical t reatments; t he set {J; will be associ-
( r - 1) [
I m;(íl.; - a·+
)2
a- a.tcd with a certain factor , say, factor B, which mig ht be a set of temp-
}, = (r - 1) r - 1
cratures.
2a2 2 For example, suppose an experimenter wants to meas ure thc effect
of different o,·ens a nd d ifferent tempera t u res on t he strength of a certain
metal. If he desires to use fo ur tcmperat ures, !et these be factor A,
n,nd Jet the set of parameters be T t , T 2 , -r 3 , T 4 , the respective addit ive
lf m; = m, this is trcat mcnt constants due to temperatures. I f, say, three ovens are
available, let t hese be factor B , a nd Jet t he set of parameters {J 1 , {J 2 , {1 3
be t he respective additi ve treatment constan ts due Lo ovens. Jf µi s a
general constant, t he model might be wri tten as Eq. ( 12.4) , where Y;;
wherc - 1 is i he obsen 'ed strength of t he materi a l when t he ith temperatu rc is
íl., = - 2: (1.¡
r ; a,p plied in the .ft h oven. ·
Sin cc thc p ower of t hc test is a n in creasing fun ction of J., this formu la Sorne might p refer to writc the molle! as
fo r }, g ivcs an expcrimenter so rne insig ht into how he can improvc his
experimcnt, i.c., by either increasing mor decreasing a 2 , or both. Y;; = (¡,t + :r. + /J.l + (T¡ - T.)
To set a confidcn cc in tcrval on any treatme nt contrast L:c;íl.;, we can which we can write as
use the fo llowing facts :
Yu = µ* + r¡' + p; + e;; (12.5)
l. 'J:.c,y¡, is distr.i butcd N(z:c,íl. 1, a2:E e¡).
ni 1 'l'his is a. repara metrization of (12.4) a nd, by Coro ll a ry 11.8. l , \\'ill leacl
2. ll'ss/a 2 is distributcd as ¡,,2(111 - r). to thc same estimates of estimable fun ctions. Notice t hat I T{ =
3. ll'ss a nd :Ec1y;. are independent.
('J:.c,y 1• - :Ec1íl. 1)(.ilf - r)l
'1;.Pt =
]
Oand t hat E(¡ Yi;)
1)
= btµ *. 1\'l odel ( l 2.5) seems to be ~uperior
Hencc, for interpretation in m a ny experimental situations, b ut \\'e sha ll some-
_ ( W55 Í:.c~/m;) 1
timcs use model (12.4) to dernlop the theo ry. Modcl ( 12.5) can of
is d istribu tcd as Studcnfs f with JJ[ - r degrees of freedom. The l - rx course be deri,·ed from (12.4) . The e11 are rando m errors, which are
confide nce in teryal is assumed to be uncorrelated wi th mean O and varia nce a 2 • Fo r tests of
hypotheses and interval estimation, the errors wi ll be assumed to be
'f.c,y¡, -
TI'SS
la/2 ( - - - 'f. -
c7 )}< L C¡O'.¡ < °f.C1Y1. + la/2 (- w-·-:E -e¡)!
SS normally d istributed.
Jlf - r m1 fil - r m; The object in t his model is generally to estímate t he contrasts L C;r;
and L,d;fJ 1 and to test t he hypotheses T 1 = Tz = · · · = T 1 ancl /3 1 =
The width is ?[ _ _s_·s_ Ir ' ), ~2)l- P2 = · · · = flb·
-' a/2 (111 - r ,&..J
Total
1
DF
bt ZYl1
ij
SS MS F E.l\IS
must find R(µ,-r,{3). To solve the normal equations, we can use the two Y~.
nonestimable conditions :E-r;= O and :EP; = O. This gives fi =Y.. , }lean
bt 1
or
(12.7) _ Y~...
- pqrs
- + -qrs- -pqrs
- +
(t YL Y~... ) (t Y~;.. -Y~···)
-
prs pqrs
+ ri. +P. +Y. +f.) + PJ + (yk -
(2~.k. (2 Y~
Yiikm = (µ (oc; - &.) + ({J¡ - 'YJ
+ (Tm - f.)+ eiikm + k ~-··) +
___ m .. m _ __ ~···)
which we write as pqs pqrs pqr pqrs
where the eiikm are uncorrelated random errors with mean O and vari-
2222
m
Y~;krn -
i ; k
R(µ,cx,p,y,-r)
ance a 2 • = 2222m (Y;;km -
i ; k
Y; ... - Y.; .. - Y ..k. - Y ... m + 3Y ... J 2
By the procedure of Sec. 11.5, the normal equations are as follows:
Suppose we desire to test the hypothesis cx. 1 = cx 2 = · · · = cx.P. This
µ: pqrs¡1 + qrsL&¡ + prsLP1 + pqsLyk + pqrL7m = Y ... . is an estimable hypothesis, since LC;cxi is estimable if Lci = O. The
ex¡: qrs¡1 + qrs&¡ + rsLP1 + qs"Lyk + qr't-Tm = Y; .. . new model is
i = 1, 2, ... p Yiikm = ¡¡, + {Jj + Yk + Tm + eiikm
wliere we have substituted ¡¡, for ex. + µin (12.7), where we let cx 1 =
¡3¡: prsft+ rsL&¡ + prsP1 + psLy,. + prL-Tm = Y.J .. (12.8)
« 2 = · · · = cxv = ex. The normal equations for this model can be ob-
j = 1, 2, ... q
tained from the normal equations (12.8) by letting &1 = &2 = · · · =
'Yk: pqsfi + qsL&i + psLP1 + pqsy,. + pqL7-m = Y .. k. &11 = Oand by eliminating the equations for rx;. The normal equations
k = 1, 2, ... r are
71'1: pqrfi + qrL&; + p1·LP1 + pqLy,. + pqr7-m = Y ... m p,: pqrsfi + 11rsL/ji + pqs"L.yk + pqrLf rn = Y ....
m = 1, 2, ....~ {31 : prsfi + prs¡ii + ps'Lyk + pr'Lf m = Y.; ..
It follows that: yk: pqsfi + ps"L¡Ji + 11qsyk + pq'Lf m = Y ..k.
l. Ld,cx., is estimable if Ld, = O, and the estimate is LdiYi ... ; thc T m: - pqrÍÍ + pr"i:.íJ¡ + 1>q"i:.yk + pqrf"' = Y ... m
variance is a 2 Ldfjqrs. Using the nonestimable conditions '"i:/i1 = O, Lyk = O, Lf m = O, we
2. Lf¡/31 is estimable if Lf1 = O, and the estimate is Lf;Y. 1.. ; the vari- solve these normal equations and get
ance is a 2 Lflfprs. _
3. LUk'Yk is estimable if Lgk = O, and the estimate is Lgky .. 1;.; the µ=Y ....
variance is a2 LgUpqs. /J; = Y.1 .. - Y ... .
4. "LhmTm is estimable if Lhm =O, and the estimate is LhmY ... m; the Yk = Y .. k. - Y ... .
variance is a 2 Lh;,1¡pqr. :¡.m = Y ... m -Y ... .
Use of the nonestimable conditions L&¡ =O, LP1 =O, Lyk =O, "Lr m = This yields
O allows for easy solution of the normal equations. The solution is
R(µ,{J,y,-r) = fiY .... + 2 iJ;Y.; .. + 2 YkY .. k. + 2 fmY ... m
i k m
fi=Y ... .
&, = Yi ... - Y ... .
_ Y~...
- pqrs
(t Y~ .•
- + -prs- -vqrs
- +
1 Y~... ) (t Y~.k. -Y~- .. ·)
pqs pqrs
P1 = Y.1 .. -Y ... .
Yk = Y ..'/c. -Y ... . 2 Y~.. m Y7...)
+ (m ---
Tm=Y ... m-Y... . pqr pqrs
26-! LINEAI't S'l'ATI S1'1CAL MODET,S 1' 1[1<; CROSS-CLA SSIFICA'l'IQ N OR l•'AC'l'On rAL MODEL 265
W e h aYe R(et. 1 µ ,T,{3,y) = R (µ,Cl. ,{3.y,T) - R (µ ,{3.y ,T) 5qnarc; a nd a li are independc nt . Thcrefore, t hc AOV Table 12.3 can
bo nsed to 1cst t.he fo ur separate hy pothoses:
=
:L YL
-;_ _ _
yz
_ _
qrs 1iqrs (1) CI. ¡ = Gf.2 = ... = (1.p Noncentrali ty = qrs? L (('/.¡ - a.f'
2cí i
LLLL
i i k m
(Y. ;.. - Y ... f (3) Y1 = Y2 = ... = Yr Noncen t ra lity = pqs L (yk - y.)2
2a2 k
de n otccl by f:ls.~ , a n<l we co uld get similar results if we wa ntecl to test y 1 =
Y2 = ... = y,O I"T¡ = 7"2 = ... = T ,.
(·J.} Tl = 7"2 = ... = T s Noncentra li ty = pqr2 L (T - f )2
It is intercsting to obserrn tha t 2a m m ·
y2
_._
... + R (et. I p ,{:J,y.T) + R (f3 I µ,r:t.,y,T) + R (y 1 µ,et.,{3,T) + R(T 1 µ ,C1.,{3,y) The 11011conLrnlity parameters are obtained from the E l\IS column a nd
11qrs Thoorern 11 . 16.
Con fidoncc inte rvals can be set on a ny contrast by using Ems as the
+ E ss = LLLL Y7ikm
i ; k "' cstimate o f a 2 •
-The q uantity Y~ ..fpqrs is sometimes called Lhc smn of squares elite lo lite
mean, a n d R (r:1. I ft,{J ,T,y) is called t he swn of squares duelo <1. (adj ustcd), 12.5 Two-way C lassification with Interactio n
etc. T hus we can get t he AOV s hown in T a ble 12.3.
Beforc d iscussing the two-way classification wi t h intcraction, wc shall
T A BLE 12.3 ANAL Ys1.s OF VAlllANCE FOR Fouit-WAY Mon E t,
defi ne what we mean by t he term inleraction.
12.5.1 Defi nitio n of Interaction. To d efino the conoept of
~ \' DF SS l\I . El\18 intoracLio 11 , conside r t he f unction of two variablcsf(x,z) . \Ve form ulatc
1 a defin it ion as fo llows .
T otal 1 JJq rs I Yiikm
.,
ijkm • Definition 12. 1 f( x,z) wi ll be defined to be a funotio n with no inter-
- actio n if a.nd only if t hcre exis t functi ons g(x) and h(z) s uch t h at
J\fean 1 -Y~-..
pqrs j(x, z ) = g(:1;) + h(z) .
For exrLmplc, Lhc fun ctions x 2 + xz, x 2 + log z + xz 2 , e::'', and e=+:i: h a ve
qrs
r.t. (udj) 1' - l « sg GCms ª2 +--I(r. t., - ci.J2
- p - l i
interaction, lrn t t he fun ctions x + z, log xz, a nd x 2 + 2x + z2 + 2z
fJ (udj ) q - 1 fJss fJms a-n+ q-vrs-I
- l j
(/J · -
'
fJ- )-.
.
ha\' C no irücracLion. T hc a bove defin it io n can be extended to any
num bcr of \'ltriab les; i. e., the fu nct ion f(x,it,v, ... ,z) h as no interaction
71qs I -. ifj(x ,u ,v , ... ,z) = h 1(x) + h 2 (u) + h 3 (v) + · · · + h 1(z), a nd has inter-
y (adj) r - L Yss Y111s a-• + - - (y. - y) -
acLion oLhe r wisc.
r - 1 k k ·
The thing to 11otice in t he model wit h no in tcraction ,J(x,z) = g(x) +
(acJj) s - L npqr I (T
a-+-- _)"
T 7 ss 7 ms s - 1 ,,,, m.
- T
.
-
h(z), is Lhat,, fo r t"·o x values x = a and x = b, t he guantity f(ct,z)
f(b ,z) i ~ indcpc ndc n t of z. Simila rly fo r two z values.
E rror pqrs - p - q - r - s +a 1
E ss Em~ a2
Tite l,wo- way classifi cation modcl can be wri ttcn as ?'J ·¡ = µ .. +e .. .. t " ,,,
whcrc µ;¡ i:; Lil e ' ·t,rne'' total e ffect of t he combin a t ions of elc ment i from
By T heo rem U .15, t h c quantit ics ('/.ss/a~ , flss/a 2 , Yss/a 2 , a ncl Tss/a2 are factor A a rH.l olc rnc 11 L.i from factor B. If t his " t rue" total effect is j ust
each d istribu tecl as a noncen tral chi -squarc; B ss/a2. is a central chi- the su n1 of Lhc ith effeet of A, which is T¡ , p lus t he jth effect of B , which
THE CROSS-CLASS1FICA1.'ION OR FACTOR(AL i\TODEL 267
266 LINEAR STATCS TlCAL l\lODELS
In the two-way classifi cation mode l without in tcraction, t he contrasts
is fJ;, p lus p, Lhcn /.t;; = µ + T¡ + {J,. To consid e r t his more carcfully,
of t he -r¡ a nd ofthe /3; we re estima ble. By examining Lhe norma l equa-
we see t hat
tions for model ( 12. 9) it can be seen t hat here the contrasts of t he T ; or
and the /J; are not estimable. In other words, t he p a.ramcter (T/3),,. cannot
bo separated from T ,,. The model can be written as
(t Y~
j •J , . ' ·
+ L(r /3);; - (-r{J);. - (rfJ). 1 + (-rf/l..] + e;;,,. R (¡t*,-r*,{3*) = Y~ .. + (:¿y¡.. - Y~..) + (:¿Y~;. - y~·-)
rbt ; rb rbt 1 rt rbt
or Y;;,.. = µ* + -ri + f37 + (-r{J)i; + e;;,,, (12. 13)
iind R[{r/3)* 1 µ* ,r *,{J*J = R[µ* ,T*,(J *,(-r{J )*J - R(µ*,r*,{J*)
It " ·ill b e s ho\\'n that t he re is no interaction in the modc l if and only if
(-r/3) 1 ; - (T/3};. - (-r/3).; + (-r/3) .. is cqual to zero for al i i and .i· By = I Y7;. - I Y; .. - I ~;. + ~..
;; r , rb J rt t·bt
Definition 12.2, \\'C m ust show that (-rf3)ú = (r{J);; - (-r/3);. - (rf3). 1 +
(-r/3) .. = O foral! iand j if a nd on ly if y* = (µ;i - p.¡) - (µ¡p - fl¡·i·) = ,¿,
°" (y d.
. - Yl.· . - !/ .J.. + !/ . .. )2
iim
= O for a li i, i' , j ,j', w hcre µ;; = fl + -r¡ + fJ; + (-r{J};;. Now, if y* = O.
'J'hc error sum of squares is
)r 2
btl (;Y * = - - µ;.
µ¡; - µ,; - + µ,.
- = O for ali i and j " 2
LYi;m
iim
- R[ µ *,T * >fJ *' (T {J)*] -_ .,¿_,!/;;,..
" ~ _ 4.
ijn¡
" -
i1
;1. -_
r
" ( Y;;,,. _
L.,
ijm.
Yii. )2
270 Ll 'J~AR S'l'A'l'fS TJ CAL l\IODELS 'J'Irn CR OSS- CLASSIFICATION OH. FACTORIAL lltODBL 27 1
This leads to t he AOV prcscntcd in Ta ble 12.4. TABLE 12.5
TADLF. 12.+ .-\ 1'A LYS I S OF VAR f :\NCE FOR TESTINC lNTERACTION ~T{3"' 1 2 1 3 y .
.J.
J N A Two-wA Y C 1.ASS1FIC ATrON -------------
3 4 3
SV ] )lo' SS E~I S :1 :l 4 l 8 :30
2
-
2 (j
Total rbt L Yf,,,. ---------
ijm
(j 6 8
Duc lo C-r{J¡• (udj) (1 - 1 )(b - 1) ¡ (Y¡,. - Y•.. - !J.1. + Y... J' a•+ (b _ l~(I _ l)¡((T/j)t¡]' 2 3 2 5 2 13 15 22 50
I
~ y
¡ 4 7 !)
Error bt(r - 1)
11 m
(Y11m - Y, ¡_)• a•
- :1
- - --- - - - -
3
4
3
G
5
G 3 Jl 16 20 47
T h e n oncentrality para.meter is 4 7 9
- - - -,------
4 :1 7
4 ij 7 8 4 13 25 !í7
4 !) 10
If it is conclu cled that t hcrc is no in teraction, modcl (1 2 .'.1,) is approp riatc
and Sec. 1 2.3 a pplies; othc rwisc model ( L2. l L) is a ppropriate.
If it is decidcd that t he i nternction model ( 1 2. J 1)
y .... 45 5!) 80 184
(a) Entrios a ro obsor vat,ions ?Jm,· (b) E n t,rios are coll totals Y;;.
~- L; Y~;- = 984.222
This is estim a.b le , a nd its bcst lin ear unui ased estímate is Y; .. - Y; .. ·
Li; L Y7;m =
tn
1,118
12.5.3 Exarnple. Assume t hat a n ex pc rimen t was run o n Lhrec T hercforc, we h ave
OYens, ea ch at four tc m peratures, to ascer tain t he s tre ngth of t hc fi na,!
product. S uppose thaL Lhree runs wcre made in each oven nt ea ch SS duc to (-r/3)* adj = L L (Yi;.
ii m.
- Y; .. - Y.;. + Y .. f
te mperatn re. The obser vations a.re g iven in Table 12.5a.
vVe sh a ll ass ume t hc rn oclcl = } L Y ;;. - "'h L YL - t L Y~;. + a11r Y~..
ij i j
( i = l , 2, 3
= 5.388
j = l , 2, 3, -1:
{ Also, the error sum of squares is
'In = J , 2, 3
(12. 14)
CD" T he norm al equ ations for t he model written as (12. 17) can be fou nd
...... froin (12.1 5). T his gives t he group of equatio ns la beled (12. 18) on
~
......
¡:::
page 27 6.
¡; o The ra.nk of this syst cm is IJ K . T hc estimat e of the p ara met ers
~ :¡;;
,;;- cil using t he normal equations (1 2. 18) is thc same as the estimate of t he
::l
+
,......,
O'
¡;q
corrcsponding clements in (12.16) by using t hc_.!)ormale quat ions (12. 15) .
:
For example, from (12. 18) we have
r;
,......, r;
li But, by (l 2.1 G),
-T7 = Y; ... -Y .. ..
1
I~ I~
~
li ...'.:,.
+ + I~
-r( = Ti - T. + (-r/3); . - .(-r/3l.. + (-ry);. - (-ry) .. + (T/3y); .. - (-r/3yl. ..
+ which can be written
r;
" -" "'! "'!
+
-
li li I~ I~ I~ I~ -r( = [¡.i + Ti+ P.+ (T/3);. + Y. + + (-r/3y); ..l
(-ry);.
li
" ·~
·~ "" ·~ -+ - [µ + f . + P.+ (-rf3l.. + Y. + (-ry) .. + (T/3yl...J
clear from t he norm al equ atio ns i n (12. 15) that contras ts of t he -r; o r
1
- 1 - the {31 or t he y,. a re not esti mable (that is to say, t here exists no u n biased
I~ - I~
·~
I~
l "'! 1 estimate of LC;T;, where LC; = O) nnless t here is no interaction. T here -
+ I! I! I~
·~
fore, it may be desirable to test the fo llowing hypotheses concerning
I~ 1
$
•:t:L + + + 1 l model (12. 14):
•
:t:L 1 ~ l. µ; 1,. is free of three-factor in teraction ; t h at is, (T/3y),j" = O for a ll
+ 1 1 11:-
1 ~ ~
I?--
~
i,,j, le
11:- ~ ~ ;>-.. CQ..
P.*, = 1 .
Y., .. - 1
.'J .. ..
A* =
y,. Y .. k. - Y ....
, . R)
._ *. . = Y u·· .. - Y··i ... - Y . ,·. . + Y . ...
+ ( 7 IJ i:;
,,..._ *
(7y);k = Yi.k. - Y; ... - Y ..h·. +Y ....
, ,.,. . .__ *' = y ., - y. - y k + y
({Jy) Jk ·'•' · .].. . . ·. . •••
+ + ....... *
(7{3y) l,.)/,; = y."l}ºk•• - y 11.... - •1'J 1... k•• - y .}/..;.
. + y . + 11 . + y • - 1'. J ....
t. . . ,/ ., . . ..,..·.
-- fl'*Y .... + .L
' A*Y + '.L FR*; Y .;. . + ''*
7'¡ i ....L Yk Y .. _j
k. T °"
L, (R)*Y
TF ;; i j ..
i i k ii
+ L
i;k· m
(Y;;¡.,,, - ?/;;k.)
2
= E ss
If we set (T{Jy) ú1.: - Oin model (12. 17) , the norma l cq1mtions are exactly
the same as (1 2.1 8} with t he equations for (T{3y)~k omitted. The rank
of t he resu lt ingsystemis J JJ( - (1 - l }(J - 1)(1{ - l }. Also , from
+ + this reduced system we have
íl* = Y ....
+ + 7;-* = Y; ... - Y ....
iJt = Y.; .. - Y ... .
r: = Y .. k. - Y ... .
,....., *
(T ¡_.R) .. = Y l}.
·· . - Y 1.· .. - y . i . . + 71
1 11 ,/ . . . .
.. .. .. .. ~ *
.,..:3_ *l-·. . ..,<:Q..
..........
*;:....""
(7y);k = ?Ju·. - Y;... - ?J ..1.-. + Y ....
,....., *
({3y);k = Y.;k. - Y.; .. - Y ..h-. + Y ....
276
278 LINEAR STA'l'IS'rICAL l\'IODELS
'l'HE ClWSS-CLA SSIFICATION OR FACTORIAL l\lODEL 279
This g ives
¡ Yk L YL ¡ Y~; . . y 2
,.,L.., y2.
1J..
,
,L,
y~
1.k.
'yz..
.L., .JI.: .
'L.,y2
'· ·· -- 'L (y'·. '' · - y •·. ·· - y .. ~·. + Y .... )2
1 and DF = (l - l)(K - 1)
= ;; JC111 + ik JM + ik IM - ~Kllf
i}km.
*:::-
';!.., ';!..,
';!..,
-'<
larlY for t he other a;b;ck.
"" ;:;:""
il'.·::-
*·-
..:. *·~
~ *""
2:: &;: *·-
;:;: *"' ;:;: 'fhe model w.i U be written
..:. ..:. ~ <c..
tA{·- ..:. + e;Jkm
<--:
:....,: l."".
..... t::> ~¡
~
• ..
tA{·...
t::>
C\l
H "-' tAt·-.,
~,
..,
~""
...t::>
~
...... 1..t::>
.-;
·-
~
w~
--- ¡..t::>
- .
~
W .""'..,
......
.-; 1..t::>
~
W "'=~
...... ,.~
Y;;¡.,,. = µ iik
where ?/;;k·m is t he m.t h o bscn ·ation of t reatme nt a ;bjck :.tnd where µ iik is
~ C\I >-, IN
...... "'' ~"" the " t ru e" cffect, c5f this part icula r trcatment. '~'e shall write
µ ;;k = µ + T¡ + f31 + (r{3);j + Yk + (ry );k + ({3y );k + (r{3y);;k
= µ * + r7 + {3j + (r{3)~ +y!+ (ry )7k + ({3y)jk + (r{3y) ~k
where thcse quantities are d efin ed by Eqs. ( 12. Jü) and (12.17). The
norma l equations are g iven in (12.15) with I = .J = ]( = 2. We shall
~
..."'I~E
*¡: ., *ªI *E"'I .,E
ce:.. r:qE :--..~ ~1
"';:-~6 ~ k1 ~"'Ir:q .f:
*61 e~ ~ E *61 kl-.,
~
...
~"
use t he model
Y;;k>11 = µ* + r( + {3j + (r{J)7; + Y: + (ry)~. + ({3y)j,. + (r{3y)~k + eiikm
--- The normaJ eq uations are g iven in J.!:q . ( 12.1 8), whcrc I = J = ]( = 2.
Table 12.8 will be nsefuJ in cxam ining the non.tal equations .
., .,
., "' *E "'
(/)
......
- ,_ *E"' "' *ª
&;: *"
;:;: ;:;: *;:;:ª ., ' l ',\DL.1!: 1 2.8 DEFr N t TtON O}' T t m EsnM ATi,;s i N 23 FACTO RIAL
~ *ª ""- *ª;'\. ... ..:. ~
<:c..
..:. ~=
1
Y 111 · I- - - -1'121·- - - -
1'211· Y221· Y112· Y212· Y122· Y222. z
- - --- - - - ---
- S.lf ({t) * + I + l ...¡ l -1 l + 1 + l + J + l Z1
4.11(Á) - 1 + l - 1 ...¡ 1 - 1 + 1 - 1 + l Zz
W(Ü) - l - l + l + l - 1 - 1 ...¡ l + l
E ...... Z3
(/)
(/)
""
c-1:::- ..·1~....,....,
('!.
~ :.:: .......
. '-.-4...,
*'...!: • :i
""-
*::;
;-.
*;:¡
&;:
..:.
~
;:...
..:.
*=
;:;:
~
~
;>...
,_
"'>.
"'
~~
4.ll(AB)
HJ(G)
-! l
- 1
- 1
- 1
- l
- L
+ l
- 1
+1
+ 1
- l
...¡ l
- 1 + l Z4
·- 4M(AG)
......
+ I - 1 + I - 1 - 1 +1 - J + l ZG
- 4M(B0)
_,,.......___
+ l + I - 1 - 1 - 1 - 1 + l + l z,
......
4M(ABC) - 1 + I + l - 1 + l - 1 - 1 + l Z9
1
;::;- 1
._
- - -
;::;-
1 1
......
1 ~ -
- ._
~
1
Note: A n exam plc of t h e use o f t he t a ble is
4í\I (B) = ( - l) Y 11 1. + ( - l )Y211 . + ( + l )Y 121. + ( + l)l'22 1.
""~ ¿ ~ ~
R
A
~
...... 1
......
1
...., ~
1
- - -
1 1
::::::
1
¿
1
~
~
+ (- l )Y112. + ( - l ) Ym. + ( -! l )l'122. + ( + l )Y222. = Z3
::::. ...... ¿ -1 ~ There a re fonr independcnt unui ascd est.ima.tes of Tz - T¡ if all the
interaetions are zero. Thcy are Y 211 . - Y 1 1 1.. Y22 1 . - Y121 .. Y212 . -
~
·~
.s
·~
"O
..::.
·~
ü
..::.
~
* . *;:;: *
;:;:
:...
o:...
E (YL. - Y1. .J = r: - ri
Ti - Tl + (T/3)2. - + (ry)2.
o &;: ;'\.
""-
~ "'
...... *,_ *
""- *;:... ..:. ..:. ~ ..:. ~ = (rfJ)1.
"" - (ry)¡ , + (r/))1) 2 .• - (r{3y )1 ..
280
282 LlNEAR STATIST I CAL i\lODELS 'l'HE CROSS-CLASSIFICA'l' ION OR FACTORIAL i\.iODET... 283
If t h e thrce-faetor inte raction is zero, t here are two indepenclcnt un.
b iased est ima.tes of 4(-r{J)-:'1- T hey a re 12.8 Using Interaction Sum of Squares for Error Sum of
Squares
Yu1. - Y2 11. - Y 121. + Y22i. and Y u 2. - Y212. - Y122. + Y222 .
In a three-way factorial m odel as explained in See. 12.6, t he total
For example,
number of observations is IJK111. Often it is impossible to repeat
E[Yw.- Y211 . - Y121. + Y221.J = 4(-r{J)ti t he t reatment eombinations, a nd then :A1 = 1. If this is the case there
since (T/3)71 + (T/J):¡ = {-r/J)tl + {-r/3):2 = Ü is 110 estimate of a 2 and no m ethod for testing h ypotheses or setting
and (-rfJ)i1 + (-rf3)i2 = (•fJ)iz + (-r{J);"!. = O confidence intervals. However , in many experiments the researcher
is willing to assume that sorne of the interactions are zero . Then the
If t he t hree-facto r interaction is not zero, t hen AB in Ta.ble 12.8 can corresponding sums of squares can be u sed asan est imate of a 2 • Sup-
be taken a s the estimate of the A B interaction averaged over thc e pose, for example, there is no three-faetor interaction; t hat is, suppose
facto r. The square a rray in T a ble 12.8 is an orthogona l mat ri x if cach (Tpy)(¡,. = O for all i, j, k. T hen the noneentra li ty fo r t his term in
elem ent is di vided by vs.
Let the z; be so defined t hat t hey are cqual Ta.ble 12.7 is zero. Therefore, (T{3y)is/ a2 is distributed as x 2(p), wher e
t o t he ind icated linear eombination in t he i th row. Th en t he z, are p = (1 - l)(J - l)(K - 1). Then (T{3y) 8~ can be used as the error
independent normal varia bles with variance 8.111a 2 . T he adj uste<l suni term.
of sq uares d ne to the m eanµ* is zi/8111, a nd s imil arly for t hc othcr If 111 > 1 a nd (-r{3y)/J1.: = O, then (T{:Jy) :',; can be pooled with E ss to
effects. This leads to the AOV shown in T a ble l 2.9. Simil ar methods give an error term with IJ K(M - 1) + (I - l )(J - l )(K - 1) degrees
a re applicab le for n > 3. of freedom. It must be remembered t h at t his is under the assumption
t ha.t (-r{3y)u1.: is zero for all i, j, k.
TABLE 12.9 ANALYS l S OF VAnrANC-E OF 2' FACTOlllAt.
Suppose t hat an experime nter does not know whether t he t hree-fac-
sv DF tor intera.ctio n is zero, but he decides to assu me t h at it is zero if t he
appropriateF test from Table 12.7 is not significant a.ta chosen proba-
To tul 8Jlf bility leve!. T hen, if he decides on the basis of the.F test that i t is zero,
he will pool (T{:Jy):'s wit h Ess a nd obtain a new error t erm for testing
~
z 1- main effects and two-factor interaction. If t his rule is followed, t he
8 .1\J tests for main effects and two-factor interaetion a re not exa ct. Tl1ese
z~ tests are of the nature of condition al tests, si nee the decision to p ool _
8M (T/1y)is with E ss depends on a preliminary test of sig nificance. The
2M reader is referred to the bibli ograp hy at the e nd of this chapter fo r
B -2 'L((Ji)2
a further reading on this subjeet .
z2
_ 4 * 2
111 'L[(-r/J)¡;)
Next we shall show t hat, in a two-way model wi th o nly one observation
AB -;;
per ij combination, if the interaction is assunied to be zero b ut is
8!\l a-
z~
really not zero, it vit iates the test of s ig nificanee. Let us go back t o the
e SM model given by E q. (12.4). Under t h is model t he error s um of
z2 squares d ivided by a 2 , thatis, E ss/ a 2 , is distrib uted as x2 [(b - l )(t - l )].
_G
AC If, however, t he model contains a two-factor in t eractio n t e rm a nd is
SM
z2
given by Eq. (1 2.9), t hen Ess/a 2 is distributed a s x' 2 [(t - l)(b - 1), ?.],
_ 7
BC l where
81lf
~2
Hence t hc error t erm is incrcased on t he average, nnd the test will not l l :J lG 20 20
reject the hypothesis as often as t he leve! of sig nificance indica.tes. 2 22 25 29 29
3 25 30 34 32
Similar situations can be shown fo r othcr m odels.
Problems (a) T csL t.ho h y poLhesis -r1 = -r2 = .,.3 = .,.4 with a t.ypc I 01-ror probabili t.y of
5 por cent..
12.1 Fine! t,h e no rmal e quatio ns for t,ho reparamot,l"i7.cd m odcl of i,ho onc- wa.y (b) Set. a 95 por cen t. confidcnce inLorval o n .,.:i - .,.4 •
classificat io n (e) Find th o csLimaLes of -r¡ - -r. (i = 1, 2, 3, 4).
Y;; = ti* + a.f + C¡; (d) Find t,ho n 1riu nco of cach of tho ol<timat es in {e ).
(e) Find 1,ho width and cxpected width in (b).
" ·h ore ¡L* = 11 + ii. and a.f = a.¡ - ii (i = 1, 2, ... t;j = 1, 2, .. . 1·). 12.8 Find Lhe nonccn Lrnli t.y pa rnmctcr for Prob. 12.7a.
12.2 In t he onc- way c lassification m otlc l 12. 1, find Lhc ;;u m of :;quarcs duo lo o: 12.9 \\"ork ouL Lho fou r 11o ncontra liLy p a 1·am c t c rs for Tablo 12.3.
{adjustcd) for t c:;t,i ng thc hypoLhesis H 0 : a.L = a. 2 = · · · = a.k. with k < t. 12.10 B y us ing D cfi n it,ion 12. l , show lhnL. iff (x,::) has fir;;L parlia l d e rinltives,
12.3 Assume t hat Lho d a ta of Table 12. 10 fit, a o nc- way clu~ificatiorr modcl: thcn ¡¡, is a fun ct.ion wi(,h 110 intc rac t.ion if off 0.1: is indc pcndcn t. of z anti off oz is
inclcpend nt o f x.
TABLE 12. 10 12.11 <..:o Lh ro ug h Lhc algebra to .find a solut,ion to Lho no n nal e quations
(12. 10) us ing t h o noncstimablo condit.ions
Treatm on t ........
f1 = ¡11 = (-rfJ )1. = (-rfJh = .. . = {;fil1. = {-r/3).¡ = ... = (:;{J).b = o
12. 12 "\\'0 1·k o ut, Lho no ncenLra lit,y pai·amct.cr for Table 12.4.
12.13 A ssumc t hltL t.ho data of T a blo 12. 12 saLis f)' Lhc t.wo-way c lassification
12.3 11 .4 15.2 1 4. :~ 15.3 rnodcl
14. l 10.9 IG.8 l ü.2
1:3.G 12.0 18.6 ) 4.~
bi li t,y of .05. 6 8 11 11
(11) Set a 95 per cenl confidcncc intcr va l on °'• - IX 2. (j 10 10 9
(e) Fincl 1¡ + &.; (i = 1, 2, ... , i;) ami tho vu ri n11co o f euch p:;t,imnle. ---- --·- - - - - - - - - -
(d) Fincl t h c powe r of t,hc tes t, in {a) if 11 11 14 l::l
2
12 12 Hi lG
for i = 1, 2
---- --- --- ---
fo r i = :3, 4, r; 13 JG 18 16
:1
12 14 lü 17
(e) T est, t.hc h y pot.hcsis H 0 : ~ 1 = °'2 = °'~·
12.4 J 11 t,h c two-wa.\· classificalion mo del g iven in A r L. 1:l. :1. 1, find tho
t reaLmen t, su m of squnres for tcs t ing H 0 : -r 1 = -r2 = · · · ~ .,.,, whc rc s < t. Test, t,ho h y pot.hci-; is t.h nt. Lhe rc is no int.c ract.ion at. th o 5 por cen t leve! of
12.5 Fi ncl tlw 11o nccn t rn li ty fo r J'ro b. 12A. signifiC'a n<:c.
12.6 ] n a t,wo -way cluss ifica t,io n m od c l as d r íinccl in A rt.. 12.3. l , work o uL in 12.14 ~ h ow t,hat, m odel (12. 14) is t he s amc as ( 12. lü).
d ot u il t,h c cx pcctcd m ean Rq11ares t,hat. are g ivcn in T i.iblo 12.2. 12.15 t;olvo t.ho no rmal cquat.ions ( 12. 18).
286 LINEAR STATIS TICAL J\CODELS
Further Readin g
1 G. \V. Sncdccor: "SLatistical Methods," Iowa Sta.te College P ross, A rncs, 13
Iowa, 1956.
2 \ V.G. Cochra11 a nd C. l\f. Cox : " Experiment al D osigns," J ohn \Viloy & Sons
lnc., New York, 1957. '
3 O. K c mpthorno: " Dcsign and Ana lysis ofExpcrim en ts," J ohn "Viley & Sons Two-way Classification with Unequal
I nc., New York , 1952. '
4 C. H . Could en : "Mothods ofStatistical Ana lysis," J ohn Wiloy & Sons, lnc.,
Now York, 1930.
Numbers in Subclasses
5 O . L. Davios: "Dosign and Ana lysis of Industl"ia l Experimonts ," Olivor &.
Boyd, J,td. , London , 1954.
6 R . L. Andorson a ncl T. A. Bancroft: "Statis tical Theory in Resoarch,"
McG raw-Hill Book Company, Inc., New York, 1952.
7 C. R. Rao: " Advancecl Statistical Mcthods in Biom etric Resea rc h ," John Chn.pter 12 was devoted to a study of various cross-classification
Wiley & Sons , I nc., Now York, 1952. modcls w it h the com mon feature that in each cell there were a n equa l
8 H. B. l\Iann: "Ana lysis and Design o f Experime n ts," D over Publ icaLions, In man y cases iL is 11ot possiule for <Ln experi-
11 umbcr of observations.
New York, 1!)49.
9 A. E. Pau ll: On a Prelim ina ry T est for Pooling l\Iean Squares in t he Analysis
mcntcr to contr ol his expcri men t so as to ensure this type of balan ce.
of Varianco, Ann. filath. Statist., ,·ol. 21, pp. 539-55Ci, 1950. Also, a resear cher may purposely design his experi ment to havc un-
10 H. Scheffó : " Thc Analysis of Va ria n ce," John "\Viley & Sons, I nc., Ncw cqual numbers in t he s ubclasses. This is t he case in incompletc block
York, 1959. dcsigns. The two-way classifi cation mode l wit h unequal numbers in
the cclls and w ith no interaction will be discussed in sorne genernlity in
this chapter. Certain special cases, s uch as incomplete blocks, wi ll be
discusscd in later chapters.
{
~_ O, ~ , 2, .. . , n;1
yijJ, = µ + T i+ {3; + C¡;k J - 1, 2, ... , U (13.1)
1: = 1, 2, ... 't
-
wherc Yiik is t he kth observation in the ijth cell ; µ, r;, /3; are u nlmown
para meters; and e ;;k are ra ndom variables with th e conven t io nal dis tri-
butional properties. The ijth cell contains nii observations, a nd, if
287
288 LINEAR S'l'A'l' ISTICAL JllODELS TWO-WAY CLASSIFICA'l'lO - \\"ITH. U""°EQUAL NUM13ERS IN SUUCLASSES 289
n .. = O, t he cell conLains no observatio ns; he nce, t he obsen ·ation y .. 13.1.2 Point Estimation. In th is section the normal equations
,, ''º will be dcri ved, esti mable fun cLions \\"ill be examined, a nd the variance
<loes not exist. The notation
b of certain estimable functions wi ll be fo u nd.
N.... = 'L., n l.1 By t he method of Sec. 11. 5, t he normal equations are
;= 1
t
l<l • J ·= ' L.,
i=l
n 11.. N.Ji. + L N¡_T¡ + L 1\ .;il; = Y ...
i i
N .. = 'L., n .. IJ NT.fi.+ N,_-TT + L; n,;P; = r ,.. r= 1, 2, .. . 't ( 13.2)
ii
w ill be uscd. Notice that
U¡¡ 'IL;¡
{J.: N_,¡1 + I n ;_,f; + N .•Ps = Y .•. s = 1, 2, ... 'b
i
Yii. = L ?/;¡1.: = L (µ + -ri + fJ; + ei;1.:) = n¡;µ + n;¡-r; + ntJf31 +e;;. To exam ine t hese eq uations in d etail we s hall m ake the following
1.: m o k= O
n¡;
assumption .
whc rc L C;;k = C;;.
k= O
• Assumptio n 13.1 The n;; are values such t lrnt -r; - -r1 is estimab le
Also, we sha ll use t hc notation fo r every i =/= j = l , 2, .. . , t and {J; - f3; is estimable for every
1 n;; - i =/= j = 1 , 2, ... ' b.
- L C;;1.: = C;;.
T hus a. des ig n s uch as Lha.t s ho\\"n in Table 13. la is admissib le, but
n;; k=O
t hat of T able 13. 1b is not. T he sym bol *_in a cell mea.ns tha.t Lhere a re
Simil ar noLaLion wi ll be used fo r e; __ , e; .. , e.; .• e.; .. etc. Also notice t ha.t
b 11¡¡ b
Y; .. = L L Y;;1.:
i= l k = O
= L n¡;(µ
i= l
+ + f3;) + L C;;i.: 7"¡
Jk
TADLE 13.l DES JC N S FOlt ExPERJME:\TS, TwO - WAY CLASSJt"I C ATJON
MODE L
Y.;. = L
t
i= l k = O
lt ;;
L Y;;1.: =
i=l
'
L n;;(µ
i
*
o 1
3
o
*
t
= N_;µ + :¿ n;;-r¿ + N.;(31 + e.;. (ri) (IJ)
i=l
t b 11¡;
a nd Y - L L L Y;;" = L L n ;;(µ + -r; + {l¡) + e ...
i=l i= l k=O i i
observations in t hat particula r cell. The symbol O means t hat t here is
no observation. In general , experiments will satisfy the cond itions
= N __ µ + _L N;.-r¿ + _LI N,lJ; + e.. .
i
set fort h in t his a.ssumption. If t he assumption is not satisfied, a,ll
parametcr d ifTe rcnces a re not estimable; in thaL case l:he experiment
Also, rnight be bro ken up inLo parts sueh that the parameter d ifTeren ce in
cach part is estimab le.
l
Y; .. = ~ LL Ym· • T h eorem 13. 1 H the n;1 i 11 model ( 13. 1) a re su e h t hat r 1 - -r,,
11 i . ; k
1
[ji' - fJj. :ire esti ma ble for a li i =/= j and all i' =/= j', t hen
Y.; . = N.
.
¿Y;;1.:
]
f (1) T here a,re exacLly b + t - L li nearly independen t esLi mable
fu nctions .
1 - (2) Í:c;-ri a nd í:d;f3; <trc estim able if Í:c; = í:d; = O.
Y .. . = 1"' LLL Y;j/,
l.V i j k Proof: It is in1 medi a.Lcly olea r t hat Lhe b + t - l esLimable fonc-
A study of t his notation will help in t he t heory Lhat follows. t io ns -r 1 - -rz, -r1 - -r3, · · · , -r1 - -r,, /J 1 - f3z, /3 1 - {33, · · · , f31 - fJb,
290 LINEAR STATISTICAL MODELS TWO-WAY CLASSlFICATION WITH UNEQUAL NUI\IBERS IN SUBCLASSES 291
and N .. µ + 2 Ni.-ri + 2 N.;P; are linearly independent. There Jf we isolate the quantity involving f r from the second term, we get
i j b 2 t b
are b + t + l parameters and, therefore, b + t + l equations in :u ...
lv r.Tr -
... " ' nri
Tr L., - -
"' L.,
L.,
"' -nr;nu ,..
-- 'T¡
the normal equations (13.2). The sum of the t equations repre- 1=1 N . i=t 1=1 :N 1
sented by -r r is equal to the equation represented by µ. Also, the Therefore, Eq. (13.6) becomes ., ¡,¡,r .
with the fact that there are at least b + t - l linearly independent The system (13.8) representa t equat.ions in the t unknowns f,. They
estimable functions, implies that the rank of the normal equations can be represented in matrix form as
is exactly b + t - 1. Since the -r i - -r; are estimable, every At=q (13.9)
linear combination of these is estimable. Consider where the t X t matrix A = (ars) has elements as follows:
t 1
2 (-ri - 'T;) - j =f=. i Diagonal elements: a=N-L.,-
~ n~¡ r = 1, 2, ... , t
;=1 t rr r. i=l N .
This gives b
·' (13.10)
t-1 1 t t-I 1 _ Off-diagonal elemen ts: a rs = _ L.,
" nr/ns; r-=/= s = 1, 2, ... , t
- - 'T¡ - - 2 'T; = - - T¡ - - (T. - 'T¡) = 'T¡ - 'T. 1==1 N. 1
t t i=l t t
i>Fi The t X l vector t has elements f r' and q has elemcnts qr, as defined by
This shows that -r¡ - 7. is estimable for i; so l:c,(r, - 7.) is ali Eq. (13. 7). If there are equal numbers in the cells (say, m), then
estimable. Butthisis LC¡riif LC¡ =O. Asimilarargumentholds nii = m, N. 1 = tm, Ni. = bm, and N .. = btm. In this case arr =
for Ld;P;· bm - bm/t and ara = -bm/t, and the system of normal equations is
easilysolved. This case, in fact, was discussed in the precedingchapter.
We shall solve the equations for contrasta of the -ri. Vve can use the There are other situations in which the ni; values are such that the
P1 equations and obtain normal equations are fairly easy to solve, for example, in incomplete
µ... + PA; =Y.;. - N1 "' n¡¡'T¡
4 " j = 1, 2, ... 'b (13.3) block models. However, in the general two-way classification model
.J i with unequal num bers in the subcells, the matrix A has no special
The -r r equation is feature to make the normal equations easy to solve.
1 nr1(P, + Pi> + Nr.fr = Yr.. r = 1, 2: ... , t (13.4) The system At = q will be studied in sorne detail, since from this we
i must obtain estimates of estimable functions of the r i and the variance
Substituting µ + P; from Eq. (13.3) into Eq. (13.4) gives of these estimates. In Art. 13.1.4 it will be shown that a solution to
this system will be needed for the AOV table used to test the hypothesis
r = 1, 2, ... , t (13.5) T¡ = 72 = • •• = 'rt•
L et A*_ = (Bª R
1
1
,) and using (13.1 4) g ives lB 2 1 = 1', o r B 21 = {l/t) l ' , which m eans
B 21 B 22 tha,t each elem entin B 21 is eq na l to l/t.
2. .To s how t ha t B 22 = O, mu lt iply Eq. (13. 13d) hy 1' . This
where Bu is a t x f. m a.trix, B 12 is l X 1, B t t = B~ 2 is 1 X f. , and
gives l ' AB 12 + 1'1B 22 = O, or tB 22 = O by (1 3. 14a).
B 22 is l x l. T he 11 , from (13 . J 1),
3. l\1ult iply (13.1 3a) b y_B 11 ; t his y ields B 11 AB 11 + B 11 1B 2 1 =
B 11 , or B 11 AB 11 = B 1 v since B 11 1 = O by (1 3. 13b).
4. By Eq. (1 3. 13a) we get AB 11 = 1 - {l/t) ll '. Thus, t hc
diagona l elem en ts are equal to (t - l ) /t, and t he off-diagonal
and ( 13. 12) elements are equa l to - l /t.
5. Eq. (13.13b) implics that t he rows of B 11 a dd to zero.
Next " ·e sh a ll p roYc a t hcorcm co11ccrn i11g t hc cle mcnts of A *- 1 . Trus concl udes th e proof of t hc theorcm.
+ T h eorem 13.4 Thc ele me n ts of A *- 1 a re s u ch t hat • Theorem 13.5 In Eq. (1 3. 12), E{f;) = T ; - f ..
(1) B 2 1 (a11d B 11!) h as a,ll its c lc me nt.s cq ua,I t o l//. P roof: S incc A-T = q is deri vcd fro m linear combinations of t hc
(2) B 22 = O. norma l e q uations, it follows t hat E(q ) = AT. H ence,
(3) B 11 AB 11 = B11 ·
(4) AB 11 is id c mpotcnt ofra nk t - l , wit h d iagona l elements each
e qual to (t - l )/t :wcl wit h o ff-d iagonal e lc mcnt s ca.ch egua.! t.o
- 1/t.
(5) The r ows of B 11 acl e! to zcro . so the ith e le ment of E(i ) g ives E(-Ti) = T; - f.
294 LINEAR STATISTICAL MODEr..s rrwo-WAY CLASSIFICATION WITH UNEQUAL NU1\1BERS IN SUBCLASSES 295
• Theorem 13.6 cov(-i') = cr2 Bw where -i' = B 11q. the nonestimable condition 7t =O. This immediately reduces the
Proof: By definition, system to t - l equations by ~rossing out the tth equation in (13.9) and
setting -Tt = O.
cov(-i') = cov(B11q) =E{['~· - E(-i')][-i' - E(-i-)]'}
= E[(B11q - B 11A't')(B11q - B 11A't')'] + Theorem 13. 7 If the nonestimable condition 7 t = O is used with
the system of equations Ai' = q, the solution is such that E( f i) =
= B 11E{[q - E(q)][q - E(q)]'}B11 ri - rt and cov(-t-) = a 2 Bj1 , where Bf1 =(A **)-1 , where A** is
Let C = cov(q) = E{[q - E(q)][q - E(q)]'}; then the matrix A with the tth row and column deleted. If 7 s = O is
used, a similar situation holds, with t replaced by s.
cu= var(qi) = var(Yi .. - InuY.J.) The proof will be left for the reader.
i=l
13.1.3 Example. Suppose the normal equations have been re-
= var(Yi.J + var(f
i=l
n,1Y.~.) - 2cov(±niiY.;.' Yi .. )
i=l
duced so that the system A-t- = q is
b 2 b 2 27¡ - f2 - f 3 = 12
= Ni.<12 + <12 2 nu - 2a2 2 nii -71 +72 = -5
;=1 N.i ;=1 N. 1·
- f1 + 73 = -7
= (Ni. - ±N.;
n~i)a2 2-1 -1)
Also, for r ::j=. s,
1=1
Then -A=
(-1
-l
l
o
O
l
crs = cov(qr,qs) = cov[( Yr .. - I rlT;Y.1.), (Ys .. - I nsl·Y.k.)]
j=l \ k=l
Ifwe use the nonestimable condition LTi =O, we get
2 -1 -1 l
= cov(Yr_,,Y11 .J - cov(Y ,Inr;Y.¡.) - cov(Yr.. ' ±nskY.rr.)
11 ••
j=l k=l -1 1 o
A*=
+ cov( .± nr;Y.i.' I nskY.k.)
i=l k=l
-1 o 1
l 1 1
= O-
11
2 "" ns;1tr;
<1 L . , - - · -
b
2 "" nsinri
<1 L., - - -
+ <12 L.,
b
"" nsinr;
--
. i=1 N. 1 ;=1 N.; ;=1 N.; 2 -1 .;._ 1 3
and - 9.i. -1
A*-1 - 5 -4 3
-1 -4 5 3
We have, therefore, shown that C = Aa2 •
= a 2 Bw by Theorem 13.4. ·
3 3 3 o
The procedure for solving the normal equations is as follows: So 71 =4 E(-91 ) = r 1 - f.
1. Eliminate µ and P;
and obtain A-t- = q. 7 2 = -1 E(-92) = 'T2 - f. (13.15)
2. Augment this by :Efi =O; this gives A*.
73 = -3 E(fa) = ra - ;¡.
3. Find A *-1 and -t- = B 11q. Then E( -T¡) = ri - f. and cov(-t-) =
~(-: =:)
a2Bu- -- -
Using the condition :E-Ti =O means that we must solve a system of (13.16)
cov(-i-) = -:
t + l equations for t + l unknowns. The work can be reduced by
using other nonestimable conditions. For example, suppose we use -1 -4 5
'º l
296 LINEAR STATISTICAL l\IODELS TWO-WAY CLASSIFICATION WITH UNEQUAL NUJ.\'IBERS IN SUBCLASSES 297
For example, After finding a solution to A-t = q, we can substituto these values
into Eq. ( 13.3) and obtain 11 + P1 , from which contrasts of the {3, can be
var(71 - 72 ) = var[(71 - f.} - (72 - f.}] estimated.
= var(f1 - f.} + var(72 - f.} - 2 cov(71 - f., 72 - f.} 13.1.4 Test of the Hypothesis Ti = -r2 = · · · = ,.,. To test
= ;a2 + i}a2 _ .¡-( _ a2) = ª2 this hypothesis we need to evaluate R(µ,T,{J) from Eq. (13.2) and then
to evaluate R(µ,{3}, the unadiusted sum of squares duetoµ and {3, from
This solution rcquired inversion of a 4 x 4 matrix.
Next let us use the nonestimable equation 7 3 =O. Thus, A-t = q
Eq. (13.2} with T1 = T 2 = · · · = T 1• Now_R(µ,T,/3) = fiY ... + ! 7\Yi ..
... - i
is immediately reduced (eliminate the third equation and set 7 3 = O) to + 2 {JjY.;., where µ, f i' /31 are any solution to Eq. {13.2). This gives
2f1 - 72 = 12
j
usR(µ,1,{3) = ¡1
(P, + /3- 1)Y. 1. + ::E7\Y;.. , and, using the value of µ + P;
-1\ + 72 = -5 from Eq. (13.3), we have
'Ve shall denote this system of equations as A'f = q. The matrix is 1
R(µ,T,{3) = Lb ( y;. . -
; =i
- L n ..7. Y · '.• +:¿.;..Y.
t ) t
A.= ( 2 -1)
-1 1
N .1 i =t u i i =i ' ' ..
But E(&.3) = -füT1 - Ta) + (T2 - 7"3) + (Ta - 7"3)] = Ta - r. µ*: N ..µ-* + .L,
"'°' N .i{J- j -- Y •••
i=l
So -3 is an estímate of -r 3 _ - r.. Also, let &. 2 = {'r 1 + &. 3 ) = 7 - 3 = 4. N.si1* + NfJ-=Y s = 1, 2, ... 'b
.s s .s.
But E(&. 2) = ,- 1 - 7.. So 4 is an estimate of T 1 - 7.. By the same
process -1 is an estimate of T 2 - 7.. These are the same estimates as Using the nonestimable equation 'LN.;p; =O gives the solution
those of Eq. (13.15). This verifies the fact that the estimates ofthese
estimable functions are the same whether the nonestimable condition fi*=Y ...
7 3 = O or ::E7; = O is used. Also, if we use f 3 = O, we get
f'Js=Y.s.-Y ...
cov(~) = a2( 1
1 1
) b y2.
2 and R(µ,{3) = y µ-*+"'°'y. ¡J.="'°' --d.:
••• ¿,,
i=l
·'· ' ,¿,,
;
N .i
so cov(f 1 , f 2) = a2 ; var(f 1 ) = a 2 • . But it can also be shown, by using
Therefore, the adj usted sum of squares due to T is
the covariance matrix in Eq. ( 13.16), that
var(71 ) = var(71
= :a2
-
+ ~a2 _
f 3 ) = var(71 )
(2 )(-!)a2 = ª2
+ var(73 ) - 2 cov(f1 ,73 )
R(.,. I µ,p) = R(µ,.,.,p) - R(µ,p) =.t.-r•( Y, .. -,tn;;YJ.)
where 7-i is the solution to Af = q with :E-Ti = O.. = ',.
t
.L, T¡qi
i=l
= "''q
't'
29 LlNEAR STATISTTCAL l\CODELS TWO-WAY CLASSIFICATlON WITH UNEQUAL NUl\lDERS IN SUilCLASSES 299
By Lhc Lheorems of preceding arLieles, we can wriLe T'q = T' Ai""' can be used to find E(T55). This gives
q ' B 11 q . \\ e ha Ye proYed the following theorcm.
• Theorem 13.8 The adjust ed sum of sqnarcs for testing the E(rss) =E(¡ Y7;k) - E(¡ Y\) - (N .. -
•ik , N .
b- t + 2
l )a
hy pothesis T1 = T 2 = · · · = Tt is L. f;q¡, whe rc f ¡ is any solution ·'
to AT = q.
The AOV is g i\·en in Table 13. 2.
'l'AllLE 1:3.2 A NALYSTS 01' VAR!A NCE FOii TWO·WAY CLASS IFICA'.l'ION - [ ~ µ
'( + '11 . ·T · + N
4-...!1...! v_,·)2 + ba2]
_ ._,.R - (N .. - b- t + l )a-.,
\\· 11·11 UNEQUAL NmrnEus IN THE SunCLASSES
, , N. 1 N_;
2
sv DF SS i\IS F ).
1 (;¿ n¡;T¡)
'l'otal ¡\' .. I Y7;k = L' N . T 21·
¡ '·
- ....
L
;
i
N
+ (t - l) a 2
;jk 3
1
¡
J
1\T~;·
'Tm ll
--;
2a·
i [ Iij (
11¡1 T¡ -
I11 '"fT
-
JJ
;V _,. 'rhe best m ethod fo r solving AT = q seems to be to set one of t he f¡'s
equal to zero (say, we set 7- 1 = O). Then t he sol ut ion f ¡ is a n unbiased
¿ ' ·" ~ '"
O u.· t o -r (ndj ) / - 1 Tm 11:
1 i= l Em1 estimate of T¡ - T 1 , a nd a 2 t imes the ijth eleme nt of t he inverse matri x
is t he eovaria.nce off; - 7- 1 a nd 7- 1 - 7- 1• The reforc, we s ha ll discuss
Error N .. - b- t+ 1 Subtrnction = E 15 Em•
1 setting confidence limits on a linear combination of T ¡ - T 1, sa.y, on
:Ec;(T; - T1). We can use t he fact that L.c¡f¡ is distributed
'l'he abbrcYiatecl Dooli ttle technique dcscribed in Art. 7.2. 2 can be
1
uscd fo goocl aclvantage to obtain T Q. Th is is seen from t he fact that
from Lhe eq uaLion X'X~ = X ' Y the Doolittlc tech nique gives ~ 'X'Y.
The p roccd ure is to lct f, = O a.ncl eli mina.te t he la st equa.tion in where b¡ 1 is the 1)th element ofthe matrix B t 1 in T heore rn 1:3.7 a.nd f; is
A-T = q . Th is ]caves (say) the solution to t he equatio ns AT = q a.ug men ted by f 1 = O. T his
gives
L.,
<ik
Q
Yi;k = I
;
__:!_:
N.1
•
+ Tss + E ss
(13.18)
y2 where ta.J-J. is t he approp riate value of Student's t with N __ - ú - t +1
or Tss =I Y¡;k - I _:!..: - E ss
;;1.; ; N.1 degrees of freedom.
300 LINEAR STATISTICAL l\10DELS -. rrwo-WAY CLASSIFICATION WITH UNEQUAL NUMBERS IN SUBCLASSES. 301
-:}
13.1.5 Example. First, suppose the reduced equations A-t =q the data ofTable 13.3b and Eq. (13.10) can be used to obtain A. The
are as givcn in Art. 13.1.3; that is, system At = q is
2f1
-71
- f1
-
+ 72
f2 -
+
f 3 = 12
73
= -5
= -7
~-(-: ~:
-3 -10
-3) (7~2
-10
13
1
'7"3
)
=
( -!)
+1~
-tí
If we put 7 3 = O and omit the third equation, we are left with Jf we let f 3 =O, we get
2f1 - 72 = 12
- f1 +72 = -5
This yields 7 1 = .¡, which is an estímate of T 1 - T 3 , and 7 2 = ¡, which
Thesolutionis f 1 = 7, 7 2 = 2; and R(µ,-r,p) = t'q = (7)(12) + (2)(-5) ¡6 an estímate of -r 2 - T 3 • The adjusted sum of the squares dueto T,
= 74. If we use the nonestimable function LT¡ = O, we can use Eq. which is R( T 1 µ,{3), is
(13.15) and get R(µ,-r,p) = t'q = (4)(12) + (-1)( -5) + ( -3)(-7) =
74. U-)(-i) + (§)( 16~> =U= 1 .06
13.1.6 Example. In a two-way classification with no interaction, The in verse of the matrix is
the data in Table 13.3 are to be analyzed.
1 3
- -- -- -
1
1 4
From Table 13.3a, the sum of squares for µ and (3 unadjusted is
- - --·- - - 66.33
2
6
2
41
1
2
16 2 2 2 2 6
LY~;.
; N.;
+ +
= (12)2
4
82
3
62 =
4
-- ---- The total sum of squares is 86. The AOV is given in Table 13.4. To
3 1 3 2 6 3 1 1 1 3
12
__s_j_6_ - - 26
--;;-4-1_3_\_4__ l_l_ TABLE 13.4 ANALYSIS OF VARIANCE OF DATA IN EXAMPLE OF ART. l:{.l.6
Y.;.
SV DF SS MS
(a) Entries are Yiik (b) Entries a1·e n;;
Total 11 86.00
The normal equations are: ))ue to I'• p (unadj) 3 66.33
Duo to T (adj) 2 1.06 .53
uµ+ 2.;1 + 672 + 3.;3 + 4/31 + 3/l2 + 4p3 = 26 Error 6 18.61 3.10
2 µ + 271 P1 Pa = 4
6fi + + 2p1 + 2p2 + 2p3 = rn
672 set a H7 .5 pcr cent confidence interval on T1 - T 3, we can use Eq.
3fi + a-r3 + P1 + P2 + Pa = 6 (l:U8) with c 1 = 1, c2 = O. We get
4,U + 71 + 272 + 73 + 4¡31 = 12
!- 2.97 (3.10)(1()) :::::: 'T - 'T :::::: ! + 2.97)(3.10)(16)
3fi + 27 + 7 2 + 3¡3 3= 8 2 9 18 ~ 1 3
~ 9 18
4ft + f 1 + 27-2 + 73 + 4¡33 = 6 which reduces to
The quantities ,U + /l; can be eliminated from these equations, or else -4.82 ~ TI - 'T3 ~ 5.04
302 LINEAR S TATIS'l'ICAL l\TODELS TWO-WAY CLASSU'ICATION WlTH UNEQUAL NUJIIBERS I N SUllCLASSES 303
with t he col u mn totals N .• as elcments; C 22 is a d iagonal matrix with
13.2 Co mputing Ins tru ctions the ro" · totals N,. as c lements. C 21 is a t - 1 X b matrix eq ual to th e
body of Table 13.5. A lso, c; 1 = C ¡ 2 . T he sum of s qua res fo r e rror
In t he p receding sections t he t hcory of t he two-way classification wit h for testi ng t h e h ypoth esis -r 1 = · · · = -r 1 can be obtained by red u cing
unequal numbers in t he su bclasses h as bcen d eveloped. Now our the system
attcntion will be turncd to t hc problc m of computing. T he proce<lurc
wrn ue to impose condit ions on the norma l eq uations to m a ke t hem ful! (
Cn C12 g1) I
C21 C 22 g2
rank , a n <l to use t he D oolittle technique. F r om t hese we sha ll obtain by t he abbreviated D oolittlc techniq uc and using Theore m 7.1. We
estim ates of estimable fu nctions, t he covariancc m atrix of t h ese csti- sha ll ill ustrate w it h t he examp le of Ar t. 13. 1.6. From d ata in Table
mates, andan AOV table. 13.3b t he fo llowing are co mputed
1 2)
~ (~ : :)
T A B LE 13.5 NUMDER 01'' ÜDSl -:RVA'l'lONS I N THE ijth CE LL
~ 1 2 ¡_3
1_ ... b
--- º" C12 =
( O 2
1 ... N J. 1 2
2
3
11¡¡
n z1
71 31
n12
n22
1132
n¡3
n23 . ..
. ..
n lb
n2b N2. F rom t hc b orde r t otals of Table 13. 3a we get
~. ~ (1:)
?133 11'3b N3.
g.
-
= (
16
4)
t - 1 N1 - 1.
- - - - - -- - - - -- - - - Thc for mat fo r the Doolittle techni q ue is given in Table 13.6. T his is
N. 1 N.2 N.3
... N .b
TADLE 13.6 DOOLl'.I'TLE T l!:C lfNI QUE FO Jt TESTINC "T¡ = .,.2 = ... = "T¡
Supposc t hat t he -r; are t hc parameters t hat a re of in te rcst in modcl Row C2 C3 e,, Check
(13. 1). The no rm a l equation s are g ivcn i n Bq. ( 13.2). T hey will be
C1
1 1 1
Cs
ªº
reprod uced here, except Lhat t he -r, a nd {3, equ ations w ill be inkr- R~ 4 o o l 2 12 19
cha ngc<l. They are n; 3 o o 2 8 13
R~ 4 1 2 G 13
¡,t: - NJt + I N.;P; + I N¡_f¡ = Y ... R'4 2 o 4 8
; i
R~ G 16 28
N_,¡1 + N .•P. + I 11 1,f, = r .•. s = 1, 2, ... ' b
R¡ 4 o o 1 2 12 19
N,.fi + 2, n,;P; + N,_f, = Y, .. /" = 1, 2, . .. ' t r¡ l o o ~- { 3 ll
4.
j
C 21 C22 g2
R ,,
:¡
-~- - 1 -~ o
r4 J - 2 - 1
o
w hc rc p = (P.) has dim ension b x l ; g 1 = (Y _,.) has d imension b X 1 ; 3 3
-
-r = (f,) has di me nsion (t - 1) x 1 (f 1 isomitted );an dg 2 = (J',..) has Rs 2 -~ ll
d ime nsion (t - l } x l. T hc C ;; have t hc n, 8 fo r ele m ents. T hc ele- rs 1 -5- 3
5
1 1
ments of C ¡; can be fou nd from Table 13.5. Cu is a d iagonal matrix:
304 L l NEAR S'l 'ATISl'ICAL l\IODELS TWO-WA y C LASSIJo'JCA'l'!ON W I T H UNEQUAL NU.MilERS I N suncr,ASSES 305
fu ll rank , a nd t he quantities R(µ,T,{3) and R( T 1 µ ,{3) are easily computed 13.13 Jn Prob. l 3.G, find Lhe value of t h o unbia sed es timatc oí r 1 - r 3 , us ing
from the 0 0 column. tho cond it.ion :Ef; = O.
13.14 l n P ro u. 13.G, find t he \·alue o í t ho unbinsed estim a.te o f' r 1 - r 3 , us ing
R(µ,T,{3) = + ( - Í)( -~") + (G)(~) + (~)(8) + (3)(12) = G7.31)
( ~)(·~ ) tho condition 7'3 = O. Compare t his value wi t h t ho vttlue obtain ed in P rob. 13. 13.
13.15 I n Prnb. 13. 13, fi nd Lhe covarianco of Lhe es t,imat.c o í r 1 - r 3 .
and R(T 1 µ,{3) = 0)(5) + ( - !)( -~) = ~ ~ = l .06 13.16 ] n P rob. 13. 14, fi nd t he covariance o í t,h e es t ímate oí r 1 - r 3 •
13.17 Using t,he equat.ions in P rob . 13.6, find t he sum o[ squaros for r (adj)
These yaJucs are, of course, t he same as t hose given in Art. 13. 1.6. by using
(a) 'J'he condit ion .E7-; = O.
Problcrns (b} The eondit ion -T3 = O.
13.18 Usi ng t.h e data in Prnb. 13.1 , test Lho h ypot.hosis r 1 = r 2 = r 3 wit.h a
13.1 The dat,a g i\·en in T able 13. 7 a re assumed to follow a two-way classifica- 5 por cen t type I e rror probability.
t.io n m o<lel w it.h unoqun l numbet'S in t ho subclnsses. 13.19 Using the <lnt,a in Prob. 13. l , set. a. 95 per cen t confidonce inwrval on
1'¡ - T3·
TABLE 13.7 13.20 S how t,h a t. A + (l/t)J hns a n invers a, where A is defined in Eq. (13. 10).
13.2 1 In Pro b. 13.20, sh ow t h at B 11 + (l /t)J is Lhe irn-erso, whoro B 11 is de -
";:-( l 2 3 4
--- --- - - - - - - - - -
finod imm e d iat,ely fo llowing Theorem 13.3.
6 7 3 7
Further R eading
1 8 l 10
4 l\l. G. Kendnll: " T he Advanced Theory of Statist ics," vols . T, II, Charlos
- - ------ --- - -- C: ri ffi n & Co., L td., L o ndon , 1946.
5 6 4 2 C. \\'. S ned ocor : "SLat ist,ical Meth ods ," Iowa Stat,e Collogo P ross, Ames,
2 8 J o wa, 1958.
7 3 O. K empt ho rne: " D es ign an<l Analysis of Expol'Íments," John vViloy & Sons,
- - - - - - - - - --- - - - Inc ., New York, 1952.
4 R. L. Ande1'Son n nd T. A. B a nernft: "St,atis t,ical 'l'heory in R esearch,"
3 5
l\fcG r a w-Hill Book Company, Inc., New York, 1952.
:-l ()
5 C. H.. Rao: "Ad vancod Sl,atistical l\foLho ds in D io met.ric R osoarch," John
' Viley & Sons, I nc., Now York, 1952.
6 !:i. Ko lodziej czy k : O n a n Important C luss of SLat.is Lieal JTy pothoses,
Find n l l' 1112• '11 ¡ 3, n t 4 • n21> n22 • n2:1' n2 4• n :11> n 32• n 33• a.nd n3 4'
13 .2 In P ro b. 13 . l , fln d N 1., Nv N a.• N. 11 N _2 , N _3 , a nd NA. JJiometrika, vol. 27, 1935.
13.3 ln P rob. 13. 1, find Y 1 ... Y 2 __, Y3 .. , Y_ 1_, Y_ 2 _, Y_3 _. a ncl Y. 4 . • 7 H. Sch e ffé : " T he Analys is of Va riance," John Ví' iley & Son s, Inc., New
13.4 In Prob. 13. 1, flnd t.he normal equat,io ns. York, 195 9.
13 .5 In P rob. 13. l , fin<l t,ho equat,io ns A ~ = q.
13.6 In a two-way cla ssification model wit,h unequa l n umbors 111 t ho s ub-
c lasses, s upposo LllflL t ho sot of oqunt,ions A~ = q is g iven by
l Oí\ - 2f2 - 8r3 12
- 2f1 + 5f2 - :3T3 = 16
- 8f 1 - 3i2 + l 1T3 = - 28
(a) Find U1 0 rank of th is s.rstem of oq uatio ns .
(b) I• ind a solut io n , us ing t,he eondit,io n ~;,. = O.
13.7 In P rob. 1:3.6. c:o ns Lrue t, Lhe A * m a trix g iven in Theorem 13.3.
13.8 In Prob. 1:3.7, find t.he rank o f A *.
13.9 In P rob. J :1. 7, find (A *)- 1.
13.10 U s ing t,ho c quat,ions in P rob. l :3.0, Ye rify a li t he statement.s m a d e in
T h eorem 13.4.
13.11 P 1·oyo Tltoorom J:l.:l.
13 .12 Using t.he data in Prob. 1:3.6, find a so lu tion to the eq uations b y us ing
t he cond it,io n f 3 = O.
JNCOJ\IPLE'l'E BLOCK J\lODELS 307
Jnstead of g iving other exarn p les, we shall talk in ter ms of t wo gener al
foctors: b locks a ncl t reatments. T he blocks "·ill be similar to ovens,
t wnshing mach ines, cte., t hat is, factors t hat are not undcr study by th e
csperi menter . On ly cases will be considcred in which each block
contains k experimental units (sometimes referred to as le plots). T he
11 0 rn ber oftreatmcnts will bedenoted by t; thcrefore, k < t. Also, onl y
those cases will be considcred in wh ich each t reatment is replicaíed
(a.ppears) in exacLly r b locks, a nd t he n umber of blocks is b.
14 There are vario us types of incom plete b locks, but wc shall d iscuss
0 11Jy what are termed balanced incomplete block.s.
" .here n;; = l if thc ith t reatme nt a ppears in t he jth block, a nd n;; = O 1 Since t hc ba la nced inco mpleto block model h as bcen defined as a
if t hc i t,h t reatment <loes not appear in the jth block. y iil will be uscd specia l case of the two-way classification with unequa l numbers in the
to rcprescnt t he o bservation in the ijth col! , a nd y 00 will m ean that subclasscs, t hc normal equations can be deri ved fro m those in Eq. (13.2)
there is no observation in that particular cell. T¡ \\"ill represent the ith by using t hc 11otation of Scc. 14.1. The normal cquations are :
treatment constant, and /3; will rcpresent thcjth block constant. The
error terms e;;,,. will be assumed to be un corrclated, with mea ns O and ¡t: bkfl + r 2: f; + k L; P; = Y ...
variances a 2 , for point esti m ation. For interval cstimation and testing
hypothcses, the e;1,,. will be assumed to be normally distributed. The 'T J): r¡í + rfv + L; nv;P; = Y,,.. p = l ,2, ... , t (14.3)
following equations are seen to hold :
1V t.. = "n" .. = r
.L., lJ
(number o[ blocks in which the ith
treatment a ppcars)
1.:13"s -- y ,5, 8 = i ; 2, . ..'b
i= l
1
(number or trcatments that appcar in Thc cqu ations reprcsented by f3. can be solved for µ + p•. This gives
N •1 = "
~ ntJ
.. = k
i= l thc jth uloek) (14.2)
N = rt = bk (total numbcr o[ observations) 8 = ], 2, ... ' b (14.4)
The quantity .¡. P will be factored out of the second term on the left of the
equals sign. This gives
A= L nPinii
i=l
ifp # i
(r - -kr) ,.
T
'P
-
l ~
-
k
~ T·
,. =
i=l '
v
.t
j)..
-
1 ~
-
y i
k jk; ; ¡ n p i ••
p = l, 2, ... 't (14.6)
(14.8)
i"ji; V
t
Thc nonestimable condition 2 f¡ =O can be used to obtain
From this we get
i=l
i#p
and (14.10)
If this is substituted into (14.6), it gives
r
( r - -k + -kÁ.) f p
= Y j),. - -k ;f;-1
""' n JI1 Y;• •
1 b
p = 1, 2, ..• 't (14.7)
The coefficient off P is (rk - r + l)/k. If the value of l in terms of r, =-- forp # p'
k, and t is used, this expression becomes J..t/k. The quantity k
1 b 1
Fron:i these facts we get the following theorem.
Y p.. --""n
k k
i=l
'Pi Y .; • =Y,, • • --B
k ,, + Theorem 14.1 In the balanced incomplete block model given by
Eq. (14.1), thc linear minimum-variance unbiased estimate of any
will be denoted by qv· This is the tot~l ~f the observations of the pth contrast of the -rP is given by (k/A.t)'Lc 11qp. The variance of the
treatment minus I/k times the total ofthe observations of ali blocks in estimate is (ka 2 /J..t)'Lc;. If the e;;ni are normally distributed, the
which the pth treatment occurs. The quantity q'P does not contain µ or estímate is maximum likelihood and is the best unbiased estímate.
312 LINEAR STATISTICAL l\IODELS INCOl\IPLETE BLOCK l\IODELS 313
4.4 Computing
14.3 Interval Estimation and Tests of Hypotheses
We ha ve considered the balanced incomplete block modelas a speci The computing for the balanced incomplete block model is quite easy.
case of the two-way classification model with unequal numbers in t nstruct a format such as Table 14.3, and for each treatment compute
subclasses. Therefore, the AOVis givcn by Table 13.2. The quantit e total Yi .. and the total of ali blocks that contain the ith treatment.
Lí\qi is (k/).tp:..q¡, since JJi = (k/A.t)qi is asolution to the system Ai' =
The labeling in the AOV table will be altered slightly. It is found TABLE 14.3 FoRM FOR CoMPUTATIONs
Table 14.2.
'.rotal of blocks
TABLE 14.2 ANALYSIS OF VARIANCE OF BALANCED lNCOMPLETE BLOCK
0
,
'1.tment
.umber
Total containing qi qi(~) = ~i - ~.
treatment i
sv DF SS MS N oncentrality
~ B1 )~
1
1 Y1 .. B1 Y1.. - -k B1 ( Y1 .. -
Total bk 2Y'f;m
ijni
Mean l
Y~.
bk
2 Y2 .. B2 Y2 .. - -:B2
k'
1
( Y2 .. - ~ B2)~
Blocks ¡Y.1_ y2
1 •••
(unadj) b - 1 -k- -bk
1 1
Treatments t Yt .. B, Yt .. -¡ Bt ( Y, - -1 B, ) -k
.. k At
(adj) t - 1 Tms 1
SV DF SS MS F N oncentrality
TABLE 14.5 COl\lPUTATIONS FOR THE BALANCEO INCO.l\IPLETE BLOCK Total 21 169.1162
Mean 1 161.8519
Total of the blocks Block (unadj) 6. 2.0347
Treatment Total containing Treatment
estimator 7
numbei· treatment i Treatments (adj) 6 3.2908 .5485 2.263 -6u2 :E('T·z - f)2
•
Error 8 1.9388 .2424
.. ~ k
i Ti - T • = lt q¡
The tabulated F at the 5 per cent level for 6 and 8 degrees of freedom is
1 7.49 25.68 -1.0700 -.459 3.58. Hence these data do not suggest rejecting the hypothesis r 1 =
2 5.63 22.74 -1.9500 -.836 .,. 2 = · · · = r 1 at the 5 per cent level of significance. The estimated
3 9.70 26.47 .8767 .376 variance of a treatment contrast LCtri is equal to
4 9.18 25.24 .7667 .329
5 8.57 23.57 .7133 .306
6 9.13 24.69 .9000 .386
7 8.60 26.51 -.2367 -.101
l l 2 3 l l 2 3 4
2 l 2 4 2 l 2 3 5
3 2 3 4 3 2 3 4 5
4 l 3 4 4 l 3 4 5
(e) (d)
TABLE 14.8
Block Treatments
15.1 Introduction
[ obtainable.
2. Estimates of the variance of ílii will be obtainable.
3. Tests of certain hypotheses about µ¡; will be obtainable.
In this chaptér the assumptions underlying model 4 will be investi- 4. Confidence intervals on µiJ will be obtainable.
gated further. The work that has been done in this area is vast, and no The conditions often used are as follows:
attempt will be made to be complete. Some of the consequences that A. E(eii) = O.
follow when the assumptions do not hold will be investigated. B. µ¡ 1 can be written µ¡; = µ + T¡ + P;; that is, there is no inter-
Only linear models will be considered. These have been defined as a.ction.
observable random variables y that are equal to linear functions of par- C. The e;; are uncorrelated.
ameters and unobservable random variables. For example, the model D. The e¡; have the same variance a2 for all i andj.
E. The e;; are normally distributed.
Y;;= µ¡;eii
As stated before, condition A, that E(eiJ) = O, is the only assumption
is nota linear model according to Definition 5.2. On the other hand, that is necessary to obtain unbiased estimates of µii and linear com-
taking the logarithm of y ii gives binations of µ¡; for ali i andj. Condition B is a condition on the param-
log Yi; = log µ¡; + log e,1 eters. If we write
which is a linear model in the logarithms. There are sorne models that µ¡;=P .. + (Pi. - P.J + (P.1 - PJ + (µ¡, - Pi. - P.1 + P,J
cannot be made linear by transformation, for instance,
= µ* + -rj + p; + (-rp)~
Yii = µz + µij' thenE(y¡; - Yi';) = -ri - -r~foralljandi -=fo i'ifandonlyif(-rp)~ =O.
These are important, but they will not be discussed here. Only models Although it may be desirable from sorne points of view for (-rP)~ to be
will be considered that either are linear or can be made linear by trans- zero, it should be noticed that -r[ - -rZ is estimable whether (-rP)~ is zero
formation. or not; the estímate is Yi. - Yk.· Condition Bis also necessary for the
analysis of variance, to give an unbiased estimate of the variance of e,1
15.2 Assumptions for Model 4 and to gi ve a test of the hypothesis that the -ri are equal. This has been
shown in Art. 12.5.2. Suppose that (-rP)ij =F Ofor sorne i andj and that
· Most ofthe ideas in this section will be discussed for a two-way classi· there is more than one observation per cell. Then the model can be
fication model. However, almost all the resulta will generalize to more written as
complex situations.
318
320 LINEAR STATISTICAL :MODELS SOME ADDITIONAL TOPICS ABOUT 1\IODEL 4 321
If the variance of eH"' equals a 2 for all i, j, and m, a 2 can be estimated by We desire a test of the hypothesis T1 = T2 = · · · = T,. Let the
the usual analysis of variance; thc variance of ftu can also be estimated. (t -1) X 1 vector Y; be
This problem has been discussed in Chap. 12. Y2; - Yu
Next we shall turn our attention to the conditions made on the ran.
Ys; - Yu
dom variables ew The normality condition E is not essential for point Y;= j = 1, 2, ... , b
estimation, bnt is ysed for interval estimation and tests of hypotheses. (
However, for the conventional estimates of µiJ to have the optimurn
Yt; -·Y1
f-~ ~:)
properties stated in Chap. 6, it is essential that conditions C and D be
rnet. \Ve have shown, however, that, if C and D do not hold but the
variances and covariances are known, then estimates of µu can be found
such that the optimum properties in Chap. 6 hold. This is stated in Weget
Theorem 6.4. E(Y,) = = y
Tt - T ¡
15.3 Tests of Hypotheses Let the covariance matrix of Y; be V. The pqth element of V is
Next we shall turn our attention to the testing of hypotheses. The vvo = E{[(Yii; - Yu) - E(Yvi - Yi;)][(Yq; - Yu) - E(Yq; - Yu)]}
normality condition E is necessary if the conventional tests arP. to be Using Eq. (15.2) and the distributional properties in (15.3), we get
strictly valid. If the normality condition does not hold, then sorne
nonpararnetric test can be used. If the norrnality condition E is satis- VM = E[(e:Pi - eu)(eai - ei;)] = ªva - ªii1 - ª1a + ª11
(15.4)
fied, but if either Cor D is not, then a nonparametric test could be used. p, q = 2, 3, ... , t
However, there are other methods available for this case. These will It can be shown that V is positive definite, since we have assumed that
be discussed next. (a11 a) is positive definite. Also, by (15.3), Y; and Y;· are independent if
For the remainder of this section we shall assume the model
j =fa j'. Therefore, we have Y; as a random sample from the density
N(y,V)forj = 1, 2, ... , bandT 1 = r 2 = · · · = Ttifandonlyify =O.
i = 1, 2, ... , t; j = 1, 2, ... , b (15.2) Therefore, we can let ·
where LT¡ = LP; =·o, and we shall assume that the random variables (b - t + 1)b Y's-ry = u
e¡; are normally distributed with mean zero. We shall call the T¡ treat- (b - l)(t -- 1)
ments and the P; blocks. We shall further assume that the errors - l 1 b -
-~
zero; that is, H 0 : LCiT i = O, where Le, = O. If the distributional prop- l 2 3 4
erties of ( 15.3) hold, then the conventional method of testing H 0 is not
exact. An alternati ve method is the subject of the next three theorems. 1 19.41 43.60 24.05 19.47
2 23.84 40.40 21.76 16.61
+ Theorem 15.3 Let the distributional properties in ( 15.3) hold for 3 16.08 18.08 14.19 16.69
the two-way classification model YiJ = µ + -r¡ + /J; + e¡; for 4 18.29 19.57 18.61 17.78
5 30.08 45.20 29.33 20.19
i = 1, 2, ... 't j = 1, 2, •.. 'b 6 27.04 25.87 25.60 23.31
7 39.95 55.20 38.77 21.15
Then the quantities 8 25.12 55.32 34.19 18.56
t
9 22.45 19.79 21.65 23.31
Z; ! CiYii
= i==l LCi =o 10 29.28 46.24 31.52 22.48
11 22.56 14.88 15.68 19.79
are independent and distributed N(LCiTi,u 2), where 12 22.08 7.52 4.69 20.53
13 43.95 41.17 32.59 29.25
The proof will not be given. Since we assume that (15.3) holds on the e, 1, we cannot use the conven-
+ Theorem 15.4 Using the assumptions ofTheorem 15.3, tional method of analysis. Instead we can use Theorem 15.2. By
Theorem 15.2 we need
u2 bz.2
u2= u2 j = 1, 2, ... , 13
is distributed as x' 2(1,A.), where Yí = (24.19, 4.64, .06)
2
A= b(LC¡Ti) Y2 = (16.56, -2.08, -7.23)
2u2 y~= (2.00, -1.89, .61)
2 2 y~= (l.28, .32, -.51)
Also, - w = _:E_(z__¡_-_z_)
u2 u2 y~= (15.12, -.75, -9.89)
is distributed as x2(b - 1), and u and w are independent. y~= (-1.17, -1.44, -3.73)
+ Theorem 15.5 Let u 2 and w2 be as given in Theorem 15.4. Then Y;= (15.25, -1.18, -18.80)
F rom t his we obtain J-Ience, if (T{J);; = O fo r a ll i andj, t he quan tity (N8sf B 88)[( b - l )(t -
1) - 1] is d istribu ted as F[ l , (b - l )(l - l ) - l ]. The proof of t hcsc
Y' = (7.13, - 2. 12, - 5.46) distri but iona l properties will be the subject of t h e n ext few t heorems .
2,151.8706 894.5435 - 274. l !.124) + Theorem 15.6 Let y,.j = µ + T; 1 fJ; + (T/3);; + eij• where L:Ti =
13
and l 2S = .2: (Yj - V)(Y 1 - Y)' = 894.543ií 542.1045 36.3449 L:/3; = ¡ (T{J)¡j = ¡ (TfJ);; = O. Let e,.j be indcpenden tly d istrib-
J= l ( ' ]
- 274. 1924 36.3449 446.3738 uted N(O ,a 2 ) . L et the q uantit ies x, u., v.J> w 'ZHI denote
1
The in verse is
X = Y ..
.0023572760 -.0040087773 .001774400 1)
[L(Y; - Y )(Y j - vn- =1
(
- .0040087773 .0086721040 -.003 16856 18 i = 1, 2, ... 't
.00 17744001 - .0031685618 .00 3588 ~215 =
vj = Y.j - Y .. j 1, 2, .. . ' b
T his g ives ti = 7 .5D, which we compa re wit h Snedecor's F Yaluc with 3 7J = 1, 2, ... ' f
and l O d egrces of freed o m. This is significant at the l per cent le Yel; so w¡)Q = Y,,q - Yv. - Y .Q + Y .. {
we h a ve cvide ncc to rejcct t hc hy pothesis T1 = T2 = T3 = T 4 • Notice = 1, 2, ... ' b q
t hat the D oolittlc procedure a nd T heore m 7. 1 can be used to obtain The n
- - ( 1) x is in<lepcndent of n,., V;, wl>Q for a li i, j, p, a n d q.
Y'S- 1 Y.
(2) n ,. and vj are ind epe nden t for ali i and j .
(3} ii¡ a nd w,, 0 are indepe nd e nt for a li i, p, and q.
15.4 Test fo r A dditi v ity (.J.) V ; a.nd w,,q a re independent for allj, p , a nd q.
Tn a two -wa.y classifi cation mod cl y ij = µ + T; + f3j + (T{J);; +e,;. Proof: S in ce cach of the quan tities x, U ¡, V;, w,,q h as a. normal d istri-
it rnay ue tk~irab l e t o h.a v e a m ethod for tcsting H o: (T/3);¡ ~ o for ali b11tio11 , we need only sh ow that they are u ncorrelated, sincc t his
i and .i· If t hesc qua nt it ies are ze ro, t hen the conventional ni ethod im plies ind ependence. We shall prove (3), a nd t he rest will foll ow
g iven in Sec . J 2.3 can be used to test T 1 = Tz = · · · = T 1 . Also, thc by simila r rcasoning. \ '\Te must show t hat cov(ii,. ,w,,0 ) = Ofor a ll
error mea n sq ua,re will b e a n unbiased est íma t e of a 2 in t his case. i, JJ, and q. But
In th is ser,t ion a meth od for testing t he hypothesis (T/3);; = O will be
ex hi bited. This tes t is dueto Tukey ( 14). In brief, the method con-
s ists in pa rtitio n ing t he error sum of squares 2,(y;j - Y ;. - Y.; + Y.Y
in to two part s : ij = B{[(y¡_ - y.J - E(y¡_ - y.)][(yJJQ - Y,,. - Y.o + Y .. n
l. S um of squ a res dueto nonadd itivity:
- E(y/)Q - Yv . - Y .o + Y..ll}
2. S um of sq ua rcs duc to ba lan ce (re mainder ): Using t hc fact t hat the eij are independe nt, we gct
A similar proofholds for (1), (2), and (3). Notice that the theo- Butthisiszero, since, by (4) ofTheorem 15.7, E(w¡;) = Oif{-r{J)ii =
rem does not say that ui and u; are independent. This is, of O. The conditional variance of z is
course, impossible, since Lui =O. Similarly for V; and w" 0 •
L w;,.a¡c;)
N ext we shall state a theorem about the distribution of these var(z 1ui = ai, v, = c1) = (
var ' 1 ,
2 2
. yLaici
quan~ties.
+ Theorem 15.7 Let the notation be the same as in the preceding ¡ W;;aic;)
=E (:.:.•i--==
2
L W¡¡U¡V¡ ·we can use the fact that La; = LC; =O to obtain
z = _i~i:====-
1Lu~LV~
'\/ 1 '
i
La;= -ai' and L ci =-e;·
is distributed N(O,a 2 ). i#:i' ;/.;'
Proof: First we shall show that z is distributed normally. We shall The variance then becomes
look at the conditional distribution of z given u¡ = a¡ and v 1 = C¡
fori = 1,2, ... ,t;j = 1,2, ... ,b. But,ifu;andv;arefixed,zis ~[(b - l)(t - 1) 2 2 t- 1
""' a~c2
~ bt L a;c¡ + -bt- ~°"" a~c2
ii
justa linear combination of wii, that is, a linear combination of nor- • i i i
Therefore, ifwe can show that z is independent of g¡; for every i and,j, it Notice that
follows that z is independent of every function of gii' and, specifically, of
Lgi. Now the conditional distribution of Yii given U¡ = ai and v 1 = c1
:¿ wr
:E u71 + z2
1 =
is normal, since, for fixed ui and v 1, the quantity gii is a linear combina- and that z2 /a2 is distributed as x2 ( l). Therefore, we have the following.
tion of w;¡, which is a normal variable. The mean (in the conditional • Theorem 15.11 Let the distributional properties and notation be
distribution) of gii is zero; that is, E(wi1) = Ofor all i andj if ( -r/3)ii = O. as given in Theorems 15.6 and 15.9. Then
rrhc variance (in the conditional distribution) of gii is
2
Í: (Yi. - YJ
. 2
E(g.' 1 1u.= a¡, v.= c1) = E (wu aiciz
- -, .......----=--
)2 (a) i;
i , 2 ....... 2
V ¿_¡(t JI ¿_¡Cq
is distributed as x' 2 (t - 1, A. 1);
= E(wr·>' - 2E( ~i,.U.iC;Z +E( La~Mz2Lc )
· • ¡...._.. 2~ 2
)
2 2 :¿(Y.; - y.J2
" ""'ª p"-'c" P " (b) ii
If we use Theorem 15.8, this becomes
From this it follows that the conditional distribution of g/a2 given it¡ = is distributed as x2 [(b - l)(t - 1) - l] if (-rP)¡; = Ofor all i andj.
a; and v 1 = c1 is x2 [(b - I)(t - l) - I]. Notice that the covariance
matrix of the y 0 is idempotent when divided by a 2 • Hence we use (e) (a), (b), (e), (d) are independent
Theorem 4.9. But, since this is chi-square for every value of ai ande;, Now we state the final theorem of this section.
330 LINEAR STATISTICAL l\IODELS 801\IE ADDITIONAL TOPICS ABOUT l\IODEL 4 331
• Theorem 15.12 Let the notation and distributional properties be The quantities Tss, B 5 s, and J.f ss are computed by the conventional
as given in Theorem 15.6. Then, to test H 0 : ( T{J) i; = Ofor all i and method of computing treatment sum of squares, block sum of squares,
j, the quantity and the mean sum of squares, respectively. To compute Nss the only
u = z2[ ( b - 1 )(t - 1) - l]
new quantityneeded is 2 Yi; Yi. Y. 1• A convenient method of comput-
ii
Í (Yi; - Y;. - Y.;+ Y.J2 - z
2
ing this quantity will be given in the next section.
ij
15.4.1 Example of Test for Nonadditivity. Suppose that the
is distributed as F[l, (b ·- l)(t - 1) - l]. If u > F«, the hypoth. data in Table 15.3 are assumed to satisfy a two-way classification model,
esis (-r{J} ;; = O can be rejected at the ex. level of significance.
TABLE 15.3 DATA FOR EXAMPLE OF ART. 15.4.l
Proof: Thc proof is immediate from Theorem 15.l l. Notice that
this can be put into an AOV table, as shown in Table 15.2. 'T Sum of cross product
Sum
{J row i with total row
(A)
TABLE 15.2 ANALYSIS OF VAUIANCE FOH. NONADDITIVI'l'Y 1 2 3 4 (B)
sv DF SS ¡ F -1 8 2
1
l 3 14 146
2 4 o 3 1 8 76
'l.'o tal bt ""'"~.
.L;Ju
3 2 1 o 4 7 63
ij
[ _'?(Y;; - Yi. - Y.;+ Y .. HYi. - Y .. HY.; - YJr 2; y31 Y.; = (2)(14) + (1)(3) + (0)(4) + (4)(8) = 63
N - z2 - 0
-------------------==--
..::::
y.J2 2 (Y.i -
ss - - Í
i
(Y;. -
J
Y.J2
Then .f: (t y,,l' J) l',, = (146)(14) + (76)(8) + (63)(7) = 3,093
2
-
[¡ t)
Yu(Yi. - Y.J(Y.; - Y.J] Also, 11-fss
(29)
=- -=
12
2
70.08
- ""'
k (yi. - y J 2,
k (y .i - y..> 2
2
i i
B 88 = (l 4 ) + (8 )2 + (7 )2 - Mss = 7.17
2 4
[ k' y u.. Y.t. Y··' - Y •• (1.'ss + Bss + 1Jfss)]
= iJ Tss = (14)2 + (3)2: (4)2 + (8)2 - j fss = 24.92
bt'l.'ssBss
332 LINEAR STATISTICAL MODELS 801\IE ADDITIONAL TOPICS ABOUT MODEL 4 333
If conditions A to E on model (15.1), listed in Sec. 15.2, are not mct, E(x) = E[h(µ) + (y - µ)h'(µ)] = h(µ)
then, as explained in previous sections, the conventional tests of since E(y) = µ. Also,
hypotheses and methods of defining confidence intervals may not be var(x) = E[x - E(x)] 2 = E[x - h(µ)] 2 = E[(y - µ)lt'(µ)]2 = a2[h'(µ)]2 .
strictly valid. In this case there are various things that might be done:
l. We might use nonparametric procedures that are valid for very since var(y) = E(y - µ) 2
= a 2• We have assumed that a2 = /(µ), where
f(µ) is known; hence, -
general assumptions (along this line the reader should investigate "ran-
domization" procedures). var(x) = f(µ)[h'(µ)]2
2. vVe might ignore the fact that conditions A to E are not met and Since var(x) is to be independent of µ, we set var(x) equal to a constant
proceed as if they were. cz. 'Ve have
3. We might transform the observed random variables Yi; in ( 15. l) c2 = f(µ)[lt'(µ)]2
so as to meet conditions A to E.
Suppose we are interested in testing a certain hypothesis H 0 in model or h'(µ) = e
4. If the test function is insensitive to the conditions A to E, it will be .JJ(µ)
said to be robust. If a test is robust, then procedure 2 above will be usc-
ful.
which gives h(µ) = cf .Jdtf(t) = cG(µ) +k
If enough information is available about the random variables y,1, it
where G(µ) is the indefinite integral of the function
may be possible to transfor!Jl them so as to meet conditions A to E (pro-
cedure 3). For example, suppose y¡ 1 = µT;eii, where log e¡; is a normal 1
variable. Then .JJ(t)
log Yii = log µ + log Ti+ log e¡¡ and k is the constant of integration. Actually, the constants c2 and k
334 SOME ADDITIONAL TOPICS ABOUT MODEL 4 335
LINEAR STATISTICAL MODELS
15.8 Prove Theorem 15.7.
are immaterial, since thcy do not depend on µ. They can be given any Prove Theorem 15.8.
15.9
convenient value (so long as they are not related to µ). 15.10 Prove Theorem 15.10.
For example, suppose/(t) = t; that is, the mean and variance of y are 15.11 Prove Theorem 15.11.
equal; then 15.12 The Poisson distribution is
Problems
Find the mean and variance of this distribution. What transformation ~akes
15.1 Suppose the data of Table 15.5 satisfy the model given in Eq. (15.2) and the mean and variance independent?
the assumptions in (15.3). Test the hypothesis Ti = Tz =Ta with a type I error 15.14 Find the mean and variance of a chi-square distribution with n degrees
of 5 per cent. · of freedom. What transformation leaves the mean and variance independent?
TABLE 15.5 15.15 Prove that the (t - 1) x (t - 1) matrix V with elements vN in Eq.
(15.4) is positiva dafinite if t.he t x t matrix with elements ai¡' in Eq. (15.3) is
Block
positiva definite.
Treatment 15.16 In Theorem 15.1, prove that u is the same if y 2 ; instead of Yu is sub-
1 2 3 4 5 6 7 8 tracted from the other Yii to form Y;.
11 \V. G. Cochran: Sorne Consequences whcn the Assumptions far the A.O.V.
Are Not Satisfied, Biomet'rics, vol. 3, pp. 22-38, 1947.
12 G. \V. Snedecor: "Statistical Methods," Iowa State College Press, Ames
Iowa, 1958. '
13 O. Kempthorne: "Design and Analysis of Experiments," John Wiley &
Sons, Inc., New York, 1952.
14 J. \V. Tukey: One Degree of Freedom for Nonadditivity, B-iometrica, vol. 5,
pp. 232-242, 1949. -
16
l\fodel 5: Variance Components;
Point Estimation
YH = µ + ai + b¡¡
where µ is an unknown constant. The thing to notice is the structure
ofthe observed random variable Yii· The object in this model is to ob-
serve the y¡ 1 and, on the basis ~f this observation, to estimate orto test
l hypotheses about µ, o;, and a¡. --
Another way to look at this is as follows: The variance of an observ-
able random variable y is
var(y) = a~+ a!
Now sup¡)ose that we can stratify the random variables represented by
y so ~hat the variance of the random variables in a given stratum is a&.
Then let YiJ be thejth observed random variable in the ith stratum, and
Yu can be written
Yu = µ + a; + b;;
For example, suppose a horticulturist wants to determine how a cer-
tain treatment has affected the nitrogen in the foliage of the trees in an
337
338 LINEAR STATJSTICAL MODELS MODEL 5: VARJANCE COMPO~ENTS; POINT ESTIM:ATION 339
orchard. He cannot examine every leaf on every tree; so he selects at The assumptions are:
random a group of trees forstudy. He cannot examine every leaf on the Case A. The a¡ are distributed N(O,a!); the b,1 are distributed
trees he has selected; so he chooses a set of leaves at random from each N(O,ul); all the random variables a¡, b¡; are independent.
selected tree. Let Yii be the observed nitrogen content of the jth leaf Case B. The ai have mean O and variance a!; the b¡; have mean O
from the ith tree, and the structure is assumed to be of the form and variance u:; ali the random variables ai, b¡; are uncorrelated.
16.1.1 Case A. In this case Yi; and Y;;· are correlated even if
YiJ = µ + a¡ + bii j =f= j'. This is given by
where a¡ is a random variable with variance a;, and bu is a random vari- a! if j =¡6 j'
able with variance ut. cov(y¡;1Yw) = ( (16.3)
Other models could be written a¡ + o! if j = j'
Also notice that
Yii = µ + ai + b¡ + e,¡ ifi:¡!:p (16.4)
Yiik = µ + ªi + b¡ + ciJ + diik The likelihood function for the random variables Yi; is
etc.
y(Y) = (J(Y11·Y12' ... 1Y111·Y21• ... ·Y2s• ... , Yr11 ... 'Yrs>
Notice that the structure of the observation is similar to model 4, the = f(YwY12· · · · ·Y1s)f(Y21·Y22, • · · ·Y2s) · • ·.f<Yr1•Yr2' · · · 1Yrs)
only change being that the a,, b¡;, etc., are random variables. Now we = f(Y1)f(Y2) · · · f(Yr)
shall define model 5.
where-y¡ = (yn,Y; 2 , ••• ,y¡ 8 } is an s x 1 vector. Y, is distributed
+ Definition 16.1 Let an observable random variable YH ... m be sucb N(lµ,V), where 1 is avectorof l's, and, ifweletv110 be the pqth element
that ofV, we have
+ · · · + eii ...m (16.l)
{
Yu ... m = µ +a,+ b,¡ a~ + a! if'p = q
vpq =
where µis a constant; a¡ is a random variable with mean Oand vari- <12
a ifp # q
ance a!; b¡; is a random variable with mean O and variance ai, ... ;
or, combining, (16.5)
and e,1... m is a random variable with mean O and variance a~. Let
all the random variables be uncorrelated and let the covariance where ~pq = l if p = q and <5 1)q = O otherwise. 'l'herefore,
matrix of the YiJ ... m be positiva definite. When these conditions
are satisfied, the relationship in Eq. (16.1) will be called model 5. !( Y.)= l exp (-12(Y. - "-*)'V-1 (Y. - "-*)]
1
(2?T)s12¡v¡112 ' r ' r
This is what Eisenhart has termed Model II in his classification [9].
In this chapter we shall discuss various iñodels within model 5. We where fL * = lµ is an s x l vector with each element equal toµ. The
shall also devote a section to a general situation. likelihood function is
For point estimation two cases \vill be considered:
Case A. All random variables are independent and normally distrib- g( Y) - l
- (2?Tr12¡v¡r12
exp [ - (Y -
2i71 i
! .f n*)'V-1 (Y -
r i
"-*)]
r
(16.6)
uted.
Case B. Ali random variables are uncorrelated but not necessarily
Before taking derivatives of the likelihood function, we shall elaborate
normal.
upon sorne of the quantities in (16.6). Notice that the matrix V has
diagonal elements equal to oi + a! and off-diagonal elements equal to
16.1 One-way Classification: Equal Subclass Numbers u:. Two lemmas that follow will be useful.
Consider the model • Lemma 16.1 The determinant of V is equal to
j = 1, 2, ... , s; i = l, 2, ... , r (16.2) (a:)s-t(o! + su!)
340 LINEAR STATISTICAL l\'IODELS l\IODEL 5: V.ARIANCE COMPONENTS; l>OINT ESTIMATION 341
The determinant of v-1 is equal to If we examine the exponent (ignore the -!) of the likelihood, we get
Next we shall provea lemma: on the inverse of V. First notice that where Yip is the pth element of Y¡ and Wpq is the pqth element of v-1.
But, by_Lemma 16.2,
V= a~l + a!J
whereJisans x smatrixeachelementofwhichisequalto l. (Through. if p ::p. q
out this chapter J will be a matrix with each element equal to unity.
The dimension of J will generally be clear from the context.) -- IX2
W 2111 IX¡+-
8
• Lemma 16.2
So (in what follows the ranges of the su bscripts are: i = 1, 2, ... , r;
v-1 = 1X1I + IX2 J
p, q = 1, 2, ... , s) Eq. (16.8) becomes
s
where IX¡= 2
1
an<l Q= 42
l p
(yiP - µ)
2
w'PP + 4 2 2 (Yiv -- µ)(yiq -
l p q
µ)wpq
O'¡, 11-:Fp
1[ 1
exp {- - 2¡(y¡p-y¡J
2+ ~+1 a:¡<Yi.-y.J
.. - µ)2])
2+ rs(y~+ ~ E[__!_
r-
¡ (Yi. -
1 o>
Y..>
2
] = _!_l
r-
E[2 i11
(µ + ai + lii. - µ - ª· - li ..>2]
J(Y) _ 2 C1 b "' h 8 a u.1 b 8 11
- (27Tr8/2(u¡)ir<s-U(cr! +su!)r/2
= -r -
1
-
1
E[z,,, (ai - ci.) 2 +z
i11
(ó,. - b.J2] (16.12)
(16.10)
since the ai and b,P are independent. To fi.nd the expected values we
We see from (16.10) that the three quantities I
ip
(Y•r> - yi.) 2 , 2 (y,,
ip
- shall use the fo1lowing lemma.
y..) 2 , and Y .. are a set of sufficient statistics. It can also be shown that + Lemma 16.3 Let z 1 , z 2 , ••• , zn be uncorrelated random variables
they are complete. If the deriva ti ves of/(Y) are taken with respect to with meanµ and variance u2 • Then
u¡, u!, and µ, the maximum-likelihood estimators (corrected for bia.s)
are E[ I (z¡ -
Í""l
z) 2] = (n - 1) var(zi) = (n - l)u2
¡1 =Y ..
The proof will be left for the reader.
Ó"~ =
L (Yi11 - Yd2 By using this lemma we get
-r -1-E[¡
_.iP_ _ __
r(s - 1)
{16.11)
(a, - á.) 2
] = -. -2 1{E[Í<ai - lif]) = L u!= su~
1 ,,, 1 - 1 P t:::1 11 -
L (Yi. - Y.J2
ª2 = __,ip'------
Similarly,
a s(r - 1) rs(s - 1)
_l_ E[¡ {b,. - ii.J2] = -
1-_¿ {E[_i (h,. - h..>2] )
This gives the following theorem. r - 1 u> r- 1 P 1=1
+ Theorem 16.1 Under the assumptions of case A for the one-way = -1- .k
"" (r - -
1) var(bi.)
classification given in Eq. (16.2), theestimatorsgiven in Eq. (16.ll) r- 1 P
I (Yi·. - y_.)2, and vrsy _ are exactly the three quantities appearing ip
r- 1 ii> r - 1 P::: 1 s
ip
in the AOV Table 12.1 for model 4. The expected mean squares are So the EMS for "between" is u: +su;. The "within" EMS can be found
by a similar process. , ·
TABLE 16.1 ANALYSIS OF VARIANCE FOR ÜNE·WAY CLASSIFICATION,
Next we shall turn our attention to finding the distributions of the
MODEL 5 "between" and"within" sums of squares. Again we shall provea lem-
ma to help decide the distributional properties of these quantities.
sv DF SS MS EMS
1 • Lemma 16.4 Let zi be distributed N(µ 0 ,u2) for i = 1, 2, ... , n.
Total rs Let cov(z,,z¡) = pu 2 for all i =I= j ( - 1 < p < l). Let
LY'f'P
1 ip 11
344 LINEAR STATISTICAL MODELS l\IODEL 5: VARIANCE COMPONENTS; POINT 'ESTIMATION 345
the off-diagonal elements are pa 2
( - 1 < p < 1). Thus V can be So the quantities in the lemma are
written
cr2 =~+BU~
V = (cr2 - pu2 )1 + pa J 2
8
p=O
Z'AZ So Yi. is distributed
Now let u=----
cr2(1 - p)
where the ith diagonal element of A is (n - l)/n and each off-diag-
( 2
N µ, ub ~suª 2)
onal element is -1/n. So and cov (Yi ..Yi.J = O for i =/= i'
1 Hence,
A= 1 - -J
n Vo = Í: (Yi. - Y.J2
Let us examine the product VA. We have ; (u~+ su!)/s
VA= [(a2 - pa")I + pa2 J](I -;J) =(a"- pa")I 2
isdistributedasx (r - 1). Butv = v 0 ;sovisdistributedasx2 (r - 1).
Now the "within" sum of squares
2 J
- -l (a2 - pu) w = Í: (Yu - y,.) 2
n, ij
L (bu - b;.) 2
=(a" - pa2)(1 -;J) = (a" - pa")A
gives us w =
ij
But, since the b¡; are distributed independently as N(O,uf), we see that
But A is idempotent of rank n - 1; hence, -VA/u2 ( l - p) is
idempotent. Therefore, by Theorem 4.9, u is distributed as k' (b;¡ - bd
- 2
·w¡ = _.1_ _ __
x' 2 (n - l, íl), where (]~
A.= l[E(Z)]' A[E(Z)] = lp.' Ap. is distributed as x (s - 1). But wi is independent of W¡· if i =F i'; hence,
2
But µ.has each element equal to µ 0 ; so µ = 1µ 0 • But l' A = O; by the reproductive property of chi-square, we see that
so íl = O, and the lemma is proved. W r
We shall exhibit the use of this lemma to find the distribution of - =.Lwi
o7, i=l
~ (yi. - Y.J
2
is distributed as x2 [r(s - l)].
V=~iv:....__ _ __
Next we want to show that v and w are independent. Let vi == Yi. -
u~+ sa! Y.. and Wpq = ypq - Yp.· These are each normal variates with zero
Let us examine zi = Yi.· means. Hence, if we can show that the covariance of vi and wP 0 is zero
E(yd =µ for ali i, p, and q, then they are independent. Also, since any functions
var(yi.) = E(Yi. - µ)2 of independent variables are independent, it follows that, in particular,
= E(a¡ + fJ¡.}2 LVf is independent of ,L
pq
w;q.
To find cov(v¡,wpq), we get
= E(ai) 2
+ E{'h¡.}2 cov(v¡, wpq) = E(viw,, 0 )
2
=a!+ ub = EL(Yi. - Y.J(ypq - yp.)]
8
= E{[(ai - cí.) + (b¡. - b_.)][bpq - hv.1}
u:+ sa! = E[(bi. - lJ.J(b 1111 - 6p.n
8
= E(b¡.b 110 - b1,/> .. + b.. h,,. - bi.bv.)
COV(Y;._, Y;J = E[(Yi. - µ)(Y;. - µ)] = c5 p;ub2 _ a,,2 + a,,2 _ ó p;a,,2
= E[(a; + b¡.)(a; + b;.)] s rs rs s
=o if i =F j =0
MODEL 5: VARIANCE COl\IPONENTS; POINT ESTll\IATION 347
346 LINEAR STATISTICAL l\IODELS
The results are given in Table 16. l. The expected mean squares for
So the "between" mean square and the "within" mean square are inde-
this table were derived under the assumptions for case A. It is easy to
pendent. By a similar method it can be shown that 'Verify that the same results hold for case B. It is clear that the esti-
rs-1.. Jllates of o! and a¡ are unbiased. However, they do not possess the
a:+ so! optimum property given in Theorem 16.1 for the model for case A.
Although the estimators(case B) are not minimum-variance unbiased
is distributed as x' (1,A.), where
2
estimators in the class of ali functions of the observations, in the class
A= rsµ2 restricted to quadratic functions only, they do have this property.
2{a~ +su!) This is the sense of the next theorem.
The distributional properties can be summed up in the following • Theorem 16.3 Let the balanced one-way classification model sat-
theorem. isfy the assumptions given for case B. In the class of quadratic
estimators (quadratic functions of the observations) the estima-
+ Theorem 16.2 Under the assumptions of case A for the balanced tors of the variance components o! and oi given by the analysis .of
one-way classification given in Eq. (16.2), the following are true: variance are unbiased and have mínimum variance.
2 (Yi11 -
ÍP
y¡,)
2
The proof of this theorem will be omitted (17].
(1)
16.2 The General Balanced Case of Model 5
is distributed as x2 [r(s - l)].
In the previous section the method of estimation by maximum Iikeli-
2
iP
(Yi. - Y.J 2 hood was demonstrated for the balanced one-way classification model
(2)
when the assumptions for case A were satisfied. The property of the
estimators was given in Theorem l 6. l. The method of estimation by
is distributed as x2 (r - l).
the analysis of variance was demonstrated for the balanced one-way
classification under the assumptions for case B. The optimum prop-
(3)
a¡+ so!· erties were given in Theorem 16.3.
Before proceeding to other special models, we shall state two general
is distributed as x' 2 (1,J.), where
theo~ems for balan~ed case.s of.model 5. Notice that the theorems say
rsµ 2
.. •
1
kth class Ík s2
k
u2ll: k Error (r - l)(s - 1) Z (Y··u - Y· -
ij
t. Y.1· +Y•• )2 Ems a2
e
1
ar
denoted by s~ (i > O). Let have f, degrees of freedom and let If the mean squares Ám 8 , Bms, Ems are equated to their respective ex-
E(si) = a; (generally a; is a linear combination of the a;, af, etc.). pected mean squares, the following equations are obtained:
Then u, = fit/<Ji is distributed as x2 (/,), and u 1 , u 2 , • • • , uk are
Ams = ~ + sú! Bms = u! + Ems = ru; a!
mutually independent.
The estimators are obtained by-solving these equations. This gives
The proof of this theorem will not be given, but a particular case has
~ = Áms -Ems ~_Bms-Ems
been proved in Theorem 16.2. u!= Ems a
8
u¡;-
r
+ Theorem 16.6 Let the model be model 5 and let it be balanced.
Also Iet the assumptions for case B hold. Then the estimators of Estimates of variance components for the three-way, four-way, ... ,
the variance components obtained by the analysis-of-variance n-way classifications are obtain~d by similar methods.
method of estimation are best (minimum-variance) quadratic un-
biased estimators. 16.4 The Balanced Twofold and Fourfold Nested Classification
of Model 5
The models that are balanced are all the models with equal numbers
in ali subclasses, such as one-way models, two-way models, latin-square, The balanced twofold nested classification components-of-variance
¡
n-way, twofold nested, etc. The theorems-·do not hold for balanced model is defined by
incomplete block models or for partially balanced incomplete block i = 1, 2, ... , r
models.
j = 1, 2, ... '8 (16.14)
where ai is the contribution d~e t~ the ith sire~ bii is the contribution due
to the jth dam mated to the ith s1re, and ciik is the effect due to the kth
offspring of the ijth dam-sire mating. The object is tq estimate o;, a¡,
anda;, since these components have special meaning in genetics.
T
"
l\IODEL
TABLE
sv
16.4
5: VARIANCE COMPONENTS; POINT ES'fll\IATlON
ANALYSIS OF
DF
V ARIANCE
NESTED ÜLASSIFICATION
SS
FOR BALANCED TwoFOLD
MS EMS
351
a: _ B ms - Oms _
[L
1 iik (Y,1. - y,,J _
_
2
L 2
(y,Jk - y,J.) ]
.;:.iJ=k_ _ _ __
Although it is usually desirable for each group to ha vean equal num-
ber of observations, this is not always possible. Suppose there are A
b - t - t r(s - 1) rs(t - 1) groups and the ith group contains ni elements. The model is
a: = A _ B [4 l (Yi .. - Y.. J
2
~ (Yíi. - Yi.J
2
]
{
ms ms = _ uk _ ;;;.ii;:.,..k- - - - - ~ = 1, 2, ... , ni
ª st st r- 1 r(s - l)
i = 1, 2, ... , A
1
352 LINEAR STATISTICAL MODELS MODEL 5: VARIANCE COMPONENTS; POINT ESTIMATION 353
A
C>OC:S
b
Let L ni= n
i=l
~
~ The AOV is given in Table 12.l and is restated in Table 16.6 with the
~
~
16.6
+ TABLE ANALYSIS OF VARIANCE FOR UNBALANCED
Olb° Ol,a
b
- ÜNEFOLD CLASSIFICATION OF l\'IODEL 5
~
~ ~ SV DF SS MS EMS
rn ~ ~
~ + +
~ C>OQ OIQ OIQ
Total n Lij Y~J
b
:e;b b
y:
~ ~ ~ Mean 1
n
~
+
c-lb"C3
+
~
++
""t:i"C3
+ +
b
:e;
C\1"1:3 c-l'!:S
~
+
b
a classes A - 1 2: ¿
i
r) n,
y•
- -··
n
Áms a: + k 0 u!
Cl1'1) Ol'I) Cl1'1) Ol 'I)
~>r, - ¡ ( ~r-)
Cl1t:l'I)
\:> \:> b b
b classes n-A Bms u2b
lJ 1 i
rn
.s .s ..e ..e s"' expected mean squares. The expected mean squares are obtained as
~ ~ ~ t:> ~ ~ follows.
E(Bms) = -l- E ( L Y~i -
y2)
L __:h
n - A ii i ni
Cl1
1
e
~
1
r:
= n
1A E[.fr (µ+a,+ b;¡ - µ - a, - 6..i•]
vl
~ ~
....... ........:.i ~ ~
E
~ ;;)
i4' í4' ~
í4'
í4' ~
l - E[2: (bi; - btJ 2]
= -n -A
1
ii
- -
-
- ---
Using Lemma 16.3, we get
~ 1 1
-~
rx. ~ 1
1 ~
A ~
1
~ ~
....... ~
~ ~
~ ;::¡-
~ ~
~ ~ Nextwe find
rn
Q)
UJ
Q)
UJ
Q)
UJ
Q) UJ
E(Ams) = -
- A -
1
-
1
E(2: l1. -
i n¡
Y~.)
n
= -
1
A - 1
- E[2: (y,_ - y_.)2]
ii
Q)
r/l UJ rn UJ
>
rn d+> ¡:::: r/l
al ~
!11
m
rn
d !
o
d
Q) o '() o o o where
~ ~ tS ..o f.) ~ Q)
354 LINEAR STATISTICAL l\IODELS MODEL 5: VARIANCE COI\IPONENTS; POINT ESTIMATION 355
This gives units of the bi; (or the a, unit contains ni units of C¡;k), according to the
2 ~ pattern shown in Table 16. 7.
E(Ams) = -
A-1
1
- E[.Lu (µ + ai + bi. - µ- .!.n112 n.,,a11 - .!: L bii)]
nu TABLE 16.7 BLOCK PATTERN FOR UNBALANCED TWOFOLD
NESTED CLASSIFICATION
= -
1 {E[¡ (a¡ - .!. 2 nf)a.,,\ +E[¡ (6¡. - .!. 2 b.,,q\ 2
]
2
]) ...
A -1 ., n 11 J " n.,,q J ª1 ~ ªA
Cin Cuu
C112 C122
...
C113 C¡23
...
= O'
2 + n2 - l:n~ O'
2
Total n LYiik
ijk
ri n(A - 1) ª
y_:_
So, in Table 16.6, Mean 1
n
n2 - l:n~
ko =
n(A - 1)
' a classes A - I L (Yi••
ijk
- Y .. )
2
Áms a; + q1u& + q2a:
The cstimates obtained from Table 16.6 do not satisfy the optimum b classes B-A L (Yi;. -yi_.) 2 Bms u; + q0u¡
properties of Theorems 16.4 and 16.6. - ijk
2
E(Bms) =_!__A E[_¡ (Yii. - Yi.J
B - .,k ]
=-
A - 1
-[¡ n~(_!_ - !)u!+ L n¡;(_!_- !)~+(A -
1
i ni n i; ni n
l)a!]
u:
'fhe coefficients of anda! correspond to q 1 and q 2 in (16.16). · If ex-
pected mean squares are equated to mean squares in Table 16.8 and if
the resulting equations are solved for a;,
uf, ando;, these u2 's are the
analysis-of-variance estimates and are unbiased. The general unbal-
anced n-fold classification is an extension of the twofold.
The analysis-of-variance estimators for the unbalanced classification
do not lead to the same estimates as maximum-likelihood estimators.
Taking expectations [remembering that E(ai) = _E(bH) = E(ciik) = O The maximum-likelihood equations are very difficult to solve in unbal-
and that all random variables are uncorrelated] g1ves anced classifications; so we shall rely completely on analysis-of-variance
estimators.
16.6.1 Example: Twofold Nested Model. Suppose the data
in Table 16. 9 are assumed to satisfy a twofold nested model. The data
are artificial, but we assume that it is a breeding experiment. The
n¡/i,2+ n¡a! + ~ n~t~ + ni~) number of sires is A = 4, the number of dams is B = 12, and n = 52.
-2i ni
t
All the quantities needed for an AOV table can easily b.e computed on
an automatic desk calculator. Quantities like
""nit
2 are easily computed. The results are
n-,¿_-
= ~ + it n¡ ~ Y ... = 1,423
e B-A b
n= 52
y2
.2 _E.:_= 40,861.201
So ij 'n¡¡
as given in {16.16).
L Y~Jk = 41,811
ijk
358 LINEAR STATISTICAL MODELS l\IODEL 5: VARIANCE CO~IPONENTS; POINT ESTJl\IATION 359
The AOV is given in Table 16.10. The estimators are
TABLE 16.9 DATA FOR EXAMPLE OF ART. 16.6.1
Si res 1
- a;= 23.744 a;= 1.767 a;= 41.416
i 2 3 4
'rABLE 16.10 ANALYSIS OF VARIANCE FOR ARTIFICIAL
BREEDING DATA
Dams j 1 2 3 4 5 6 7 8 9 10 11 12
- -- - - - - -- ---- ---- - sv
--~~-i.--~-fs
DF SS
32 30 34 26 22 23 21 16 14 31 42 26 __, EMS
31 26 30 20 31 21 21 20 18 34 43 25 Total 52 41,811.000
23 29 26 18 20 24 30 32 16 41 40 29 l\{ean 1 38,940.942
26 28 34 21 26 17 40 35 40 a classes 3 1,669.943 556.648 1 a: + 4.410a: + 12.679a!
18 32
31
18 29 37 b classes
e classes 40
8 250.316
949.799
31.289
23.744 1 a2
a: + 4.270a:
26 1 e
Total of jth
-- - - - - -- ------ -- - - -
dam in ith yii· 112 131 213 64 94 112 72 68 65 146 189 157 16.7 The Unbalanced Two-way Classification in Model 5
sire The model for this case is
- -- -- - - - - - - -- - - -- - - -
{
Number injth ~ = O, 1, ... , n¡¡
dam in ith n¡; 5 7 3 4 5 3 3 4 4 5 5
si re 41 J = 1, 2, ... , B
1
i = 1, 2, ... , A
Total of ith
si re Yi .. 456 342 133 492 where 2 n¡; = n¡. .2 n¡¡ = n.; .2 n¡; = n
; i ii
The ai, b¡, ciJk are random variables with zero means and variances a;,
Number in ith
si re ni 16 15 7 14 a¡, anda:, respectively. We find that the maximum-likelihood equa-
tions are very difficult to salve. Hence we shall rely on analysis-of-
variance estimators. This, however, presents a difficulty. In the un-
2 11.. = 40,610.885
balanced cross classification discussed in Chap. 13 we saw that
i ni R(µ, -r,p) = R( -r 1 µ,{J) + R(µ,{J)
2
.2 n;; = 17.844 and - R(µ,-r,{J) = R(íJ 1 µ,-r) + R(µ,-r)
ii n¡ So we can partition the total_ sum of squares in two ways:
2 1,.2
¿nis= 4.615 :Ey¡;k =---:.:.:. + R(T 1 µ,{J) + R(µ,{J) + (error)
ii n n
2 y2
¿ni= 13.961 and :Ey¡Jk =--:.:.:. + R(fJ I µ, T) + R(µ, T) + (error)
i n n
y2 If this is done for the model in this section and if mean squares are set
-··· =·38,940.942 equal to expected mean squares, we shall obtain two sets of estimators.
n Each· set is unbiased.
So - 52 - 17.844 - 4 270 - 17.844 - 4.615 -- 4 .410 Another method that is sometimes used is to partition the total sum
qo - 8 - . q¡- - of sq uares as follows:
3
q2 = 52 - 13.961 = 12.679
3
};!fi;. = ~·· )
( y2 + [ ~ (y;
..::·) - y2~·· + J [t (..:;· - ~-·J+
y2. ) y2
(remainder)
360 LINEAR STATIS'l'ICAL l\'IODELS l\IODEL 5: VARIANOE COM:PONENTS; POINT ESTIM:ATION 361
This can be put in an AOV table such as Table 16.11. The values of r 5 and r6 can be obtained from this expression. Also,
E(Bms) can be written down with symmetry, replacing i with j and
TABLE 16.11 ANALYSIS OF VARIANCE FOR UNBALANCED interchanging a! and af in E(Ams). We get
Two-WAY CLASSIFICATION MoDEL
To find the estimates, we equate observed mean squares to expected - (A - l)E(Ams) - (B - l)E(Bms)]
mean squares and sol ve the resulting set of three equations for the three
unknowns a-;, a-¡, and D;. The quantities on the right have been computed, except for EC~Y~·1c)·
~~ u
It should be pointed out that the remainder mean square obtained
by subtraction can be negative.
To find the expected mean squares and the r iJ we proceed as follows:
1 yz
( 2 __.!.:.:. y2 ) Hence,
E(Ams) = - - E - _ ...
A - 1 i n.i. n
l
E(Rms) = - - - - - -
n - A - B + I
2 2
X [ n(µ2 + a! + a: + o!> - nµ2 - 2 ni. o! - 2 n.; a:
i n i n
- a~ - n a!- 4--il
(n - ¡-..!:.
2
) n2) a~ -
n2 - 2--=1 (
(A - l)a~
z n ,, ni. n J
=a~ - 1 2
( In;; - 2n.i a: 2)
1i - A - B +l i; ni. ; n
2
= -1- ( n - n7 ) a! + -l- ( 2 -n7; - 2 ~
2i ~ n 1) ªb2 + <1~ 9
_ 1 <)
( " n¡; , n¡,<))
_¿,- - _¿,- C12
A - l n A - l iJ J n n¡,
n - A - B + 1 ;; n.; i n ª
362
This gives r 1 and r 2 •
LINEAR STATJSTICAL MODELS
ri = n -
We have
1
A - B
(In~; L n7:1)
+ 1 :1 -;- - ;; n;.
T
.,,
sv
MODEL 5:
TABLE
VARIANCE COl\IPONENTS; POINT ESTil\IATION
DF SS MS
5
l) FOR
EMS
363
r2=
1 ""'ni.2 2)
""'n;J Total bk = rt :E Y'fsm
n- A - B + 1( i k--~-
n ;; n.;
Y~.
2 Mean 1 -¡;;;
ra = -1- ( n -
B- I
I; n
n
.)
__!!
Li Y~.. yz
... t-k
= - - ( Ln;:12- 2)
1 Treatments t - 1 -r- --¡;;; T ms a; + - - a& + ra;
t - 1
r4 Lni.
B - l ii n.; i n ""'y2.;.
-4- y2 b - r
Blocks b - 1 _J__ - _.:.:.; Bans a2
,,
+ ka2,, + - - a2t
b - 1
k bk
1
1
2 (t - k)ac + (b - r)uf
n~) Error .bk - t - b + 11 Subtraction Rms
r6 = -1- ( n
A -1
-1_!.:.
ni
O"
"
- -------
bk - b - t+l
16.8 The Balanced Incomplete Block in Model 5 ni. = r; n.; = k; n = bk. Also, cr; = cr!; u! =a~; and cri is the same in
both tables. For example,
The balanced incomplete block classification in model 4 has been
defined and discussed in Chap. 14. The definition of this classification r ¡_- 1 ( - """'n~;
'-- + ""'n~;)
~-
in model 5 is the same, except that the block (3 1 and treatment -r¡ com- n - A - B +1 i; n 1• ; n
ponents are now random variables. Therefore, we shall write the
= 1 (- 1 n¡; + 1 k2)
modelas bk - t - b +1 i; r ; bk
i = 1, 2, ... , t; j = 1, 2, ... , b
where m = ni; and n;; = Oor 1. Also, there are k treatments per block,
=
bk - t - b
1
+ l (- ~r + k)
and each treatment is repeated r times. Each pair of treatments
appears together in exactly Ablocks. Also, as in Cha p. 13, when m = O t-k
the corresponding observation is not present; that is, Y;;o does not exist. bk - t - b + l
Now t¡, b1, eum are uncorrelated random variables with zero means and 1 ""'n.21 ., )
variances a;, ~, cr;, respectively. r2 -
- bk - t - b +1 (
-
k
iJ n .J
+ ""no:
'
k
i ~
'·
Since the maximum-likelihood estimators are difficult to obtain, we
shall use analysis-of-variance estimators. However, this presenta the
same problemas did the general two-way classification model discussed = i
bk - t - b +1
(- 2: 'li¡; + I
¡; k i
r2)
bk
in Sec. 16.7.
There are at least three ways to find analysis-of-variance estimators, 1
justas there were for the model in the preceding section. - - -- - ( - b + r)
bk - t - b + 1
16.8.1 Method 1. If we use the method of Table 16.11, the
coefficients ri simplify somewhat. This AOV is given in Table 16.12. b-r
The correspondence in notation is as follows: A beco mes t; B = b; bk - t - b + l
364 LINEAR STATISTICAL MODELS 365
r.
3
= __.]:_
b-1
(n - ¿; n~;)
n
= _l_(bk - L k2)
b- l i bk
=bk-k=k
b-1
r4 = -1- (""
L.- - n¡; """ n¡·)
L.- Problems
b- l a n.; í n
16.1 Evaluate the following determinant by any method, and verify Lemma
b-r 16.l.
=-- 8 2 2 2
b-I
2 8 2 2
r¡; = _l_(¿n71 _ Ln~;)
t- 1 ii ni. i n 2 2 8 2
t-k 2 2 2 8
t- 1 16.2 Evaluate the inversa of the following matrix by any method, and verify
Lemma 16.2.
1- (
rG=- n~)
n-¿...!:
(:. : :)
t - 1 n i
2
= -1- ( r t - I -
r )
t- l i rt 16.3 Prove Lemma 16.1.
16.4 Prove Lemma 16.3.
rt - r 16.5 Assume that the data shown in Table 16.14 satisfy the assumptions for
=--=r
t- 1 case A given in Eq. (16.2). Find the estimators of a! and ~ given in Eq. (16.11).
16.8.2 Method 2. Another method for constructing an AOV TABLE 16.14
for the balanced incomplete block classification is that used in Chap. 14.
a effects
TABLE 16.13 ANALYSIS OF VARIANCE (METHOD 2) FOR
BALANCEO !NCOMPLETE BLOCK IN MODEL 5 1 2 3 4 5 6
_4_1~
8 F. A. Graybill and A. W. Wortham: A Note on Uniformly Best Unbiased
~ 1 2 3
Estimators for Variance Components, J. Am. Statist. Assoc., vol. 51, pp.
266-268, 1956.
1.3 7.2 9.1 11.2 9 C. Eisenhart: The Assumptions Underlying the A.O.V., Biometrics, vol. 3,
1 1.6 7.5 pp. 1-21, 1947.
7.6 10 S. L. Crump: The Present Status of Variance Component Analysis,
--- - - - - - - - - - - - - - - - Biometrics, vol. 7, pp. 1-16, 1951.
4.2 10.1 12.2 14.l 11 E. F. Schultz, Jr.: Rules of Thumb for Determining Expectations of Mean
2 4.3 10.3 12.4 14.3 Squares in A.O.V., Biometrics, vol. 11, pp. 123-135, 1955.
12.6 12 H. Nagler: On the Best Unbiased Quadratic Estímate of the Variance,
--- --- --- --- ------ Biometrika, vol. 37 ,_ pp. 444-445, 1950. ·
6.1 8.4 12.1 16.l 13 J. A. Nelder: The Interpretation of Negativa Components of Variance,
3 6.3 8.2 16.3 Biometrika, vol. 41, pp. 544-548, 1954.
17.2 14 J. W. Tukey: Variances of Variance Components: I, Balanced Designa,
Ann. Math. Statist., vol..27, pp. 722-736, 1956.
16.13 F~~m the data in Prob. 16.12, estímate u!, and u:,
by constructing a; l 15 S. R. Searle: Matrix Methods in Components of Variance and Covariance
Analysis, Ann. ..1.Wath. Statist., vol. 27, pp. 737-748, 1956.
Table 16.11. 16 C. R. Henderson: Estimation of Variance and Covariance Components,
16.14 Assume that the data of Table 16.17 satisfy a balanced incomplete Biometrics, vol. 9, pp. 226-252, 1953.
17 F. A. Graybill: On Quadratic Estimates of Variance Components, Ann.
TABLE 16.17 llfath. Statist., vol. 25, pp. 367-372, 1954.
Block Treatments
17 Let g = i=l
k
ÍYiX.:
•
Proof: The theorem can be proved by straightforward application improved approximation given in Theorem 17 .3 can be used. If the
of the method of using a chi-square variate to determine a con- correction term
fidence interval. 0 11 = j(2z¡ + l)(ys-2t - 1)
Although Theorem 17.2 can be used to set approximate confidence is small, the method described in Theorem 17 .2 is adequate.
intervals on variances, a better method in most experimental situations 17.1.1 Example. To illustrate the theory in this section, we
has been given by Welch; this is the text of the next theorem. shall work an example of a twofold classification. The model is
• Theorem 17.3 Let nixi/ur (i = 1, 2, ... , k) be independently k = 1, 2, 3
distributed as x2 {n¡), and let j = 1, 2, 3, 4, 5
{
k i = 1, 2, ... , 12
g = _2g,x,
i=l The distributional properties are those given for case A in Chap. 16,
and 'Y= _2giui2
k
where the variances of a¡, b¡¡, and ci;k are o;,
u:, and ~, respectively.
i=l
The original data will not be given, but the AOV is shown in Table 17 .1.
Then the probability statement TABLE 17.1 ANALYSIS OF VARIANCE OF TWOFOLD ÜLASSIFICATION
MS (xi) EMS
is approximately true. This can be used to give the 1 - IX con- Total 180
fidence interval · Mean 1
a classes = n1 ai = a~ + 3aC + 15a!
p( ng
Aa:12
~ '}' ~ --~--.) = l
A1-a:12
- IX b classe,s
e classes
11
48
120
= n2
=na
3.5629 =X¡
1.2055 = X2
.6113 = x 3
ai = a: + 3a:
ai =a:
Aa.12 and A 1 _a.12 are computed by using
The estimates are
A 11 = xi(n) - 0 11 = xj(n) - f(2zi + I)(gs-2t - 1)
For a proof of this theorem, see Welch [7]. If n > 30, the confidence and :E g¡x¡ = (0) 2 + (.40180) 2 + (-.20375) 2 = _
003709
interval given in Theorem 17 .2 is probably adequate. If n < 30, the n; 11 48 120
372 LINEAR STATISTICAL MODELS MODEL 5: VARIANCE COl\IPONENTS; INTERVAL ESTIMATION 373
The limits are
'l'ABLE 17.2 TABLE OF QUANTITIES FOR lNTERVAL 1,
EXAMPLE OF ART. 17.1.l p(2.179 ~ ª2 & 2.179) = .95
1 25.41 ~ b ~ 7.31
i \ ni 1 Yi xi gixi 1 g¡xf/ni iM/ni
--i-- or P(.086 ~a: ~ .298) = .95
1 i
2
11
48
1 .0000
.!l333
3.5629
1.2055
.00000
.40180
1
1
1
·ºººººººº
.0033630 ·ººººººººº
.000028150 There seems to be an improvement resulting from use of Theorem 17 .3
-.3333 .6113 -.20375 .0003459 -.000000587 instead of Theorem 1 7 .2.
3 120 ¡
To compute interval 2, the confidence interval for
.1981 =g 1 .003709 =s .0000276 =t
y= ªª +2
ab
2 + <Je=
2 ...1.
2
15G1
+ _4 2
i-1ra2 -r• J.O 2
1s<J3
ng
p (-- ~ y ~ --
ng) = .95 y = u= 'Lgixi = a! + a; + ~ = .9665
X~o2s X~975 To demonstrate confidence lirriits on y, we shall use only Theorem 17.2.
2.1791 .- 2 . - 2.1791) -
_( 9r.: This gives
which gives P--:::::-...<1::::-....--
( 21.92 ~ b ~ 3.82 .O
374 J...INEAR STATISTICAL MODELS MODEL 5: VARIANCE COMPONENTS; INTERVAT.1 ESTll\IATION 375
The power of the test, denoted by {J(A.), is
17.2 Tests of Hypotheses
In this section we shall discuss two general tests. One is a t~st that {J(Á) = f 00g(u;Á) du
a variance component is zero, for example, that u! = O. The other is JFa
that a linear combination of u!, a~, a;, ... , is equal to zero. The nota- To evaluate {J(A.), we make the transformation 'W = 1u and sub-
tion in Chap. 16 will be used, and references to various special models stitute this into g(u;A.). Now g(u;A.) becomes the density func-
and AOV tables will be made. tion of w, say h(w). But w is distributed as F(ni,ni) for every
First we shall considera general analysis-of-variance model and the value of ). (O ~ A. < ex:>); so k(w) is the central F distribution.
quantities in Table 16.2. To evaluate the limits on the integral, we note that, when 'lt =Fa.,
In many cases, the difference u'f - aj in Table 16.2 is equal to a w = AF'a, and when u = ,ex:>, w = ex:>; so
multiple of one of the variance components u!, u¡,.... For example,
in Table 16.1, wegetai - a~= su;; in Table 16.3, wehave ai -ai =su;; P(A) = f a:i h(w) dw
in Table 16.4, we have ~ - oi =tu~; etc. JAFa
A test ofthe hypothesis H 0 : ar = aj will be given in the nexttheorem. and the power can be evaluated by using only the central F tables.
+ Theorem 17.4 Let the model satisfy the conditions of Theorem 17.2.1 Example: One-way Classification. For this example
16.5, and let the AOV be represented by Table 16.2. To test the we refer to Table 16.1. The correspondence in notation is
hypothesis H 0 : ar = aj (i ~ j) against the alternative hypothesis _2 + Baa2
HA: aj > ui, the quantity <Ti = u.,,2
u =-1
82
s~ = Bms
8~1
can be used, and H 0 is rejected at the « level of significance if
u;= o7,
u >Fa., where Fa. is the upper percentage point of Snedecor's P s:= Wms
distribution with n; and ni degrees of freedom. That is to say, F'X We shall test H 0 : ui =a~; that is, u! =O, and HA is a; > O. The test
is such that criterion is to reject H 0 at the ex level of significance if u > Fª, where
00
«= f /i(u) du Bms
JFa U=--
Wms
where h(u) is the F(n;,ni) distribution. The power of the test
depends on the quantity Also A= uf
a;+s~
A=~< 1
u¡ As another example, suppose we wish to test H 0 : a! = O using the data
in Table 17.1. We get
and is f lt(u) du = {J(A.)
00
);.¡.'~ u= 3.5629 = 2.956
Proof: Let g(u;J.) be the frequency function of u for various values 1.2055
of )~. If ,l = 1, u is distributed as F(n1 ,n¡); so g(u;I) is Snedecor's The tabulated F value for 11 and 48 degrees of freedom for « = .05 is
F distribution with n¡ and ni degrees of freedom. Of course, if
F. 05 = 2.00. Hence we reject H 0 •
,l ~ 1, then g(1t;J.) is not the F distribution. The test procedure
is to reject H 0 if u >Fa., where Fa. is the upper « point of the F
17.2.2 ·Test of a Linear Combination of ar.
In the previous
article a t_est criterion was de,rised for testing the hypothesis that
distribution; that is, where Fa. is given by ar - aj = O. This is equivalent to testing the hypothesis that af/aj =
00
l. This test will satisfy many experimenter's requirements. How-
«= f g(u;l) du
JFa ever, a more general test may be desired. That is to say, we may need
1
376 LINEAR STATISTICAL MODELS MODEL 5: VARI.ANCE COMPONENTS; INTERVAL ESTIMATION 377
to test H 0 : í:,gia¡ = O, where the gi are known constants that are either further approximation. Also, m and n will not be integers, but their
plus or minus unity. Of course, a more general test would be one in values can be approximated in any F table.
which gi were any known constants, but the case in which they are plus It should be pointed out that sorne care should be exercised in using
or minus unity covers many important situations. In most cases an any of these approximate methods where no bounds on the error of
exact test for this hypothesis has not been found, but an approximate approximation are known.
test will be the subject of the next theorem. 17.2.3 Example: Testing a Linear Combination of Variance
Components. The details will not be given, but suppose a three-way
+ Theorem 17.5 Let the model satisfy the conditions of Theorem classification model with equal numbers and all interaction is given by
16.5, and let the AOV be represented by Table 16.2. To test the
hypothesis a; + · · · +a: =a; + · · · + df against the ~lternative i = l, 2, ... ' 10
hypothesis a; a;
+ · · ·· + > a; + · · · + Uf, the quant1ty Yiik = µ + ai + b; + (ab)i:i + ck + (ac}ik + (bc);k + eiik j = l, 2, ... , 8
82 + ... + 82 k = l, 2, ... '4
u= P "
8; + · · · + 8~ Let the distributional properties satisfy the requirements for case A in
can be considered to be approximately distributed as F(n,m,), Chap. 16. Let the variances be u;, u:, u;b, a;, <I~c' u¡c, and a 2 • The
where (s2 + ... + 82)2 AOV will be as given in Table 17.4. This should be compared with the
n= P q
notation introduced in Table 16.2.
s!/f + · · · + s!/fq
11
m = (s; + ... + s~)2 TABLE 17.4 ANALYSIS OF VARIANCE FOR THREE-WAY ÜJ.-ASSIFICATION
s:ffr + · · · + s1fft
sv DF MS E:MS
1
and the hypothesis will be rejected at the approximate (/.. signifi-
Total 320
cance level if u > F ex· Mean 1
This theorem is justan extension of Theorem 17 .1; i.e., we are assuming a class Í1 = 9 sf =47.54 af = a2 + 4a!b + Ba!c + 32a!
that
+ · · · + s:)n
(s;
b class
ab interaction
12 =
f3 =
7
63
8~ =
sf =
74.54
14.31
ªªaf = a2
= a2
+
+
4a:b
4a:b
+ 1OuCc + 40ac
V=..;....:.:.~~-~_.._ e class f4 = 3 s¡ =99.26 a: = a2 + Ba!c + 1OaCc + soa;
a;+···+ a! ao int.eraction f5 = 27 si= 7.29 ªE = ª22 + sa:c
be interaction 16 = 21 s:= 18.01 a: = a + IOatc
is approximately distributed as x2 (n). The first two moments of vare , Remainder 11 = 189 s1= 6.01 a~ = u2
is approximately distributed as x 2 (m), where m is similarly fou~1d by Theorem 17.4 provides an exact test. However, for any ofthe follow-
equating the first two moments of w to those of x2 (m). Then smce v ing hypotheses:
and w are independent, the quantity a:=oa!=O ~=0
s3 + s¡ 21.60
Solving for a';/a: gives
n = + 4) = 2,867.6025 = 11 .41
(s~
2
Also,
s:/J + s~/f7 251.3074 E[Bms r(s - l) - 2 _ ~] = a!·
1
Wms r(s - l)s s a~
m = (s~ + si) = 466.5600 = 89 .40
2
and so Bms r(s - 1) - 2 1
s:ffa + s~/f5 5.2187
lVms r(s - l)s s
The 5 per cent significance level of Snedecor's F for 11.41 and 89.40 is an unbiased estímate of a;¡o¡.
degrees offreedom is approximately 1.88. Since u> 1.88, the hypoth-
esis is r~jected. Next we shall give a theorem concerning confidence limits on O-:/
(o!+~).
17.3 Ratio of Variances • Theorem 17.7 Let the conditions and the model be the same as in
Theorem 17.6. Then a confidence interval on ~/(0-: + ~) with
A problem of sorne importance in applied work such as genetics, · t · . b a b
breeding, and certain industrial work is estimation of the ratio of vari- ~ coe ffi 01en 1 - IX is g1ven by
anee components. For example, in the one-way model given in Table
ms
W sF1-a/2 &. ab2 &. W msS F a/2
16.1, it may be desirable to estimate a!/af, and to set confidence limits on
Bms + TVms(s - l)F1-a¡2 ~a!+ a!~ Bms + Wms(s - I)Fª12
this quantity or on aif(a! + ai). This will be the subject of the next
theorem. Pmof: By Theorem 17.6, the quantity w is distributed asF[r - 1,
r(s - l)]; so
+ Theorem 17.6 Let the model be a one-way classification with
equal numbers in the subclasses. Let the distributional properties
satisfy the requirements for case A in Chap. 16. The AOV is given
in Table 16.1. The best (minimum-variance) unbiasedestimateof or p (1' 1-a./2
&.
~ ~ a: _2 -w
Bms ~ ·Fa./2) = 1- IX
a'fi/a: is u, where b +su; ms
Bms r(s - 1) - 2 1 After manipulating the quantities within the parentheses, we
'U=-- --
Wms rs(s - 1) s arrive at
ms
Proof: Sin ce y~, Bms, and W is a set of complete sufficient statistics p[ WmssF1-.:r/2 &. a: :S.::: WmsSFatz ]
(see Theorem 16.1), it is necessary only to show that u is an un- Bms + Wms(s - l)Fl-«/Z "-:a!+ a!~ Bms + Wms(s - l)Fa/Z
biased estímate of a;¡ar,. The minimum-variance unbiased prop-
crty then follows. Let
= 1- IX
18
Mixed Models
18.1 Covariance
Covariance can be used with any case of model 4. However, only
the one-way classification will be discussed here.
The model is
j = l, 2, ... , r
{ (18.l)
i = l, 2, ... 't
where YH is an observed random variable; X;J is an observed fixed
quantity, usually called a concomitant variable; µis a constant; -ri is the
ith treatment constant; fJ is the regression coefficient (unknown); and
383
384
the eii are unobserved normal variables that are independent and have Tl
The fJ equation can be written
¡ l
(P,
MIXED MODELS
L = j(e) = 1
(2'"a2r /2
exp[- - 1
- L (Yii -
2a2 iJ
µ - 7i - fJxi;>
2
] Substituting the value of µ + f¡ into this equation gives
olog
- 1 ~(
- -L = -=2 ¿_, Y'Pi -
..
fl -
...
T'l) -
tJ
pXp;
)
= O p = 1, 2, ... 't To assist in simplifying the results the following notation will be used:
OT,., <1 i
(18.2) Exa: = L (X¡; - xd2
ii
olog
o{J
L 1 ""
= ¡jZ f; (Yii -
..
µ -
...
Ti -
...
{jX¡¡)XiJ = Q Ea:11 = L (x;; - x,.)(yii - y,.)
il
2
Notice that we have used 0:2 • This is because we reserve 8 for the
un biased estimator.
The first three equations can be used to give the following normal T 1111 =~(y. - y.• )2
¿_, •.
equations: ii
These quantities will be quite useful, and tl}ey can be tabulated in the
µ: rtP, + form shown in Table 18.1.
..
T/l + ...
TT 11 + {Jr.X p. =
y p. 1' = 1, 2, ... 't (18.3)
X .. µ"+~X. f.+ fJ"'"''x~. ~ Y·¡X·· TABLE 18.1 Co:a1PUTATIONS FOR CovARIANCE MoDEL
{J: ¿_, 1. l ¿_, IJ = ¿_, 1 IJ
i u u
sv DF SS:x2 SS:xy SS:y2
There are t + 2 equations, and the rank is t + 1 ; so there is one linear 1
N otice the identity in the SS :x2 column, v~I ue of the ~ariance of f i - -?;, a veraged over all the t( t - 1) treatment
d1fferences, is
t(t -
1
1)
2: [~ + (x;. -
i; r E
x;J
2
]cr-
i?i :z:x
Similar identities hold for the other two columns. Hence Ea:z, E:z: 11 , and 2a2
= - r + t(t -
(]2 .
E can be obtained by subtraction. The treatment-plus-error row ~ [(x. - x ) - (x1 x )]2
1111
will play an important part in covariance models. This will be ex- 1 )E
:Z::I:
f;
i7é;
'· ·· • - ••
if LAi =O (18.7) Also, from the last derivative of Eq. (18.2), we get
Since xi. and x;. enter into this variance, a different standard error is 1 ~ ,
required for the estímate of each treatment difference. The average =-
rt
[E 11v - 2/)E:r;u + fJ E:u:]
. .
388
By the definition of p, we get
LINEAR STATISTICAL MODELS
a2 = !:_(E
rt
- E'E;")
VII
:r:i:
1 \J
1\lIXED ~IODELS
We shall omit the proof, but it can easily be shown that of a 2 , p, LA¡T¡, respectively; hence, they are the best (minimum-
variance) unbiased estimators.
E(<12 ) = t(r - 1) - 1 ª2 Next we state an important theorem on the distributional properties of
rt the above estimators.
Therefore, we shall define • Theorem 18.2 Let the covariance model be given by Eq. ( 18.1)
fl· =
t(r -
l (E E!.,)
1) - 1
11
11 -
E:i::i:
with the distributional properties noted there. Then
(1) /j is distributed N({J, a 2 /E=:)·
(2) LA¡f i is distributed N(LA;T,, ai), where
so that E(a2) = a2 ~= a2[Li..7 + (l:,1.,x¡.}2]
N ext let us examine the exponent of the likelihood equation r Ex:x:
if LA,= O.
- 2~ ¿ (YiJ - µ - T¡ - px¡¡)2 (3)
[t(r - 1) - 1)82
a
2 is distributed as x2 [t(r - 1) - l].
If we add and subtract the quantity µ+ f, + /jx¡¡, this becomes (4) ú2 is independent of /j and LA.,fi.
Proof: The proofs of (1) and (2) are clear, since Pand LA.,7, are each
- _1_ 2 [(y¡¡ - µ- f¡ - Px¡;) + ({t + fi + íJx,, - µ - Ti - {3x¡¡)]2 equal to a linear function of normal variables YH· The mean and
2a2 u variance of both have been given elsewhere in this article. The
proofs for (3) and (4) follow from general theorems in Chap. 11.
= - ~[2
2a..
(y¡¡ -
H
µ ~ 7-¡ - íJxu)2 + 2iJ ({t + fi + íJxu - µ- T¡ - {Jx¡¡)2] W e are now in a position to state a theorem about confidence intervals
on a 2 , {J, and Ll,T,.
= - 1-[{t(r - 1) - 1 }á2 + 2 ({t + fi + PxiJ - µ - Ti - px¡1)2] 18.1.2 Confidence lntervals. From Theorem 18.2 it is clear
2a2 · iJ that confidence intervals can be put on the parameters a 2, p, LA¡Ti.
The cross-product term is zero, by virtue of the normal equations. The • Theorem 18.3 A confidence interval on a2 with confidence coeffi-
quantity px,. will be added and subtracted to the second term, and cient· 1 - a is
µ + f, will be replaced by y,_ - /jx,.. This gives us [t(r - 1) - 1]02 ~ 2 _. [t(r - 1) - 1]02
2
Xa.12
-...:::::. ª ': : : 2
X1-a.12
- _!__2 {[t(r - 1) - l]Ó'2 + !. [(/j - fJ)(xii - x¡J + (y,_ - µ - T¡ - {Jx¡_)J1.} A confidence interval on Pwith confidence coefficient 1 - a is
2a ii a- Ó'
/3 + ta./2
A A
a2 p- ta./2 ,- ~ {J ~ /.
'V E:i:x '\ Ex:x:
= - - [t(r - 1) - l]
2a2 A confidence interval on LA;T¡, where LA¡ = O, with confidence
2 1 .coefficient 1 - a is
P> 2 (x .. - 2
A
- ({J - 11
x. )2 - - (y¡ - µ - Ti - ¡Jx¡_)2 .. ,. [LÁ¡ + (LA¡X; .)2] A ....
2a2 iJ t. 2a2 iJ · LA¡T¡ - ta.¡2<1 - ~ LA¡T¡ ~ LA¡T¡
r Exx
Therefore, thet + 2 quantities &2,p, y 1 _, y 2 ., ••• ,yt. areasetof sufficient + (:El,xtJ ] 1 2
statistics. They are also complete. The foregoing can be summed up + ta12á[LA¡
r
E' :i::i:
in the following theorem. The proof is straightforward and will be left for the reader.
[
MIXED l\IODELS 391
390 LINEAR STATISTICAL :MODELS
18.1.3 Tests of Hypotheses. From Theorem 18.2 it follows that lf the value Yi. - Pxi. is substituted for ¡1 + 7-i, this becomes
P~/a is distributed N({J~/a, 1). Then u = í) 2 Exx/a2 is distrib-
uted as x' 2 (1,A.), where l
V [t(r_- ___;_
= .;._.,;_ I]ci2
1) _- _;;.._
If we let
a2
k"" y~i.
then, by (3) and (4) ofTheorem 18.2, visdistributed as x2 [t(r - 1) - l] = - ir-+PE:c11
and is independent of it. So
u p2E ¡Y~.
w = - [t(r - 1) - l] = - E
V (¡2 = _,__ + E2___El
r E:cJ:
is distributed as F' ( 1, t(r - 1) - 1, A.). This is summed up in the next
theorem. Also, Y'Y - (3'X'Y of Theorem 11.15 is
+ Theorem 18.4 Let the covariance model be given by Eq. (18.-1)
with the distributional properties noted there. To test H 0 : fJ = O f""y2i. E2 E2
L;; y¡¡- -r- -fiExu = Iii (Y·¡-Y· )2 -
A
2
-E!.= E - .......=!! = [t(r - 1) - l]Ó'
at the ex level of significance, a test criterion is i •· E :cz uu E :i::i:
f,
00
g(w) dw
Fa µ*:. rtP,! + Px.. = Y ..
1
392 LINEAR STATISTICAL MODELS l\IIXED l\IODELS 393
So, referring to Theorem 11.15, So H 0 is rejected if v > Fa.. A computational procedure is given
+-
- {J ~Yii X ii
~
- -
rt rt (SS:xy) 2
SV DF SS:yz - SS:x2 MS F
1
y2 X y ) s!v
= -··
rt
+ {J-(LYuX;; - -··-·-·
rt
Trcatmcnt 1
plus error
rt - 2 S
1111
--=A
szz
E;,, B A - B t(r - 1) -1
Error rt - t - l En-E=B V=--
zz t(r - 1) -1 B t- 1
A-B
Difforcnce t- 1 A-B
1
e::-T
Proof: The proof has been given above with the aid of Theorem
11.15.
So noting that s of Theorem l l.15 is t - 1, that k = t + 1, and that
n ~ k is the degrees of freedom for error t(r - 1) - l, we find that the The noncentrality parameter A. can be found by using Theorem 11.16.
quantity v in Theorem 11.15 here is '1Ve need the expected value of the treatment mean square, i.e., the
treatment mean square Tms = (A - B)/(t - 1). This gives
1 i • L y¡ + E2 __!!! _ _•• _
y2 (T :r.11 +E XII >2] f
v = (t - l)a2 - r -
[ Exz rt Txx + Exx L E(T ) = E-1-[T
ms t _ l 1111
+E
1111
- (T:r;y + Ez,,)2 -
T +E
xz :r:i:
E + E~j
E1111
:r
(18.9)
¡ 1'7. y2
Sin ce - ' - - - · · = T 1111
r rt But, by the de_finition of a2 , we get
we can write vas
[t(r - 1) - l]E(G2) = E(E - E~) = [t{r - l) - l]a2 (18.10)
_ 1 [T +E _ (Tx 11 + Ex 11 )
'1'x:c + E x:i:
2
-'--(E _E E!.11)]
1111
E:r:r
V -
(t - l)a"2 tlll 1111 1111
xx
(18.S) So we nced only evaluate
This is the text of the following theorem.
_1_ E[T +E - (T:r11 + E:i:11)2] (18.11)
+ Theorem 18.5 Let the covariance model be given by Eq. (18.1) t- 1 1111 1111 T :i::i: +E:i::i:
with thc distributional properties noted there. To test the hypo-
thesis Ho: Ti = T 2 = · · · = Tt, the function v in ( 18.8) can be used. Using the notation of (18.4) and Table 18.1, this can be written
vis distributed as F'[t - 1, t(r - 1) - l, J.], where
2a2
•
2
2a ( T xz + E :i::i:) --=-1
t
¡(Y;; - Y.J -
u
T
~
+ E~
394 LINEAR STATISTICAL MODELS l\fIXED MODELS 395
Substituting by model (18.1), we get 18.1.4 Example. The model for the following artificial data is
Eq. (18.1). The problem is to estimate {3, estimate contrasts of the
t l l E{~(µ+ .,., + px., +e,, - µ - T. - Px.. - <..>' l treatment effects T¡, and test the hypothesis Ti = T 2 = T 3 = T 4 •
TABLE 18.3 DATA FOR EXAMPLE OF ÜOVARIANCE MODEL
= - -{2 (T' -
1
t - 1 ¡;
f. + f3xii - fJx.J 2 + (rt - l)a
2 5.0
5.3
2.0
3.0
6.2
6.7
11
2.1
3.3
9.4
11.5
2.7
6.3
13.5
13.4
2.3
3.1
5.2 3.5 7.0 3.5 10.9 6.4 14.0 4.0
7.2 6.1 7.1 4.1 13.0 8.3 14.5 5.1
8.5 8.3 6.9 4.2 13.1 10.2 15.0 6.0
1
1
31.2 22.9 33.9 17.2 57.9 33.9 70.4 20.5
1 1
:Ex¡¡ = 94.5 :Ey;¡ = 193.4
+ 2{J 2 ('Ti - 7.)(xii - x.J + (rt - l)a2 From the row of totals in Table 18.3 and from Pas calculated in
ii
Table 18.4, the quantities ¡1 + -T, = y i. - Px,. can be computed. The
P'[.fr (x,1 - x. .>'J" 2{J z[( 'Ti
ii
- f.}X¡¡] _¿(x110 - x.J2
flO
results are as follows:
¡1 + '91 = 3.77 ¡1 + -92 = 4.92 ¡1+-93 =7.93
Txx+E:r::r: Tx:r:+E:r::x:
¡1 + '94 = 11.87
The average variance is .46la2• From these values any contrast ofthe
T:r::r:+E:i::r: TABLE 18.4 ANALYSIS OF VARIANCE FOR THE DATA IN TABLE 18.3
If we use the fact that ! (x¡; - x..) 2 = T :u +E=, this can be written sv 1
DF SS:x2 SS:X1J 1 SS:y 2
ij 1 1
Total 20 547.13 977.94 2,107.66
Mean 1 446.51 913.82 1,870.18
Treatments 3 31.43 = T:r::r: 26.89 = T:r:v 216.06 = Tvv
Error 16 69.19 = E:r::r: 37.23 = EZfJ 21.42 = Ev 11
By using Theorem 11.16, the value of A can be evaluated. .,. i can be estimated. The auxiliary table for testing
A covariance model for a one-way classification has been presented
in sorne detail. Other covariance models can be analyzed in similar Ho: Ti = 'T2 ='Ta= 'T4
fashion. corresponding to Table 18.2 is Table 18.5.
1
396 LINEAR STATISTICAL MODELS l\IIXED l\IODELS 397
TABLE 18.5 AUXILIARY TABLE FOR TESTING 'T1 = 'T2 = T3 = T4 discussed. If an experimenter desires distributional properties other
than those given in Eq. (18.12), tests and confidence intervals must be
(SS:xy)2 altered accordingly.
sv DF 1 SS:y 2 -
SS:x 2
MS F
1
The two-way classification mixed model \Vith interaction and with
Treatment
equal numbers in the subclasses will be defined by
-
plus error
Error
19
16
196.62
1.39 .087 748+ ri =l. 2 •...• t
Difference 3 195.23 65.077
1 j = l, 2, ... 'b (18.12)
The result indicates that the treatment effects are certainly different; where µ and -r i are assumed to be fixed unknown parameters su ch that
t k = 1, 2, ... , 1n
i.e., we reject H 0 •
t
.2
¡::::::¡
Ti= 0
18.2 Two-way Classification Model with lnteraction and with
Fixed and Random Effects and where p1, ( -rP),.;, and e,.;k are random variables such that
In Chap. 12 the two-way classification model with interaction was E(f.J;) = E( -r{1)¡1 = E(ew.:) = O
discussed. In that chapter model 4 was assumed; that is to say, µ, -ri, the eiJk are independent and distributed N (O,a;) ; the eiik are independent
f.J ;' and (-r{J) i; were assumed to be fixed unknown constants. In Chap. of the {.J1 and ~he (-r{.J)i;; the P; are independent and distributed N(O,o¡);
1 7 various types of model 5 were discussed. In these models ali effects the {.J1 are mdependent of the (-r{J)ii; the {-r{J);; are distributed
were random variables except the O\Ter-all mean µ. t - 1
In man y experiments the model that fits the situation is neither 4 nor N(O, - t- tr;p);
5 but a combination of the two. In a mixed model such as this, there
could be many specific types; that is to say, in an n-way classification E[( -r{.J),.1( -rf.J)i·;] = - ~ a;p if i =I= i,
model, any of the main effects could be fixed and any could be random.
However, in a realistic model, the interaction terms are fixed or random E[ ( -r{.J)¡¡( -rf.J)w] =O if.i =I= j'
depending on the main effects that occur in them.
Since so many different types are possible, all cannot be described. and .2i (T{.J)¡; = O for allj.
Only one will be discussed: the two-way classification model. From · If all the terms were fixed except em.:• the AOV would be as given in
this model the general procedure can be inferred. Chap. 12. This AOV is also given in Table 18.6, except for the EMS
The distributional properties for model 4 and model 5 were fairly column, which is different for the mixed model ( 18.12). The object in
straightforward, but this is not true of the mixed model. If any this model is to test H 0 : a~ = O, to test H 0 : a;tJ = O, to test H 0 : Ti =
element in an interaction term is random, it may be realistic to assume -r 2 = · · · = -rt, and to set confidence intervals on contrasta of the Ti.
that the interaction term is random also. However, in that case we· To accomplish this, we shall find the distribution of the sums of
must postulate the distributional properties of the interaction terms. squares in Table 18. 6.
The proper error term to use to test certain hypotheses and set certain The following theorem will be proved concerning the distributional
confidence intervals depends on what distributional properties are proper~ies of the sums of squares in Table 18.6.
assumed to hold. 1 The model and distributional properties given in
Eq. (18.12) will be assumed, andan AOV for this situation will be • Theorem 18.6 Let the model and distributional properties be as
given by Eq. (18.12), i.e., a mixed two-'''ªY classification model.
l There are many sets of distribut.ional properties that have been advancecl as Then
realistic for various situations [3, 4, 5, 6]. A discussion of these various
assumptions is beyond the scope of this book. The concem hero is to explain (1)
methods of analysis when various assumptions hold.
1
: '
l=
mb 2 (r, - f.} 2
i
1
·1
Further, for i =:p. i',
cov(xi,x,.) = uii' =
MIXED MODELS
Y•• J 2
But AVA= (1-Hv(1-H =V-HJV+vJ-fJVJ)
ijl.:
Let C = JV; then
Dueto {J b-l Bss = Í (Y.¡. - Y•. J 2 t
ijl.:
cpq = .L asq = 2s usq + uqq
lnteraction (t - l)(b - 1) Isa = L (Y11. - Y; •• - Y.¡. + Y..J 1 a:+ mu;p s==l
S:/:q
ijl.:
= _!__[~ + m~ + m(t -
mb 0
p t
1) ~J
T/J
or C = J
2
<16 +. m tUp2
mb
1
·~ .
400
Also, 1
-JVJ=-
t
LINEAR STATISTICAL MODELS
1 JJ a!
t
+ mta~- _- J a! + mto]
mb mb
'T and
l\HXED l\IODELS
Á. = mb :L (T, - -r.)2
401
u mtb mt
b
So we see that
u =
2 (X¡ - x.)2
.....
1=__1______
D (1 - !.t J) a~ +mbma~p = A<J! +mbm~p
= a:+ <i;/mt
then, clearly, u is distributed as x2(b - 1); but
Therefore, b
L (x; - x.)2 2
and BVB=B
u= mt ;=1 = _2 (Y.;. - Y.. J
a! + mta¡ iJk o! + mtaj
We can now invoke Theorem 4.9 and conclude that X'BX is a
To prove (3) we shall use the fact that, if random independent
noncentral chi-square variate. But quantities X;; are distributed N(O,u2), then
mb mb ""' )2 _
Í (Yi .. - Y.. J2 t b
X'BX=
o!+ mG~p X'AX = -2
U~+
. -2
maT/J i
k (Yi .. -Y ... - 2
(1C
+ mO-:T/1
•;::.;.3k_·_ _ __
k""' """
k (x11
·· - X·a. - X .,. + X •• )
2
i=l i=l
t - 1 2)
and z¡; is distributed N ( O, - t- u , E{ziizii') = O if j =f.= .i and
E(z¡;Zi·;) = -u2 /t if i -=F i'. This says that, if Z¡; has these distribu-
where -r is a t x 1 vector with elements Ti and where p. is a t X 1
tional properties, then
vector with elements µ. So
L (zd - Z¡.}2
t (. -, - T.)2
T'(1 - ~J )..- =
¡
V = =iJ_.- - - -
(fL*)'A(fL*) = u2
402 LINEAR STATISTICAL MODELS l\IIXED l\IODELS 403
and vis distributed as z2((b - l}(t - l)]. This result will be used + Theorem 18.8 To test H 0 : 'Ti = 'T 2 = · · · = Tt, use Tms/lms as
L (Y¡¡. - Yi .. - Y.i. + Y.. J2 The power of these tests can be found by using previous theorems.
lss _ !!;iil!:....:- - - - - - - - - It can be shown that the best (minimum-variance) unbiased estí-
a!+ m~p - o!+m~11 mate of the treatmcnt contrast LAi'Ti is "i:.Ji.iYi .. and that the variance of
this estimate is
If ,ve substitute model (18.12) into this, we get
lss = L [(T{J)ii - (T/1)1. +e¡¡. - e¡ .. - é.1. + é .. .12
iik
So a confidence interval on LAiT i with coefficient l - <X is
which can be written
2
~ {[(.,.fJ)·¡j +e, u.
k
.. - e·]
·'·
- [('Tp)i
.
+ ei .. - e... ]}2 = L
iik
(w11 - w;.)
ijk
E(wiiW¡•¡) = - t a,p -
2
mt <1e - t m t and the elements µ, T¡, and p, were assumed to be unknown constants.
However, in many experimental situations it is quite realistic to assume
The wiJ satisfy the distributional properties of zii above; hence, that the ~~ock effects, p1, are random variables. If the P1 are random,
L (wii - w¡.)2 L (Yu. - Yi .. - Y.i. + Y.J
2 it may be possible to obtain more information about the treatment
u = iJ = !l,ii!i.,k- - - - - - - - - constants -ri than when the {J; are fixed.
u2 a!+ ma;p 18.3.1 Definition and Notation. The balanced incomplete
block model with blocks random will be defined as follows: Each treat-
is distributed as x2 [(b - l)(t - 1)]. . ment appears r times, and there are k plots in each block; every pair of
The proof for (4) is straightforward and w1ll be Ieft for the reader.
treatments appears together in A. blocks. The model can be written
Also, the proof for (5) will not be gi ven.
From Theorem 18.6 we have the following theorem. i. = 1, 2, ... 't
(18.13)
+ Theorem 18.7 (1) To test H 0 : ~=O, use Bms/Ems as_anF statistic {J = 1, 2, ... 'b
with b - 1 and bt(m - 1) degrees of freedom. . . .
(2) To test H 0 : a;p = O, use lms/Ems asan F stat1st1c w1th (b - 1) ~ where m = n¡1 and where n¡¡ = 1 iftreatment i appears in blockj and
x (t - 1) and bt(m - 1) degrees of freedom. = Ootherwise; y i;o does not exist. The µ and -ri are fixed unknown
n¡¡
404 LINEAR STATISTICAL MODELS MIXED MODELS 405
2
parameters; b; is distributed N(o,ai); e¡; 11i is distributed N(O,a ); and all Howevcr, by Table 14.2,
random variables are independent.
18.3.2 Point Estima tes of T v - f-.. If the model in Chap. 14 ¡ y~i. k k
with b; fixed is used, the normal equations are as given in Eq. (14.3).
E ss -- ~
"""A.2
Vm
'!lijm -
J
- k- -1t
A
L qp2 = ¡ (y iim -
um
Y.;J 2 - ,-;
M
L~
From these an unbiased estimate of T :r> - -r. was formed. This is given
in Eq. ( 14.9) and is
If we substitute the model into 'L(yiim - Y.;.)2, we ~t
(18.14) Ess = L
iim
[-r¡ - T.+ eiim - é.;.] 2 - ..!::._ "Lq2
J.t P
w here q 11 is defined as But, by ~q. (14.8), it was shown that qJ> <loes not involve {3;; hcnce Ess
1 b does not mvolve fJ;· Thus Ess/a 2 is distributed as x2 (bk - b - t + 1)
qp =Y'P·· - -k """n
k
i=l
1'1 .Y;
••
(18.15)
under the assumptions of model ( 18.13), that is, where the b. are random
variables. Therefore, the following theorem has been pr¿ved.
The variance of qP is given in Eq. (14.10), and is
+ Theorem 18.9 Let the model be a balanced incomplete block
r(k____
- 1)
_;..... (}'
2 (18.16)
model with blocks random, i.e., model (18.13). Then
= (k/J.t)qP is distributed
(1) :Ec,f• = ~ :Ec,( Y•·· - ~ t n.,Y.,.) = :Ec.q. i
is an unbiased estimator of LC;-r; if ~ci = O.
N(T - :P
T,
•
~ t -
iJ, t
l a2) (2) 1:c 11.r11 is distributed
and
....
COV''T ... -
\ ..
e
. .,
...
T , T..,.• -
~)
'T
.
k 1
= - -i..t -t O'
2 (18.17) 1 ;J ~
N('....¿.,CP T P' ka2 .,)
¿.,C;
The estimator of a treatment contrast LCi-ri based on the q11 is given (3) Ess/cr2 is distributed as x2 (bk - b - t + l), where (see Table
by Theorem 14.l and is 14.2) . .
k 2 Y~;. k
-At Lc.n.
Ess=LY~· --'----Lq211
í'.11
iim iim k J.t
So (4) Ess and LC 117 11 are independent.
1
t
t
i::::l
TI var("Lc B")
11
=
MIXED MODELS
= L (µ + Ti+ b¡ + e;;m)
im
But, sin ce b; is independent of bi' ifj =F j' and since the e¡;m are indep~nd
X [Y. - ! '7~ n ·
1.. k iv
Y .v. - E( Y.1.. ) + !k E(~
7 n."' Y .v.) ]}
ent, this becomes (n;i = n1,i)
= Ter( u2 + ku!)
The covariance of B 11 and Bv' is needed.
¡
= k2la: they do not appear in Ess· So in this case the B¡ and Ess are still inde-
pendent.
= kl( u2 + ka!)
408 LINEAR STATISTICAL MODELS l\HXED l\IO D ELS 409
This proves ( 1) and (2) of the following theorem; the proof for (3) and Before discussing methods for combining the interblock and intrablock
(4) will be omitted. estimators in the balanced incomplete block, we shall provea general
theorem about combining two independent unbiased estimators of a
• Theorem 18.10 Let the model be a balanced incomplete block parameter.
model with blocks random, i.e., model (18.13). Then
+ Theorem 18.11 Let 61 be an unbiased estimator of 8, an:d let the
:Ec B .2 n,,;Y.;. variance of Ó1 be denoted by ai- Let 02 be another unbiased esti-
(l) LC
p
;¡-
JJ
= _'P_P
r-Á.
= ~:e
,, P
'!.....;- - -
r-A. mator of (), and let the variance of 02 be denoted by o¡. Let 61 and
62 be uncorrelated. Then the best (minimum-variance) linear un-
is an unbiased estimator of Le,,.,. P if Lci> = O. biased estii:nator of B is
(2) l:c,,-rJJ is di~tributed
(J = (a:Ó1 + oiB2Hoi + U:)-1
N(,. . ~ 2k ª2
LiC'P'TP'"""Cv
+ka!)
r-A.
Proof: By a linear estimator of Owe mean a linear combination of (J 1
and ó2 ; that is, y = a. 161 + a/J 2 . The problem is to determine cx 1
L Y~;.
; k
~-.
bk
[2 B¡
; k(r - Á) . tk(r - A)
B~ ] and cx 2 such that y is unbiased [E(y) = O] and such that var(y) is a
minimum.
(3) Rss
If O = E(y), this gives
a2 + ka~ a 2 +ka;
is distributed as x2 (b - t) if b > t. O = E(y) = E(«161 + rk¿Ó2 ) = a. 10 + a.z{J = (a.1 + rk¿)8
(4) Lc,,-TP, Lc,;rP, Rss, and Ess are mutually independent. So ª1 + rk¿ = l or ª1 = 1 - rk¿
..
The estimator l:cPBP/(r - A.) will be termed the interblock estimator of var(y) = var(a181 + a 2o2} = a 1 var(01 ) + a 2 var(02 )
JI 2 .. 2 "
1 N ow
the treatment contrast LCP'T ,,.
Sufficient machinery has now been developed so that methods of utiliz- l = aiai + (1 - a 1 ) 2 ai
ing interblock estimators can be explored. These will be discussed in The only unlmown in var(y) is a 1 • The value of a 1 that minimizes
the next two articles. var(y) is found by
18.3.3 Combining Interblock and Intrablock Estimators.
Instead of working with estimators of general contrasts LCP'T P' we shall d var(y) .-2
_ _; ;. . ;. = 2cx¡a¡ - 2(1 - <X¡}a22 = o
consider estimators of Ti - :r_. From these any contrast can easily da1
be estimated. From Theorems 18.9 and 18.10, the intrablock and This gives
interblock estimators of T¡ - :r. can easily be found by setting C¡ =
(t - l)/t and the remaining ci equal to -1/t. 'l.'hen -- l « 1 = oi<ai + 0-:>-1 and a 2 = ai(ai + U:)-1
t- 1 So y = ( a:61 + a~Ó2 }( a~ + ai)-1
l:c!=--
t and y= Ó.
This gi ves us Itis important to notice that ifthe variances ai and oi are not known,
Ó cannot be constructed. In the balanced incomplete block model,
Intrablock:
,.. ,.. k .... ~ k(t - 1) 2
where we shall make use of this theorem, the variances are not known.
Ti - T. = lt qi var(r. - r} , .= u2 a Therefore, the theorem cannot be used directly, but a slight variation
will be employed. The variation consists in replacing the variances
Interblock: ui and ~ with their unbiased estimators. The full impact of this
_ ::
7'.· - ' T = Bi - B. var(-fi - .fJ = (t - l~k (a2 + ka~) (18.21)
variation has not been studied in great detail at the present time. In
' · r - A. t(r - A) order to get sorne insight into this, the following theorem will be useful.
410 LINEAR STATISTICAL MODELS T"~ ",. l\IIXED MODELS 411
• Theorem 18.12 Lct fJ 1 and Ó2 be unbiased estimators of8 with vari- Of course, this quantity cannot be calculated, since the variances are
ance ui and u~, respectively. Let Gi and ~be unbiased estimators ..,1
not known. The estimators of a 2 and ~can be used; but, before thisis
of ui and ai, 1~spectively. Let 01 , 02 , a¡,~ be mutually independ- discussed, various forms of Tf will be examined.
Let Ti denote the total ofthe observations that rcceive the ith treat-
ent. Then . - l t
J ._
IT·,. Then
roen t ; t l iat is, Ti= Y-i .. · and let T = -ti=l
is an unbiased estimator of 8.
Proof: The theorem can be proved by using the fact that, if the ran- 1 - l
q¡ =Ti - - Bi =(T.1 - 1.1 ) - - (B. - B)
k k ' .
dom variables x and y are independent, then E(xy) = E(x)E(y). •
w-w'
But, because Ó1 and ~/(di + á~) are independent, thc first term can
where V=----------------
t(k - l)w + (t - k)w'
(18.24)
be written 1 l
~ a;) = OE e~ ~ ~)
and W=- w'=--- (18.25)
E(0 1)E (;, u2 u2 ka~ +
Similarly, the second term can be written -rf in Eq. ( 18.23) is the best (minimum-variance) linear unbiased
combined estimator of Ti - 7_. However, since v is unlmown, this
mators in Eq. (18.21) can be used. If we let Tf denote the combmed t(k - I)w + (t - k)fv'
estimator of T¡ - f., this gives and where wand íi/ are estimators of l/u2 and l/(a2 + ka2), respectively,
~ . (t - l)k (u2 + ku2)11 + jji - ~- k(t - 1) 0'2 obtained from an AOV table.
* Át q, t(r - A.) r - A .iJ
2 There are many waysof estimating l/u 2 and l/(a2 + koi), but we shall
present the method given by Yates [11], who was the first to consider
Ti = k(t - 1) u2 + k(t - 1) (u2 + ka!) the incomplete block model with blocks random.
).t2 t(r - Á) To obtain an analysis of variance, the model will be considered with
If each term is multiplied by ).t 2 (r - .A.)/k(t - 1), this simplifies to blocks fixed in order to use the theory in Chap. 14. Then, when an
AOVtableisobtained, themodel will beconsidered with blocksrandom,
Ti
*
=
kq¡(a 2
(r -
-\- ka~)
.A.)a2 +
+ (B¡ - B.)u
}.t(a2 +
ka!Y
- 2
(18.22) 1 and the appropriate mean squares and expected mean squares will be
equated, so as to give rise to estimators of u 2 and ~·
MIXED :M:ODELS 413
There are two methods of analysis that could be used:
M ethod A. This is often called the metliod of treatments eliminating
(adjusted for). blocks. The sum of squares dueto 'T and {J is broken into
R( -r,{J) = R( T 1 {J) + R({J)
= [SS due to treatments (adj)] + [SS due to blocks (unadj)]
M ethod B. This is often called the method of blocks eliminating (ad-
j usted for) treatments. The sum of squares dueto 'T and{Jis broken into
R( 'T,{J) = R({J j -r) + R( 'T)
= [SS due to blocks (adj)] + [SS due to treatments (unadj)]
Method A is used in Table 14.1. From the above it follows that
R({J j -r) + R(T) = R(T 1{J) + R({J)
or, in words, that the sum of squares dueto blocks (adj) plus the sum of
squares due to treatments (unadj) equals the sum of squares due to
treatments {adj) plus the sum of squares dueto blocks (unadj).
These two methods of analysis are given in Table 18. 7. Method A in
Table 18. 7 is exactly the analysis gi ven in Ta ble 14. 2 except that the su m
of squares dueto blocks (unadj) has been further partitioned into two
parts (A 31 and A 3 - A 31). However, these quantities are not used
directly for estimating r, -
f-..
To compute 7-f in Eq. (18.26), Yates used the analysis given under
.method '.B. The procedure will be explained by rewriting method B of
Table 18. 7 in Table 18.8 and putting in the EMS column. The expected
mean square for intrablock error is as shown because A 4 = B and A
4 4
has the same expectation whether the model has hlocks fixed or blocks
random. Th~ expectation of B 3 /(b - 1) will be examined in sorne _
l detail.
J E(B3) = E(A2 + Á3 - B2) = E(~ l:q7 + z Y~¡.
Al i k
- I YL)
ri
(18.27)
Also,
) and
~ E(~,q~) = (t - l)a2 +~:E(.,, - :;'.)" (18.28)
412
1
414 LINEAR STATISTICAL MODELS MIXED l\IODELS 415
Now, TABLE 18.8 ANALYSIS OF VARIANCE FOR METHOD B OF TABLE 18.7
l Total
sv DF
bk B = ~Yl;m
SS MSI EMS
y2
Mean 1 Bi = -
...
bk
= L.-r--bk
y2 y2
Treatments t - 1 B i.. •..
2
(unadj) i
t(r - 1)
Blocks b - 1 Ba = A2 + Aa - B2 Eb a 2 + b - 1 ub
2
(adj)
Intrablock bk - b - t +1 B4 =B - B1 - B2 - B3 Ee a2
error
- L (rµ + r-r¡)2 - ta: - tu2
i r We get from the EMS column in Table 18.8
Eb
_
-
~
u-
+ t(r - l) ,. 2
ab (18.29)
b-I
E(B3) = .¡ l:(-r¡ - -r.) 2 + (t- l)a2 - ~ :E(-r¡ - i=.) 2 + (b - t)a2 + t(r - l)u;
by
'
r
r
.
square of blocks (adj) and intrablock error in Table ~8.~ ~o obtam est1- where .. (b - I)(Eb - Ee)
V=------------'-----
mates of a2 ando¡. These estimates are used to obtam .,, m Eq. ( 18.26) · t(k - l)(b - l)Eb +
(t - k)(b - t)Ee
1
-
l~
~
<...
1
1
1
MIXED MODELS
+
~ + Theorem 18.14 Let the model be given by Eq. (18.13), i.e., a bal-
1 ancea incomplete block model with blocks random. In addition,
~ let b > t. The notation will be that given in Table 18. 7 for method
........
11 A. If r 1 = r 2 = · · · = rt, the following distributional properties
hold:
""
I~ ( 1) A22 is distributed as
(f
x2 (t - 1).
1
+ ~~a¡;! is distributed as x (b
,.....OlOt-.-1COCO 4 1
~ ~ º~.....: ~ ~ (4) A: - 2 - t).
( N . - l ( N . . - 4 0 ...... <N
,....¡ .............................. ...... a
1 1 1 1 1 1 1
(5) The quantities in (1), (2), (3), (4) are mutually independent.
- 11
From this we obtain the following theorem.
e A 2 is distributed as
(!) " = (bk - b - \+!)A
~ z- (t - 1 4
lltl 1 F(t - 1, bk - b - t + 1).
1 -
~+ (2) v = (b - t) A Aa1-A is distributed as F(t - 1, b - t).
(t - 1) 3 -
=~
31
1 ;:;-
(3) u and ?J are independent.
l--+>~~~-7-~~~~~~-1 e' :::.
(4) Z -2 log ptlpV is distributed as x2 ( 4), where PU is the signifi- "
=
s
~
~ C) ~
ti] Ol~OOOl-C':IO
"o:llCOt-..-11Q,.....CO
....,:.¿a)a)oc)oioci
......
canee level of ii u.ndPv is similarly defined forv. So the hypothesis
r 1 = r 2 = · · · = r t is rejected at the a level of significance if z >
a;> +>
;...
E-t x!< 4).
For a complete discussion of this combined test the reader may consult
Zelen [7] and Weeks [l].
18.3.4 Computing Instructions for the Balanced Incomplete
Block Model with Recovery of Interblock Information. Com-
puting instructions for f{ in Theorem 18.13 will be given. A format
416 such as that ofTable 18.9 will be useful. From this table anAOV can
1
418 I ...INEAR STATISTICAL MODELS l\:CIXED .l\fODELS 419
r
-
p2
---=-
bk
l 1 2 3
y X 1 y 1 X 1 y X
SS for treatments-(adj):
k
Á2=-~~
At 'li (18.32)
------------
4.0 LO 7.1 1.3 7.0 .3
8.0 2.1 9.0 .2.0 8.3 .8
Y2. y2 9.4 3.0 13.2 3.3 9.2 1.1
A3 - ~-± - --:..:..!
SS for blocks (unadj): - 7 k ble 12.8
15.7
4.2
5.0
12.8
15.5
3.4
4.1
10.4
11.5
1.5
1.8
1
SS for blocks (adj):
18.8 In Prob. 18.7, find 82
and set a 90" per cent confidence interval on a2.
18.9 In Prob. 18.7, find -91 - '92 •
The example given in Art. 14.4.1 will be used to illustrate the above
18.10 In Prob. 18. 7, set a 95 per cent confidence interval on T¡ - T2 •
procedure. A 2 and A 3 can be obtained from Table 14.6; the other 18.ll In Prob. 18.7, .find the average variance of the estimates of the
quantities are easily computed from (18.32). difference-of-treatment means.
18.12 In Prob. 18. 7, test the hypothesis T¡ = T2 = Ta with a type I error
B 2 = 3.7880 A2 = 3.2908 Á3 = 2.0347 B 3 = 1.5375 probability of 5 per cent.
18.13 Suppose the data of Table 18.12 satisfy the assumptions for a balanced
The AOV is given in Table 18.10. incomplete block with blocks random. Find the intrablock estimates of T. - -r •
The numbers in parentheses are treatment numbers. ' ·
TABLE 18.10 ANALYSIS OF VARIANCE FOR EXAMPLE OF ART. 14.4.l TABLE 18.12
SV DF SS MS Block Treatments
1
Total 21 169.1162 1 (1) 1.2 (2) 2.2
1 161.8519 1
Mean 2 (1) .8 (3) 1.8
Treatments (unadj) 6 3.7880 3 (1) 1.1 (4) 7.1
Blocks (adj) 6 1.5375 .2563 = Eb 4 (2) 2.2 1 (3) 4.2
Intrablock error 8 1.9388 .2424 = E6 5 (2) 1.6 (4) 6.7
6 (3) 4.5 (4) 6.3
1
18.18 Use Theorem 18.15 and the data in P roblem 18.13 to test the hypothesis
18.4 Show that var(P) = a 2/E32 •
T1 = Tz = T 3 = T 4 with a 5 per cent type I error.
18.5 Show that 18.19 Prove parta (3) and (4) of Theorem 18.10.
18.20 If b > t, show that Ras and Esa in Theorem 18.10 can be used to form
a combined estimator of T 1 - -r.•
18.6 Show that
E(ii2) = t(r - 1) - 1 a2
Further Readinf!
rt l D. L. Weeks: An Exact Test of Significance in the Balanced IncompJete
18.7 The covariance model given in Eq. (18.1) is assumed to fit the data in Block Design with Recovery of Inter-block Information, unpublished M.S.
Table 18.l l. Find fi. thesis, Oklahoma State University, Stillwater, Okla., 1957.
1
420 LINEAR STATISTICAL l\IODELS
P= J
x~(v)
g(u)du=P[u ~ x~(11)]
for probability values P = .0001, .001, .005, .O 1, .025, .05, .1, .25,
.5, .75, .9, .95, .975, .99, .995, .999, .9999, and for degrees of
freedom values v = 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 12, 15, 20, 24, 30,
40, 60, 120.
Table A. 2 Student' s t
~ for probability values p = .0001, .001, .005, .O 1, .025, .05, .1, .25,
.5, . 75, .9, .95, .975, .99, .995, .999, .9999, and for degrees of
421
freedom values v = 1, 1.2, 1.5, 2, 3, 4, 5, 6, 7, 8, 9, 10, 12, 15, 20,
24, 30, 40, 60, 120, oo.
~ ~ ~ ;~ ~ ~ ! : ~ ! s ~ 55! ; ~ ~
0. CO • N H
~ ~ ~ ; ~ i ~ ~ l
H---'
!~ ~ 5! ~ ª : ~ ~
+,:t.++
f~(v1 , v2)
where v1 is the numerator degrees of freedom and 112 is the de- a.!"l __ . ~++.t.•
<tO co
~
422 423
Table A.2 Student's l Table
V
p•0.0001 p•0.001 p•O. 005 p .0.01 p•O.OZS p. o. 05 p•O. I p•O. ZS p•O. 5 p "º· 75 p• o. 9 ps0.95 p•0.975 p• 0.99 p•O. 995 p. o. 999 p• o. 9999
6. 3138 z. 414l. l. 0000 (-114.1421 (-11 l. 5838 (-ZI 7. 8701 1-Z) 3. 9Z90 (•2) l. 5709 (-317.8541 (-3) 1. 5708 ( ·41 l. 5708 1
1 (+31 6. ló6l. jtZ) 6. 366l. (+ZI l. Z732 1+116.3657 (+112..5452 (+111.270!.
1.2 (+31 l. 5434 (+Z) Z.1.654 (ti) s. 9240 l+l) J. 3239 1+111. 5468 8. 6488 4. 7958 z. 0897 (-119.3358 (-1) 3. 9768 (-111.5305 l·Z) 7. 6117 ( ·Z) 3. 8008 ·ZI l. 5198 1·3) 7. 5984 1-3) l. 5197 {-·I) l. 5197 l. z
9. 6353 6. 0166 3. 7051 1.8230 1-1) 8. 7ZS9 (·113.8131 (•I) 1.4763 (·Z) 7. H81 (-Z) 3. 6699 ·Z) l. ·1675 (·3) 7. 337Z ( ·ll l. 4674 1-41 l. 4674 l. 5
l. 5 {+Z) 3. 8f5B (+l) 8.Z847 (+!) Z.8317 (+ 11 l. 78ZO
4. 30Z7 l..9ZOO l. !.0)6 1-118.1650 (-1) ). 6514 (-1) l. 4Zl3 HI 7. 0799 (·Z) 3. 5367 (·Z) l. 4143 (·3) 7.0711 1·3) l.4142 1-4) l. 4142 l.
z (+I) 9. 999Z (+1) 3.1599 (ti) 1.4089 9.9Z49 6. Z053
)
3 (H)Z.8000 (+1} 1.2.924 7.4533 s. 8409 4.1765 3.18l5 l.. H34 l. 4ZZ6 1-1) 7.6489 (-11 l.49ZI (·I) l. 366C (·Z) 6. 8087 (·Z) 3. 4018 (·Zl l. 3604 (-3) 6. 801! (-31 l. 3!.04 (·4} l. 3603
4 (+I} l. 5544 8.6103 s. 5976 4. 6041 3.495'1 2. 7764 2.1319 l.H44 (-1)7.4070 1-1) 3.413! (·11 l. 3383 1-Zl 6. 67Z'l (-l.) 3. 3341 (-Zl l. H34 {•ll 6.ó6b6 (•ll l. 3333 (-4) l. 3333 4
5 1+111.1178 6. 8688 4. 7734 4. 03Zl 3. 1634 z. 5706 2. 0150 l. 3010 (-11 7.Z.668 (-11 3. 367Z 1-111.3u·1 (-Zl 6.5915 (·ZI 3. Z.936 (-Z} l. JI 7Z 1-316.5958 (·3) 1.3112 (-4) l. 317Z 5
l. Z733 (-117.1756 (-113.3365 (-1) l. 3108 f ·ZI 6. 5374 (-Zl 3.2666 (-ZI l.J06·1 (-3) 6. SJZI (-31 1. 3064 1-41 l. 3064 6
6 9. 0823 s. 9588 4. 3168 3. 7074 2. 9687 Z.4469 •l.9HZ
l. ZS43 (·ll 7. 1114 (•ll J. 3145 (·I} l. 30Z9 (-ll 6. 4988 (•2) 3.Z475 (-ZI 1. Z988 1-316.4'36 (·31 1. 2987 (·41 l. Z987
7
7 7. B846 S.407c;. 4. OZ94 3.4994 }:.!!B : ..~~6 l.8946
l. 8595 l. 2403 (·117.0639 (-1) J. Z983 (-1} l.Z971 (·Z) 6. 4701 l·ZI 3.Z3JI 1-z.1 1.z.930 (-316.4651 (•l) l. Z930 (-4} l.Z930 8
8 7. ll.00 s. 0413 3. 83ZS J. 3554 z. 7515 z. 3060
l. 2Z97 (•I) 7. OZ7Z 1-11J.2851 (-11 l.29ZS (·ZI 6. 4477 (•2) 3. Z2ZO (·Zl l. Z886 (-31 6. 4427 (·31 l. 2,885 1•41 J. 2885 9
.p.. 9 6. 5937 4. 7809 3. 6897 3. 1498 2. 6850 2. Z6Z2 l. 8331
N') 10
.p.. 10 6. 2110 4. 5869 J. 5813 3.1692 z. 6338 z. ZZ8l l.81Z5 l.l?ll (-1) 6.9981 (-11 3.Z755 l-ll l.Z889 {•ZI 6.4Z99 (-ZI 3.Zl3l 1-211.2sso (-JI I>. 4ZSO (-31 l. Z8SO {-4) 1. ze50
IZ
IZ 5. 6945 4. 3178 3.4284 J.0545 2. 5600 Z..1788 l. 78Z3 1.Z088 (-1) 6. 9548 (-1) 3. Z.605 (-111.283! 1-z16.4030 ( ·ZI 3. 1997 (-2) l,Z797 1-31 6. )Q84 1•3) l.Z797 1·41 l. Z797
1.1967 1-11 6.9119 (·11 ).2451 1-11 1.2781 1-ZI I>. 3764 (•2) l. 1865 1-ZI 1.2744 (-31 6. 3119 (·31 J. 2744 (·411.2744
u
15 s.2391 ¡ 4. 072.7 3. 2860 z. 9467 Z.4899 2.1315 l. 7531
1.1848 1-116.869& 1-11 3. 2306 1-11 l. 2127 {•ZI 6. 3499 1-Z) 3. 1732 1-21 1.26'11 1-3) 6. 3454
l·l) l. 2691 (-4) l.2691 zo
zo 4.837] 3. 8495 3.1534 2. 8453 2.4231 z. 086G l. 7247
1-41 1.2665 24
u 4.6544 3. 7454 J.0905 z. 7969 2.3910 2. 0639 l. 7109 l. 1789 1-11 6. 8485 (•l} 3. ZZJZ (•l) 1.2700 l·ZI 6. 3166 (·ZI 3. 1667 1-21 l. 2665 1•3) 6.H2l 1-31 1. 2665
30
30 4. 4824 J. 6460 3. 0298 2, 7500 2..3596 2. Q42l 1.6973 l. 1731 1-116.8276 1-11 3. ZI 57 l·ll l.Z61l (·ZI 6. 3234 (·Zl 3.1601 (-21 l. 2638 (-l) 6. 3190 (·3} 1. 2638 1-4} l. 2638
(-4) l.261Z 40
40 4. 3206 J. 5509 Z.9712 2. 7045 Z. JZB9 2.02.11 1.6839 l. 1673 C·l) 6.8066 (-1} 3. 2084 1-111.2646 (-ZI 6. 310Z. 1-ZI 3. 1535 (-211. 2612 (·3) 6. 3059 (-JI l. Z61Z
(·2) l. 2.586 1-JI 6. n28 (-3) 1.2585 (·41 l.Z.585 60
60 4.1681 3.4602 2.9145 2. 6603 Z..2991 z.. 0003 1.6707 l. 1616 (-1) 6. 7862 1-1} l. 2011 (-1) J.2619 l·Z.) 6. Z969 (·21 3. 1469
(-4) l. 2559 IZO
IZO 4. 0254 3. 37l4 2. 8599 2. 6174 2.2699 l. 9799 1. 6576 1.1559 (-1) 6. 7656 (-1) 3.1937 (-l) 1.H93 1-ZI 6.2839 {-21 3. 1404 (-Z.} l. Z.559 (·31 E..Z796 (-l) l.Zs59
CI) 3.8'106 3. 2905 2. 8070 z. 5758 2. 2414 1.9600 1.6449 1.1503 (·116.7449 1-11 3.1864 (-11 1.2566 (-21 6. Z707 1-21 3. 1338 1-21 •• Z.533 (·JI 6. Z.666
(-3) l. 2533 1-41 l.BJ3 ...
v =degrees of freedom.
The numbers in parentheses inclicate the power of ten by which. the number following is to be multiplied; e. g., (-1) 1.2345
.12345.
Table A.2 is reprinted from Lewis E. Vogler and Kenneth A. Norton, Graphs and Tables of the Significance Levels F(v1, va, p)
for the Fisher-Snedecor Variance Ratio, from NBS report 5069, National Bureau of Standards, Boulder Laboratories, by the
kind permission of the authors and the director of the Boulder Laboratory, F. W. Brown.
J
Q)
1 ~+71 4. os29 (+5) 4. 0528 +41 1.621 l (+31 4. 0522 (+Z) 6. 4779 (+21 l. 6145 (+JI l. 9864 5. 8Z85 1.0000 -11 l. 7157 1-21 2. 5085 (·l) 6. 1939 1-ll l. 5437 (-4) 2. 4678 1-5) 6.1687 1-6) 2.4674 (·8) 2. 4674 1
1.z. (6) 2. 3821 (4} 5. 1319 (31 3. 5094 (JI 1.1048 {21 2. 3927 111 7. 4802 111 2. 3000 4. 3669 ·l) 8. 7158 (·11 l. 581' ( •2) Z. 34Z4 1-3) 5. 7938 (·31 l. -1446 1-4) 2. 3097 {-51 5. 7736 (-6) z. 3094 (·81 2. 3094 l. 2
1 5 151 l. 4790 131 6. 8637 121 8.0184 (Z) 3. 1756 fil 9.2839 fil 3. 6200 111 l. 3728 3. ll35 ·ll 7. 6142 1-11 1.4540 1·21 2.1794 (-31 5. 3994 1-31 1. 3468 (·41z.1536 (•51 S.3834 (-6) z. 1533 (-B)Z.1533 l. 5
2 +3) 9.9985 l+ZI 9. 9850 1+21 l. 9850 1+11 9. 8503 (+11 3. 8506 (+11 l. 8513 8. 5263 l.. 5714 1-116.6667 1-11 l. 33)3 (·ZI 2. 0202 1•31 5. OIZS (-31 l. 2S08 (-41 2. 0002 1-51 5.0000 {-6) 2. OOOll (·812.0000 2
J +ZI 7. 8401 (+21 l. 6703 1+11 s. 5552 (H) 3.4116 l+ll 1. 7443 1+11 1.0128 S. 5183 2. 0239 1-11 s. 8506 {-11 l. Z.195 l·ZI 1. 8659 1-ll 4. 6359 (-3) l. 1572 1-41 l. 8507 1-51 4. 6264 (-61 l. 8506 (-811. 8505 3
4 +2) 2. 4162 i+I) 7. 4137 1+1) 3. llH (+I) Z.1198 (+ll 1. 2218 7. 7086 ... 5448 l. 8074 1· U s. 4863 (-1) l. 1654 (·l.) l. 7911 1-3) 4.4SZ8 1-311.1116 (•4) 1. 7779 (·SI 4,4444 1-6} l. 7778 (-8) 1.7778 4
5 +21 l. Z494 (+ll 4. 7181 (+I) z. 2.785 (+I) 1.6258 (+I) J. 0007 6. 6079 4. 0604 l.69ZS 1-1) 5.Z.807 1·11 l. 1338 (·Z) l. 7470 (-31 4.3448 (-31 1.0848 (•4) l. 7350 1-51"·3373 1·6} l. 7349 1-8) 1. 7349 5
6 + 11 8. 2489 !+I) J. 5507 (+I) l. 8635 (HI 1.3745 8. 8131 s. 9874 l. 7760 1.6214 1-11 5.1489 1-11 l. 1132 (·Z) l. 7181 1·31 ... 2737 (-3) 1.0671 1-41 J. 7068 (· 5) 4.Z668 (-61 l. 7067 (-8] l. 7067 6
7 +I} 6. 2167 (+I) Z.9245 (+l) l.6ZJ6 (+11 l. ZZ.46 8. \i7Z7 5. S914 l. 5894 l. 57ll (-1) 5. 0572 1-11 l. 0986 (-2) 1.6976 (-314.2235 (-31 l. 0546 (·41 l. 6868 (-514.2167 1-611. 6867 (·81 l. 6867 7
8 + 11 s. 0694 (+1) z. 5415 1+11 l. 4688 (+11 l. ll.59 7. 5709 5. 3177 J. 4579 l. 5384 1-114.9898 (· 11 l. 0879 (-Z) l. 6824 1-31 4.1862. 1-31 J. 0453 (·41 l. 6718 (·SI 4.1797 1-6) l. 6718 ( ·8) l. 6718 8
.p.. 9 +11 4.3477 +11 2.2857 (+1) l. 3614 (+I) 1.0561 7. Z093 s. 1174 J. 3603 l. 51ZI 1-114.9382 1-11 1.0796 l·ZI 1. 6706 (·3) 4.1573 t-JI 1. o3u (·41 l. 6604 (·5) 4. 1509 (-6) l. 6603 (·811. 6603 9
~
C11 10 + 11 3. 8577 +I) Z.1040 1+111.28z.6 1+1) l. 0044 6. 9367 4. 964~ J. Z.850 1.4915 l· 11 4.897l 1·11 l. 0729 1·21 t.6613 1-3) 4.1343 (-3) l.OJU 1-4) 1. 6513 1-514.1181 (-611.6512 1-8) l. 6512. 10
12 + 1) 3. 2427 +I} 1.8643 (+1) l. 1754 9. 330Z 6. 5538 4. 7472 3.176! 1.4613 1·11 4.8369 (-11 1.0631 (•21 l.647l 1-314.0999 (-31 l. 02)8 (-41 l. 6377 1-51 4.0940 {·61 l. 6376 (•8) l. 6376 IZ
15 +1) z. 7448 (+ 11 l. 6587 (+ 11 l. 0798 8. 6831 6.U9l ... 5431 J. 073Z 1.4321 l·ll 4.7775 (·111. 0534 (·l.) l. 6335 1-ll 4.0659 (-31 l. 0154 (-4) l. 62.41 (·SI 4. 0601 1-6) l. 6240 (·8} l. 6240 15
20 +I} Z. 3399 tll l.4819 9. 9439 8. 091>0 s. 8715 4. 3513 2. 9747 1.4037 (-"l) 4.7192 1-11 l. 0437 (-Z) 1.6197 1-31 4. 032. 1 (-3) l. 0069 1·411.6106 (-5) 4.0264 (-611.6105 (-8) l. 6105 20
Z4 (+11 2.1663 (+I) l. 4028 9. 5513 7. 8229 s. 7167 4. 2597 z. 9271 1. 3898 1·11 ... 6902 1-1) l. 0389 (·2) l. 61l.9 1-Jl 4. 0153 (-31 1. 0028 HI 1.6040 (·SI 4.0096 (·61 1. 6039 (-8) l.603'! 24
30 1+11 z. 0092 (+11 l. 3Z93 9.1797 7. 56l.5 5. 5675 4. 1709 2. 8807 l. 3761 (-114.6616 1-111.0341 (-21 l. 6060 (·31 J. 9986 1-4) 9. 9860 1-41 l. ~973 (· 51 l. 9930 (-611.5972 (•81 l. 597Z 30
40 +I) l. 8668 1+11 l.Z.609 8. 8278 7. 3141 s. 4239 4.0848 2. 8354 l. 3626 (-1) 4.6330 1-11 l. 0294 1·21 l. 5993 (·31 3. 9818 (·419.9443 1·41 l. 5906 (·SI J. 9765 (·61 l. 5906 (-8) l. 5906 40
60 1+111.nn 1+11 1.1913 8.4946 7. 0771 s. 2857 4. 0012 2. 7914 1.3493 (-11 4. 6053 1-111.0247 1-21 1. 5925 1-3) J. 9651 1-419.9030 (·41 l. 5840 (- 51 3. 9599 (·61 l. 5839 (-811. 5839 60
120 +I) 1.6?04 (+I) l.1380 8.1790 6. 8510 S.1524 3. 9201 z.. 7478 l.3362 <-ll ... 5774 1-111.0200 l•ZI l. 5858 1-31l.9487 (·41 9. 8619 1-41 1. 5774 (·SI 3. 9434 (·61 l. 5174 1·8} l. 5774 120
... +11 l. 5131 (+11 1.0828 7. 8794 &.U49 5.0239 3. 8415 2. 7055 l. JZ.33 l·ll 4. ~94 1-111.0153 1·21 l. 5791 (·l) 3. 93ZI (·41 9.1120) (·41 l. S708 1-51 3. 9270 1-611. 5708 1-81 l. 5708 CI)
1 +7) 5.0000 (tSI 5. 0000 (t4) 2.0000 1+314,9995 ltZ) 7. 9950 li-2) 1. 9950 (ti) 4. 9500 7. 5000 l. 5000
(·IJ l. 8889 l·ll 1.1728 (·ZI 5. 4016 l·ZI 2. 5970 1·21 l. 0152 (•JI S.Ol78 1·31 1. 00)5 (-4) l. oooz 1
1.2 (6) 2. 7850 (4) 5.9999 131 4.1033 (3) 1.2921 (Z) 2. 8011 (1) 8. 7817 (1) 2. 7250 5.4476 1.3049 (-1) 3. 6913 (·l) 1. 1518 1·2) 5. 3550 (·2) 2, 5860 (·21 l. 0135 (·3) 5.0335 (•31 1.0013 (·41 l. 0001 l,2
1. 5 (5) l. 6158 131 7.4992 121 8. 7646 121 3.4737 (21 1.0185 (1) 3.9966 (1) l. 5408 4. 0122 1.1399 ( -11 3. 5064 (-1) 1. 1312 (·2) 5. 3088 (•2) 2. 5750 (·2) l. 0118 (•31 5. 0293 (·3) 1. 0012 1·4) l. 0001 1. 5
2 (+3) 9. 999C ltZ) 9.9900 (+2) l. 9900 (+I) 9. 9000 l+I) 3. 9000 (+11 l. 9000 9. 0000 3.0000 1.0000 1-11 3. 3333 1-11 l. 1111 (·ZI 5. 2632 (•212.5641 1·2) 1.0101 1·31 5.0l51 1-3) 1.0010 1-4) 1.0001 z
3 t2) 6.947<1 (+Z) l. 4850 (HI 4.9799 (+I) J, 0817 (+111.6044 9. 5521 5. 4624 2. 2798 (-1) 8. 8110 (-1) 3, 17 IZ 1·11 l. 0915 (·Z) 5.2181 1·21Z,5533 1-2) 1.0084 1·31 5.0208 (-3) 1.0008 (·<11 l. 0001 3
4 (+2) 1. 980C (+l) 6. 1246 (tll 2.6284 (ti) l. 8000 (tl) l,01>49
8. 4336
6.9443
s. 7861
4, 3246
·), 7797
2. 0000 (-1) 8.2843 (-1) 3, 0941 1-11 l. 0819 (-21 s. 1956 (-2) 2, S479 (·21 1.0016 1-31 5. 0188 (·31 1.0008 (-4) l. 0001
l. 8528 1·117. 987'1 l·ll 3, 0489 (-l) 1.0761 (·Z) 5, 1824 (-Z) 2, 5447 1·2) l. 0071 1-3) s. 0176 (·3) J.0007 (•4) l. 0001
•
5 (+I) 9.7021 (ti) 3. 712Z (+I) 1. 8314 (+1) l. 3274 5
6 (+1) 6. 1633 (+112.7000 {ti) 1.4544 (ti) l. 0925 7. 2598 5, 1433 l. 46)) 1. 7622 (-1) 7. 7976 -11 3.0192 (-1) 1. 07Z3 1-21 5. 1733 ·Z) Z. 5425 ·21 J.0067 ·3) 5,0168 (-3) l. 0007 (-4) l. 0001 6
7 (ti) 4. 5132 (+1) Z, 1689 (+I) J. 2404 9. S466 6. 5415 4. 7J74 l,Z574 l. 7010 (.1)7.6655 ,.1) 2,9983 1·11 1.0696 (·21 5. H>'IZ ·2) z. 5410 (·21 1.0065 (·31 5. 0161 (·31 1. 0006 1·411.0001 1
8 (ti) 3.600C (+l) 1.8494 (+IJ 1. 1042 8.6491 6.059' 4.4590 l. llll 1.6569 (·1) 7. S68J (-1) 2. 98U l·ll 1.0676 (-2) 5.1624 (·2) 2. 5398 (•21 1.0063 (·3) 5.0158 {·3) 1. 0006 (·4) 1. 0001 8
9 +1) 3.0342 tl) 1.6387 (+JI l. 0107 a.ou5 5. 71 .. 7 4. 256! ),0065 1.6236 1-l) 7.4938 (·112.970S l·ll 1.0660 1·21 5. 1586 l·ZI z. 5389 1·2) 1. 006Z (-3Í s. 0153 (-3) 1, 0006 (-4) l. 0001 9
~
t..:>
::n 10 (ti) 2.6541 (+!) 1.490! 9. 4l70 7. 5594 5.4564 4.1028 z. 9245 l. 5975 1·11 7.4349 1·11z.9612 l· 11 1.0648 (·21 5.1557 (·Z) 2. 5382 (•Z) 1,0060 (-3) 5.0150 (-3) 1. 0006 f-4) l. 0001 10
12 {ti) 2.1851 (+11 1.2974 8. 5096 6, 9266 5, 0959 3. 8853 Z.8068 l. 559! 1-11 7.1477 1·112.9469 (·l) 1.0629 l·ZI 5. 1512 1·21 z. 5371 (·21 l. 0059 (·3) 5.0145 (-3) 1. 0006 (•4) l. 0001 u
3,6823 1, 5227 (-1) 7.261l (·11 2. 93Z7 1·11 1.0610 (·21 5. 1469 H!l z.5361 (•21 l. 0057 {·3) 5. 0143 (-3) 1. 0006 1-411.0001
15 {+I) l.810l (+I) l. 1339
6. 9865
6, 3589
s. 8489
4. 7650
4,4613 3.4928
2. 6952
20
24 (tl) l. 395; 9.3394 6,6610 5,6136 4.3187 3, 4028 1 z. 5383 1,4695 1-117.1356 HI Z.9116 (·l) 1,0582 (·Z) 5.1403 (·Z) 2. 5345 l•Z) 1,0054 (·3) 5. 0135 (•3) I, 0005 t-4) l. 0001 24
30 (+11 l.2711 8. 773~ 6.nn s. 3904 4.1821 3,3158 z. 4887 1. 4524 (-l) 7,0941 (·ll z. 90'6 1-11 l. 0513 (•2) 5. ll82 l·Z) 2, 5339 (•211,0054 (·31 s. 0133 1-31 1, 0005 1-41 l. 0001 JO
40 (ti) 1.1691 8.250! 6,066<1 5.1185 4.0510 3. Z317 z. 4404 1.4355 (-1) 7.0531 l·ll 2. 8976 (.1) l. 0564 1·21 5. l358 (·ZI z. 5334 {•Z) 1,0053 (·SI 5. 0133 (·3) 1. 0005 (-4) l. 0001 40
60 (tl) 1.078: 7. 7671 5. 7950 4. 9774 3. 9253 3, 1504 Z,3932 1.4188 (•l) 7.0IU l· 11 2.8907 1· ll l. 0555 l•Zl 5. 1337 l·21 Z, 5329 1-21 1.oos2 (-31 s.ouo (-3) l. 0005 (-4) 1.0001 60
IZO 9.954' 7. 3211 5. 5393 4. 786! 3.8041 3.0718 Z.3473 l. 4024 l·ll 6. 9717 1·11 z.8ua (-1) 1.0545 (·Zl 5. 1316 l·ZI 2, !llU (•2) ¡, 0051 1·31 s. 0128 (-3) l. 0005 (·41 l. 0001 IZO
CD 9.ZIO' 6.9071 5, 2983 4,6051 3,688l 2.99S1 1. lOZ6 l. 3863 l·ll 6. 9315 l·ll 2. 8768 C·ll 1.0536 (·ZI 5.1293 l-2) 2. 5318 (•Z) l. 0050 (·31 5.0UJ (·3) l. 0005 1-41 l. 0001 CD
Vi= 3
"2 p•0.0001 p•0.00,1 p•0.005 p•0.01 p•0.025 p• o. 05 p•O, I p• 0,25 p•O. 5 p•O, 75 p•0.9 p•0.95 p•O. 975 p•0.99 p•O, 995 p•O, 999 p•0.9999 "z
1 (+7) 5, 403l (t5) 5, 4031 (t412.1615 (+3) 5. 4033 (tZI 8.6416 (+Z) 2. IS71 {+l} 5, 3593 8.1999 l. 7092 1-114,9410 (•l) l.8056 1·21 9. 8736 1-21 5. 7330
l·ZI 2. 9312 (-Z) 1. 8001 1-3) 5. 9868 (·31 l. 2755 1
1,2 (6) 2. 9549 141 6. 3660 (31 4. 3538 (l) l. 3710 (2) Z,9731 11) 9.3286 (1) 2.9023 5,8883 1.4842 (-1) 4, 7352 l•IJl.8079 (-1) l. 0030 l·Z) S. 8709 l·Z) 3. 0192
1·211.8586 (•J) 6.1978 (-311. 3218 1.2
l. 5 (5) 1.6727 (3) 7.7635 (2) 9. 070 (2) l. 5973 (2) 1.0557 (1) 4.1506 (1) 1.6083 4.2806 1.2947 (-1) 4. 5502 Hl l,8158 l·ll 1.0225 (·216.0349 1·21 l. uzz (·21 l. 9269
1-31 6. 4432 (·3) l. 3756 1. 5
2 (+31 9. 9992 (+2) 9.991' (+Z) 1.9917 (ti) 9. 9166 (ti) l, 9165 (ti) 1. 9164 9.1618 3, 1534 1.1349 l• 1) 4. 3863 (.1) l. 8307 l·l) 1.0469 (·Z) 6, 2329 1·2) l. Z450 (·ZI 2.0081 (·3) 6. 7340 (•J) 1.094 2
3 +Z) 6, 5934 (+2) 1.4111 (+I) 4, 7467 (ti) 2. 9457 (+I) l. 5439 9, 2766 5, 3908 2. 3555 1.0000 1-114,2454 (·I) 1.8550 1-1) 1.0780 (•Z) 6,4771 {·Z) l. 3948 (·2) Z. 1067 (·3) 7.0868 (·3) l. 5167 3
4 {+Z) 1. 8102 {ti) S.6177 (tll 2. 4259 (tl) 1.6694 9. 9792 6, 5914 4.1908 2.0467 1-119.40S4 1-ll 4. 1839 (·I) 1.8717 (.1) 1.0968 l·Z) 6.62ZI (·21 3.4831 1·2) 2. 1647 (-3) 7.2939 1-3) l. 56ZO 4
5 1+11 8. 6292 (ti) 3. 3202 (+I) l. 6530 (ti) 1, Z060 7. 7636 s. 4095 3.6195 1.8843 1-11 9.0715 (•l) 4. ISOZ 1-1) l. 883! (•I) l, 1094 (·Z) 6. 7182 (·2) 3. 5415 1-2) 2.Z030 (-3) 7.4305 (-3) l. 5919 5
6 kt11 5. 368< (+1) 2. 3703 (+l) l. Z917 9. 7795 6. 5988 4. 7571 l. l888 l. 7844 1-l) 8, 8578 (-1) 4. 1292 l·ll 1.8923 1-_I) 1. 1185 (·2) 6, 7866 (·ZI 3, 58Z8 (•2) 2. 2303 (·317.5275 (•3) l.61JZ 6
7 (+l) 3.8671 (ti) l. 8772 (ti) 1, 0882 8.4513 5. 8898 4, 3468 l.0741 l. 7169 l· 11 8. 7095 l·ll 4. 1149 1·11 J. 8989 l·I) l. IZ53 (·Z) 6, 8381 (·Z) 3.6138 l·ZI z. ZSOS 1·31 7. 5998 {-3) l. 6290 7
8 l+I) 3, 0456 1+11 l. S8Z9 9. 5965 7. 5910 5.4160 4.066Z 2. 9238 l. 6683 (•l) 8. 6004 1·114.1044 1-111.9041 1-11 1.1306 (·2) 6.8776 l·Z) 3, 6378 (•2) z. Z66Z 1-3) 7. 6559 1-J) l. 6413 8
~ 9 ti) 2, 5404 (+1) l. 3902 8, 1111 6. 9919 S,0781 l. 86Z6 2. 81Z9 1.6315 1-11 8, 5168 l-11 ... 0967 1-11 l. 9084 1-1) 1.1348 l·Z) 6, 9094 (·2) J, 6570 (·21Z.2788 (•3) 7. 7006 (-3) l. 6511 9
~
..:J 10 +11 2. ZOJ8 (+1) l. 2553 8. 0807 6. ssu 4.8256 3, 7083 2. 7Z77 J,60Z8 1-118.4508 l· ll 4, 0905 (· I) l. 9119 l·I) l, 1382 (·2) 6, 9353 C-ZI 3. 6726 (·21 Z, Z891 1-l) 7. 7371 ( ·3) l. 6590 10
IZ k+ll l. 7899 (+I) l. 0804 7. Z258 5. 9526 4, 4742 3.4903 2. 6055 l. 5609 1-11 8. 3530 l·ll 4,0816 l·ll l. 9173 l-11 1. 1436 1·2) 6.9750 1·21 3,6966 (·Z) z. 3048 (·JI 7. 7933 1-Jl l. 6713 12
IS ~ti) 1.4635 9. 3353 6, 4760 5. 4170 4.1528 3. 2874 z. 4898 l. 5202 1-1) 8. ZS69 (-1) 4,0730 (•I) l. 9230 1-11 1.1490 l·ZI 7,0161 (-2) 3. 7Zll l·ZI 2. 3210 1-31 7. 8509 (-3) 1.6839 15
zo ti) l.Z050 8.09841 5. 8117 4. 9382 3. 8587 l. 0984 z. 3801 J.4808 l•I) 8.16ZI k-ll 4. 0647 l·ll l. 9Z88 l·ll 1.1547 1-217.0587 1·2) 3. 7467 1·21 z. 3377 (-31 7.9103 1-3) l. 6969 zo
24 (ti) l. 0964 7. 5545 5. 5190 4, 7181 3. 7Zll 3,0088 z. 3274 1.4615 l· 11 8. 1153 1-114.0607 (·I) J. 9318 (-1) J.157E (-Z) 7. 0801 (·2) 3, 7597 (·2) 2, 3~62 (-317.9406 l·ll l. 70J6 Z4
30 9.9942 7. 0!45 5. Z388 4. 5097 3. 5894 2. 92Zl Z.Z761 1.4426 l•I) 8.0689 1-11 4. 0568 1-11 l. 9349 1-11 J.1606 (·Z) 7.1018 (·Z) 3. 7729 1·~1 2. 3548 1-317.9714 ( ·l) l. 7103 JO
40 9. l.Z78 6. 5945 4. 97 59 4. 3126 3.4633 2. 8387 Z.ZZ61 1.4239 1-118.0228 1-11 4. osze 1-11 l. 9381 1-1) 1.1635 (-2) 7, IZ40 l·Z) 3. 786) (·Z) Z, 3636 (·31 8. 0026 (·311.7171 40
60 8. 3526 6.171Z 4. 7Z90 4, 1259 l. 3425 Z.7581 z. 1774 1.4055 l·ll 7. 9770 1-11 4. 0491 (·ll 1.9413 (-11 1.16€.6 1·21 7. 141>9 l·ZI 3. 8000 (·Z) l.. 3725 (•l) 8. 0343 (•l) l. 7241 60
IZO 7. 6584 5. 7814 4.4973 3. 9493 3, 2Z70 2. 680Z 2. llOO l. 3873 (-1) 7. 9114 1-11 4.0453 (-11 l. 9446 l·ll l, 1697 (·ZJ 7, 1700 1·21 3. 8137 (·ZI Z. 3816 (·31 e. o66s l•l) l. 7311 IZO
CD 7. 0358 5,4221 4. Z794 3. 7816 l, 1161 z. 6049 2. 0838 l. 3694 1-1) 7.8866 (.1) 4.0417 l·ll l. 9479 1·11 1.1728 1-ZI 7.1932 (·213,8Z78 l·ZI z. 3907 ,_,, 8. 0991 (·31 l. 7383 CD
V1 =4
p•O. 25 p•O. S p•0.15 p•0.9 p•O, 95 psO. 975 p•O. 99 p•O. 995 p•O. 999 p•0.9999 "z
p•0.001 p=0.005 p •0.01 p•0.025 p•0.05 p•O. I
"z p•0,0001
8. 5810 1.8zz7 (.1) S. 53Z8 (-112,2003 l·ll l.Z97l {•Z} 8. 1846 1·21 4. 7174 l•ZI 3.1915 1-21 l. 3488 1-3} 4. 1387 1
1 1+7} 5. 6ZSC (+5} S. 6250 IHI Z,2500 (+31 5.6246 l+ZI 8. 9958 (+2) Z. 2458 (+I) 5. 5833
11 z. 9990 6. IZ65 l. 5810 1-11 5. 3271 1-l)Z.ZZ49 l-11 l. 3347 (·Z} 8. 5083 1-214.9438 (·Z} 3. 357Z 1-211.4256 1·314.3849 l. 2
l. z 6} 3.0HS (4} 6. 5663 31 4.4908 (3) l. 4l•tz (ZI 3. 0671 {11 9.6Z74
4.4235 l. 3780 1-11 5.1487 (-1) 2.Z608 (-1) l. 3811 (·21 8. 9012 (·ZI 5. 2170 (-21 3. 5571 (·ZI l. 5181 1-3) 4. 68Z 1 J. 5
l. 5 ~SI l. 70l8 ~31 7,9077 Z) 9.2437 (2) 3. 6648 (21 l. 0759 (1) 4.2343 (1) 1.6446
1.2071 l· 1) s. 0000 l·ll z.3124 1-11 l. 4400 (·ZI 9. 3906 C-ZI s. 5555 (·ZI 3. 8046 (-Z) l. 63Z8 1·31 5. osos z
9. 2434 l,ZJZO
z (+31 9. 9992 1+2} 9. 9925 1+21 l. 992S (ti) 9. 9249 (+I) J. 9248 1+111. 9247
3
9.1172 s. 3427 z. 3901 1.06lZ 1-11 4. 8859 1-11z.3862 1-11 l. 5171 1-11 l. 0021 {.21 s. 9902 1·21 4. 1222 HI 1.7801 {·l) 5. 5Z4J
l +ZI 6. 4019 1+2) l. 3710 (+l) 4. 6195 1+112,8710 1+11 l. 5101
4
9. 6045 6.3883 4.1073 l.,064.l. 1.0000 (-1) 4. 8445 1•11 Z.4347 1-11 l. 5654 (-1) 1.0412 1·21 6. Z.590 1-z.14.3181 (-21 l. 8114 1-31 s. 8183
4 (+ZI l. 7187 (+11 s. 3436 l+ll z. 3155 1+11 l. 5977
5
l. 89Z.7 1-1} 9. 6456 (-1) 4. 8Z56 1·11l..4688 1-11 l. 5985 (·11 l. 0679 l·ZI 6.4U~ 1-Z.I 4. 45)2
1-211.9338 1·31 6. 0193
7. 3879 S.19ZZ 3. 5.l.OZ
5 (+ 11 8. 0527 l+I) 3. 1085 1+11 l. SSS6 1+11 l. 139Z
1·2) l. 9792 l·l) 6. 1657 6
9.1483 6,ZZ.71. 4. 5337 3. 11108 l. 787.l. (-ti 9.4191 (-1)4.8156 1-11z.4939 1·11 1.6226 1-11 l. 0873 ~-ZI 6. 5759 {-21 4. 5506
6 +1) 4. 1Hl9 (ti) Z.1924 1+111. zoz8
(-212.0138 (-3) 6. 2771 7
7.11467 5. 5ZZ6 4.1203 2. 9605 l. 7157 1-119.2619 1-114,8100 l· 11 2. suz l·ll 1.6409 1-111.1020 (·ZI 6.6774 1-21 4.6.l.49
7 j+ll l. 5Z2Z +111. 7198 1+11 1. 0050
l·ZI 6. 7572 1·2) 4, 6834 1·21Z.0410 {·316,3648 8
3,8378 2. 8064 1.6642 1·11 9.1464 1-11 4. 806S 1-11 z. 5285 (•11 l,6SS4 1-111.1136
8 1+11 2. 7493 (+11 1.4392 8.8051 1. 0060 s.0526
(·31 6. 4357 9
4. 7181 3.6331 2. 6927 l. 6ZS3 l· 11 9. 0580 (-11 4, 8045 2. 5408 (·IJ 1. 6670 (·1) 1.1230 1·2) 6. 8211 1-21 4. 7306 (·ZI Z. 0629
1-11
9 +11 2. 2766 1+1) 1.2560 7. 9559 6. 4221
i;:..
l·ZI 6. 8747 l·Z) 4. 7694 l·Z} Z. 0811 (-3} 6. 4941 10
NI
7, 3428 5. 9943 4, 4683 3, 4780 2. 6053 l. 5949 1·118.9882 1-11 •• 8031 (•l)Z.5Sll (-1) 1,6766 1-11 1.1307
CX> 10 +11 1.9630 (+I) 1.1283
(·ZI 6, 9570 l·ZI 4, 8298 (-ZI z. 1092 1-3) 6. 5849 12
6. 5Zll 5,411' 4, 1212 3. ZS92 z. 4801 l. 5503 (-1) 8. 8848 1-11 4. 8017 1-112.S611 1-11 1,6916 1-111.1427
IZ +I) l. 57)3 9.6327
l·ZI 7,0432 l·Z) 4, 8928 (·Z) Z.1385 (-3) 6. 6796 IS
s. 8029 4. 8932, 3. 804 1 3,0551 z. 3614 l. 5071 1-11 8. 7830 1-11 4. 8010 l·ll z. 5847 1-111. 7071 1-11 1.1ss2
15 1+11 l. Z783 8. 2527
(-11 1.1682 1-21 7.1327 (-ZI 4, 9586 1·212.1692 {·31 6. 7786 z.o
4,4307 3. 5141 2. 8661 z.1.48' l. 465. 1-11 8. &831 1·114.8012 1-11 2. 6013 1·111. 7234
zo (+11 1.041! 7. 0960 5. 1743
1-111.1750 1·21 7. 1793 l·ZI 4. 9~Z5 1•2) Z.Ul50 1-31 6. 8298 24
4, 8898 4.2184 3, 3794 2, 7163 Z.1949 l. 4447 l· 11 8. 6335 1-114.8015 (-11 2. 6103 (-1) l. 7318
l.4 9. 4Z.46 6. 5892
(-11 1.1819 l·ZI 7,2265 (·21 5.0271 1-212.2013 (-31 6.88Z2 30
4. 6233 4.0179 3.2499 z. 6896 2, 1422 1.4244 1·118.5844 1-114.8019 1-11 2.6196 1-11 1. 7404
30 8. 5437 6.1245
{-3} 6.9358 40
HI l. 7492 (-1) 1.1889 1-217.2754 (·21 5, 0628 1·21 2.2.179
4. 3738 3. 8213 3.1261 2.6060 2. 0909 l. 4045 (.1) 8. 5357 (·ll 4. 8028 (-11 2. 6291
40 1. 759Z 5. 6981
1·3} 6. 9907 60
l·ll 1.1961 l·ZI 7, 31.49 1·2) 5. 0992 1·21 2.Z349
5, 3067 4, 139~ 3,6491 3.0077 z. 5252 Z.,0410 l. 38411 (-1) 8, 4873 (-ti '· 8038 1-11 Z.6388 1-11 l. 7581
60 7. 0599 (-3)7.0470
l·ZI 5, 1366 l·ZI 2, 2523 IZO
), 9207 3. 4796 2. 8943 2. 4472 l. 9923 l. 3654 1-11 8.4392 1-114.8049 (-1) z. 6488 1-11 l. 7674 l·ll 1.Z035 l·ZI 7. 3757
IZO '" 4357 4.9472
{-217.4278 1-21 5. 1746 1-21 2 • .no1 (-317.1046 m
5. 8782 4, 6167 ), 7151 3. 3192 2, 7858 2. 3719 l. 9449 l. 3463 (-11 8. 3918 l-11 4. 8063 l· 11 2. 6591 1-111. 7768 HI l. 2110
m
~.J
------- ""º ---~-
V1 =5
"z p•0,0001 p•0,001 p•0.005 p•0.01 p•0.02.5 pa 0,05 p•O. I p• 0,25 p•O, 5 p•0, 75 p•0.9 p•0.95 p•0,975 p•O. 99 p•O. 995 p•O. 999 p•0.9999 ~l.
1 (+71 5. 7640 1+51 5. 7640 1+412.3056 1+3) s. 7637 1+21 9,ZlllS (+212.3016 1+11 5. 7241 8.8198 1.8937 1·11 5. 9084 1-11 2. 462.8 1·111. 5133 (·Zl 9. 9930 l·Z) 6.1508 (·214,3889 (-21 2.1195 (·318.0038 1
1.2 161 l.106Z 141 6. 6921 (3) 4. 5769 131 1.4413 (21 3. IZ62 {I} 9. 8152 111 3. 0596 6.2153 1:6415 1-1) s. 7046 (·1) z. 5048 1-111. 5689 1-11 l. 0483 l·ZI 6. 51•9 {·214.6708 1-z1 2.2101 1-31 8, 6059 1.2
1. 5 (5) l. 7233 131 7. 9984 12) 9. 3500 121 3. 7072 (Z} l. 0887 111 4,Z867 111 1.6673 4. 5121 1.4298 1-11 5, Sll.7 1-11 z. 5636 HI 1.6385 (·1) 1.1011 l·ZI 6.96U l·ZI s.01111 f·ZI z. 4561 f•l) 9. 3515 l. 5
1
z (+31 9. 9993 1+2) 9. 9910 l+ZI l. 9930 !+11'·9299 (+l) 3. 9298 (+1) 1. 9296 9, 2926 l,2799 1.2519 1·11 s. 3972 1-11z.6457 1-111. 7283 1-1)1.1851 (·217,5335 1·21 s. 4603 f ·ZI z. 6938 (·ZI l. 0306 2
3 (+Z) 6. Ul 1 1+2) 1.3458 l+l) 4,5392 (+I) Z.BZ37 (+l) 1.4885 9. 0135 5. 309Z z. 4095 1.1024 l·lt 5. 3070 1-11 z. 7628 l•ll 1,8486 l-11 l. 2881 f·ZI 8, 2919 1·21 6.0496 1·21 3. 0118 l·Zll.1589 l
4 (+Z) l. 6613 (+11 5.1712 l+l) Z, 2456 1+11 l. 55ZZ 9. 3645 6, 2560 4. 0506 2. 07Zl 1.0367 (·11 5.2835 1-11z.8407 l•ll 1. 9260 1·111.3536 {·21 8.1781 l·Z) 6. 4284 (·2) 3.2170 (·2) l,Z·U8 4
5 (+ 11 7. 6911 1+11 2. 9752 l+l) l. 4940 1+11 1. 0967 7.1464 sloso3 3. 4530 l. 8947 1.0000 (-11 5. 2779 1-11 z. 8960 1-11 l. 9801 1-1) l. 3993 l·ZI 9.1183 (·21 6,6934 1-2) 3. 3611 (·ZI l. 3002 5
6 (+11 ... 6741 l+ll 2. 0803 (+I) 1.1464 8. 7459 5, 9876 4. 3874 3, 1075 l. 7852 l· 11 9. 7654 1-11 5, 2784 (-11 2. 9373 f·ll z.0201 (·11 1.4331 l·ZI 9. 3703 (·~) 6.8904 ( ·Z) 3. 4681 (·21 1, 3436 6
7 (+ 1) l. 3 OSI (+l) l. 6Z06 9. 5ZZI 7. 4604 s. 2852 3. 9715 z. 8833 l. 7111 (•11 9,60Z6 l·ll 5.2811. 1-11Z.969Z f·ll Z.0509 l·ll 1.4592 {·Z) 9. 5639 (·ZI 7.0423 1·ZJ l. 5508 ( ·21 l. 3772 7
8 (+11 z. 563! l+ll 1.3485 8. 3018 6. 6318 4.8173 3.6875 2. 7Z65 1.6575 l·ll 9.4831 (-1) 5. 2846 1-112.9946 1-11 z. 07S4 l·ll 1.4799 1·219.7191 (•2) 7.1628 l·ZI 3. 6167 1·21 1. .. 040 8
1,6170 l· 11 9. 3916 (·11 5.Zl!l79 (.1) 3, 0154 1·11z.0953 1-11 1.4968 (·ZI 9, 8445 l•Zl 7,2.611 1-21 3. 67os l·Zl 1 ... 259 9
t
<D
9 1+1) z.1112 1+11 1.1714
6.11723
6, 0569
s. 6363
4,4844
4,2361
3.4817
l. 3Z58
2.6106
l.. 5216 l. 5853 (-11 9. 3193 (-11 5, 2913 1·11 l.03Z7 1-1) 2.1119 l·ll l. 5108 1·219.9493 (•ZI 7,3432 1·213.715Z (·Zl 1.4440 10
IZ (+11 1... 471 8. 8921 6.0711 5,060 3. 8911 3, 1059 2. l940 l. 5189 1-11 9.2124 1·11 5, 2975 1-1) 3. 0598 1-11z.1378 l·ll l.5327 l•IJ 1.0113 (·Z) 7.4716 1·2) 3. 7853 l·Z) 1.1726 IZ
IS 1+11 J.1621 7. 5674 s. 371.1 4. 5556 l, 5764 z. 9013 2.2730 1.4938 {·11 9.1073 1-11 5. 3041 1·11 l. 088) 1•11 Z.1651 1-111. 5558 1-11 l. 02116 l•Zl 7. 6069 l·Z) 3. 8594 (·ZI J. 5028 IS
l.O 9. 388( 6.4606 4. 7616 4. IOZ7 3.2891 z. 1109 Z.158Z 1,4500 (.1) 9, 0038 1·11 5. 3135 l· 11 3, 1185 HI 2. 1939 1-11 l. 5802 1-11 l. 0468 l·ZI 7. 7501 1·Zl 3. 9378 l·Z) l. 5348 zo
l.4 8.4578 5. 9768 4,4857 l. 8951 1. s¡s48 z. 1>207 2.1010 1.4285 1-11 a. 9s21 1-11 5. 3186 (-1) l, U•U 1-11 2 . .1089 (•I) l. 5919 f•ll 1.0564 (·.ti 7. Bl.47 1-ZI 3. 9788 1-ZI l. HU u
lO 1. 6322 5. 5339 4, ZZ76 l. 6990 l, OZ65 2. 5336 2,0492 1.4073 <· 11 8. 9019 1-11 s. 3ZJ7 (-113,1505 l· 11z.2243 l-11 l. 6059 l•l) 1,0662 (•ZI 7. 9014 (·Z) 4. OZll -ZI l. 5687 30
40 6. 8981 5. IZ83 3, 9860 ), 5138 z. 9037 2. 4495 1.9968 1.3863 l-11 8.8516 1-11 5. 3296 (-1) ), 1673 l•ll 2.2402 (-111.6194 1·11 1. 0763 (·ZI 7. 91108 ·Zl 4. 0647 -21 1, 5865 40
60 6.246! 4. 7565 3. 7600 3, 3389 z. 7863 2,3683 l. 9457 l. 3657 1-118.8017 (-1) s. 3356 (·113.1845 1-11 2.2566 1-11 1. 6333 1-111.0867 {·Zl 8.0632 ·21 4.1097 l·ZI l. 6049 60
IZO S.6661 4. 4151 3. S48Z. 3.1735 Z.6740 2. Z900 l. 8959 l. 3453 1-11ll.7521 1·11 5. 34ZZ (• l} 3, 2023 l• I) Z.2736 {-1) l.6476 1-11 l. 0975 l·Zl 11.1473 ·214.156Z ·Z) l. 6.l.39 120
m S. 149C 4. 1030 3. 3499 ), 0173 z. 5665 Z.2141 l. 8473 l. 3251 1-1) 8. 70Z9 1-11 s. 3493 l-ll 3, ZZ06 (·112.2.910 l· 11 1. 6624 1-11 l. 1086 (OZI 8. 2345 -2) 4. Z043 -21 l. 6435 m
Vi= 6
p• o. 25 p•O, 5 p• o. 75 p•O. 9 p•O, 95 p• o. 975 p•0.9 9 p•O. 995 p• o. 999 p•0.9999 '2
p.0.005 p•0.01 p•0,015 p• o.os p•O, I
"z p•0.0001 p• 0.001
1
(tl) s. 8204 a. 9833 1.9422 1-ll 6.1675 1-11 2. 6483 (-1) l. 6702 (-1) 1.1347 (-2) 7.2154 (·ZI 5, 3662. {·Z) Z, 8163 l·Z) 1. 2123
l 1+7) 5. 859< (+5) 5. 8594 (+4) 2. 307 (+3) 5. 8590 (+z) 9, 3111 (+Z) Z, 3399
1.68!8 (-1) 5.9653 (-1) 2. 7040 (-1) 1.7403 1-1) 1.1975 (·Zl 7. 7613 1·2) 5, 7558 l·Z) 3, 0441 (·21l.3171 1.2
l,Z (6) 3, 1463 141 6. 7785 (3) 4.6360 (3) 1.4599 (2) 3, 1668 (1) 9,9439 (1) 3, 1012 6. 3770
(l) 4. 3226 11) 1, 6BZ9 4, 5724 1.4652 (-1) 5. 7991 1-112.7806 (·1) 1. 1288 1-111.2760 (·ZI 8, 3679 )·Z) 6.ZOO (·2) 3. 3302 (·ZI 1.4492 l. 5
l. 5 15) l. 7367 131 a. or.06 (2) 9.4230 (Z) 3, 7363 (Z) l. 0974
9. 3255 3.3Ul l. 2824 1-11 5.6747 l•ll 2. 8874 (-1) l. 9443 1-11 l. 3774 t-2) 9.1533 (-21 6. 8757 (·ZI 3. 7037 1-2) l. 6225 z
z (+3) 9. 9993 l+Z) 9.9933 (+zl l. 9933 (+l) 9. 9332. l+ll 3,9331 (+11 l. 9330
8. 9406 5.2847 Z.4218 1.1289 (-l) 5. 6041 1-ll 3. 0406 1-11 2.1021 1-ll l. 5154 l•ll 1,0225 (·2) 7. 7417 (·21 4.2188 (-21 l. 8629 3
3 (+2) 6.1991 1+21 l. 3285 (+ll 4. 4838 (+l) 2. 7911 (+ll l.4735
9.1913 6.1631 4.0098 2. 0766 1.0617 (-11 5. 5953 C·ll 3.1439 1-11 2. 2057 C·ll l. 6059 l•ll 1.0931 1-2111.3139 l·Z) 4. 5613 (•Z) 2. OZ.35 4
4 (t2) 1.6219 (+11 5. 0525 (ti) 2.1975 ltll l. 5207
6. 9777 4. 9503 3. 4045 l. 8945 1.0240 (-1) s. 6016 1-11 3. 2180 1-ll Z.2793 C·ll 1.6101 (· ll 1. 1434 l·ZI 8, 7230 1-214.8071 1-2) 2. 1392 5
5 l+l) 7.44ZI (tl) 2,81134 (ti) l.45U (tl) 1,0672
1 l.0000 1-11 5.6114 1-ll 3. 2738 1-ll 2. 3343 1-111. 7183 l•ll l. llUZ 1·21 9.0310 (•2) 4, 9926 1-212.2261 6
8.4661 5. 8197 4. 2839 3,0546 1. 7821
6 (+ll 4. 4909 (+11 z. 0030 (+11 l. 1013
9.1554 7.1914 5.1186 l,11660 z. 8274 l. 7059 1-11 9. 8334 l-l)5.6Zl5 l·ll l. ll'U 1-11 z. nn (· ll l. 7558 1-111. Zl07 l·ZI 9,nu (·2) 5. 1378 l·ZI 2. 2954 7
7 (+I) 3. 1561 (+ll l. 55Zl
7. 9520 6, 3707 4. 1>517 3, 5806 z.6683 l. 1>508 l·ll 9. 7111 C·ll 5. 6306 l•ll 3.3513 (-1) 2.4115 (-l) l. 7858 l·ll l. 2343 (•Z) 9. 4643 1·2) 5. 2548 1-Z) 2, 3507 8
8 (+l) z. 4351 l+ll 1.2858
7, U38 5.8018 4.3197 3.3738 z. 5509 l. 6091 (-1) 9, 6175 C· ll 5. 639Z 1· ll 3. 3810 l·ll z.4396 (-1) l. 8105 1-111.2.537 (•ZI 9. 623? l·Z) 5. 3510 1-21 2. 3963 9
9 l+ll 1.9974 (+I) 1.112.8
t
o 10 ltl) l. 7081 9.9251> 6. 5446 5. 3858 4,07ZI 3, Z172 2. 4606 l. 5765 (·119,5436 (·ll 5.64U C•ll J.4050 l-ll 2.4631 1-ll l. 8311 C·ll l.noo (•Zl 9. 7561 l·Z) 5. 4316 (·ZI Z,045 10
4,8206 3.7283 2.9961 z. UlO l. 5286 1-11 9.4342 1-ll 5.6600 (·113.442.7 C·ll 2. sooo 1-111. 86S5 1-11 l. 2956 (·Z) 9, 9661 l·ZI S. 5590 (·21 Z.4950 u
lZ K+ll l. 356C 8.3788 5. 7570
4,JlU 3.4147 2. 790! z. 2081 1.4820 (-l) 9. ll.67 HI 5.6750 l·ll 3.4829 l·ll 2,5393 (·ll l. 8910 ~-l) l,JU9 c-111.0190 C-21 5. 6952 (•2) 2. 5597 15
15 ~+ll 1. 0819 1. 0917 5. 0701
6.0186 4.4721 3.8714 3.1283 z. 5990 z.0913 1.066 (-1) 9. ZZlO 1-11 5. 6918 (-1) 3, 5Z57 1-11z.5812 (·l) l. 9341 k-ll l. 3521 l-11 l.0429 1-2) 5. 8411 1·2) z. 6291 20
20 8. 6789
s. 5504 4, 2019 3.6667 z. 9946 2. 5082 2.0351 1.4143 1-11 9.1687 1-115.7013 l·ll l. 5482 l·U z.6031 1-111. 9542 ~-o 1. 3675 l·ll l. 0555 (·21 S. 9181 (·212.6658 24
Z4 7. 7896
(•2) z. 7039
5.1223 3. 9492 2.8667 2. 4zb5 1.9803 l. 3923 1-119,1169 C·ll 5.7110 l·ll 3, 5?14 C·ll z,6Z59 1-11 1. 9743 l•ll 1.3834 (·11 l. 0686 (-2) 5.9980 30
30 7.0017 '· 4735
4. 7306 3. 7129 3, 2910 2. 7444 z. 3l59 l. 9269 1.1706 (·ll 9. 0654 C·ll s. 7218 1-ll 3. 5956 l·ll z.6495 1-11 l. 9950 1-11 l. 3999 1-11 l.oazz (-2) 6. 0811 (·ZI Z. 7435 40
40 6. 3031
2. 6Z74 z. 2540 l. 8747 l. 3491 C-11 9.0144 1-11 5. 7330 1-111. 6206 1-11 z.6739 (•ll 2,0166 l·ll l.4171 1·11 l. 0963 (-Z) 6, 1674 l·Z) 2, 7847 60
60 5. 6830 4. 3721 3, 4918 3.11117
uo 5.1323 4.0437 3. 2849 2. 9559 z. 51:54 2. 1750 l. 82.l& l. 3278 1-ll 8. 9637 l·ll 5. 7448 1-11 3,6466 (·11 2.6993 {·I) Z,0389 (-1)1.4349 l·ll 1.1109 (-21 6. 2574 (·Z) 2.8ZT7 120
(·Z) Z, 8725 CD
3. 7430 3,0913 2. 8020 2,408Z 2,0986 1,7741 1.3068 (• I) 8. 9135 1-11 5. 757'1 l·ll 3. 6735 1-112.7ZS7 1-ll 2.06U 1-111.4535 l•ll l. ur.z (·Z) 6. 3511
CD 4.6421
~ ----~
4
V.= 7
"z p•0.0001 p:o0,001 ps0,005 p•0.01 paO.OZS p•0.05 p•O, I p.o. 25 p•0.5 p•O, 75 p•O. 9 p= 0.95 p• o. 975 p•0.99 p•O, 995 p•0.999 p•0.9999 •2
l 1+7) 5. 928~ (+SI 5. 9Z8~ 1+4) z. 3715 (+ll 5.9283 l+ZI 9.4BZ2 (t2) 2. 3677 1+11 5. 8906 9, 1021 l. 9774 l· IJ 6. 3565 l• 11 Z. 7860 (.1) l. 7885 1-111.2387 (·21 8. 1659 (·21 6. 1592 1·21 3.4194 1·21 1. 6086 l
l. 2 (61 3.1755 (41 6.8413 (3) ... 67'10 13) l. 4735 (2) 3. 1963 121 1. 0038 (1) 3.1314 6. 4509 l. 7128 1-116.1560 l·ll 2, 85Z3 l•l)l.870Z (-111. 3130 1·ZI 8. 7554 (-2) 6.6435 1-21 3, 7203 1-21 l. 7609 l. 2
l. s 151 1.7465 (3) 8. 1059 121 9. 4761 (21 3, 7574 (2) 1.1037 (ll 4. 3487 (1) 1. 6941 4. 6161 1.4908 (-11 5. 9944 (-112.9432 (-1) l. 9741 (-111. 4064 1·219.4990 1·217.2556 (·ZI 4. 1029 1-2) l. 9557 l. 5
2 (+l) 9. 9994 k+z1 '9. 9936 (+Z) l. 9936 (+ll 9. 9356 1+11 3. 9355 1+11 1. 9353 9. 3491 3. 3352 l. 3045 l· 11 5, 11789 1-113.0699 l·ll Z.1109 (·l) 1. 5287 (-11 l. 0475 (•Z) 8. 0619 C·Z) 4. 6106 (·212.2157 2
3 C+21 6. 1388 k+ZI l. 3158 1+11 4.4434 1+11z.7672 (+I) l. 4624 8. 8868 S.1662 z. 4302 '
1.1482 (• I) S. SZ.U (· 11 3, ZSJO 1-11 2. 3005 1-1) l. 6979 (·11 1.1832 (·21 9. 1895 C·21 s. 3270 l·Zl 2, SU6 3
4 (+Z) 1.5931 l+ll 4. 9658 (ti) 2.16U 1+11 1. 4976 9.0741 6.094l 3, 9790 2. 0790 1,0797 <·11 s.aus l• 11 3. 3778 (•11 Z.4270 1-11 1. 8107 1-11 1. 2744 1-21 9, 9502 l-21 s. 8146 l·Z.12.8391 4
S l+I) 7.Z611 "+11 2. 8163 l+l) l. 4200 l+ll l. 0456 6.8531 4. 8759 3. 3679 1.893! 1.0414 (·ll 5.11442 1-11 3. 4681 l·I) Z. 5179 l-1) 1.8921 1·11 l. 3404 1-11 1. 0502 l•ZI 6. 1706 1·21 3. OZSI 5
6 (+11 4.3566 k+11 l. 9463 (+11 1. 0786 8. 2600 S.695! 4. 2066 3. 0145 l. 7789 l.0169 1-ll 5.8620 l· 11 3. 5368 l· 11z.5867 l· 11 l. 9537 1-11 1. 3905 1-.11 l. 09ll 1·2) 6,4430 (·Z) 3. 1679 6
7 1+11 3. 0471 (+11 1.5019 8. 8854 6. 9928 4. 9949 l. 7870 2. 71149 l. 7011 l.0000 1-11 s. 878! 1-11 3. 5908 1-11 Z.6406 1-11 z.0020 C·ll 1,4300 1-ll l. IZS4 l·ZI 6. 6584 1·21 3. ZSIZ 7
8 K+ll 2. 3421 K+ll l.Zl98 7.6942 6, 1776 4. 5286 l. 500! 2, 6241 J.6441 1-11 9. 8757 1-ll 5. 8931 f-11 3, 634Z 1·11 Z.6841 1-11 2. 0411 1-111.4620 1-ll l. JSZJ l·ZI 6. 8334 l·ll 3. 37)3 8
~ 9 "+11 l. 9140 +l) 1.0698 6,8849 5,6129 4. 1971 3, Z9Z7 2. 5053 1, 602l 1-11 9. 780! 1-11 5. 9063 HI l.6701 1·112.7198 f-11 2. 07J3 1-11 1. 4884 1-11 1.1746 l·ZI 6. 9784 l·Z) 3. 4498 9
~
l...& JO +11 l. 6319 9. 5175 6, 30ZS s.2001 3, 9498 3.135! Z.4140 l. 5688 1-119.7054 1-11s.9179 1-11 3. 7003 l• 1) z. 7499 1-1) z. 1004 c-111. s106 l·ll 1.1933 l·Z) 7. 1006 1·21 3.5143 10
12 1+11 l,Z89Z e. 0009 5. 5245 4.6395 3,606! 2. 9134 Z. Z8Z8 l. 5197 1·11 9. 5943 1-11 5. 9372 1-11 3. 7480 1-11 z. 7974 !-11 2. 1433 l• I) l. 5458 (•11 l. ZZJO 1-:1 ·¡, ¿954 l·ZI 3.6174 12
IS +11 l. 0231 6. 7408 4. 8473 4.1415 3. Z9l4 z. 7066 z.15a2 1,4718 1-11 9. 4850 1-11 s. 9591 1-11 3. 7991 (•I) 2. 8484 (-11 Z. 189Z 1·11 l. 5837 1·111. Z~51 (·Z) 7. 5054 (·ZI l. 7287 1s
20 8.1577 s. 6920 ... 2569 ], 6987 3. 0074 z. 514C Z.0397 l.4ZS2 1-11 9. 3776 1-11 5. 9837 1-ll 3. 8540 1·11z.9032 l·I) Z. Zl88 1-11 l. 6246 1·11 l. 2897 l·ZI 7. 7330 1·2)3.805 zo
24 7.2980 5.2349 l. 9905 3. 4959 2. 8731 2.4226 1. 9826 1.4022 1-119.324! 1-11 s. 9970 (·11 3.8830 l·ll 2.9321 (· 11 z. Z650 1-11 l. 6463 f-11 1.3080 (·2) 7. 8541 1·21 3.9139 u
30 6. 5375 . 4.8173 l.1416 l. 3045 z. 7460 2.3Ml 1.9269 l. 379! 1·11 9.271~ 1-0 6. 0114 l-ll l.9UI l· 11 l. 9623 1-112.29Zl 1-11 1.6689 1-11 1. 3272 1·21 7. 9806 1·213.9813 JO
40 5. 8640 4.4355 3. 5088 l.1238 2.6U! Z. Z49C l. 87Z' l. 3571 1-11 9. Zl97 l• 11 6.0266 1-113.9446 1·11z.9937 1-11z.3Z08 1•11 l.69Z! 1·11 l.H73 l·ZI 8. 1129 l·ZI 4. 0518 40
60 S.2672 4. 0864 3. 2911 2. 9530 2. 5061 z. 166• l. 819~ l. ll49 1-119.167' 1-ll 6.0430 1-11 3. 9774 (.1) 3.0Z64 l· ll Z. l50S 1-11 l. 7172 1-ll l. 3682 l·ZI 8. 2516 ( ·ZI 4. 1258 60
IZO 4. 7381 3. 7670 3. 0874 z. 7918 z. 3941 2,086 l. 767! l. 3128 (-1) 9. lH•· 1·11 6.0599 1· ll 4. 0116 1-11 3. 0605 l·ll Z, 3816 (.11 l. 7430 1-11 l. 3902 l-218.3970 ( ·21 4. ZOlS IZO
CD -1.Z68Z 3.4746 2. 8968 Z.6393 Z.287! 2. 009! l. 7167 l. Z910 1-11 9. 0654 {-JI 6, 0783 l• 114,0473 (-1) 3, 0962 1-11z.4141 t-11 l. 7701 (.11 l. 413Z 1-218.5499 (·21 4.Z85Z CD
. ...
N ~
... . ... ... "' ... ... "'
... 2 : ~ 2 .... ...o ~ ...o ~ 9
.......
N
..:
..~ .. ... ... ~ !
... ...... ... ... ,,.oo ...o ... ... .. ....... ... ... .... o
; !
... s : ...... ~
.... N
."' .... .
N
.;
N
..; ..; "' ... ... ..; ..; .,¡ ..; ..... ..; .; .;. ~ ..o .D ~ ..o r: ,.:
CIO o N N
... ... ... "' "' "'N ..o
..
N N
..
N N N
..
N N N
11
... ......... .......... ... ......... s .. ..... ..."' ~ ... ... ....... ~ ... ...... : ... ... "'
...
... ... .....: 1 ...... "' ,,.
. ..
o
:::.. .... o o o
....o
. .. ....
.. . . o
o .... o
o ~ ~ o • o ...o ...o ...
o. ... ... ... ... "'
..,; ..,; ,.; .,; ,¡. ,¡. ,¡. ..: ..; ..: ...: ...: ...: ...: ...: ..:
... ... ! ... ... ... ... ... ... ...
"' ..... ...
....
- . -...... ...
~ ... 5 ~
..
... ..
...
...
... s
N
...
...
N
s :. - ... -
:!
... .
"' .....
...
!
N
N N
"'
o
s; .... o"'
...
.....
.... .......
o
°'
.. .,
...... ~ o ~
."' ... .... ..t a.... -... -... ...- -... -..... ...... -... ...,,.. -. .... ...- - - ...... .... -
.... o
"'
. : : ~ -; ¿~
N
~ :;; N
"'
..,; ,.; ..: ..: ..: .: o ,.: ..: ...: ..:
o
N
2- - - - - - - - - - 2
...... .....
"'
..... ... ... ; .... ... .. .,,... . ...... ! a ~ ..... ~
N N
.!.
o .... N o co ...
"'
::: .,...
."'
:!\
..
N
~
N
.... ~ .... .o o o~·
~
,¡. ..: ..: .: ... N o ,,:. ..:
~
! - - - ' -. -. - - - - - -... -...... -. i -.:... -......"'... ...~
...:;: "' ...... -~ ~... ...... .......... ~ ! . ~ ...... ...... .... "'.....
"'
N
- ..!. "'
:.,, *
N N
~ ~-~ ~ ~~ ~~ ~~ E : g~ :~ ~Ñ ?
."' ' .' ..:' ... -... - ... -.. -. ...- .-... ......- - - - "'- ..- -... . - -..
~ o :; 2 "' "' ~ :il : ..· _: -= ... :
.: ,.¡ .; N Ñ Ñ Ñ ..; ..; N N N ~ Ñ ·~ ·• ~ N ~ N ~
o
ª.
g o ~
..."'o ... "' ~ ~
."' -... - - ~
.;
N
... ..¡ Ñ N N ...
-... -...... -... -....... -...... -. ....-... .:. - -... ..:....... !:- - ... ... ......
Ñ ..; ..; ..; ..; ... o
... - - -- - -
.;
=== - =
.~ - - - -
.. - - ......- ......- -
..!. .!. .!. .!. .!. ~ ~ ~
.....
... . ... ...... . . i :... i .. . .. !.. ..~ .. .. .... : "' o
-....... -...... -...... -........ -... -........ =...... -....... -...... i- =..... -:g -....... ...- ...-...... -......... -~ -..........
..!. .!.
... - - - - =
. o ....
~- o ... .,;g ... ...~
o
......
... .,;
ID
"' "'
....
....... .
...
..:..
; s! ~ ~
o
."' -
"' ~
..,;
o
....
..,; ..,; ..,; .,¡
o o
...
.,;
o
..,; ..,; ..,; ..,;
- ......- ...- "'- .... - ......- = .......- ......- ...- ........- ...,,.- ......= .....- ...=.......- . - ...-
N
.,; .,; .,; ..,; .,;
."' g N
o o
N
~ .... "',¡. "',¡. ..: ..:
o o
..:
o
..:
o o
"' ..: ,¡. ,¡. ,¡. ,¡. ,¡. ,¡. ~ ~ ~ ~ ~ ~ ~ ~ ~ ~
- - =-
o Ñ
"'..: ~ "'
... ... ... N
......
"'
...
...:
...
..:
~ ~
...:
i ~
o ,¡. ..; ... N N ..: ..: "'
11
"' ....... ... ... ...... ... ...... o ... ... ...... ... ... "'.. .. .. . . ..
:..; ...... ~ ...,¡. ..;"' ..;~ .....; Ñ..... ......... Ñ~ ~... Ñt; ~..¡ Ñ g: :
s ... ,,. ....... o.. ...o...
- "'"'. ~ i ! ~ : ~ i i ......o
N N
N
... ... ... N
.
N
,..¡ ,..¡ ~ ,.,, "" ,..¡ ...: _: ..:
o
.= ;
..;
"' ~ - -
..
.. . .. .. .... .. .
... ...... ... i o... ...oo ...o .. a ..,,. .... ... .. .
.... ....
... ... ~ ~ ~ ~ ~ ~ o "'
"'
~ .... ...... i ...o
o
E ...
... 5 ...o : :so i """'o .. ~· : ~ .. : .. ~ ~ o o o ...
o
o
"'
.; . .,; ... ..; ... Ñ Ñ ,.¡ Ñ
"'
Ñ Ñ .¡ ..: . ... ~ ~ ~ ~ ~ ~ ~ N ~ Ñ
~
~
o ... ..
... - + ..
.. .. .. .. ... .. ..
:'..
.. ª .... .. .
... ... ... ... ...... ... ... ...... ...
5..; ~ ~.; .,;E ....,¡... ... ...... ~ !..; ..;"' ..;... ...Ñ ...... ....¡ Ñ~ ...... ......Ñ .....¡
.... ~ .... "' o ... "'
"' ..."'
"' ......
. ..
o ,;.
o
o
..: ... . ..
: ti.
... .......
... . ..
! ::!.
"' ... ... ...... .... ...
+
,,..
+
.. .. .. .
...... ... ... o ... ... ... ... "' ; ......
j ... ... ... ! !O: o o .... ~ ...... ~ ~ oo :!! ......
...
..."' ....¡ ; "'Ñ
o ...
..
a.
..
+
........ ...
....."'
. -
o
o ..;
...
.., ,¡. .¡ ..,; ..
..,; .,¡ .... .
.... ..; ... ..; Ñ ...
ª
o
o ...
"a.
:::
...±.
-r ~ :
o
"'
....
... + + + +
"' + ...
"' "'
;;
..
... "' ......
... . . ..
... ... ....... ... ... ... o ...... ::; .... ... o ... o ..."' ...
..
... ... .., :¡;..: ~ ... ~,.: ...... .,; ... ~ ~ :..; ..;:g ~..; ..; ... ......
N
..
... E~~; .. ~ ...o
m ~
coi.02:;$
~ • • ~
.
...o
...
:
g
:
g
o
=
.
g
o Ñ
"' !...
..; ,¡. ..:
... "'+ + + + +
N
...
...
...
..
"' ... =
... ... o ... ... .. .. ª ..
•
-
,¡.
N ~
•
+
~ ..: ..:
..
.....
ai ...= ..o ti\ "' ... "" "" .. ,.¡
.. ....... o ...
...... i o... ~,;. ~ ......... ~.; ~ ...... ¡ o ~,¡. ~,.; .o~ !.,; ......... ......... ......o ....,... ..;"' ..;
o
o
.-
..
o : ~ s Et
.
o
o
ó
...
..,; ...
... ...... + ª
o
a.
..,;
.. ~
CD o
- - + +
...N : 5~ : ~ ~ i...; ~ ~ ~ :
o ... ....... ~ "'
... ,.: ... .o .,¡ . ... ..;
o o
.
CD
o
"' ,.: ... ... ... ~ ~ ...: _: ..: ,¡. ,.: ~
o
."'
o .,;
..;
-"' :..
¡:::
:!;. ...
..;. ..,;
i ...+ ~ ~ + :!;. ~ ;. ;. :'.. +
-
~ - +
N
+ + +
-
l +
..... - ... ... ... "' ... ... CD ... ~ : ~ ... :.:
o ~ .
o
~ ...
o
8
...
"'
432 433
V1 = 10
"2 p•O. 0001 p• 0.001 p•0.005 p•0.01 pa0.0Z5 p= o. 05 p•O. l p• 0.25 paO. 5 p•O. 75 p• 0.9 pa O. 95 p• o. 975 p•O. 99 p•O, 995 p•0,999 p•O, 9999 "2
1 (+7) 6. 0562 (+51 6. 0562 (+41 Z.4224 (+ll 6. 0558 (+Z) 9,b8b3 (+21 2.4188 (+1) 6.019! 9. uoz Z.0419 (·116.7047 1-l) 3. 0441 1·11 2.0143 1-ll 1.4416 (·219.9562 1·217.7967 (-21 4. 7529 (·ZI z. 5922 1
1.Z (6) 3.Z291 (51 6. 9569 (3} 4. 7581 (ll l. 4984 121 3.2506 (Z} 1.0210 (1) l.1870 6. 5863 l. 7677 :-116.5084 1-11 3.1319 1-11 2.11'18 (·11 1, 5397 (-1) 1. 0772 (-21 8.4934 (·ZI 5, 2325 l ·ZI 2. 8772 l. 2
l. 5 (5) l, 7644 (3) 8. 1893 (21 9, 5739 (21 3. 7964 (2) I, 1154 (11 4. 3967 (11 l. 7148 4. 6961 l. 5378 1-11 6.3565 (-11 3. Z516 (•I) 2.2555 1-11 l. 6651 1-111. 1819 (-219.3921 l·Z} 5. 8562 ( •21 l. 2508 I, 5
2 1+31 9. 9994 1+2) 9. 9940 ltZI l. 9940 (ti) 9. 9399 l+ll 3, 9398 l+ll 1. 9396 9, 3916 3. 37°70 l. 3450 (-11 6, 2598 (-11 3.4194 1-11 2.4374 (-1) l. 8327 1-11l.32Z9 (-1) 1.0608 (·21 6. 7090 (•ZI l, 7668 z
3 (+21 6. 0275 (+21 1,2925 (+114,3686 (+ll z. 7229 (+ll 1.4419 8. 785' s. 2304 2. 4447 l. 1833 1-11 6. Z391 (-1) 3.6661 (-1) 2. 6967 (-1) 2.0723 (-1) l. 5262 1· ll l. Z375 1-ZI 7. 9664 (·21 4. 5377 l
4 ~+2) l. 5398 (+11 4. 8053 (+112.0967 1+11 l. 4546 8. 8439 5.9644 3. 9199 2. 0820 1.112& 1-11 6, Z700 (-1) 3. 8383 l•l) Z.875Z 1-11 Z.2380 (-1) 1. 6683 1-1) l. 3619 1-21 8.8630 (•Z) S. 094Z 4
5 °+1) 6. 9250 (+l) 2.6917 (ti) 1. 3618 1+11 1.0051 6.6192 4. 7351 3. 2974 l. 8899 l. 0730 1-116.3080 l·ll l. 91>57 1-11 3.0068 1-11 z. 31>01 (-1) 1.7742 1-111.4551 1-11 9. 5413 •-21 5. 5187 5
6 ~+1) 4. 1071 (+11 1.8411 (+ll l. 0250 7. 8741 5.4613 4,060C 2. 9369 l. 7708 1.0418 (-11 6. 3432 1-ll 4.0640 (-1) 3.1083 1-11 2. 4557 1-11 1.8567 (·t) l. 5280 1-11 l. 0075 1-21 5. 8546 &
7 +l} 8. 3803 6, 6201 4. 7611 3.636! 2. 702! 1,6898 l.0304 1-116.3743 1-1) 4, 1425 1-11 3.1893 (-11 2. 5318 l·ll 1.9UO 1-11 l. 5867 1-11 l. 0507 (·2) 6.127'> 7
8 +11 2. 1683 ~+11 1.1540 7. 2107 5.8143 4.2951 3. 3472 z. 5380 l.6310 1.017! 1-11 6.4016 1-114.2066 l· 1) 3. ZS55 1-11 z. 5941 1-11 1. 9776 1-11 1.6~51 1-111.0865 (•216.3549 8
~ 9 +11 l. 759C 9. 8943 6.4171 5. 2S6! 3,963! 3.1373 2.4163 l. S86l 1.0077 1-116.4255 (-114.2602 (-11l.3108 <·ll 2.6462 1-11 2. 0233 1-111.6757 1-11 l. 1166 1·21 6. 5467 9
(.:1
~ 10 ~+11 1.4901 8. 7539 s. 8467 4. 8492 3. 7168 2. 9782 z. 3226 l. 5513 1.0000 1·11 6.4462 1-114.3055 l·l) 3.3577 1-112.6905 1·11 Z.0622 1-11 1. 7104 1-11 l. 1424 1·21 6. 711 l 10
1Z ~+1) l. 1647 7. 2920 5. 085! 4. 2961 3.3736 z. 7534 2.1878 l,499! 1-119.8856 1-1) 6.4809 (-1) 4. 3781 1-ll 3.4329 l• ll z. 7617 1-11 2.1250 1-1) l. 7664 1-1) 1.180 l·Z) 6, 9783 u
IS 9. 1309 6. 0808 4. 4236 3. 8049 l.0602 2. 5437 2. 0593 1.4491 1-119.7732 1-116.5198 1-114.4573 1-113.5149 (-ll z. 8395 l·ll 2.1938 1-11 l. 8279 (-1) 1.2302 1•21 7. 2744 15
20 7.1805 5, 075Z 3. 8470 3. 3682 z. 7737 2. 347! 1, 9367 l. 399! (·JI 9, 66Z6 1-1) 6. 5638 (· 11 4. 5440 1·11 3.6049 l· 11 z. 9252 1-11 .•.2699 l•I) 1.8961 1-1) 1.2814 1-2)7.6049 20
24 6. 3750 4. 6379 l. 5870 3.1681 Z.6396 z.zsn l. 8775 l. 3750 1-11 9.6081 1· ll 6. saao (.1)4.5905 l-11 l.6534 1-ll 2. 9714 1-ll 2. llll 1-11 1. 9330 1-11 l. 3093 1-21 7. 7852 24
30 s.6641 4,2388 3, 3440 2. 9791 2.5112 2, 1646 l. 8195 l. 3507 1-ll 9. 5540 1·11 6.6142 1-1) 4,6395 1•113,7043 1-11 1. 0202 1-112.3541 1-111.9722 (·l) l.3389 1-217.9770 30
40 5. 0363 l. 8744 3. 1167 2, 8005 z. 3882 z.0111 l. 7627 l. 3266 1-11 9. 5003 1-ll 6. 6419 1-11·4.6911 1-ll 3. 7581 1-11 3.0718 (-11 Z.4008 1-11 z. 0137 1-11l.3'104 l·ZI 8. 1816 40
60 4.4815 3. 5415 2. 9042 2. 6318 2.210• l. 992! l. 7070 l, 3026 1-119.4471 1-11 6. 6711 l-114.7456 1-1) 3.8152 1-113.1266 l·ll 2.4498 (· 11 2. 0580 1-11 l. 4040 1-21 a. 4001 60
120 l. 9907 3.2372 l.. 7052 2. 4721 2.1570 1.910! 1.6524 l. 2787 1-119.3943 1-116.7029 1-114.8035 1·113.8759 1-11 3. 1848 1-ll 2. 5022 1-112.1052 (-ll 1.4400 (-21 8. 6358 IZO
CD l. 5564 2. 9588 2. 5188 2. 3209 2.0483 l.8301 l. 5987 l.254~ 1-11 9. 3411 1-11 6. 7372 1·114.8652 1·113.9403 1-11 3. 2470 1-11 2. 5582 (.1) 2.1559 1-111.4787 1-2) 8. 8891 CD
"1 = 12
"2 p .. 0.0001 p•0.001 p•0.005 p•D.01 p•0.025 p•O, OS p•O, I p•O.ZS p•O. S p•O. 75 p•0.9 p•O, 95 p•0.975 p•O, 99 paO, 995 p• o. 999 p• o. 9999 "2
1 +71 6. 1067 l+SI 6. 1061 1+41 2.4U6 1+316.1063 (+219.7671 (+21 2,4391 1+116.0705 9. 4064 z. 067~ 1-116,1432 l·ll 3.1481 1-11 2.1ou 1·11 l. 5258 1-1) 1.0718 1-218. son 1-21 5. 3638 1·21l.0538 1
1.2 (6) 3.2504 141 7. 0027 131 4. 7894 131 1.5083 IZI 3.2721 121 l. 0278 111 3,2090 6. 6399 l. 7894 1-116.6485 1-11 3.2449 (·11 1.222! (·11 1, 6344 1-11 1.1636 (·219.3036 1-21 5. 9318 l·Zl 3. 4412 l.Z
l. 5 15) l. 7715 UI 8.2U3 (Zl 9. 6126 121 3.8119 lll 1.1200 Cll 4.4157 111 l. 7230 4. 7276 l. 5563 1-116.5009 1-11 3. 3771 1-11z.3719 1-11 l. 7741 1-111.2824 (·11 l. 0338 (·ZI 6. 6769 1-21 3. 910 l. 5
2 +31 9.9994 +21 9.9942 (tZI 1. 9942 c+u 9. 9416 l+ll 3. 9415 (+l) 1. 9413 9. 4081 J, 3934 l. 3610 1-116.4123 (· 11 l. S628 1-11 z. 5738 1-1} l.9624 1-11 1.407 1-111.1751 (-21 7. 7080 (·2) 4. 5768 2
l 1+21 s. 9833 (+2) 1.2832 l+l) 4, 3387 (+112.7052 l+l) 1.4337 a. 7446 s. 2156 2. 4500 l.197Z 1-11 6.4066 l· 1) 3. 8380 1-11 2.8651 1-11 2.2350 1-11 1.6799 1-11 l. 3839 ~-219.2557 (-21 5. 5867 3
4 (+2) l. 5186 l+ll 4. 741Z (+I) 2. 0705 1+11 1. 4374 8. 1s1z 5. 9117 3. 8955 2. 0826 l.1255 1· ll 6, 4504 l· ll 4,0321 1-11 3. 0682 l•ll Z.4265 1-11 1. 8478 l•ll 1. 5135 1-11l.0381 1·21 6. 3321 4
5 l+I) 6. 7908 1+11 2. 6418 1+11 l. )384 9. 8883 6. 5246 4. 6777 3, Z68Z l. 8877 l.0855 1· 11 6.4981 1-ll 4.1771 1-ll l. Zl97 1-11 z. 5700 1-11 l. 9746 1-11 l. 6471 •ZI 1.1246 l·ZI &, 9101 5
6 (+I) 4,0081 l+ll 1.7989 (+l) l. 0034 7. 7183 5. 3662 3, 9999 z. 9047 l. 71>68 l.0600 1-ll 6, 5419 l-1) 4.2900 1-11J.3377 l•l) 2.6822 1-ll 2. 0744 1-ll l. 7370 -11 1.1935 ~-z11. m5 6
7 l+I} z. 7644 l+ll 1.3707 8.176-1 6. 4691 4,¿659 3. 5747 2, 6681 1,6843 1.0423 1-116.5802 (-11 4. 3806 1-11 3. 4324 1-ll z. 7728 1·11 Z.1554 l·l) 1.8101 -11 l. 2499 ~-21 7. 7566 7
8 (+I) Z.098' l+l) 1.1194 7. 0149 5. 6668 4. lm 3. 2840 2. 5020 1.6244 l.0293 1-ll 6.6138 (-114.4551 1-113.5105 (·I) Z.8475 -11 2. Z2Z5 1-111. 8709 1-11 l. 2970 ·Z) 8. 0771 8
~
(.:1
9 (+11 l. 6961 9. noo 6. 2274 5. 1114 l. 8682 3.0729 2. 3789 l. 5788 1.0194 1-11 6,6428 "1) 4, 5177 1-113,5760 1-112.9105 1-112.2792 l• ll l. 9223 ·ll l. 3369 -218.3503 9
c.n 10 (+1) 1.433~ 8.4452 5. 6613 4. 7059 l. 6209 2. 9130 2. 2841 l. 5430 1.0116 (·11 6.6684 1-1) 4, 570! (-1) 3.6319 l·ll 2. 9642 1-112.3277 l· ll l. 9664 ~-11 l. )7lot -Z) 8. 5861 10
IZ 1+11 1.1144 7. 0046 4. 9063 4. 1553 3. 2773 2.6866 2.1474 l. 490Z 1.0000 1-ll 6. 7105 l-1) 4,6568 1-ll l. 7222 1-11 3. 0513 1-11 Z.4066 1•11 2,0l8Z -1) l.4276 -2) 8. 9732 1Z
IS e. 6859 5.8121 4. Z498 3. 6662 z. 9633 z. 4753 2.0171 l. 4383 1-119.8863 1-11 6. 7586 1-11 4. 7508 1-11l.8213 1-1) 3. 1474 l·I) Z.494C. l•ll 2.1180 -11 1.4905 ·2) 9. 4075 15
zo 6. 783 4. 8229 3. 6779 3. 2311 2. 6758 2. 2776 l. 8924 l. 3873 1-11 9. 7746 1-116.8129 t-11 4.8551 1-l) 3. 9314 1-11 3. 2544 l· ll z. 5917 1-ll t. 2076 -11l.5613 ·ZI 9. 8996 20
Z4 5. 9992 4. 3929 3.4199 ), 0316 z. 5412 z. 1834 l. 8319 l. 3621 l·ll 9. 7194 1-ll 6.8432 1-ll 4. 9116 1-113.9912 (·I) 3. 3127 l· ll Z. 645Z 1-11 z. 2566 -1) l. 600) -11 l. 0171 Z4
30 s. 3075 4. 0006 3.1787 Z.8431 Z, 41ZO 2. 0921 l. 77Z7 l. 3369 1-11 9. 6647 1-116.8757 1-114.9714 1-114.0547 1-11 3. 3746 (·IJ Z. 70ZI l· ll z. 3090 -11l.6421 (•l) l. 046<1 JO
40 4. 697l l. 6425 z. 9531 Z.6648 Z, 2882 z. 003! l. 7146 l. 3119 (-l) 9, 6104 (-1) 6. 9104 (-11 5. 0350 l·ll 4. IZZZ Hl l. 4408 1-1) 2. 7630 l· ll z. 3651 ·11 l. 6870 1-111.0778 40
60 4. 158' 3. 3153 2. 7'119 z. 491>1 2. 1692 l. 9174 l. 6574 l. 2870 1-11 9. 5566 (-1) 6. 9478 1-1) 5. 1029 1-ll 4. 1943 ~-113.5115 l• ll Z, 8Z8S l· ll z. 4255 1-11 l. 7354 -11 l. 1119 60
IZO 1., 5139 Z. 331>3 z. 0548 1.8337 1.6012 l. 2621 1·11 9. S03Z 1-ll 6. 9881 1-ll 5. 1752 1-114.2717 1-1) 3, 5876 l·ll z. 8991 1-11 2. 4907 -11 l. 7879 (•I) 1.1490 11.0
3. 68ZJ 3. 0161
m l.2612 Z. 7425 z. 3583 z. 1848 1. 9447 l. 7SZl 1. 5458 1. ll7l 1-119.4503 1-1) 7. 0319 l-11 s. 2531. 1-1) 4, 3550 1-ll 3. 6699 HI 2.9755 1-1) z. 5615 (·ll l.845Z -11l.1896 a>
IJ1 = 15
p110, 5 p•0.9 p1t0,95 P"º· 975 p=0.'19 p•O, 995 p=0.999 p•O. 9999 ...z
p•O.OOS p•0.01 p• o.ozs p• 0,05 p•O. l p•O,ZS p• 0.15
"z p•0.0001 p•0.001
1
9.4934 2.0931 l·ll 6.98Z.8 l·ll 3. Z539 l• I) Z. ZOll (.11 1. 6UO (-11 1.1517 (·ZI 9, 2610 l·ll 6. 0287 1-ZI 3. 64lZ
1 1+71 6.157< 1+5) 6. 1571 1+4) 2.4630 1+31 6, 1573 l+zl 9. 8487 l+ZI z. 459! (+!) 6. IZZO
6. 6939 1,8112 1-11 6. 7903 1-1) 3. 3603 (-1) z. 3Z78 1-11 l. 7328 l-1) 1.2546 (-1) 1.0166 l·ZI 6.6968 1-Zl 4. 0868 1.2
l. z 161 3. 2718 141 7. 0489 131 4.8210 131 l. 5183 IZ) 3. 29~8 121 l. 0347 (1) 3.Z31Z
1·214.6800 l. s
l. 5 15) l. 1787 131 8.US7 121 9. 6518 12) 3. BZ.75 IZI 1.1247 (1) 4.4349 lll l. 1313 4, 7594 l. 575G HI 6.6473 1-11 3. 5055 1-112.4922. 1-111. 8878 1-111. 3887 1-11 1.1350 1-217.5809
9.4247 l.4098 z.
l. 3771 1-11 6. 5673 1- ll 3. 1103 (-l) 2.7157 (-11 2.0986 l·ll l. 5726 1-111.2986 1·21 8.8190 1-2) 5. 5221
2 1+31 9. 9994 1+21 9. 9943 1+21 l. 9943 1+11 9.902 C+ll 3.901 1+11 l. 9429
3
3 1+21 5, '1384 (+2) l. 2737 t+ll 4, 3085 (+112.6872 1+111. 4253 8. 7029 s. 2003 Z.4552 l. Zll l 1-ll 6. 5781 e-ti 4. 0164 1-1) 3.0419 1-11 2. 4080 ·ll l. 8460 l·ll 1.544.t (-11 l. 0712 1·21 6. 8329
Z.G829 1.1386 1· ll 6. 6353 1-114.2348 1-ll 3. Z.127 1-1) 2.6286 l·ll z. 0437 (-1) 1.7233 k-ll 1.U17 ~-Z) 7.8228 4
8. 6565 5. 8576 3, 8689
4 IUI 1.4971 (+I) 4. 6761 (+I) 2.0438 (+11 l. 4198
1.8851 1.0980 (-11 6.6943 (-11 4. 3995 1-1) 3,4467 1-ll 2. 7961 HI Z.1951 1-111.8615 (•l) 1.3215 ~-2) 8. 6050 5
5 1+11 6. 6544 l+ll 2. 5911 (+1) l. 3146 9. 7ZZZ. 6,4277 4.6198 3. 2380
l.0722 1-116,7476 1-11 4, 5288 1-ll 3. 5836 1-112.9285 1-112.3157 1-11 l. 9721 1-ll 1.4101 1-21 9.2427 6
9.8140 7. 5590 5. 2687 3. 9381 Z.8712 l. 76ZI
6 1+11 3. 9061 (+11 l. 7559:
1.0543 1-116.1944 1-11 4. 633! 1-11 3. 6947 l-1) 3.0364 -ll 2. 4146 1-11 2.0630 1-1) 1.4835 C-219.7743 7
7. 9678 6,310 4. 5678 3. 5108 Z. 632Z 1.6781
7 1+1) Z.681~ C+ll l. 3314
1.0412 1-11 6.1346 (-1) 4. 7203 l·l) 3. 7867 1-1) l. 1Z63 -11 2.4972 (-2) Z, 1394 1-11l.5454 1-1) 1. 02Z5 8
6. 8143 5. 5151 4.1012 l.21H 2. 4642 1.6170
8 1+11 z. OZ7~ 1+11 1.08"1
1.0311 1-1) 1) 3,2024 1-11 2. 567! 1-l) z.zou 1-1) l. 5985 1-111.0614 9
6. 0325 4. 9621 3, 7694 3,0061 Z. 3396 1. 570! 6,870~ 1-1) 4. 7934 1-11 3.8646 ( 0
9 1+1) 1.6331 9.2381
~
10
~
5.4707 4, 5582 3, 5217 2. 8450 z. 2435 l. 5H8 1.0l.32 1-1) 6,9008 (-114.8560 (-11 3. 9311 l•ll 3. Z678 1-ll 2. 6282 l·ll z. 2606 c-11 1. 6445 1-11 l. 0952
(j) 10 1+1) l. 374i 8.1288
12
4.1214 4.0096 3, 1772 Z.6169 Z.1049 l.4796 1.011! 1-11 6, 9527 1-114.9576 1-ll 4,0399 1-11 l. 3746 1-ll 2, 7276 1-11 2.3531 1-111.1206 1·1) l.1513
12 l+I) l,0631 6. 7092
1.0000 C·ll 7.0111 1-11 5. 070: 1-ll 4.1606 l-11 l.4939 1•11 Z.8391 1•1) Z.4571 (-1) l.8067 1·11 1. Zl52 15
5,5351 4. 0698 3, szzz z. 8621 2,403! l. 9722 1.4263
15 8.U91
zo
3. 5020 3,0880 2, 57ll z. Z033 1.8449 l. 3Jl6 1-1) 9. 8870 1-11 7,0786 1-1) 5.1967 l-1) 4. 296! C·ll 3.6286 1-11 z. 9657 1·112.5756 1-11 l. 9053 c-11 1. ze90
zo 6. 3741 4. 5618
1-11 1. 9604 (-JI J. 3304 24
4, 1387 3. 2.456 z. 8887 z. 4374 Z.1077 l. 7831 l. 3474 1-11 9. 8312 1-117,1164 l·I) 5.2659 (-1) 4. 3710 1-11 3. 702.9 1-1) 3. 0358 1-1> 2.6414
24 5. 6112
30
30 3, 7527 3. 0057 z. 7002 2, 307Z 2. 0148 l. 7223 1. 3213 1-11 9. 7759 1-11 7, 1567 1-1) 5. 3396 1-11 4. 4508 1-11 3. 7826 1-1)3.llll 1-11 z. 7lZ5 l·I) 2.0201 (-11 l. 3754
4.938!
1-1) J. 4246 40
40 4. 345' 3, 4003 z. 7811 z. 5216 Z.1819 l. 924! l. 6624 1.2952. 1-119.7211 1-111. zoo• 1-11 5. 4189 1-114.5366 1·11 3.8685 1-11 3. l9Z9 1-112.7894 l·l) z. 0851
(-11 l. 4787 60
60 3. 8ZZI 3.0781 2. 5705 2. JSU 2. 0613 l. 8364 1.6034 l. 2691 1-11 9. 6667 (-1) 7. 248 1-11 5. 504Z -11 4,6294 1-11 3.9617 l-11 3, Z818 1-11 z. 8733 l-llZ.1562
1-1) l. 5386 IZO
IZO 3.36~ 2. 7833 2. 3727 2.191S l. 9450 l. 750! l. 5450 l. 2428 1-119.61Z8 1-117,3003 1- ll s. 5')6<i 1·114.7301 1-11 4.063Z 1-11 3. 3789 l· ll z. 9654 (•11 2,Z347
1·11 1. 6055 co
Z.950' z. 5132. Z.1868 z. 0385 1.8326 1,666~ l. 4871 l. 2163 1-11 9. 5593 1-111. 3578 1-11 5. 6977 1-11 4.8407 1-ll 4.1748 1-11 3.4863 1-11 3. 0673 C·I) z. 3Zl8
CD
------"\·-
,____________. . ._____ --~ d
IJ1 = 20
"2 p•O, 0001 p•0.001 p•0.005 p•0.01 p•O.OZS p• o.os p•O, l p•0,25 p•O, 5 p• o. 7S ' p•O, 9 p• o. 95 P" 0, 975 p•0.99 p•0.995 p• o. 999 p•O, 9999 "2
1 +7) 6.2091 (+SI 6.2091 1+4) 2. 4836 1+31 6. Z087 l+ZI 9, 9310 l+ZI 2. 4801 1+11 6. 1740 9. 5813 z. 1190 1·1)7.1240 1-11 3. 3617 1-1) Z.Z98l 1-11 1. 7031 0 1) l.2352 1-11 l. 0056 (·ZI 6. 7482 1-214.2736 1
l,Z 16) 3. Z93S (41 7. 0956 131 4. 8530 (JI l. 5283 IZI 3.31S7 12) l. 0417 (1) 3.2536 6, 7484 l. 83)) 1-116.9338 l-l) 3.4780 (-11 z. 4362 HI 1. 8348 1-11 l. 3502. 1-111.1080 1-ZI 7, SZ89 (-l) 4. 8187 1.Z
l. 5 (SI 1.7860 (31 8,Z895 121 9.6914 121 3,803 (ZI 1.1294 11) 4. 4543 fl) l. 7396 4. 7914 l. 5939 1-11 6. 7958 1-11 3. 6370 l·l) 2,6164 1-11z.0063 1-111. 5009 ·l) l.Z4Z9 ·2) 8. 5708 -ZI 5. 5549 l. s
z +ll 9. 9995 (+2) 9, 9945 ltZI l. 9945 l+l) 9. 9449 (ti) 3. 9448 l+l) l. 9446 9. 4413 3. 4263 l. l9Jl 1-11 6. 7Z49 l· 11 3. 86ZO 1-11 z. 8630 1-11 Z.Z415 (.1) l. 7097 1-1) l. 431 1-111.0048 l·ZI 6. 614.Z 2
3 +Z) S, 893C 1t21 l.Zt..12 (ti) 4.2778 (+l) Z.6690 1+11 l. 4167 8.660Z S.1845 Z. 460Z l. Zl5Z l·ll 6.7531 1-11 4.2015 (-1) J.Z27S 1-1) Z. S91 S l· 11 2. 0250 1-11l.7189 r-11 1. z34s HI 8.Z988 l
.. +Z) 1.4152 l+I) 4. 6100 l+I) z. 0161 1+11 l. 40ZO a. 5599 s. aozs 3.8443 l. 08Z8 1.1517 1-116.8250 1-11 4.4466 1-113.4891 1•11 2..84SZ 1-11.2. 2510 l•ll l. 93Z6 1-11 l,4092 1-21 9. 6018 ..
5 I+ 11 6. 5157 (+11 2. 5395 l+ll l. 2903 9. SSZ7 6, JZ85 4. 5581 3. Z067 l.88ZO 1.1106 1-116,8966 1-114.6335 1-11 3. 6888 1-11 3. 0403 1·11 Z,074 1-11 z.1001 1-11 l. 5479 C·l) l.0652 5
6 +l) ), 8036 (+11 l. 71ZO 9. S888 7. 3958 5.1684 3, 8742 z. 8363 l. 7569 l. 0845 1-11 6, 9609 1-114.7817 1-11 3. 8476 1-11 ], 1966 (· ll z. 5830 1-ll z. 2361 1-11 l.66lS l•ll l. 15ZZ 6
7 C+ll 2.5977 (+11 1.Z932 7. 7540 6. 1554 4. 4667 J,4HS l. 5947 l. 6712 l. 0664 1-il 7.0166 l·ll f,90Z7 HI 3.9777 l· 11 3. lZSI l•l) 2,7037 l•ll Z.H91 l·I) l. 7569 (·I) l.Z258 1
a r+11 1,9547 l+I) l. 0480 6.6081 5. 3591 3, 9995 3, 1503 z. "146 l. 6088 1.0531 1·117.0656 l· 11 5. 0038 (·1)4,0865 '-11 3,4331 (-1) 2. 8055 1-11 2. 4450 1-111.8382 1-11l.2891 8
.¡i.. 9 l+ll 1, 568C 8,8976 5. 8318 4. 8080 l. 6669 z. 9365 2. 1983 l. S611 1.0429 1-11 7.1078 1-11 5. 0893 1-11 4.1792 (-11 3. 5255 l· ll 2. 89Z9 1-ll z. 5Z76 1-11 l. 9087 (-1) l. 3442 9
~
-.:t 10 +ll l. 31SO 7,8037 5. Z140 4. 4054 J, 4186 2, 7740 2, ZOO? l. 5235 1.0349 (·ll 7.145-1 (.l) 5.1634!(-l) 4.2591 (-1) 3, 6053 -11 2. 9689 {·112.5994 1-11 1.9703 (-11 l. 3927 10
12 k+ll l. 0101 6,4048 4. SZ99 3. 8S84 3. 0728 2. 5436 2. 0597 1.4678 l. OZll 1-11 1. zoe1 l•ll S. 2843 1-114. 3906 (-1) 3. 7372 1-11 ]. 0949 (·11z.7189 (·1} 7., 0734 l·ll 1.4741 IZ
15 7, 7582 5.2484 J, 8826 3. 3719 Z.7559 Z. ll7 5 l. 9243 1.4127 1.0114 (· 11 7.2801 (•ll 5.-1203 1-11 ... S386 (-1) 3. 8864 1-11 3,2383 1· 11 2. 85SS l·ll z. 1921 1-1) l. 5687 IS
20 s. '1516 4.2900 3. 3178 z. 9377 Z.41>4S 2. IZ42 l. 7938 l. 3580 1.0000 1-1)7.3638 1-115.57-48 l·ll ... 7077 1-11 4. 0576 1-113.4040 1·113.0140 1-1) 2. 3310 (-1) l. 6802 zo
Z4 S.2084 3.8'13Z 3.0624 2. 7380 2. 3273 z. 0267 l. 7302 l. 3307 1-119.9436 1·11 7.4107 l· 11 s. 6603 1-114.8019 (-1)4.1535 1-11 ). 4972 1-113.1037 1-11z.4100 1-111. 7441 z.t
JO 4. 5540 3. 4928 z. 8230 z. 5487 l. 19SZ l. 9317 1.6673 1.3033 1-11 9. 8877 1-11 7.4f>Zl l· ll 5. 7531 1-114.9041 1-11 4. Z579 1-11 3. 5991 1-11 3. ZOl6 l•ll Z.4969 1-1) 1.8147 30
40 l.977Z 3. 1450 l. 5984 l. 3689 z. 0677 hU89 1,605Z l. Z758 1-119.83Zl 1·117. H8Z 1-11 5. 8S38 l· l) S.0155 l• 11 4, l7ZO 1-ll 3. 7110 1-11 3, 30'16 l·I) z. 5931 1-1) l.8933 40
60
"º ).4681 z. 8266 2. 3872 Z.1978 l. 9445 l. 7480 l. 5435 l. 2481 1-119.7773 (-117.5798 1-11 5. 9637 l· 11 5. 1377 1-114.4976 1-11 3. 9346 1-11 3, 4Z95 1·11 2. 7005 1-111.9817
IZO ), 0180 2, 5344 z. 1881 z. 0346 1, 8Z49 1.6587 1.4821 l. zzoo 1·119.7ZZ8 1-11 7,6488 1-11 "· 085J 1-1) S.Z734 1-ll 4. 6378 1-11 3. 9733 C-1) ], 5640 (•!) 2,8Zl8 1-1) 2. 0820 IZO
Q)
z. 6193 Z, 2657 l. 9998 l. 8783 l. 7085 l. S705 l. 420(¡ 1.1914 1-119.6687 l· 11 7. 7l62 l· 116.221Z 1- ll s. 4253 1-114.7955 1-11 ... 1302 1-11 3. 7169 1-112.9605 {-11 2. 1976 co
111 = 24
p•O.OS p•O, l p• o.zs p•O, 5 p:oO. 7'• p•O, 9 p•O. 95 paO, 975 p•0.99 p.o. 995 p• o. 999 p• o. 9999 "z
"z p•0,0001 p•0.001 p•0.005 p•0,01 p•0.02.5
ltll 6, 2002 '1.6255 2.1321 1-ll l. 0470 1-Z) 7. 1286 1-2) 4. 6161 1
l 1+7) 6. 235~ (+SI 6. USO (+t) 2.4940 (+3) 6. Z346 (ti) 9. 97ZS 1+z1 z.4905 1-1) 7. I°'~ (-1) l. 4164 1-112.3476 1-11 l. 7493 1-1) 1.2783
(1) 3.2649 6. 7759 1.8444 1-1) 7. 006Z 1-11 3. 5378 (-1) 2. 4916 (-1) l. 8872 (-1) l. 3996 (-1) l.1556 l-2) 7.9704 1·2) 5, 2180 l,Z
1.2 16) 3. 3044 (4) 7.1192 (3) 4.8691 131 l. 5334 IZ) 3.U67 (2) 1.0452
(1) l. 7438 1. 8075 l. 6034 (-1) 6. 8707 1-1) 3. 7039 (-1) 2. 6799 (-1) z. 0673 (•I) l. 5592 (•I) 1.2.994 1-2) '1. 0987 (·2) 6. 0349 l. 5
l. 5 (5) l. 7897 Ul a. 3065 121 9.7113 (Z) 3, 8512 12) l, 1318 (1) 4,4641
9. 4496 3.4345 1.4014 (-1) l. 5013 (-1) l. 0707 l·Z) 7, 2185 2
2 (+3) 9. 999' (+2) 9. 9946 l+Z) 1. 9946 l+ll 9, 9458 l+l) 3. 9456 l+ll l. 9454 1-ll 6. 8050 l· 11 3. 9396 Hl 2.ua8 (-1) 2. 3155 1-1) 1. 7814
8,6385 s. 1764 z. 462& 1, 2322 (-1) l. 3237 l·Z) 9, 12.08 3
3 1+2) s. 870~ 1+2) 1. 2.593 l+l) 4. Z.622 l+ll 2.65911 l+l) 1.4124 1-1) 6.8423 (-1) 4.2966 1-ll 3. 323& (-1) 2.6874 1-ll z. 1195 1-ll l. 8119
s. 7744 3. 8310 Z,0827 l. 1583 1·1) 1. 5116 (·l) l.0&11 4
4 K+2) l.4642. (tl) ... 5766 (+l) Z.0030 1+11 l. 3929 8. 5109 1-1) 6.9219 1-11 4. 5560 1-ll 3,6019 1-1) 2. 9591 1-1) 2.3706 1-11 z.001
1.1170 1-1)7.0004 1-114.7551 (-l} 3.8158 (-1) 3. 1698 1-1) 2. 5673 (•l) 2.2293 (•1) 1. 6732. (•1) 1.1823 5
s l+ll 6.445! l+ll 2. 5133 l+I) l. 2780 9. 4665 6. Z780 4. 5Z7Z l. 1905 1.8802
l.0907 (-l)7.0706 l·ll 4. 9138 1-ll l. 9869 l•ll 3. )393 1-ll z. 7Z7l. l·l) 2. 3799 l·I) l. 8017 (·1) 1.2838 6
6 l+I) l. 7512 (+ll l,6891 9. •741 7. 3127 5. 1172 3. 841S Z.8183 l. 7540
7
7 ~+l) 2. 555( l+l) 1.2.732 7. 6450 6,0743 4.4150 3,4105 2. 5753 l.667S 1.0724 (-1) 7. 1317 1-1) 5.0439 1-11 4.1278 1-l) l.4797 1-11 2. 860! (-1) 2. 5060 1-111.9103 (-111. 3702
8
8 l+l) 1.917 (+I) 1. 0295 6. 5029 5. 2793 3. 9472 3.1152 2,4041 1.6043 1.0591 1-1) 7.1849 1-11 5. 1528 (-11 4.2461 1-11 3. 5983 (·l) 2. 9736 (•11 Z.6134 l·l) 2. 00)5 1-11 l. •451
9
4. 7290 l. 6142 z. 9005 Z. l768 l. 5560 1.0489 (.117. 2317 (· 1) 5. 2458 1·114.3474 1·11 3. 7000 (·ll 3.0713 1-11z.7064 l·l) z. 0847 1-11 l. 5106
t
00
9 (+l) l. 5349
10 t+I) 1.284'
8. 7239
7.6376
5. ?il.92
s. 1732 4. 3269 3. 3654 Z. 737Z z. 1784 l. 5179 1.0(08 1-11 7.2727 (-11 5. 3262 1-11 4.4352 1·113.7885 1-11 l. 1565 (-1) 2. 7878 l•l) Z.1561 1-11 l. 5686
10
IZ
lZ 9. 8314 6.2488 4. 4315 l. 7805 3.0187 z. 5055 z. 0360 1.4613 l. 0289 l·ll 7, 3416 1-11 5.4588 1-11 4. 5800 l·ll l. 9351 1·113.2986 1·11z.9241 (•l) 2. 2.764 l·l) l. 6669
15
15 7. 5168 5, 1009 3. 7859 3. 2940 2. 7006 2. 2878 l. 8990 l. 4052 1.0112 l·I) 7.4Zl7 1-11 5,6082 l·ll 4. 7445 (·114.1027 1-11 3.4618 l·ll 3. 0811 1·1) Z.4162 1-11 l. 7821
20
20 5. 7336 4.1493 3. zzzo 2. 8594 2. 4076 z.o8Z5 l. 7667 1, 3494 l. 0057 1·117. 5148 1-1) s. 7797 1-114.9341 1-114. 2968 (· ll 3. 6523 l· ll 3. 2654 (-l) 2. 5918 l·l) 1.9200
24
Z4 5. oooz 3. 7354 2. 9667 z.6591 2, 2693 1,9838 1. 7019 l. 3214 l. 0000 1-ll 7. 5(>71 l·ll 5,8158 l·ll 5,0408 1·1) 4.4066 (·11 3. 7607 (· ll 3. 3707 1·1) z. 6771 1-11 1.9999
JO
30 4. 3545 3. 3572 z. 7272 2.4689 2, 1359 1.8874 l. 6377 l. Z.933 1·119.9438 (•I) 7.6260 (-11 5. 9805 (·I) S. 1Sl3 l·ll 4.5i69 l· 113,8800 (-1) 3.4869 1-1) 2. 7828 (-1) 2. 0891
40
40 3. 785Z 3. 0111 z. 5020 2. 2880 2. 0069 1. 7929 l. 5741 1. 2649 1· ll 9. 8880 1-117.6899 1-11 6, 0950 1-ll 5,2854 1-11 4,6598 l•l) 4.0lZ4 l-IJ l.6161 1-11 2.9012 l·l) 2.' 897
60
60 3.ZBZ5 Z.6938 z. 2898 z. 1154 l. 8817 l. 7001 l. 5107 l. 2361 l·l) 9.8U8 1-1)7.7610 l· ll 6, Z216 (.l) 5.4277 1-1) 4, 8079 1-n 4.1606 (-1) 3. 7615 1-113.0352 1-11 Z. 30of3
UD
120 z. 0890 1. 9500 l, 7597 1.6084 l.447Z l. 2068 1·119.7780 1-ll 7, 8407 l· 116.3633 1-11 s. 5875 l·ll 4, 9754 (·ll 4. 3292 1-ll 3. 9273 (·11 l. 1890 (·I) Z.4368
2.8373 2.4019
(1)
l. 3832 l.1767 1-ll 9. 7236 1-117. 93Zl (.11 6. 5244 1-11 5. 7700 (.1) 5. 1672 (-11 4. SU! 1-U 4.1193 1-113.3687 l·l) 2. 5929
(1) Z.4422 2. 1324 l. 8983 1. 7908 1.6402 l. 5173
v.=
- 30
"z p•0.0001 p•0.001 p•0,005 p•O. 01 p•0.025 p•0.05 p•O, I p.o. Z5 p•O. 5 p. o. 75 p•0.9 p. o. 95 p• o. 975 p•O. 99 p•0.9~ p•O. 999 p• o. 9999 "z
1 1+7) 6. Z611 (t5) 6.2610 (+.) 2. 5044 l+ll 6. Z607 ltll 1. 001~ ltZI z. 5009 1+11 6. ZZ65 9. 6698 Z.145Z (-1} 7. 2669 l-lll.4714 1·11Z.3976 1·11 l. 7961 (-11 l. 3ZZ3 1-11 1.0894 1-211. 52ze (·Z) 4. 9771 1
l.Z 16) 3. 3154 141 7. 14Z8 131 4. 8852 (l) l. 5385 IZ) 3. 3378 121 l. 0487 111 3.2763 6.80)'. 1.8555 1-11 7. 0790 (-1) 3. 5981 1-112.5476 1-11 l. 9405 l-ll l.'450Z 1-11 l.2<H6 1-Z} 8, 4290 (-2) 5. 6401 l.Z
l. 5 (5) l. ¡934 (3) 8. JZ36 IZI 9, 7314 12) 3. 859Z IZ) 1.1342 111 4. 4739 (11 l. 7480 4. 8237 1.6129 1-1) 6.9·!61 1-113.7715 (•112. 74H l•I) Z. ll95 l•I) 1.6190 (-1) l. 3577 1-21 9. 6•87 1·21 6. 5442 l. 5
.1. K+ll 9. 999• (+Z) 9. 9941 (+21 l. 9941 l+ll 9. 9466 (ti) l. 9465 (+11 l. 9462 9. 4579 l. 4428 1.4096
l· 116.8852 (-1) 4. 0182 (-ll l.0159 1-11 Z.l911 l· 11 1. 8551 l· 11 l. 5736 (-1) l. 1398 1-21 7. 86)1 2
3 ~+2) 5. 8469 l+ZI l. Z545 (ti) 4. 2466 (ti) z. 650! 1+111. 4081 9.6161 5. 1681 Z.465C l. 2393
1-116.9319 1-ll 4. 393! (-1) l.4220 1-11 2. 7860 (·ll Z. ZIH 1· ll J. 9088 l•ll l.·U7S 1·11 l. 0006 3
4 l+Z) l. 4531 1+11 4. 5429 ltll l. 9891 (+11 l. 3838 8. 4613 s. 7459 3. 8174 2. 082 5 1.1649 4
(-1) i.0205 1-11 4. 6681 (-1) l. 7180 1·1) l. 0770 1-ll 2. -1889 1·11 2.1630 (-11 l. 6328 (·ll l. 1705
5 1+11 6. 3741 (+I) Z, 4869 1+111.206 9. )793 6. 2269 4. 4957 3. 1741 1. 8784 1.1234
l·ll 7.1058 1-114.8800 1-11 3. 9470 1-11 l. 3041 l· 11 z. 7034 (· ll 2. 3654 1·111.8070 l•I) l. 310Z 5
6 1+11 3. 6984 (+11 l. 6672 9. 3583 7.2U5 5. 0652 1. soez 2. 8000 l. 7510 1.096~
1-117.1824 l· ll s. 0497 l• 1) 1.1314 1-l l l. 4883 l• 11Z.,8789 l· 11 2. 5322 1-11 l. 9523 1-111.4282 6
7 l+I) 2. 5111 1+11 1.2530 7. 53"5 5. 9921 4. 3624 3. 37 58 2. 5555 l. 6635 l. 078l
(· 117.2490 (·11 s. 1897 l·ll 4.21139 l·l) 3,6417 t-11 3,026Z 1-11 z.. 6727 l·I) Z. 0759 1-111. 5296 7
8 {+I) l. 8803 l+l) l. 0109 6. 3961 S, 1981 3.8940 3. 0794 Z, 3830 l. 5996 1.0651
1·117,3073 (-1) 5. 3076 (-1) 4.4127 1-113.7717 1-11 3. 1520 1-11 2. 7932 1-11z.18Z7 1·11 l. 6181 8
9 1+11 l. 501~ 8. 5476 5. 6zt8 4, 6486 3. 5604 2. 8637 2. 2547 l. 5506 l. 0548
tco 10 (+11 l. 2531 7.-t688 5. 0705 4. 2469 3. 3110 2.6996 l.. 1554 l. 5119 1.0467
1·1) 7.3584 1-115.4083 1·11 ... sus 1-113.8841 1-11 3. 2610 (-1) z. 8981 1-112.2763 1-11l.6961
10
9
l·ll 7. 4036 (-1) 5.4960 l•ll 4.6198 1-11 l. 98ZZ l·ll 3. 3S67 1-11 2. 9904 (-1) Z. 359Z 1-11 l. 7655
12 9. 557C 6. 0898 4. 3309 3. 7008 2. 9633 2.4663 z. Oll 5 l. 454~ l. 0347 lZ
1·1)7.4800 1-11 5. 641l t-114.7799 1·114.1459 1-11 3. 5173 (-1) 3. 1459 l•I) Z. 4996 (·111.8841
15 7. 210, 4. 9502 3,6867 3.Zl41 Z.64~7 z. 246E l. 8728 l. 397l l. ozz~ 15
(-1) 7. 5683 1-11 S. 806Z 1-11 ... 96ll 1-114.3343 1-IJ J. 7034 1-11 3 . .3270 (·11z.6647 (-1) 2. 0249
20 5. 510! 4. 0050 3. ll.34 z. 7785 z. 3486 2.0391 l. 7382 l. l401 1.0114 20
l•ll 7.67Z8 1-11 5. 9977 1-ll 5.1768 (•IJ 4. 5554 1·113,9236 {• l) J. 54Z3 1-112.8630 1-112. 1959
24 4. 786 3. 5935 z. 8679 2. 5713 2,2090 1. 9390 l. 67l.I 1. llll 1.0057 Z4
1-1)7,7322 1-116.1061 l• l) s. 298l (·1)4.6819 1-11 4.050-1 (.1) 3. 6668 1-112.9787 (-1) Z.Z965
lo 4.1492 3.2171 Z. 6Z78 Z. 3860 z.0139 l. 840~ 1.606 1 l,28Z3 1.0000 JO
l-ll 7.7985 l-116. 2Z47 1· ll S.4321 1-1)4.8218 l·ll 4. 1911 (· l) 3. 8055 (·I) 3, 1081 (-1) 2.4101
40 3. 5861 Z,87Zl 2.4015 2. Z0l4 1.9429 l. 744~ l. 5411 1, Z5Z9 (-1) 9. 94"1 40
1·117.8722 l· 1) 6. 3565 1-11 s. 5810 l·ll 4. 9778 (-11 4. 3493 l·l) 3. 9618 l·I) 3, 2556 (-1) 2. 5400
60 3. 0894 ' 2. 5549 z. 1874 z.OZ85 l,81SZ 1.6491 1.47~5 1. zzz9 1-1) 9. 8884 60
l·l) 7.9548 1-11 6. 5036 1-11 s. 7484 1-11 s. 1546 l•ll 4. 5Z9Z 1-11 4.1406 1-1) 3.4251 l·l)Z.6908
IZO z. 6480 l 2.2621 l. 9839 l. 8600 1.6899 l. 5543 1.4094 1.1921 1-11 9. 8333
(-1) 8.0489 1-116.6716 (-1) s. 9400 (-11 5, 3579 l• 11 4. 7378 1-114.3484 (-1) 3. 6238 1-11z.8693 IZO
(1) z. 2544 1.9901 l. 7891 1.6964 l. 5660 1.091. l. 3419 1.1600 (· 11 9, 1781
l·ll 8.1593 1·116.8662 1·11 6.11•41 1· ll 5. 5969 l·ll 4. 9845 1-ll 4. 5956 1·11 3, 8627 (•ll J. 0860 11:)
111 =40
"2 p•0.0001 p• 0.001 p1rO, 005 p •O. 01 p•O,OZ5 p=0.05 p•O, I p• o. u p•O, 5 p•D. 75 p•O, 9 p•O. 95 p•O, 975 p•O, 99 p•O, 995 p• o. 999 p•O. 9999 "z
1 +71 6. 2811 (+51 6. 2811 (+4) z. 5148 (+31 6, 2868 1+31 l. 0056 (+Z) Z. 5114 1+116,25Z9 Z, 1584 (-1) 7. 3389 (· I) 3. SZ68 1-11 z. 4481 (-1) 1.8437 1-11 1. 3612 1-11 1. 1328 l·ZI 7. 9306 1-21 s. 3566 1
'· 7144
l.Z 161 3. 3264 (41 7. 1666 131 4. 9015 13) l. 5416 IZ) 3, 3490 (2) l. OSZZ 111 3.2877 6. 8310 1.8677 1-11 7.15Z2. (-11 3. 6589 l·ll z.. 6044 (-11 l. 9946 (-1) l. 5019 (·l) 1.ZS48 l·Z) 8. 9046 (·216,0850 l. z
l. s (SI l. 7971 111 8. 3408 IZ) 9. 7515 IZ) 3, 8672 12) 1, 1366 111 4. 4837 (1) l. 75ll ... 8399 l. 6ZZ5 l·l) 7, 0221 (-1) 3. 8399 l•I) Z.8099 (-11 z. 1929 (·l) 1.6803 (-l) 1.4176 1-11 l. OZZI l·ZI 7. 0830 1. s
z +31 9. 9995 ~+21 9. 9948 l+ZI l. 9947 (+119.9n4 (+I) 3, 9473 (+11 J.9471 9, 4663 3. 4511 1.4178 1-116.9662 1-114,0977 1-11 3.0943 1-11 z. 4685 1-111.9311 1-11 1,6484 (·11 l.Z.lZ.O 1-z.18.5485 z
3 +ZI 5, 8236 (+Z) l. 2496 (+l) 4. Z308 l+ll l.6411 1+11 l. 4037 8, 5944 5. 1597 Z,4674 l. Z4b4 l· 117.0230 1-11 4. 49Z2 1-1) 3. 5227 1-11z.8874 (-11z.3188 1-11 Z.0097 Hl 1.5164 1-11 l. 0956 l
4 +ZI 1.4418 1+11 4. 5089 (+111. 975Z 1+11 l. 3745 8. 4111 s. 7170 3, 8031> Z.08ZI 1.1116 1-117. IZOO 1-114.7826 (-1) 3. 8373 1-111. 1989 (-!)Z.61ll 1-ll Z.Z86J (-l) 1.7550 1-ll l.Z88B 4
s +ll l>.3031 l+ll Z,460Z l+I) l. 2530 9. Z91Z 6.1751 4.4638 3, 1573 1.8763 1.1Z97
l• 11 7. Zll4 l·ll 5.00110 1·114.08Z5 1·11 3.4439 1-112.8459 l·ll Z.5088 (-l) 1.9500 (-ll l.H95 s
6 +11 3. 6450 K+I) l. 6445 9. 2408 7. lOZ 5,0125 l. 7743 z. 7812 J.7477 l, 1031 6
l•ll 7.2961 1·11 5.1897 l•ll 4.Z810 1·113.6418 l-ll l.0386 1-11 2.6933 1-11 2. 1139 1-111.5865
7 +11 z. 4680 1+11 I, 23t6 7. 4225 5. 9084 4. 3089 3. 3404 z. 5351 1.6593 1.0846 1·117.3687 1· 11 5. 3405 l· 114.4464 (·113.8113 C·ll l. ZOIZ 1-11 2. 8500 1-11 z.Z545 1-111. 7053 7
8 +11 1.8425 9.9194 E.. 287S 5. 1156 3. 8398 3. 0428 1.. 3614 l. 5945 1.0111 1-117,4JZZ l· 11 5, 4678 1·114. 5867 1-1) 3. 9543 1-11 l. 3411 l· ll z. 98S3 (·11z.3770 HI 1. 8097 a
t
o 10
9 +I) 1.4672
(+11 l.ZZZi
8. 3685
7.2971
5, 5186
4. 9659
4, 5667
4.1653
3. 5055
3. 2554
2, 8Z59
z. 6609
z. uzo
2.1317
l. 5450
l. 5056
1. 0608 (·117,4884 (· 11 S, 5776 (-114.1081 1·11 4.0785 1-113.4631 1-ll 3, 1037 (-112.4849 1-11 l. 9025
l. 05Z6 1-ll 7, 5381 1-1) 5,6731 1-1) 4.814Z (-11 4, 1873 (-1) 3, 5708 1-11 3. 2085 1-11 2. 58-11 (-1) l. 9856 10
9
zo 5. Z817 3.8564 3.0ZIS 2. 6947 l., Z87l 1.9938 1.1083 1.3301 1.0171
1-117.8382 l·ll 6, 2298 1-11 5.4380 1-11 4.8l63 1•11 4.ZZl4 l• 11 3, 8485 l·Jl 3.1797 (-1) 2. 5143 zo
2. 7654 Z.49U Z.1460 1.8920 1.6407 1.3004 1.0113 24
24 4. 5669 3.4468 1·117.9058 1-116.3528 l·ll 5. 5776 (-1) 4. 98Z8 1-ll 4, 3706 l· I) J. 9968 (-1) 3.3210 (-11 2.6418
z. 5241 z. 2992 Z.0089 l. 7918 l. 5732 1.2701 1.0056 JO
30 3, 9370 3.0716 l-l) 7.981! 1-116.4á89 (.1) 5. 7326 1·11 S. 1469 (-1) 4. 5384 l•I) 4.1641 (-1) 3.4818 1-11z.7880
IZO z. 4471 Z.1128 1. 8709 l. 76Z8 l. 6141 l.495Z l, 3616 1.1752 1-11 9. 8887 (-l) 8,2781 l· 11 7,018! 1-116.3428 1-11 5, 7998 (-1) 5,U59 1-ll 4, 8461 (-114.1489 (-1) 3. 4073 IZO
m 2,0516 l. 8350 l. 6691 1,S9U 1.4835 1.3940 l. Z951 1.1404 1-11 9. 8339 1-11 8. 4154 (-1) 7.Z6Z7 (.1)6,6Z73 l-1) 6.1084 (-11 s. 5411 1-11 5, 1165 HI 4.4791 (·ll 3. 7208 m
V1 = 60
-·
-....--
-"z
1
P" .0001 p•0.001. p•0,005 p•0,01 p• 0.025 p•O.OS p•O. I
l•"l 6. 313·1 l+SI 6. 3134 1+41 2. 5Z5l 1+31 6, 3130 l+ll 1.0098 1+21 z. SZZO (+11 6.1.794
p •0,25
9. 7591
p•O, S p•O, 75 p• o. 9 p•0.95 ps O, 975 p•0.99 p •O. 995 p• o. 999 p•0.9999
Z.1716 1·11 7.4113 1·11 3. 5824 1· ll 2.4993 (-1) 1.8919 1-1) 1.4130 1-111.1772 1-ZI ·s. l5Z 1 l·ZI s. 7548
"z
1
l. z 16) 3. 3375 (4) 7.1901 131 4.9178 !31 1.5488 IZI 3. 360Z IZI l. 0558 111 3.2991 6. 8588 1.8779 1·1) 7.2258 (· 11 l. 720-1 1·1) Z, 6619
1·11z.0497 (-1) 1.55"7 l·l) l.3064 1-ZI 9. 3973 l·Zl 6. 5529 1.2
l. 5 151 l,8008 131 8. 3581 IZI 9. 7718 IZI J. 8753 IZl l. 1390 (11 4. 4937 (1) l. 7565 4. 8562 1.6321 1-1) 7, 0985 1-113.9090 1•11 2. 8763 1-11 z.2575 ·111,700 1•11 1.4792
1-111.0815 l·ZI 7, 6516 l. s
l 1+3) 9. 999! (+Z) 9. 9941 1+21 l. 9948 l+ll 9.9483 1+11 3. 9481 l+I) 1. 9479 9. 4746 3. 4594 1.4261
1-117.048l 1·11 4.178! 1-11 3. 1741. l·ll z. 5476 •JI Z.0091 1-1) l. 7251> 1·11 l.ZBM (-Z) 9, 1.758 z
l +ll 5. 800Z ~+Z) l. Z447 1+114.2149 1+11 2. 6316 C+I) 1. l99Z 8. 5720 5. ISll 1..4697 l. ZSl6 1-117,1149
l·ll 4. 5926 (-1) 3. 6257 1·11Z.9918 l· 11 z. 4237 <-11 2. 1146 1-11l.6204 1·11 l. 1972 l
4 +ll 1.4305 (+11 4. 4746 (+111. 9611 l+I) 1. l65Z 8. 3604 s. 6878 3. 7896 Z.0817 1. 1782
1-ll 7.1.Zll 1-114.8996 l· 113,9601 1·11l.31.48 1-112.7404 1·11 Z.41 SS 1·111.8844 (-1) 1.-1165 4
s +11 l>.2309 l+ll 2,4333 11'111.1.402 9, ZOZO 6, llZ5 4.4314 3, 1401. l. 8742 l. 1361
1·11 7. lZZl 1-11 5, 1395 1·114,22Z.f l· 11 3, 5890 1·11 Z,9950 1-11 z. 1>596 (·11 2.1024 1-111.6009 s
6 ,.11 l. 5910 l+ll l. 6214 9.12.19 7. 0568 4, 9589 l. 7398 2. 7620 l. 7443 l. 1093
1-ll 7. 411.l 1· I) 5. 3341. (-114.4366 l·ll 3. 8060 l·ll 3.2065 l• I) 2.81>39 1·1)2.Z873 (-l) 1.7596 6
7 +11 2. 4Zl8 1+11 1.2119 7. 3088 5, 8236 4. 2544 l, J04J z. 5142 1.6S4e 1.0908 l·I) 7, 4912 1-11 5.496\ 1-114.6157 l· 1) 3. 9891 l· 11l.381>4 (· 11 3.0385
1·11 2.4471 (-11 l. 8985 7
8 +11 l. 8041 9. 7Z7Z 6.1772 5,0316 3. 7844 3.005l z. Jl91 l. 5891. l. 0771 1-11 7, 5609 l· 11 s. 6344 1-11 4. 7687 1-11 4.1465 l·ll l. 541.0 l·ll 3.1904
1·ll z. 5874 1-11 Z.0216 8
~ 9 +11 l.43Z7 8.1865 s. 4104 4. 4831 l.449l z. 7872 Z.1.085 l. 538~ 1.06&7 l·ll 7.&ZZ5 (·11 5.1537 1-114.9017 l· 11 4. ZB38 1·11 3.6785 1·11 3. U41
~ 1-11 Z. 71ZO 1·11 Z.1319 9
1-1 10 ~+11 l. 1904 7.122-t 4. 8592 4. 0819 3.1984 Z.61.11 Z.1071. l.499Q 1.058' 1-117.6770 l· 11 S. 8SllZ 1-11 5.0186 l·ll 4.4049 1-11 J. 7997 l· 11 J. 4433 1·112.8237 1•11 Z.2ll4 10
12 8. 9933 s. 7623 4, 1229 3. SJS! 2. 8476 2. 3841. l. 9597 1.09; 1.0464 1-117.7700 1·116,0JH (.1) 5, 2154 <·ll 4.6100 1-11 4. 006Z 1-111. 6471 1-113.0163 1·11 Z.4017 IZ
15 6. 76Z6 ol,6377 3, 480l 3, 0471 z. 5Z.4Z z. 1601 l. 8168 l. 3796 l. Ol4!
1-117,8796 1-116,2367 1-11 s. 4454 l· 11 4. 8513 l-ll •J. Z.512 1-111. 8903 (-1) 3.Z.488 1-11z.6164 15
zo s. 0463 J. 7030 z. 9159 z. 6077 Z. 2ll4 J, 9464 I, 6768 l. 3193 l. ozzs 1-11 8, OIZ2
1·116.4788 l· 11 5. 7208 1-11 5:1427 (· 11 4. 5500 1·114,1890 (-11 J. 5378 1-11 Z.8834 zo
1.4 ~. 3397 l. Z946 Z.6585 z. 4035 Z.0799 l.84H l. 6073 l. 2885 l. 0170
30 3. 7163 z. 9196
1-11 8, º'ºº 1-1> 6,6194 (.1) 5. 8820 l-11 5. llfl l· 11 4. 7272 (-11 4, l67Z l•ll l. ?IZl 1·113.041.S l4
z. 4151 l, 2079 l. 940( l. 7396 l. 5376 l. 2571 1.011 1 1· ll 8. 1773
(-1) "· 7714 1·116.0639 l· 11 s. 5090 l· 114.9298 1-114.5716 1-1) 3.9140 (-11 3, 2369 30
40 3, J64Z z. 5737 l. 1838 Z.0194 1.8028 l. 6373 l. 467Z l.Z.Z49 1.0056 l·ll 8, Z77' l-116,9575 l· 11 6. Z72l 1-11 5. 7339 1·11 ~. 1653 1-114.8102 (-1) 4. 1518 1-11J.4638 40
60 Z,671.3 1.. 25ZJ l. 961.Z l. 8363 1.6668 l. 5343 l. 3952 1.1911. 1.0000 1-11 8,
IZO Z.2301 1.9502 l. 7469 J,6557 l. SZ.99 J.4290 l. 3203
394~ l· 117.1674 1-116.5176 (· 11 s. 9995 1-1) S.4457 1-11 5. 09&3 l-1) 4.4400 1-113,7420
1.155! 1-11 9. 9443 1-1) 8, 5361 1-11 7. 4Z06 1-116.815• 1•11 6, JZSI l• 11 5. 791.7 l•ll 5.45Zl 1·114.80Z8 l·l) 4, 0975
"º
IZO
m l.8Z50 1.6601 l. 5l1.! l. 4730 l. 3883 1.3180 l. Z.400 1.1164 l·ll 9. 8891 1-11s.71 s~ 1-ll 7. '74ZI 1-ll 7.1971 1-116. 7461 (-11 6. Z.417 1·11 S.91.1.4 1-1} 5.2897 1-1) 4, 5828 m
- .... N ... .
... ... ... ... .... 2 :::: '.!: o ... .. ~ ~ ...o e
"'"' - ...o
..!
N
..
..N N
... ......
.............. ... .. ... ... :: ... ... ... ...... = g;... ...
¡ ª~ ~ :~
o
.:
g ...o ...... ~ :: o "' ~ ~ "'... ...... ... .... ~...;
Q
~ N
i: ..."' o
:
o
g
.... ... ~
... ... ...
N
- -
Ñ Ñ ..; ..;
--
N 8 N• ,.J Ñ .-.¡ -O ...:
..."'o ...... ...-.... -... ~... ~... o......... ............ ~ : ....~ -......... -....... ............ ......
N
! o
......... ...... N N o
o
.. .. .... . . ....
N
.... ."'
.,.
o ..; ... .¡ ... ...
..; ..; "' N
...; ...; ..o ..o
N
-...... - -...... - -...; - -....... -:ll...o -... ...o- - -!;;o o... - -... ~ ! ........... ...... s =~ : : : : ; : o
r
o
~ 5~ ¡ ; a
N
.... - - . ª . - -
o
..; ..; ... ... ... ..; ..; ..; ...
N
- - - - - - -
...
. -...... -::: .......... ..... ......
.... .... .,; .,; .,;
.
o
a.
T ..O CD """'
~
..O
Ñ
O--
N
N
,..; ..; "" "' "' "' ..;
.,;
.. .. .
..!.
... ... ... o .... oco
... ... ..!.
.. .. .. .. 5 .........
... ... ...... "' ...... ~ ... ..O Oio ID
~ § ~
f'lol llft ..
- --
::;
...;
o
.,;
::!
.,;
....
o ..: ,.¡ ..
.. ...-
.... . ... ............. ; ¡ .... ........:
.:.
!..; ......... .....; ......o..; ..."'...... ............ "'..;...... ~ ~.. ......... ...... ......; .......,; ......;
co .... o
.... ......
o N
E ~ ~ Ñ~ Ñ~ ¡..; !..; i ; ; ; : : o
. . ª.... .. ..
:;:: : g
o
."' -
12: .,; .,; .,;
o
.""----..!.-..!.-
..
=
o -o ...-... - ... ... .!.co -§;... -...... -... ..!.:::2 ..!.... -... -...... .:....... .!.
;! ... ~ ... ~ ~ ... : ~ ..."' ... o . s ........ ...-.."'... ...-....~ -E..: -...
CD
"':;:: : ~ ~ ~ : ~. .~ ; ~ ;- ...- o~ - - ...-... ...-... -... -
.....
... o
... ;
.... - . -.. .. ..- . - .
... ...
... o o '°
o
"'
..; N Ñ ..; ... ..; ... ..;
"'
.,;
"' .... ... ...
.!. - - - - - ..!. - ....... ¡-
.,; o
. -- --------
"'
.. ~ ~ ~ ~ "'
g
.,¡ ..o
8;
.o ..o"' 1: ~
... ..
' ... -...... ...... ..."':;; ... ... ... a oco ......... :.;!; ~ .......~
N N
.. ..!.
~
a. -
~ ~ ~ °" ~
~ ~ ~ ~
- -- - - - -
.... ......
~
~ ... ...... o ,,; ~ ~ ~ ~
o
5 ~ : : i : : :
."' - . .. . .. . ~
... .,; ,,; .,; ...; ... ~ .,; ... ... ... ...
;;; lñ ~
:,.~----
~ ~ • ~ • ~ "' 11\
...... ;
o
.... . ..:
...... -...
. ...
º..!.
... ..: ..: ... ... ... "'..: ... ... ... ..: ..:
' ..!.
..; .. ... . .. ..
.
..;
N
CD
o
,,;.
~
~
i,..: ~
,.:
~
~
a ¡ i ; ;e
~ ~ ,.: G
¡
~
ª: ;::
~ ~ CO ~
§
s ...... ~ . .....
... ... ... ... .. :: ~ ...... ..... ..."' .!.
.., ......
. ... ......... e.. ......
.... :; CD
....
.... oco "' '# ~
N
o
:¡¡ ::: ~ i §· :;; ~ ...co ...
.... !; CD
o o
o ,,;. "'CD ..; ..; N ..: ..: ..: ..: ..: ..: ..: ...: ...:
...o .. . . ......... ... ...... ... ....... ........... ...... "' ;...... ;...... ...~ ~ ......
... ... ... ... ... ...
....
......o... ... ....,; ...¡: ...... :::~
co ."' .... .........
; ~....¡ ~,.,¡
-.
o .,;
~
... ~ .
... ... ... ..: ,.¡ ..:
..;
o
N
l ~
o .,;
N
,.J
... ~
- ... ...
~
.. . ...... ...-:.... "'......... ..... ? ...... ......... ....... ...... ........... ......"' - ........
...o "'::: ............ ......~ i ~ ! .......... o...
...
... +
~~-~~i~~=! ~
o o
o
.....:
N
.. =.. ..
.. o o
.. g;
..."' .;.... ...
N
. .. ..
CD
~ lñ "' o
o
o
.;. ..; .,; ... ..; ..:
,.¡ ..: ,.¡ ..: ...; ,..; Ñ Ñ ..: ....: ....: ..;
... l !, . . -
- ~
... ...... ....... ... .......... .... ... .. ...... .........
... ...
... .. : ... . ! =
.. o
a.
°" ..... .....
+
.... .. .... o
.. .. ... ... ~
o o
N :¡; o ~ ; :; :: o : o
. .. ..
o ... ... ..:
... :t :t ~
..:
........... i ............ ....... ..;.... ~ ...... ¡... ... ........ ...... ::: ;
..; N N ..; ..:
~
..,.
+ ~
co .o .. .;.
........
...... ... .......
o o ... ... o
... ...co co
:... ... ...... .,:. i.,;
o ...
...o ... o ... o o
~ 5 ... ..
g: ~
.:
::! "' .. oO "' o
. .
...
o
.,;
...
"'
...; ... ... "'
.,;
= ..; ,.¡ ..; N ..; ~ ...
CO
Ñ
,.-
N
N
N N
o
...... .
o
.... ... ... o
o
o
...... o N
...... ~ o
~ ~ ....,; ... ...
...
....
o
o
! ... ..
.... "' ; :; o :; ~ ; M CD o
N
~ 't. 't. ~
442 443
Table A.4 Noncentral Beta Table
{31'(11,, 113, cp\
T
Entries are (31,(v 1 , v 2 , $ ), where P = J g( (3) d/3 1
1
f2 E~os
1 1.5 2 2.5 3
rp
4 5 6 7 8
l f2 E~os
1 1.5 2 2.5 3
cp
4 5 6 7 8
2 .903 .862 .763 .643 .517 .395 .200 .083 .028 .008 .002 1 2 .950 .881 .803 .704 .595 .484 .286 .140 .064 .024 .008
4 .658 .805 .631 .428 .247 .120 .016 .001 - 4 .776 .824 .661 .460 .272 .135 .020 .001
6 .500 .777 .570 .343 .164 .061 .004 6 .632 .789 .579 .340 .153 .052 .002
7 .444 .768 .552 .319 .144 .050 .003 7 .575 .777 .551 .304 .124 .037 .001
8 .399 .761 ,537 .302 .129 .041 .002 8 .527 .767 .530 .277 .104 .027 .001
9 .362 .756 .526 .288 .119 .036 .001 9 .486 .759 .513 .257 .090 .022
10 .332 .751 .517 .278 .111 .032 .001 10 .451 .752 .498 .241 .080 .017
11 .306 .747 .510 .269 .105 .029 .001 11 .420 .• 747 .486 .228 .072 .015
12 .284 .744 .504 .262 .100 .027 .001 12 .393 .742 .476 .217 .066 .013
13 .264 .741 .499 .256 .096 .025 .001 13 .369 .737 .468 .208 .061 .011
14 .247 .739 .494 .251 .093 .024 .001 14 .348 .734 .461 .201 .057 .010
15 .232 .737 .490 .247 .090 .023 15 .329 .730 .454 .195 .054 .009
16 .219 .735 .487 .243 .087 .022 16 .312 .727 .448 .189 .051 .008
17 .207 .734 .484 .240 .085 .021 17 .297 .725 .443 .184 .048 .008
18 .197 .732 .481 .237 .084 .020 18 .283 .722 .439 .180 .046 .007
21 .171 .729 .476 .231 .079 .019 21 .248 .717 .428 .170 .042 .006
22 .164 .728 .473 .229 .078 .018 22 .238 .715 .425 .168 .040 .006
23 .157 .727 .471 .227 .077 .018 23 .229 .714 .422 .165 .039 .006
24 .151 .726 .470 .226 .076 .018 24 .221 .712 .420 .163 .038 .005
25 .145 .725 .468 .224 .075 .017 25 .213 .711 .417 .161 .037 .005
26 .140 .725 .467 .223 .075 .017 26 .206 .710 .415 .159 .037 .005
27 .135 .724 .466 .222 .074 .017 27 .199 .709 .413 .157 .036 .005
28 .130 .723 .465 .221 .073 .017 28 .193 .708 .411 .155 .035 .005
29 .126 .723 .464 .220 .073 .017 .t 29 .187 .707 .410 .154 .035 .004
30 .122 .722 .463 .219 .072 .016 30 .181 .706 .408 .153 .034 .004
60 .063 .715 .450 .205 .065 .014 60 .095 .692 .384 .134 .027 .003
00 .707 .437 .193 .058 .011 00 .678 .362 .117 .021 .002
v1 = numerator degrees of freedom; v2 = denominator degrees of Creedom.
Table A.4 is reprinted by permission of University College, from P. C.
Tang, The Power Function of the Analysis of Variance Tests with Tables and
lllustrations of Their Use, Statist. Research Mem., vol. 1.
444 445
<P
cp
f 2 E~os E~os
1 1.5 2 2.5 3 4 5 6 7 8 la 1 2 2.5 5
1.5 3 4 6 7 8
2 .966 .888 .817 .726 .624 .519 .324 .177 .084 .035 .013 2 .975 .892 .824 .738 .640 .537 .345 .195 .097 .043 .017
4 .832 .830 .670 .468 .278 .139 .020 .001 4 .865 .833 .673 .471 .279 .139 .020 .001
6 .704 .791 .574 .326 .139 .044 .002 6 .751 .791 .567 .314 .128 .038 .001
7 .651 .776 .540 .283 .106 .028
7 .702 .774 .529 .265 .092 .022
8 .604 .764 .513 .251 .084 .018
8 .657 .760• .497 .229 .069 .013
9 .563 .754 .491 .226 .068 .013
9 .618 .748 .471 .201 .054 .008
10 .527 .745 .472 .206 .057 .010
10 .582 .738 .449 .179 .043 .006
11 .495 .738 .457 .190 .049 .008
11 .550 .729 .430 .161 .035 .004
12 .466 .731 .444 .178 .043 .006
12 .521 •721 .414 .148 .030 .003
13 .440 .726 .433 .167 .038 .005
13 .494 .714 .401 .136 .025 .002
14 .418 .721 .422 .158 .035 .004
14 .471 .708 .389 .127 .022 .002
15 .397 .716 .414 .151 .032 .004
15 .449 .702 .378 .119 :019 .002
16 .378 .712 .406 .144 .029 .003
16 .429 .697 .369 .112 .017 .001
17 .361 .709 .399 .139 .027 .003
17 .411 .693 .361 .106 .016 .001
18 .345 .705 .393 .134 .025 .002
18 .394 .689 .354 .101 .014 .001
19 .331
20 .317
.702
.700
.388
.383
.130
.126
.024
.022
.002
.002
J 19 .379 .685 .347 .097 .013 .001
20 .364 .681 .341 .093 .012 .001
21 .305 .697 .379 .123 .021 .002
21 .351 .678 .335 .089 .011 .001
22 .294 .695 .375 .119 .020 .002
22 .339 .675 .331 .086 .010 .001
23 .283
24 .273
.693
.691
.371
.367
.117. .019
.114 .019
.002
.001 -
. 23 .327 .672 .326 .083 .010-
24 .316 .670 .322 .080 .009
25 .264 .689 .364 .112 .018 .001
25 .306 .668 .318 .078 .009
26 .255 .687 .361 .110 .017 .001
26 .297 .665 .315 .076 .008
27 .248 .686 .359 .108 .017 .001
27 .288 .663 .312 .074 .008
28 .240 .684 .356 .106 .016 .001
28 .279 .661 .309 .072 .008
29 .233 .683 .354 .105 .016 .001
29 .272 .660 .306 .071 .007
30 .226 .682 .352 .103 .015 .001
30 .264 .658 .303 .069 .007
60 .121 .662 .320 .083 .010 .001
60 .144 .632 .265 .049 .004
00 .642 .289 .067 .007 00 .604 .227 .036 .002
446
447
/2 E~os
'P
l
t·
f 2 E~os "'
j 1 =6
'
1 1.5 2 2.5 3 4 5 6 7 8 1 1.5 2 2.5 3 4 5 6 7 8
2 .980 .894 .828 .745 .649 .549 .359 .207 .106 .048 .019 2 .983 .895 .831 .749 .656 .557 .368 .216 .112 .052 .022
4 .887 .835 .675 .473 .280 .138 .020 .001 4 .902 .836 .677 .473 .280- .138 .019 .001
6 .785 .790 .561 .304 .119 .033 .001 6 .811 .789 .556 .296 .113 .030 .001
7 .739 .772 .519 .251 .082 .018 7 .768 .769 .510 .239 .074 .015
8 .697 .756 .483 .211 .059 .010 8 .729 .753 .472 .198 .051 .008
9 .659 .743 .454 .181 .04~ .006 9 .692 .738 .440 .166 .037 .005
10 .625 .731 .429 .158 .033 .004 10., .659 .725 .412 .142 .027 .003
11 .593 .720 .408 .140 .026 .002 11 .628 .713 .389 .123 .020 .002
12 .564 .711 .390 .125 .021 .002 12 .600 .702 .369 .108 .016 .001
13 .538 .703 .374 .113 .017 .001 13 .574 .693 .351 .096 .012 .001
14 .514 .695 .360 .103 .015 .001 14 .550 .685 .336 .086 .010 .001
15 .492 .689 .348 .095 _.012 .001 15 .527 .677 .323 .078 .008
16 .471 .683 .338 .088 .011 .001 16 .507 .669 .311 .071 .007
17 .452 .678 .328 .083 .009 17 .488 .663 .301 .065 .006
18 .435 .673 .320 .078 .008 18 .470 .657 .291 .061 .005
19 .419 .668 .312 .073 .007 19 .454 .652 .283 .056 .004
20 .404 .664 .305 .069 .007 20 .438 .648 .276 .053 .004
21 .390 .660 .299 .066 .006 21 .424 .644 .269 .050 .003
22 .377 .656 .294 .063 .006 22 .410 .639 .262 .047 .003
23 .365 .653 .288 .060 .00~ 23 .397 .635 .257 .045 .003
24 .353 .650 .284 .058 .005 24 .385 .632 .252 .043 .003
25 .342 .647 .279 .056 .005 25 .374 .629 .247 .041 .002
26 .332 .644 .275 .054 .004 26 .363 .625 .242 .039 .002
27 .323 .642 .272 .052 .004 27 .353 .623 .238 .037 .002
28 .314 .640 .268 .050 .004 28 .344 .620 .234 .036 .002
29 .305 .637 .265 .049 .003 29 .335 .617 .231 .034 .002
30 .297 .635 .262 .048 .003 30 .326 .615 .228 .033 .002
60 .165 .604 .219 .031 .001 60 .184 .576 .181 .019 .001
00 .567 .177 .019 .001 00 .532 .138 .010
448 449
.fi= 8
cp cp
f 2 E~os f2 E~os
1 1.5 2 2.5 3 4 5 6 7 8
1 1.5 2 2.5 3 4 5 6 7 8
2 .986 .896 .833 .753 .660 .563 .374 .222 .117 .055 .023 2 .987 .897 .835 .'755 .664 .567 .380 .227 .121 .05'7 .024
4 .914 .837 .678 .474 .280 .138 -:019 .001 4 .924 .838 .678 .474 .279 .137 .019 .001
6 .831 .788 .552 .289 .108 .028 .001 6 .847 .787 .548 .284 .103 .026 .001
7 .791 .767 .503 .230 .068 .013 7 .810 .765 .497 .222 .064 .012
8 .754 .749 .462 .187 .046 .007 8 . . 775 .746 .454 .178 .041 .006
9 .719 .733 .427 .154 .031 .004 9 .742 .729 .417 .144 .028 .003
10 .687 .719 .398 .129 .022 .002 10 .711 .714 .386 .119 .019 .001
11 .657 .706 .373 .110 .016 .001 11 .682 .700 .359 .099 .013 .001
12 .630 .695 .351 .094 .012 .001 12 .655 .688 .336 .084 .009
13 .604 .684 .332 .082 .009 13 .630 .677 .316 .072 .007
14 .580 .675 .316 .073 .007 14 .607 .666 .298 .062 .005
15 .558 .667 .301 .065 .006 15 .585 .657 .283 .055 .004
16 .538 .659 .289 .058 .005 16 .564 .648 .269 .048 .003
17 .518 .652 .277 .053 .004 17 .545 .641 .257 .043 .003
18 .501 .645 .267 .048 .003 18 .527 .634 .247 .039 .002
19 .484
20 .468
.639
.634
.258
.250
.044
.041
.003
.002 t
'
J 19 .510
20 .495
.627
.620
.237
.228
.035
.032
.002
.001
21 .453 .629 .243 .038 .002 21 .480 .615 .220 .030 .001
22 .439 .6:'.4 .236 .036 .002 22 .466 .609 .213 .027 .001
23 .426 .619 .230 .034 .002 23 .452 .604 .207 .025 .001
24 .414 .615 .224 .032 .001 24 .440 .600 .201 .024 .001
25 .402 .611 .219 .030 .001 25 .428 .595 .196 .022 .001
26 .391 .607 .215 .028 .001 26 .417 .591 .191 .021 .001
27 .381 .604 .210 .027 .001 27 .406 .588 .186 .020 .001
28 .371 .601 .206 .026 .001 28 .396 .584 .182 .019
29 .362 .598 .202 .024 .001 29 .386 .581 .178 .018
30 .353 .595 .199 .023 .001 30 .377 .578 .175 .017
60 .202 .550 .150 .012 60 .219 .527 .125 .008
CIO .498 .105 .005 CIO .466 .081 .003
450
l 451
E~01 and the Corresponding Values of P(II)
1
f 2 E~Ol "' ·¡ <p
i
20 .369 .895 .701 .417 .170 .045 .001
21 .276 .898 .732 .488 .249 .093 .004 21 .355 .894 .697 .410 .165 .042 .001
22 .265 .897 .730 .484 .245 .090 .004 22 .342 .893 .693 .404 .160 .040 .001
23 .255 .896 .728 .481 .241 .088 .004 23 .330_ .891 .690 .399 .155 .038
24 .246 .896 .72~. .477 .238 .086 .004
1 24 .319 .890 .686 .394 .151 .036
25 .237 .895 .724 .474 .235 .084 .004
26 .229 .894 .722 .471 .232 .082 ---.003 ~ 25 .308
26 .298
.889
.888
.683
.680
.389
.385
.148
.144
.035
.034
27 .221 .894 .720 .469 .229 .081 .003 27 .289 .887 .678 .381 .141 .032
28 .214
29 .212
.893
.893
.718
.717
.466
.464
.227
.225
.079
.078
.003
.003
í 28 .280
29 .272
.886
.886
.675
.672
.377
.373
.138
.136
.031
.030
30 .201 .892 .716 .462 .223 .077 .003 30 .264 .885 .670 .370 .134 .029
60 .106 .885 .696 .430 .194 .061 .002 60 .142 .873 .637 .324 .102 .019
00 .877 .675 .400 .169 .048 .001 00 .860 .601 .279 .076 .011
452 453
- f t= 4-
cp
cp
f2 E~Ol
f2 E~Ol 1 2
1.5 2.5 3 4 5 6 7 8
1 1.5 2 2.5 3 4 5 6 7 8
2 .995 .978 .962 .942 .915 .884 .810 .724 .631 .536 .444
2 .993 .977 .961 .939 .911 .878 .800 .709 .612 .515 .421
4 .941 .960 .1109 .822 .700 .557 .280 .102 .027 .005 .001
4 .926 .959 .907 .818 .695 .552 .276 .100 .026 .005 .001
6 .859 .943 .849 .685 .475 .277 .053 .005
6 .830 .943 .850 .691 .486 .290 .059 .006
7 .818 .936 .821 .624 .389 .191 .018
7 .784 .936 .825 .636 .408 .210 .025 .002
8 .778 .928 .796 .571 .322 .136 .010
8 .740 .929 .803 .590 .347 .158 .014
9 .741 .922 .773 .526 .269 .098 .003
9 .700 .923 .783 .550 .299 .120 .008
10 .706 .916 .752 .487 .227 .073 .002
10 .663 .918 .765 .517 .261 .094 .004
11 .673 .911 .733 .453 .195 .055 .001
11 .629 .913 .749 .487 .231 .075 .002
12 .643 .906 .716 .424 .169 .042 .001
12 .598 .909 .735 .463 .206 .062 .001
13 .616 .901 .700 .398 .148 .034
13 .570 .906 .723 .441 .186 .051 .001
14 .590 .897 .687 .376 .131 .028
14 .544 .902 .711 .422 .170 .044 .001
15 .566 .893 .674 .357 .117 .022
15 .520 .899 - .701 .406 .156 .038 .001
16 .544 .890 .662 .340 .106 .018
16 .498 .896 .692 .391 .145 .033
17 .523 .886 .652 .325 .096 .015
17 .478 .893 .683 .378 .135 .029
18 .504 .883 .642 .312 .088 .013
18 .459 .891 .676 .367 .126 .026 1
19 .486 .880 .633 .301 .081 .011
19 .442 .889 .669 .356 .119 .023 J
20 .470 .878 .625 .290 ·.075 ..010
20 .426 .887 .662 .347 .112 .021
21 .454 .876 .618 .280 .070 .009
21 .410 .885 .656 .339 .107 .019
22 .440 .873 .611 .272 .066 .008
22 .396 .883 .651 .331 .102 .017
23 .426 .871 .604 .264 .062 .007
23 .383 .881 .646 .324 .098 .016
24 .413 .869 .598 .257 .059 .006
24 .371 .880 .641 .318 .094 .015
25 .401 .867 .593 .250 .056 .006
25 .359
26 .349
.879
.877
.637
.633
.312
.307
.090
.087
.014
.013
l
(
26 .389 .865 .588 .244 .053 .005
27' .378 .864 .583 .239 .050 .()05
27 .338 .876 .629 .302 .084 .012
28 .368 .862 .578 .234 .048 .005
28 .329 .875 .625 .297 .081 .012
29 .358 .861 .574 .229 .046 .004
29 .319 .874 .622 .293 .079 .011
30 .349 .860 .570 .225 .044 .004
30 .311 .872 .619 .289 .077 .011
60 .196 .837 .509 .165 .024 .001
60 .171 .856 .571 .233 .050 .005
00 .810 .443 .115 .011
00 .836 .519 .182 .030 .002
454 455
i -
11= 5 1
cp
f2 E~o1
1 1.5 2 2.5 3 4 5 6 7 8
J f 2 E~o1
1
cp
1.5 2 2.5 3 4 5 6 7 8
2 .996 .978 .964 .944 .918 .888 .817 .733 .642 .549 .458 2 .997 .978 .964 .945 .920 .891
4 .951 .961 -
.910 .824 .702 .559 .282 .103 .027 .005 .001 4 .958 .962 .911 .825 .704 .560
.821
.283
.739
.104
.650
.027
.558
.005
.468
.001
6 .879 .943 .848 .679 .466 .266 .048 .004 6 .894 .944 .847 .675 .459 .258 .044 .003
7 .842 .935 .818 .614 .394 .177 .014 7 .860 .935 .815 .605 .362 .166 .011
8 .806 .928 .790 .556 .301 .121 .007 8 .827 .927 .784 .543 .285 .109 .006
9 .771 .920 .764 .505 .245 .083 .003 9 .795 .919 .756 .488 .226 .071 .003
10 .738 .914 .740 .461 .201 .058 .001 10 .764 .912 .730 .441 .181 .048 .001
11 .707 .908 .718 .424 .168 .042 11 .734 .905 .706 .400 .147 .033
12 .679 .902 .699 .391 .141 .031 12 .707 .899 .683 .365 .120 .023
13 .652 .897 .681 .363 .120 .023 13 .681 .893 .663 .334 .100 .017
14 .626 .892 .664 .339 .104 .018 14 .656 .888 .645 .308 .084 .013
15 .603 .888 .649 .318 .090 .014 15 .633 .882 .628 .286 .071
16
.009 -
.581 .883 .636 .299 .079 .011 16 .612 .878 .612 .266 .061 .007
17 .561 .880 .624 .283 .071 .009 17 .591 .873 .598 .249 .053 .005
18 .541 .876 .612 .269 .063 .007 t 18 .572 .869 .585 .233 .046 .004
19 .523 .873 .602 .256 .057 .006 1 19 .554 .865 .573 .220 .041 .003
20
21 .490
.506 .870
.867
.592
.583
.245
.234
.052
.047
.005
.004
' 20 .537
21 .521
.862 .562 .208 .036 .003
.858 .552 .198 .033 .002
22 .475 .864 .575 . .225 .044 .004 22 .506 .855 .542 .188 .029 .002
23 .461 .861 .567 .217 .040 .003 23 .492 .852 .533 .180 .027 .002
24 .448 .859 .560 .210 .037 .003
-
- 1
1
24 .478 .849 .524 .172 .024 .001
25 .435 .857 .553 .203 .035 .003 25 .465 .846 .517 .165 .022 .001
26 .423 .855 .547 .196 .033 .002 26 .453 .844 ~ ~510 .159 .020 .001
27 .412 .853 ·.541 .190 .031 .002 27 .442 .• 842 .503 .153 .019 .001
28 .401 .851 .536 .185 .029 .002 28 .430 .839 .497 .147 .017 .001
29 .391 .849 .531 .180 .027 .002 29 .420 .837 .491 .142 .016 .001
30 .381 .847 .526 .176 .026 .002 30 .410 .835 .486 .138 .015 .001
60 .218 .819 .452 .116 .011 60 .238 .801 .401 .081 .006
00 .784 .373 .070 .004 00 .755 .311 .042 .001
1
¡
-
456 457
lndex