Professional Documents
Culture Documents
2016 Bookmatter ACourseInMathematicalStatistic PDF
2016 Bookmatter ACourseInMathematicalStatistic PDF
Standard Distributions
Let Xi denote the number of failures between the (i − 1)-th and i-th successes
(i = 2, 3, . . . , r), and let X1 be the number of failures before the first success. Then
X1 , X2 , . . . , Xr and r independent random variables each having the distribution
N B(1, p) with probability mass function
Also,
X = X1 + X2 + · · · + Xr .
Hence
∞
∞
E(X) = rEX1 = r p x(1 − p)x
= rp(1 − p) x(1 − p)x−1
x=0 x=1
∞ ∞
d d
= rp(1 − p) − (1 − p)x = −rp(1 − p) (1 − p)x
x=1
dp dp x=1
∞
d d 1
= −rp(1 − p) (1 − p)x = −rp(1 − p)
dp x=0 dp 1 − (1 − p)
=p
r(1 − p)
= . (A.1)
p
λx
f (x) = e−λ (x = 0, 1, 2, . . . ),
x!
where λ > 0 is the mean. To see this let X be a random variable with this
distribution. Then
∞
∞ ∞
λx λx λy
E(X) = xe−λ = e−λ = λe−λ
x=0
x! x=1
(x − 1)! y=0
y!
= λe−λ · eλ = λ. (y = x − 1)
Also,
∞
∞
λx λx
E(X(X − 1)) = x(x − 1)e−λ = e−λ x(x − 1)
x=0
x! x=2
x!
∞
λx−2
= e−λ λ2 = e−λ λ2 · eλ = λ2 ,
x=2
(x − 2)!
so that
E(X 2 ) = λ2 + E(X) = λ2 + λ,
var(X) = E(X 2 ) − (E(X))2 = λ2 + λ − λ2 = λ. (A.3)
[See below for the relation B(α, β) = Γ (α)Γ (β)/Γ (α + β).] Here if X ∼
BB(α, β, n),
nα
E(X) = EE(X | Y ) = EnY = nEY = ,
α+β
E(X 2 ) = EE(X 2 | Y ) = E[nY (1 − Y ) + n2 Y 2 ]
α(α + 1) α
= (n2 − 1)EY 2 + nEY = (n2 − 1) +n ,
(α + β)(α + β + 1) α+β
var (X) = E(X 2 ) − (E(X))2 .
(See below for the computation of the moments of the beta distribution.)
I. The Uniform Distribution U (α, β) on [α, β] has the probability density function
(p.d.f.)
1
f (x) = for α ≤ x ≤ β,
β−α
=0 elsewhere.
Here α > 0, β > 0 and B(α, β) is the normalizing constant (beta function)
1
B(α, β) = xα−1 (1 − x)β−1 dx.
0
We now prove
Γ (α)Γ (β)
B(α, β) = , ∀ α > 0, β > 0. (A.6)
Γ (α + β)
Now
∞ ∞
Γ (α)Γ (β) = e−x xα−1 dx e−y y β−1 dy
0 0
∞ ∞
= e−(x+y) xα−1 y β−1 dxdy.
0 0
x=z−y
Change variables: z = x + y, y = y, to get , with
y=y
∂x ∂x
∂t ∂y
Jacobian = det ∂y ∂y = 1
∂z ∂y
∞ z
−z
Γ (α)Γ (β) = e (z − y) α−1 β−1
y dy dz
0 0
∞
−z α−1 β−1 y α−1 y β−1z
= e z z 1− dy dz
0 0 z z
∞ 1
= e−z z α+β−2 z (1 − u)α−1 uβ−1 du dz u = yz , du = 1z dy
0 ∞ 0
But
1 1
B(β, α) = uβ−1 (1 − u)α−1 du = xα−1 (1 − x)β−1 dx
0 0
= B(α, β), (x = 1 − u). (A.8)
Hence (A.7) and (A.6). Using (A.4)–(A.6), one gets the k-th moment of a beta
(Be (α, β)) random variable X as
In particular,
α (α + 1)α αβ
E(X)= , E(X 2 ) = , var(X)= .
α+β (α + β + 1)(α + β) (α+β)2 (α+β+1)
where α > 0, β > 0. Here α is a scale parameter, i.e., if X is ∼ G (α, β), then X/α
is G (1, β), Note that
k ∞
X 1
E = z k e−z z β−1 dz
α Γ (β) 0
Γ (β + k)
= = (β + k − 1) · · · (β + 1)β,
Γ (β)
so that
E(X k ) = αk (β + k − 1) · · · (β + 1)β. (A.10)
Hence EX = αβ, var(X) = α2 β.
has p.d.f.
1 2 2
fμ,σ2 (x) = √ e−(x−μ) /2σ , −∞ < x < ∞, (A.11)
2πσ 2
where μ ∈ (−∞, ∞), σ 2 > 0. The standard normal distribution N(0, 1) has p.d.f.
1 2
f (x) = √ e−x /2 , −∞ < x < ∞. (A.12)
2π
x−μ
To show that (A.12) (and hence (A.11), by transformation y = σ ) is a p.d.f.,
one needs to show that ∞
2 π
e−x /2 dx = (A.13)
0 2
For this use the transformation z = x2 /2 to get
∞ ∞
2 √ 1 1 1 1
e−x /2 dx = e−z 2 z − 2 dz = √ Γ . (A.14)
0 0 2 2 2
1 1 1 1 1 ∞
Γ2 = Γ (1)B , =B , (since Γ (1)= 0 e−x dx= − e−x |∞ 0 = 1, )
2 2 2 2 2
1 1
= x−1/2 (1 − x)−1/2 dx = z −1 (1 − z 2 )−1/2 2zdz (z = x1/2 , dx = 2zdz)
0 0
1 π/2
2 −1/2 cos θdθ
=2 (1 − z ) dz = 2 (z = sin θ, dz = cos θdθ)
cos θ
0π 0
=2 = π.
2
Hence
1 √
Γ = π, (A.15)
2
which when used in (A.14) yields (A.13).
348 A Standard Distributions
∞
X −μ 1 2 2
E = E(Z) = √ xe−x /2 dx = 0 (since xe−x /2 is odd ),
σ 2π −∞
2 ∞ ∞
X −μ 1 2 2 2
E = E(Z 2 ) = √ x2 e−x /2 dx = √ x2 e−x /2 dx
σ 2π −∞ 2π 0
∞
2 1 2
= (2z) 2 e−z dz (z = x2 , dz = xdx)
π 0
∞
2 −z 1/2 2 3 2 1 1 2 1 √
= √ e z dz = √ Γ = √ Γ = √ π
π 0 π 2 π 2 2 π 2
= 1.
Hence
E(X − μ) = 0, or E(X) = μ,
2
(A.16)
E X−μ
σ = 1, or, var(X) = σ 2 .
Comparing (A.19), (A.20) the constant ck (= volume of the unit ball in Rk ) may
k/2
also be obtained (ck = kΓ2π(k/2) ). Note also that χ2k is G (2, k/2) and hence Z/2 ∼
2
G (1, k2 ). In particular, if X ∼ N(0, 1), X 2 ∼ χ21 , X2 ∼ G (1, 12 ). This also shows that
the sum of independent chi-square random variables Z1 , Z2 , . . . , Zm with degrees of
freedom k1 , k2 , . . . , km , respectively, is a chi-square random variable with degrees
of freedom k = k1 + · · · + km (Exercise A.2). Also, if Z ∼ χ2k then using (A.10)
[with α = 1, k = 1, β = k/2 or the fact that EX 2 = 1 if X ∼ N(0, 1)],
Z k
EZ = 2E = 2 = k.
2 2
VI. The Student’s t-Distribution
tk with k Degrees of Freedom is defined to be
the distribution of T = X/ Z/k, where X ∼ N(0, 1), Z ∼ χ2k , and X and Z are
independent. The (cumulative) distribution function of T is given by
P (T ≤ t) = P X ≤ t Z/k = EP X ≤ t Z/kZ
∞
t √ 1
= P X≤ √ z Z = z · k/2 z k/2−1 e−z/2 dz
0 k 2 Γ (k/2)
∞ √z/k t
1 1 x2 /2
= k/2 √ e dx z k/2−1 e−z/2 dz
2 Γ (k/2) 0 2π
Differentiating w.r.t. t under the integral sign one gets the p.d.f. of T as
∞
1 1 zt2 k z z
f (t) = k/2 k
√ e− wk z 2 −1 e− 2 dz
2 Γ 2 0 2π k
∞ 2
1 − z2 1+ tk k 1
= √ k+1 √ k
e z 2 − 2 dz
k 2 2 πΓ 2 0
350 A Standard Distributions
− k+1
k+1
2 2 1 + tk
2 2
k+1
= √ k+1 √
Γ
k 2 2 πΓ k2 2
− k+1
Γ k+1 2 t2 2
β 1 1 1
f (x) = · 2 2
= · 2 , −∞ < x < ∞,
π β + (x − α) β
π 1 + x−α
β
which is the p.d.f. of a Student’s t with k = 1 d.f. Note that the first moment of
the Cauchy distribution does not exist (and, therefore, no higher moment exists).
The final example in this section is important in the theory of testing hypotheses
considered in Chap. 5.
VIII. Fisher’s F Distribution Let U and V be independent chi-square random
variables with degrees of freedom r, s, respectively. (That is, U and V are in-
dependent gamma random variables G ( r2 , 2) and G ( 2s , 2).) The F -statistic with
degrees of freedom (r, s) is then defined by
U/r
F = . (A.22)
V /s
Its distribution Fr,s , say, is called the F -distribution after R.A. Fisher. Its distri-
bution function is computed as follows:
r
G(x) := P (F ≤ x) = P U ≤ xV
s
rs xV
1 r
−1 − u
= E
r u 2 e 2 du
Γ 2r 2 2
∞ r
1 1 s v s xv r u
−1 − −1 −
= r
r s
v2 e 2 u 2 e 2 du dv.
Γ 2 2 ν 0 2 2 Γ 2s 0
r
r2 r+s
Γ
=
s
2
r+s . (A.23)
Γ r2 Γ 2s 1 + rs x 2
A Standard Distributions 351
London.
352 A Standard Distributions
The mean of Z is μ and its variance is h2 π 2 /3. The logistic regression in its
simplest form used in the student projects in Chaps. 4 and 14, has a binary response
variable, say Y , with values 0, 1, and a continuous regressor Z. When Z = z,
then Y = 1 with probability FZ (z) = Prob(Z ≤ z) and Y = 0 with probability
1 − FZ (z). One may also apply this in the case of a non-stochastic Z using the
same relations formally. For many applications other than logistic regression see,
e.g., Balakrishnan and Nevzorov (2003).
x − μ, Σ −1 (x − μ) =
By, B −1 B −1 By =
By, B −1 y =
y, BB −1 y (since
k
B = B) =
y, y = y2 = 1 (y (j) )2 . Thus,
k
det B 1 2
f (x)dx = k/2 1/2
e− 2 j=1 yj dy1 . . . dyk ,
Rk (2π) (det Σ) Rk
This proof also shows that if X is distributed as N(μ, Σ), then Y ≡ B −1 (X− μ) is
distributed as N(0, I) where I is the k ×k identity matrix (having 1’s on the diago-
nal, an zero’s off the diagonal). The distribution N(0, I) is called the k-dimensional
standard normal distribution. Notice that Y = (Y (1) , Y (2) , . . . , Y (k) ) is distributed
as N(0, I), if and only if Y (1) , Y (2) , . . . , Y (k) are k independent 1-dimensional stan-
dard normal random variables.
Conversely, given any positive definite matrix Σ (k ×k), and any vector μ ∈ Rk ,
the random vector X = (X (1) , X (2) , . . . , X (k) ) defined by
X = BY + μ (A.25)
E(X) = BEY + μ = B0 + μ = μ,
k
k
(i) (j) (r) (i) (r ) (j)
cov(X , X ) = cov bir Y +μ , bjr Y +μ
r=1 r =1
k
k
(r) (r )
= cov bir Y , bjr Y
r=1 r =1
k
k
= bir bjr cov(Y (r) , Y (r )) = bir bjr
r,r =1 r=1
This justifies the name mean vector for μ and dispersion (i.e., covariance matrix )
for Σ.
β = EY = ν + CEX = ν + Cμ,
Γ = covY = cov(CX) = CΣC . (A.27)
Note that if the d × d matrix CΣC is of rank k0 (≤ min{k, d}), then Y may
be represented as Y = BZ + β, where B is the non-negative definite symmetric
matrix satisfying B 2 = CΣC , and Z is a d-dimensional standard Normal N (O, I).
354 A Standard Distributions
One may prove this directly using (A.25), or using moment generating functions
that we discuss in Appendix B.
Let now X be N(μ, Σ), a k-dimensional Normal distribution with Σ non-
singular. Then, for any given m, 1 < m ≤ k, Y = (Xm , . . . , Xk ) has the Normal
distribution N(μkm , Σm
k
), where
μkm = (μm , . . . , μk ) , k
Σm = ((σij ))m≤i, j≤k . (A.28)
(A.29)
fX (x1 , . . . , xk )
fW |xm ,...,xk (x1 , . . . , xm−1 ) =
fY (xm , . . . , xk )
1
m−1
ij 1
ij
− σ ((xj − μj ))(xi − μi ) + γ (xi −μi )(xj −μj ) (A.30)
i=1 j=m
2
m≤i,j≤k
⎧ ⎫
⎨
⎬
1
= c1 (xm , . . . , xk ) exp − σ ij (xi − i (xm , . . . , xk ))(xj − j (xm , . . . , xk )) ,
⎩ 2 ⎭
1≤i,j≤m−1
1 1j
k
1 =
1 (x2 , . . . , xk ) = μ1 − 11 σ (xj − μj ). (A.33)
σ j=2
Consider all vectors x in the following as column vectors, with x as the transpose
of x, a row vector.
Lemma A.1. Let Γ be a k × k symmetric and positive definite matrix. Then there
exists a k × k nonsingular matrix B such that Γ = BB .
Ex. A.2. Show that the sum of independent chi-square random variables
Z1 , Z2 , . . . , Zm with degrees of freedom k1 , k2 , . . . , km , respectively, is a chi-square
random variable with degrees of freedom k1 + k2 + · · · + km .
Ex. A.3. For the case of the Normal distribution (A.24), with k = 2,showthat
1
σ 11 = σ11 σσ2222−σ2 = σ11 (1−ρ 2) , σ
12
= σ 21 = − σ11 σσ2212−σ2 = − √σ111 σ22 1−ρ ρ
2 , so
12 12
that the conditional
3 distribution of X 1 , given X 2 = x 2 , is Normal with mean
2
1 (x2 ) = μ1 + σσ11 22
ρ(x2 − μ 2 ) and variance σ11 (1 − ρ ).
Appendix B
Moment Generating Functions (M.G.F.)
Theorem B.1. Suppose the m.g.f. ϕ(ξ) (or, ϕ(ξ)) of a random variable (vector) X
is finite in a neighborhood of zero (origin = 0). Then ϕ determines the distribution
of X.
∞
zk
(b) ϕ(z) = 0 μk k! , where μk = EX k .
Proof.
n
n x n−x n
n
zx
ϕ(z) = e p q = (pez )x q n−x
x=0
x x=0
x
= (pez + q)n < ∞, −∞ < z < ∞.
where
∞
∞
p
ϕ1 (z) = EezX1 = ezx pq x = p (qez )x = <∞
x=0 x=0
1 − qez
1 1
if ez < , i.e., z < ln .
q q
1
(ln > 0).
q
Therefore, r
p
ϕ(z) = .
1 − qez
Its m.g.f. is
∞
1 1
ϕ(z) = EezX = ezx f (x)dx = β
e−( α −z)x xβ−1 dx
0 Γ (β)α
1
=∞ if z ≥ .
α
B Moment Generating Functions (M.G.F.) 359
1
1
For z < α, change variables y = α − z x, to get
∞ 4 β
1 −y β−1 1
ϕ(z) = e y dy −z
Γ (β)αβ 0 α
β
α 1 1
= · Γ (β) = , −∞ < z < .
Γ (β)αβ (1 − αz)β (1 − αz)β α
Example B.7. Consider X ∼ N (μ, σ 2 ). By Example B.6 above, for the case μ = 0,
σ 2 = 1, expanding the m.g.f. ϕ of X in a power series,
2 3 4
z 2 /2 z2 1 z2 1 z2 1 z2
ϕ(z) = e =1+ + + + + ··· ,
2 2! 2 3! 2 4! 2
In general, the term-by-term differentiation of the power series for ϕ(z), evaluated
at z = 0, yields
⎡ k ⎤
z2
2k ⎢ d 2 ⎥ (2k)!
EX = ⎣ zk · ⎦ = = (2k − 1)(2k − 3) · · · 3.1. (B.1)
dz k! k!2k
z=0
μk = 0 ∀ odd integers k,
μ2k = E(X − μ)2k = σ 2k EY 2k = (2k − 1)(2k − 3) · · · 3.1.σ 2k , (k = 1, 2, . . . ). (B.2)
Theorem B.4. Suppose that the m.g.f. ϕ(ξ) of a random vector X = (X (1) , . . . ,
X (k) ) is finite in a neighborhood of ξ = 0 (i.e., for |ξ| < δ, for some δ > 0).
(a) Then all moments of X are finite and, using multi-indices ν,
5 (1) (2) (k)
6
μν ≡ EX ν ≡ E (X (1) )ν (X (2) )ν · · · (X (k) )ν
ν (1) ν (2) ν (k)
∂ ∂ ∂
= [D ϕ(ξ)]ξ=0 ≡
ν
··· ϕ(ξ)
∂ξ (1) ∂ξ (2) ∂ξ (k)
ξ=0
for all ν = (ν (1) , . . . , ν (k) ) ∈ (Z+ )k (= the set of all k-tuples of nonnegative inte-
gers).
μν
(b) ϕ(ξ) = ξν |ξ| < δ,
ν
ν!
(1) (2) (k)
where ξ = (ξ (1) )ν (ξ (2) )ν · · · (ξ (k) )ν , ν! = ν (1) !ν (2) ! · · · ν (k) !.
ν
(1)
ϕ(ξ) = (1)
eξ x dQ(x)
∂ξ ∂ξ Rk
∂ (1) (1) (i) (i) (k) (k)
= (1)
eξ x +···+ξ x +···+ξ x dQ(x)
k ∂ξ
R
(1) (1) (i) (i) (k) (k)
= x(1) eξ x +···+ξ x +···+ξ x dQ(x).
Rk
Hence
Dν ϕ(ξ)
(1) (k)
= (x(1) )ν · · · (x(k) )ν dQ(x) = μν = EX(ν) .
ξ=0 Rk
(b)
∞
k r
k 1
ξ X i=1 ξ
(i)
X (i) (i) (i)
ϕ(ξ) = Ee = Ee = E ξ X
r=0
r! i=1
∞
1 r r − ν (1) ν (k)
= E · · · (k) ×
r! ν (1) ν (2) ν
r=0 |ν|=r
ν (1) ν (2) ν (k)
(1) (1) (2) (2) (k) (k)
× ξ X ξ X ··· ξ X ,
where |ν| = ν (1) + · · · + ν (k) and the second sum is over all ν ∈ (Z+ )k such
that |ν| = r.
(k)
r r − ν (1) ν r!
· · · (k) = (1) (2) .
ν (1) ν (2) ν ν !ν ! · · · ν (k) !
k $
k
ξ (i) Y (i) (i)
Y (i)
ψ(ξ) ≡ Eeξ Y = Ee i=1 = Eeξ
i=1
$
k
(i) 2 2
= e(ξ ) /2
= e|ξ| /2
.
i=1
Example C.1. Consider the UMP test ϕ∗ of size α ∈ (0, 1) for H0 : μ ≤ μ0 against
H1 : μ > μ0 based on X = (X1 , . . . , Xn ) where
√ Xi ’s are i.i.d. N(μ, σ02 ), with
σ0 > 0 known and μ ∈ R = Θ : ϕ (x) = 1 if n(x − μ0 )/σ0 > zα and ϕ∗ (x) = 0
2 ∗
√
n(X − μ0 ) zα
γ(u) = Pμ
σ0 > 2
√
zα n(X − μ0 ) zα
= 1 − Pμ − ≤ ≤
2 σ0 2
√ √ √
zα n(μ − μ0 ) n(X − μ) n(μ − μ0 )
= 1 − Pμ − − ≤ ≤ zα/2 −
2 σ0 σ0 σ0
√ √
n(μ − μ0 ) n(μ − μ0 )
= 1 − Φ zα/2 − − Φ −zα/2 − (μ = μ0 ).
σ0 σ0
(C.2)
where kn is the p.d.f., and Kn the cumulative distribution function of the chi-
square distribution with n d.f.
Example C.3. Consider now the test for H0 : μ ≤ μ0 against H1 : μ > μ0 based
on i.i.d. Xi (1 ≤ i ≤ n) which√are N(μ, σ 2 ) with σ 2 unknown. The UMP unbiased
test of size α rejects H0 iff n(X − μ0 )/s > t1−α (n − 1) [the upper α-point of
2
Student’s t with n − 1 d.f.]. Here s2 = (Xi − X)2 /(n − 1); and (n−1)s has a
√ σ2
chi-square distribution with n − 1 d.f. and is independent of n(X − μ)/σ which
is N(0, 1) under Pμ,σ2 . Thus the power of the test is
√
n(X − μ0 )
γ(μ; σ 2 ) = Pμ,σ2 > t1−α (n − 1)
s
% √ &
n(X − μ0 ) s
= Eμ,σ2 Pμ,σ2 > t1−α (n − 1)s
σ σ
% √ √ &
n(X − μ) n(μ − μ0 ) s
= Eμ,σ2 Pμ,σ2 > + t1−α (n − 1)s
σ σ σ
∞ √
n(μ − μ0 ) (t1−α (n − 1)) √
= 1− Φ − + √ u kn−1 (u)du. (C.4)
0 σ n−1
In obtaining (C.4) we used (i) the independence of X and s and (ii) the fact that
2
U ≡ (n − 1) σs is a chi-square random variable with n − 1 d.f. (As before, kν is
the p.d.f. of a chi-square with ν d.f.)
Replacing t1−α (n − 1) by a general argument t > 0, and differentiating (C.4)
w.r.t. t (and changing the sign of the derivative) one arrives at the p.d.f.
√
of the
0)
so-called non-central t-distribution with the noncentrality parameter Δ = n(μ−μ
σ
and d.f. ν = n − 1, as
∞ √
t √ u
fν,Δ (t) ≡ f (t) = ϕ −Δ + √ u · √ kν (u)du (C.5)
0 ν ν
where ϕ is the standard normal density. Simplifying a little, this density is (Recall:
u ν
kν (u) = 2ν/2 Γ1 v e− 2 u 2 −1 )
(2)
⎧ √ ⎫
∞ ⎨ u 1 + t2
+ Δ 2
− 2√
Δt
u⎬
1 ν−1 ν ν
f (t) = √ ν
√ u 2 exp − du
2π 2ν/2 Γ 2 ν 0 ⎩ 2 ⎭
⎧ 2 ⎫
νΔ2
− 2(ν+t ∞ ⎨ ⎬
e 2)
ν−1 1 √ ν +t 2 Δt
= √ √ v
u 2 exp − u −√ du.
ν 2π 2ν/2 Γ 2 0 ⎩ 2 ν ν + t2 ⎭
C Computation of Power of Some Optimal Tests: Non-central t, χ2 and F 365
ν ν/2
cν = √ ν
.
πΓ 2 2(ν−1)/2
σ2
≡ (n−1)s
σ2
> χ21−α (n − 1). Its power is
0 0
(n − 1)s2 2 (n − 1)s2 σ02 2
Pμ,σ2 > χ 1−α (n − 1) ≡ P 2 > χ (n − 1)
σ02 μ,σ
σ2 σ 2 1−α
= σ2 kn−1 (u)du (σ 2 > σ02 ).
0
σ2
χ21−α (n−1)
∞
Δ
f (v) ≡ f (v; n, Δ) = p j; kn+2j (v) (C.8)
j=0
2
where j
Δ −Δ Δ 1
p j; =e 2 (j = 0, 1, 2, . . . ) (C.9)
2 2 j!
is the probability (mass) function of a Poisson random variable with mean (pa-
rameter) Δ2 , and kn+2j is the chi-square p.d.f. with d.f. n + 2j.
Proof. The p.d.f. of Vi := Yi2 is (use a two-to-one change of variables {y, −y} →
v = y 2 , or differentiate (w.r.t. ν) the cumulative distribution function of Vi , namely,
√v 1 − (y−μi )2
√ √ e 2 dy)
− v 2π
366 C Computation of Power of Some Optimal Tests: Non-central t, χ2 and F
1 1
1 1 2 9 √ √ :
fi (v) = √ v − 2 e− 2 v− 2 μi e−μi v + eμi v
2 2π
⎧ ⎫
1 1
1 1 2
⎨ ∞
2(μ 2 j⎬
v)
= √ v − 2 e− 2 v · e− 2 μi i
2 2π ⎩ (2j)! ⎭
j=0
j! j j− 1 v j− 12 − v
∞ j
1 − μ2i μ2i 1
= e 2 ·√ 2 ·2 2 e 2
j=0
j! 2 2π (2j)! 2
∞ 1 v j− 12
μ2i 1 2j− 2 v
= p j; ·√ e− 2
j=0
2 2π (2j − 1)(2j − 3) · · · 3.1 2
∞ v 2j−1
μ2 1 2 v
= p j; i 1
e− 2
j=0
2 2Γ j + 2
2
∞ 2
μ
= p j; i k2j+1 (v).
j=0
2
The p.d.f. of Y12 + Y22 is the convolution of f1 (v) := f (v; 1, μ21 ) and f2 (v) :=
f (v; 1, μ22 ), namely,
∞ ∞
μ21 2
μ2
(f1 ∗ f2 )(v) = p j; p j; (k1+2j ∗ k1+2j ) (v)
j=0
2 2
j =0
∞ ∞ 2 2 j 2 j
e−(μ1 +μ2 )/2 μ21 μ2
=
k2+2(j+j ) (v).
j=0
j!j ! 2 2
j =0
For the convolution of two (central) chi-square p.d.f. s is a chi-square p.d.f. whose
d.f. is the sum of the d.f.’s of its two components. This equals
⎧ ⎫
∞ ⎨
2 2 j ⎬
2 2 1 μ1 μ2
e−(μ1 +μ2 )/2 · k2+2s (v)
⎩
j!j ! 2 2 ⎭
s=0 (j,j ):j+j =s
∞
2 2 (μ21 + μ22 )s
= e−(μ1 +μ2 )/2 k2+2s (v) = f (v; 2, μ21 + μ22 ). (C.10)
s=0
2s s!
For the last step in (C.10), one uses the combinatorial identity:
j j (a+b)s s
{(j,j )∈(Z+ )2 :j+j =s} a b /j!j ! = s! . (Binomial expansion of (a + b) ).
The general result follows by induction.
Remark C.1. The non-central chi-square appears in the power function of F -tests
in linear models.
The final example in this section enables one to compute the power of F tests
in linear models.
Example C.6 (The Non-Central F Distribution Fr,s (Δ)). Recall the F -distribution
Fr,s in A.2, VIII. This is the distribution of the F -statistic
U/r
F = , (C.11)
V /s
C Computation of Power of Some Optimal Tests: Non-central t, χ2 and F 367
where U and V are independent chi-square random variables with degrees of free-
dom r, s, respectively. If we let U be a noncentral chi-square distribution with d.f.
r and noncentrality parameter Δ, then the Proposition above (Proposition C.1)
says that U may be thought of as a chi-square random variable with a random
degree of freedom r + 2γ, γ having the Poisson distribution P(Δ/2) (with mean
Δ/2). Therefore, conditionally given γ = j, the distribution of (U/(r + 2j))/(V /s)
F
r+2j,s , with p.d.f. fr+2j,s (see Sect. A.2). Hence
is the (central) F -distribution
the p.d.f. of (U/r)/(V /s) = r+2jr [(U/(r + 2j))/(V /s)], conditionally given γ = j,
is
r r
fr+2j,s x (j = 0, 1, . . . ). (C.12)
r + 2j r + 2j
Therefore, the density of the non-central F -statistic (C.11) is
∞
j
−Δ/2 Δ 1 r r
fr,s,Δ (x) = e fr+2j,s x , 0 < x < ∞.
j=0
2 j! r + 2j r + 2j
(C.13)
By using the formula for fr+2j,s one gets
∞ j
Δ
r s
1 r 2 +j s 2 Γ r+s +j
r
x 2 +j−1
−Δ/2 r 2
s
·
fr,s,Δ (x) = e r+s . (C.14)
j=0
2 j! Γ 2 + j Γ 2 (s + rx) 2 +j
For the use of the N–P Lemma to an observed random variable X given
by (C.11), one considers the ratio
∞ j
j
fr,s,Δ (x) Δ 1 Γ r+s + j Γ r2 rx
−Δ 2
=e 2 . (C.15)
fr,s (x) j=0
2 j! Γ 2r + j Γ r+s
2
s + rx
kn
Xj,n L
−→ N (0, 1). (D.1)
j=1
sn
kn
Eε2j,n 1[|εj,n |>η] −→ 0 as n → ∞ (D.2)
j=1
kn
L
εj,n −→ N (0, 1). (D.3)
j=1
L
Polya’s Theorem Suppose Fn −→ F , and F (x) is continuous on (−∞, ∞).
Then supx |Fn (x) − F (x)| → 0 as n → ∞.
Proof. We need to show that, under the given hypotheses, given any ε > 0 there
exists an integer n(ε) such that
1 For proof of Lindeberg’s and Liapounov’s CLTs, see Bhattacharya and Waymire (2007),
pp. 99–103.
© Springer-Verlag New York 2016 369
R. Bhattacharya et al., A Course in Mathematical Statistics and Large Sample
Theory, Springer Texts in Statistics, DOI 10.1007/978-1-4939-4032-5
370 D Liapounov’s, Lindeberg’s and Polya’s Theorems
Find Aε > 0 such that F (−Aε ) < ε/4, 1−F (Aε ) < ε/4. Since Fn (−Aε ) → F (−Aε )
and Fn (Aε ) → F (Aε ), there exists a positive integer N1 such that ∀ n > N1 ,
|Fn (−Aε ) − F (−Aε )| < ε/4, |1 − Fn (Aε ) − {1 − F (Aε)}| = |Fn (Aε ) − F (Aε )| < ε/4.
This implies,
ε ε ε ε ε ε
Fn (−Aε ) < + = , 1 − Fn (+Aε ) < + = ∀ n > N1 . (D.4)
4 4 2 4 4 2
Since F is continuous, it follows that it is uniformly continuous on [−Aε , Aε ].
Hence there exists δ > 0 such that |F (x) − F (y)| < ε/2, whenever x, y ∈ [−Aε , Aε ]
and |x − y| < δ. Choose points x0 = −Aε < x1 < x2 < · · · < xk = Aε such that
|xj+1 −xj | < δ. There exists, for each j, an integer Mj such that |Fn (xj )−F (xj )| <
ε/2 for all n > Mj (j = 0, 1, . . . , k). Now let
Solutions of Selected Exercises in Part I
Chapter 1
1 1
N k Ni
v= (xi − m)2 = (xij − m)2
N i=1 N i=1 j=1
1
k Ni
= (xij − mi + mi − m)2
N i=1 j=1
1
k Ni Ni Ni
2 2
= ξ (xij − mi ) + (mi − m) + 2(mi − m) (xij − mi )
N i=1 j=1 j=1 j=1
1
k k k
Ni Ni
= Ni , vi + Ni (mi − m)2 + 0 = vi + (mi − m)2
N i=1 i=1
N i=1
N
k
k
= wi vi + wi (mi − m)2 . (S.2)
i=1 i=1
k
vi 1
k
E(Y − m)2 = wi.2 = wi vi ,
i=1
nwi n i=1
1 1
k k
v 2
E(X − m) = = wi vi + wi (mi − m)2
n m i=1 n i=1
> E(Y − m)2 ,
Thus, ni is proportional to the size of the stratum and to the standard devia-
tion of the stratum.
Chapter 2
(n − 1)s2 n
2
= Yi2 (Yi i.i.d. N (0, 1)).
σ i=2
Then
2 2
(n − 1)s2 (n − 1)2 s4 (n − 1)2
n
E =E = Es4 = E Yi2 . (S.3)
σ2 σ 4 σ4 i=2
Solutions of Selected Exercises in Part I 373
n
= EYi4 + (EYi2 )(EYj2 ) = 3(n − 1) + (n − 1)(n − 2)
i=2 2≤i
=j≤n
= (n − 1)(3 + n − 2) = (n − 1)(n + 1).
σ4 n+1 4
Es4 = 2
(n − 1)(n + 1) = σ ,
(n − 1) n−1
so that
2σ 4
Eθ (s2 − σ 2 ) = Eθ s4 = σ 4 − 2σ 2 Eθ s2 = Eθ s4 − σ 4 = .
n−1
Chapter 3
By Example 3.5, the Bayes estimator for the beta prior Be (α, β) under loss
L(θ, a) = C(θ − a)2 (c > 0 does not depend on θ) is given by
r+α
do (x) = .
n+α+β
Its risk function is
n 2
1 Xj + α
R(θ, d0 ) = cEθ ε −
n+α+β
= c varθ (d0 (X)) + (Eθ d0 (X) − θ)2
2
nθ(1 − θ) nθ + α
=c + −θ
n + α + β2 n+α+β
% &
nθ(1 − θ) α + θ(α + β)2
=c +
(n + α + β)2 (n + α + β)2
c
= [nθ(1 − θ) + (α − θ(α + β))2 ]
(n + α + β)2
374 Solutions of Selected Exercises in Part I
√
n
For α = β = 2 , this simplifies to
c n4 c
R(θ, d0 ) = √ 2 = √ ,
(n + n) 4( n + 1)2
a Bayes rule which is an equalizer rule. Hence it is minimax (Theorem 3.7). Also,
the property P1 (p. 29) holds with Θ ; = (0, 1). Hence d0 is admissible,
. √
n √ X + 2√1 n
d0 (X) = Xj0 + /(n + n) = .
1
2 1 + √1n
n $
n
−nθ
f (x | θ) = e θ 1 xj
/ xj! (x = (x1 , . . . , xn ) ∈ Zn+ = H ).
j=1
θ
(a) Let L(θ, a) = eθ (θ − a)2 . For the gamma prior τ = G (α, β), the Bayes risk of
an estimator d is
∞
1 eθ
r(τ, d) = (θ − d(x))2 e−nθ θ xj
xj! 0 θ
x∈X
∞
1 1 α
−θ/ (n−1)α+1
= (θ2 − 2θd(x) + d2 (x))e · θ xj +β−2 dθ.
x∈X
xj ! 0 Γ (β)αβ
If x = (0, . . . , θ) then the integral diverges (at 0) if β ≤ 1 (β > 0), unless one
sets d((0, . . . , 0)) = 0. For all other x (and for x = (0, . . . , 0) if β > 1), one has
the integral equal to
∞
1 β
· Γ (Σxj + β − 1)α · (θ − d(x))2 g(θ | α , β )dθ,
Γ (β)αβ 0
α n
d0 (x) = xj , x ∈ X = Zn+ , (S.4)
(n − 1)α + 1 1
d0 (x) = x.
Solutions of Selected Exercises in Part I 375
θ
Hence x is admissible under the loss function L(θ, a) = eθ (θ − a)2 . [Note that
P1 holds with Θ; = Θ, since Pθ (x) > 0 ∀ x, and Pθ0 (A) = 0 implies A is empty,
whatever be θ0 ∈ Θ.]
(c) We now show that d0 = X is admissible under squared error loss
; a) = eθ (θ − a)2 ≡ eθ L(θ, a). Suppose d0 = X
L(θ, a)=(θ−a)2 . Let L(θ, θ θ
is not admissible under L. Then there exists an estimator d1 such that
; d1 ) ≤ R(θ,
R(θ, ; d0 ) ∀ θ, ; 0 , d1 ) < R(θ
R(θ ; 0 , d0 ).
;
But this implies d0 is inadmissible under loss L—a contradiction.
(d) We now show that X is minimax under the loss function L(θ, a) = (θ − a)2 /θ.
The risk function of X is
1 1 1
R(θ, X) = Eθ (θ − X)2 /θ = varθ (X) = nθ = ∀ θ ∈ Θ = (0, ∞).
θ θ θ
[Note that the variance of X1 is θ.] Thus X is an equalizer rule. We will now
apply Theorem 3.6 to find a sequence of priors τN such that the Bayes risks
of the corresponding Bayes estimators dN , say, satisfy
1
lim r(τN , dN ) = . (S.5)
N →∞ n
Now for the gamma prior G (α, β), the Bayes estimator is obtained by mini-
mizing the Bayes risk r(τ, d) over the class of all estimators d:
∞
1 (θ − d(x))2 −nθ ni xj θβ−1 e−θ/d
r(τ, d) = <n e θ dθ
1 xj! 0 θ Γ (β)αβ
x∈X
∞
1 1 α
= β
< n (θ − d(x))2 e−θ/ nα+1 θ xj +β−2 dθ
Γ (β)α 1 xj! 0
x∈X
∞
1 β 1
= · Γ (β )α < n (θ − d(x))2 g(θ | α , β )dθ,
Γ (β)αβ 1 xj! 0
x∈X
Let us choose β >1. Then, as in (b), the last integral is minimized (for each
α( n
1 xj +β−1)
x ∈ X ) by α β = nα+1 . Hence the Bayes estimator for the prior G (α, β),
β > 1, is n
α( j=1 xj + β − 1)
d0 (x) = .
nα + 1
We now show that for α = αN , β = βN with αN ↑ ∞ and βN ↓ 1 as N ↑ ∞,
the Bayes estimator dN for the prior τN = G (αN , βN ) has Bayes risks r(τN , dN )
satisfying (S.5). Now for the prior G (α, β), β > 1, the Bayes risk is
376 Solutions of Selected Exercises in Part I
n 2 % &
α( 1 Xj + β − 1) d20 (X)
r(τ, d0 ) = E − O /O = E O − 2d0 (X) +
nα + 1 O
% & % &
α(nθ + β − 1) varO (d0 (X)) + (EO d0 (X))2
= αβ − 2E +E
nα + 1 O
% &
α(nα + 1β − 1)
= αβ − 2
nα + 1
2 2
α nO α (nO + β − 1)2
+E +
nα + 1 O nα + 1 O
2α[(nα + 1)β − 1]
= αβ −
nα + 1
2 2 % &
α α 2 (β − 1)2
+n + n E(O) + 2(β − 1) +
nα + 1 nα + 1 O
2
2α α
= αβ − 2αβ + +n
nα + 1 nα + 1
2 % &
α 2 β−1
+ n αβ + 2(β − 1) + .
nα + 1 α
1
∞ 1 −θ/α β−1 Γ (β−1)αβ−1 1
Note that, for β > 1, E(1/O) = Γ (β)αβ 0 θ
e θ dθ = Γ (β)αβ
= α(β−1) .
1 −2 2
Also use (1 + nα ) = 1 − n−α + θ( α12 ) as α → ∞ to get
2 2 % &
2α 1 1 2(β−1) β−1
r(τ, d0 ) = αβ−2αβ+ +n + 1+ αβ+ +
nα+1 n+ α1 nα n2 n2 α
2 1
= αβ − 2αβ + (1 + o(1)) + n + o(1)
n n2
2β 2(β − 1)
+αβ − + + o(1) (as α → ∞)
n n2
2 1 2β 2(β − 1)
= + − + + o(1) as α → ∞.
n n n n2
1
−→ as α → ∞, β ↓ 1.
n
Chapter 4
4.3
T (T − 1) . . . (T − k + 1)
d∗ (X) = .
[n(n − 1) . . . (n − k + 1)]
(b) (i) For X1 , . . . , Xn i.i.d. uniform on (0, θ), it is shown in Example 4.6 that
M = max{X1 , . . . , Xn } is a complete sufficient statistic. Hence to find the
UMVU estimator g(M ) of sin θ, we may solve for g satisfying the following.
(Note that the pdf of M is (n/θn )tn−1 .1{0<t<θ} .)
n
n n−1 θ
sin θ = Eθ g(M ) = n g(t)t dt, or, sin θ = g(t)tn−1 dt.
θ (0,θ) n (0,θ)
1
n
n
= nμii − Eθ Xij Xi j − (Eθ Xij )(Eθ Xi j )/n
n j=1
j=1
j = j
= nμii − μii − (n − 1)μi μi
= (n − 1)(μii − μi μi ) = (n − 1)E(Xij − μi )(Xi j − μi )
= (n − 1)σii .
[Note: E(Xij − μi )(Xi j − μi ) = EXij Xi j − μi μi − μi μi − μi μi + μi μi =
μii − μi μi .] Hence the sample covariance
1
n
s ii ≡ (Xij − X i. )(Xi j − X i . )
n − 1 j=1
⎡ ⎤
1 ⎣
n
= Xij Xij − nX i. X i . ⎦
n − 1 j=1
n
= mii − X i. X i . = g(T ), say,
n−1
is the UMVU estimator of the population covariance σii . It follows that
4.8 Let X1 , . . . , Xn be i.i.d. N (μ, σ 2 ), θ = (μ, σ 2 ) ∈ R × (0, ∞). Find the UMVU
estimator of μ/σ.
X
Eθ = (Eθ X)(Eθ (1/s)) = μ/Eθ (1/s).
s
Now one may write
% &1 1
1 (n − 1)/σ 2 2 (n − 1) 2 1
= = · 1/2 ,
s (n − 1)s2 /σ 2 σ U
where U has the chisquare distribution with n−1 degrees of freedom, i.e., a gamma
a distribution G (2, n−1
2 ). Hence
1 1
1 (n − 1) 2 1 (n − 1) 2 ∞ − 1 1 n−1
Eθ = Eθ 1 = u 2 n−1 (n−1)/2
e−u/2 u 2 −1 du
s σ U2 σ 0 Γ ( 2 )2
1 ∞
(n − 1) 2 n−2
= n−1 n−1 e−u/2 u 2 −1 du
σΓ ( 2 )2 2 0
Solutions of Selected Exercises in Part I 379
1 n−2
(n − 1) 2 Γ ( n−2
2 )2
2
= (n−1)
σΓ ( n−1
2 )2
2
1 1
(n − 1) 2 Γ ( n−2
2 ) dn (n − 1) 2 Γ ( n−2
2 )
= √ n−1
= , dn = √ n−1
.
σ 2Γ( 2 ) σ 2Γ( 2 )
Hence Eθ dn1 s = 1
σ, and 1
dn (X/s) is the UMVU estimator of μ/σ. This requires
n > 2.]
4.11 Let X1 , . . . , Xn be i.i.d. beta random variables Be(θ, 1), with common p.d.f.
1
EYj = .
θ
n n
Hence j=1 Yj /n ≡ − j=1 ln Xj /n is the UMVU estimator of θ1 .
Chapter 5
j=1 j=1
i.e.,
T1
P(0,π2 ) > η(t2 ) | T2 = t2 = α ∀ π2 ∈ (−∞, 0),
T2
where η(t2 ) = t1 (t2 )/t2 . Now the distribution of U ≡ TT12 under ΠB , i.e., under
π1 = 0 (or, θ1 = θ2 ) is that of a ratio U = V1V+V
1
where V1 = m1 Xi /θ1 is G (1, m),
n 2
V2 = 1 Yj /θ1 is G (1, n), with V1 and V2 independent. This distribution of U does
not depend on π2 (i.e., it is the same for all π = (0, π2 ) ∈ ΠB ). Hence, by Basu’s
Theorem, T1 /T2 is independent of T2 under ΠB . Therefore, the UMPU test is to
reject H0 iff
T1
> η,
T2
where T1 /T2 = V1 /(V1 +V2 ) has the beta distribution function Beta (m, n). (See the
next Exercise.) Hence η is the (1 − α)th quantile of this Beta (m, n) distribution.
5.10 Let U1 and U2 be independent gamma random variables G (θ, m) and G (θ, n).
Prove that Z1 ≡ U1 /(U1 + U2 ) and Z2 ≡ U1 + U2 are independent random vari-
ables with Z1 having the beta distribution Beta(m, n) and Z2 have the gamma
distribution G (θ, m + n).
Solutions of Selected Exercises in Part I 381
5.18 In this two-way layout with one observation per cell, the model is given
by (5.147) and (5.148), but with ηij = 0 ∀ i, j, i.e.,
I
J
X − EX2 = IJ(X .. − μ)2 + J (X i. − X .. − δi )2 + I (X .j − X .. − γj )
i=1 j=1
I
J
+ (Xij − X i. − X .j + X .. )2 , (S.9)
i=1 j=1
382 Solutions of Selected Exercises in Part I
I
J
min X − EX2 = (Xij − X i. − X .j + X .. )2 , (S.10)
EX
i=1 j=1
m̂ij = X .. + (X i. − X .. ) + (X .j − X .. ) = X i. + X .j − X .. (1 ≤ i ≤ I, 1 ≤ j ≤ J).
(S.11)
(a) To test H0 : All the IJ means θij ≡ EXij are equal, that is, δi = 0 ∀ i and
ηj = 0 ∀ j one has [see (S.8)]
The sum over i, j of the product of the two terms on the right vanishes, so
that the minimum of
I
J
X − EX2 = IJ(X .. − μ)2 + (Xij − X .. )2 , (S.13)
i=1 j=1
I
J
min X − EX2 = (Xij − X .. )2 . (S.14)
EX under H0 i=1 j=1
2
J
ˆ m̂2 =
m̂− (X i. −X .. )+(X .j − X .. ) =J (X i. − X .. )2 + I (X .j − X .. )2 .
i,j i=1 j=1
I
J
+ (Xij − X i. − X .j + X .. )2 , (S.18)
i=1 j=1
I
I
J
2 2
min X−EX − J (X i. − X .. ) + (Xij −X .i − X .j +X .. ).
EX under H0 i=1 i=1 j=1
(S.19)
Hence the numerator sum of squares of the F -statistic is the difference be-
tween (S.19) and (S.18), namely,
I
J (X i. − X .. )2 . (S.20)
i=1
One may obtain (S.20) also as m̂ ˆ − m̂2 , where m̂ij ’s are given by (S.11),
ˆ
and m̂ij = X .. + X .j − X .. = X .j , (1 ≤ i ≤ I, 1 ≤ j ≤ J). Hence the UMPU
invariant test for H0 : δi = 0 ∀ i is
J Ii=1 (X i. − X .. )2 /(I − 1)
Reject H0 iff I J
2
i=1 j=1 (Xij − X i. − X .j + X .. ) /(I − 1)(J − 1)
> F1−α (I − 1, (I − 1)(J − 1)). (S.21)
N Behrens-Fisher, 87
Nadaraya-Watson estimator, 273 two-sample, 84
natural parameter, 49 problems, 371
natural parameter space, 47
negative binomial distribution, 61
negative unimodal, 176 R
Neyman structure, 80 random, 4
Neyman-Pearson Lemma, 69 random sample, 6
nominal coverage, 257 simple, 4
non-uniqueness, 24 stratified, 5
nonparametric, 3 rank test, 214
normal Rao’s scores test, 222
asymptotically, 131 Rao–Blackwell, 44
normal correlation model, 161 Theorem, 42
normal distribution, 347 raw moments, 359
normal equations, 146 rejection region, 68
normal scores test, 209 response variable, 121
nuisance parameters, 3 risk function, 11, 12, 68
O S
observation, 12 sample, 4
observation space, 12, 95 sample correlation coefficient, 132
odd function, 77 sample mean, 286
One-Way Layout, 103 sample quantiles, 126
orbit semi-parametric, 3
constant on the, 92 sequential probability ratio test, 337
order statistic, 209 Shapiro and Wilk test, 228
ordinary least squares estimators, 122 sign test, 204
Ornstein-Uhlenbeck process, 60 simple alternative hypothesis, 69
simple null hypothesis, 69
simple random sample, 4
P with replacement, 4
p.d.f., 165 without replacement, 4
p.m.f., 165 size of the test, 68
Parameter identifiability condition, 337 SLLN, 119
parameter space, 12 Slutsky Theorem, 130
parameter vector, 146 solutions, 371
parametric, 3 space
parametric statistical inference, 3 observation, 95
partition function, 325 SPRT, 338
Pearson’s chi-square statistic, 223 squared error
Pitman ARE, 203, 204 mean integrated, 269
Polya’s Theorem, 369 stability conditions, 241
pooled estimate, 86 standard normal distribution, 347
population, 3 statistic, 39
distribution, 3 Anderson-Darling, 227
posterior density, 26, 325 Cramér-von Mises, 227
posterior distribution, 189 statistical decision problem, 6
power, 214 steady state distribution, 234
power of the test, 68 strata, 5
PRDS, 317 stratified random sample, 5
preshape sphere, 308 Strong Law of Large Numbers, 119
Principal Component Analysis, 335 strongly consistent estimator, 119
prior Student t test, 213, 287
distribution, 22 sufficient, 39
improper, 32 minimal, 44
probability density function, 165 superefficient, 195
probability mass function, 19, 165 symmetric convex sets, 176
problem symmetric difference, 44
Index 389
T two-sided alternative, 74
test Two-Way Layout, 104
α-similar, 80 one observation per cell, 112
Fisher-Yates, 97 Type I Error, 14, 68
goodness-of-fit, 222 Type II Error, 14, 68
Kolmogorov-Smirnov, 231
level of significance, 68
Mann-Whitney, 97, 208 U
most powerful of size α, 73 UMA, 89
nonrandomized, 67, 68 confidence region, 91
randomized, 68 UMP, 68, 73
size of the, 68 invariant test, 102
UMP unbiased, 91 unbiased test, 91
uniformly least powerful, 74 UMPU, 75
uniformly most powerful, 73 UMVU, 45
uniformly most powerful of size α, 73 unbiasedness, 73, 75
Wilcoxon, 97, 208 of test, 204
Test of Absence of Variety Main Effect, 104 uniform distribution, 46
Test of Equality, 104 uniformly most accurate, 89
theorem uniformly most powerful, 68
asymptotic joint distribution of regression
tests, 73
coefficients, 136
uniformly most powerful unbiased, 75
Basu’s, 84
univariate normal, 181
Central Limit, 128
unobservable errors, 97
Gauss-Markov, 98
Glivenko-Cantelli, 267
Polya’s, 369
Rao–Blackwell, 42 W
time series, 295 Wilcoxon
transformation, 92 rank test, 7
Treatment group, 88 test, 97, 214
Tukey Model, 152
two-sample
multi-dimensional problem, 230 Y
two-sample problem, 84, 94 Yule-Walker method, 244