Professional Documents
Culture Documents
Mathematics
Edited by A. Dold and B. Eckmann
879
IIIIIII IIIIIIII I IIIIII ill
Vja(~eslav V. Sazonov
Normal Approximation -
Some Recent Advances
I I I IIIIII I
Springer-Verlag
Berlin Heidelberg New York 1981
Author
VjaEeslav V. Sazonov
Steklov Institute of Mathematics
42 Vavilov Street, 117966 Moscow, USSR
This work is subject to copyright. All rights are reserved, whether the whole or
part of the material is concerned, specifically those of translation, reprinting,
re-use of illustrations, broadcasting, reproduction by photocopying machine or
similar means, and storage in data banks. Under § 54 of the German Copyright
Law where copies are made for other than private use, a fee is payable to
"Verwertungsgesellschaft Wort", Munich.
© by Springer-Verlag Berlin Heidelberg 1981
Printed in Germany
Printing and binding: Beltz Offsetdruck, Hemsbach/Bergstr.
2141/3140-543210
Preface
the normal probability measure with the same first and second mo-
(*)
for the absolute value of (~). This topic in the finite dimensional
case was also covered in the recent monograph L8] . The difference
between our approach and that of [8~ is that we use mainly a method
the pseudomoments are small enough - less then some absolute con-
have the usual order ~-'~. Special attention is paid to the con-
estimate.
all the known achievements in the area. Our main aim ~s to present
the main directions and methods (with the emphasis on the method of
convolutions).
The Notes were written at the time the author was lecturing
red (at UCLA Chapter I only). The content of Chapter II, ~ 2 was also
and for his suggestion to write these Notes. While I was at UCLA
a part of the first draft and to ~r. V.P.Sharma of ISI, Delhi and
the manuscript.
V.V.Sazonov
Table of Contents
Comments on Chapter I . . . . . . . . . . . . . . . . . 67
REFERENCES . . . . . . . . . . . . . . . . . . . . . . . . .102. . .
INDEX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
The Main Notations
~Elall/~xal ak
1 "'" ~Xk
xcE or x#gE
Ec complement of a set E
closure of a set E
~E boundary of a set E
E~ = U{S (x):x~ E}
c
E- ~ = U{x: S (x)cE}
c
NT,T> 0 is N0,T-2 I
N = N1
c (resp. c(.)) with or without indices denote positive constants (resp. positive
If the area of an integration or the set over Which max, inf, etc. is taken is
x x
not indicated it is understood that the integration is over the whole space and
Let XI,X2, ... be i.i.d, random variables with values in R k such that EIX112 < co
and suppose that the covarianee matrix V of X is nondegenerate. Then by the central
Sf(x)Pn(dX) + S f(x)N(dx) as n ÷ co
for any bounded Borel measurable function f on R k such that the set of points of its
discontinuity has N-measure zero. We are interested in the estimation of the speed of
this convergence. In this Chapter we shall construct bounds for Sf(x)(Pn-N)(dx) for
any bounded Borel measurable function on R k, which depend in general on n, the function
particular case when f is the indicator of a set E, i.e. f(x) = XE(X). The weak
convergence Pn => N itself imply the uniformity of the convergence (Pn-N)(E) ÷ 0 over
large classes E of sets E, e.g. over the class C of all convex Borel sets~ and some
of our bounds for I(Pn-N)(E)I for convex sets will be independent of particular sets.
The two main methods and the main directions of extensions and improvements of
classical results.
We start with an illustration of the two main methods employed in the estimation
of the speed of convergence in the central limit theorem in R k, i.e. the method of
compositions and the method of characteristic functions. To this aim we shall give
two different proofs of the classical Berry-Esseen theorem (one dimensional case) using
n
Fn(X) = P(c -1 n -½ ~ ( X r - p) < x )
1
and l e t ¢ be t h e s t a n d a r d normal d i s t r i b u t i o n function on R. Then for all n = 1,2,...
suplFn(X) - ~(x)l I ~ B3 °-3 n-½ (i)
X
lemma.
Preof. Take a T > 0 and put t = vT -I. Clearly either ~ : - inf D(x) or ~ = sup D(x).
X
We assume the former case. The latter case is treated similarly. For any e > 0
there exist x such that D(x ) < - 6 + ¢ . Note that if IYl < t, then
C
<-6 + e + 2Mt .
Therefore
< (-~+e+ZMt)X + 6 ( i - ~ ) ,
X ~ u (see ( 2 ) ) , we have
In this case M = (2~) 2 and we may take e.g. u = 3/4. Then v = 1.16 and we have for
these G and H
M > O. Then for any u:2 -I < u < 1 there exist a positive number v(u) such that for
all T > 0
T
Sup[F(x)-G(x)l ~ ( 2 u - l ) - l [ (2~)-1 5 I ( ~ ( t ) - G ( t ) ) / t [ dt+2uv(u)NT-z] (5)
x -T
interval [-T,T].
to deal with the case p = 0,o = i) (To reduce the general case to this one it is enough
to consider Yj = (Xj-p)/o).
~,,2~3/2
Noting t h a t ~3 ~ (LAI) = i , we have
IFl(X)-{(x)[ 2 6} ,
thus (1) is true for n : i. We shall show that if (1) is true for all values of n
less then some fixed value, then it is also true for this fixed value of n with the
n-1
(i0)
= (i:1~ Ui + n {To)*H1
where
U.=H.*{ , H. = r i - {~½ ' %i = ((n-i-1)/n+T-2)-½
i i ~i i (n)
Ui(x-y) = ~ Hi(x-y-u){Ti(u)du
= S Hi(x-z){Ti(Y-z)dz
where {T is the density function of {T' we see that i t is a smooth function of y and
we can write
2
Ui(x-y) = ~ (-1)J(jz)-lu~J)(x)yJ-6-1ul',(x+ey)y 3, (ll)
j=O i
yJ dHl(Y) = O, j = 0,1,2,
since F(n ) and ~n ½ both have mean value 0 and variance l/n. Therefore
IUi*Hl(X)l = I S Ui(x-y)dHl(Y)l
= 6 -I ~U"'
- i (x+ey)y3dH1 (y) I
2 6 -1 sup [U~'(x)lSlyl3(dF(n)(y) + d{ ½(y))
X n
The definitions of F(n ) and ~T imply that
= IFi((n/i)½z) - ~((n/i)½z)l
I cB3 i ~ (12)
Using the above inequalities and noting that
n-2 3 i-½ n-1 dx
i=l (x_l)~((n_x_l)/n+T-2) 3/2
: 2n½T ((n-2)/n) ½
((n-2)/n) + T-2
< 2v~'n ½ T, (13)
we obtain
n-2
tUi*Hl(X)l ~ c oB~ Tn-1 (14)
i=1
For i = n-1 by the same argument but using representation
For the term l~To Hl(X)I we have (using again the above arguments)
* Hl(X)l = I ~{~ (x-y)dHI(Y)I
0 e
1
=~ I~' (x+Oy)Y 3 dHI(Y) J
O
3 supj{, ,
sup
y
I~"'To (Y) I = %
y
'(~oY) l
and
To = ( ( n - 1 ) n -1 + T-2) -½ <_ v~
Assuming now t h a t c i s chosen large enough to s a t i s f y c(c -1 + 2c -½) < 1 we obtain (1).D
For the second proof of the Berry-Esseen theorem we will also need the following
lemma.
Lemma 2. Let X be a random variable with distribution function F and suppose that
then
^ e-t2/21 -~ -t2/4
I rn(tn-½) - 2 (7/6) I i t i 3n z 83 e
and, consequently, If(tn-½) l ~ 24/25 > O. For such t we can thus write
Fn(tn -½) = exp(n log F(tn-2)}.
where ^ ^ ^ ^ ^ ^
6B3(25/24) 3
7B 3
Remark 1. The speed n -½ is "the true" speed of convergence in the central limit
theorem . Indeed let {Yj} be a sequence of i.i.d, real random variables with P(YI=O)=
n
P n ( m n ) ~ 7 2 n ½ -*i as n ÷ oo (20)
1 = 2 1
On t h e o t h e r hand, d e n o t i n g Xi = Yi - 2 ' we have EXi = O,o 2 EX_ =
> (1/4)(2/~) ½ n -½
The classical results of Berry-Esseen type were extended and improved in several
1. One can construct esLimates with less restrictions on the existence of moments.
Denote
-
P3n = o -3 b o-211x-P~2dF(x )
Ilx-pl~dF(x)+n2
{x-p~<_dn ½ ix-pl>on 2
=o -3 I Ixi~dr'(×) +n ½o -2 ; Ix12dr'(×)
l×]~on~ Ixl>an 2
sup t F n ( X ) - { ( x ) j £ c~3 n n
X
Note that if 0 < 6 ~ l, then since Ixi 3 % Ixl 2+6 01-6 n(1-6)/2 when Ixl%an½ and
i.e. -P3n n-½ ~ 62+6 n - 6 / 2 where 62+6 = ElXl-ml 2+6 . Thus ifi EIX112+6 < ~ , we obtain
an estimate with the speed n -6/2. One can show that in general the speed can not be
3. In the Berry-Esseen estimate the absolute moment can be replaced by the so-
s u p I F n ( X ) - { ( x ) I ~ c max(~,v~(n))n ~
X
where ~(n) = m i n ( n / 4 , 1 ) . This estimate i s b e t t e r than Berry-Esseen one when the d i s t -
(recall that 63/g 3 ~ i always). Note that ~(n) =min(n/4,1) is unlmprovab]e here.
~f(R) = sup{If(x)-f(y)l,x,y ~ R}
f (x) = f(x+y).
Y
If f is Berel measurable and ~ is a distribution function on R, denote
~f(~,G) = 5wf(x,c)dG(x)
simplicity)
if we take f: f(z)=l For z< x, = 0 for z > x, then the left side in (23) is
bounded Borel measurable function f and at the same Lime it is precise (it implies
the Berry-Esseen estimate). The right side of (23) depends essentially on f and this
is something one should expect since in general we have only weak convergence of F
n
to {.
independent real random variables and let E X = p, E(Xj-pj) 2 = o2.,j=1,2,j .... Denote
k J
s n2 = ~ o2J . If Sn ~ 0 then
n
IP(Snl i (Xi-~i) < ×) - {(x)i -< eSn-3 i=i
~ ElXi_Pil3
Denote EX I = p and let V be the covariance matrix of X 1 with elements vij = E(Xli-Pi).
states that there exists a constant e(k) such that for all n = 1,2, ...
suPlPn(A)-N(A)1 _<e(k)P3n-½
A~C
Remark 2. Since, as it is easy to see, for any two real random variables X,Y
able t o make clear the main ideas behind general estimates. Some general estimates
§ 2. ~he method of compositions. Estimates for convex and arbitrary Borel sets.
the multidimensional central limit theorem using the method of compositions. Estimates
will be given for deviations on convex sets and (under some restrictions) arbitrary
Berel sets. To make the proofs somewhat more simple and transparent we assume the
existence of the third moment (the estimates become trivial when it does not exist).
Theorem i. Let XI~ X2, ... be a sequence of independent random variables with values
in R k with the same distribution P. Suppose that EIX112 < ~ and that the eovariance
n-½ i v-½ (Xi-P), where p=EXt, and let N be the standard normal distribution in Rk.
1
Then
^ I vifv >-I'
~= ~ = flxl31#-N[(d×) (2)
max(v,v nk/(k+3)) if v < 1 ,
~n = /Pn N[(Rk)
- ~ c2vn-½ (3)
Remark i. Note that in the definition (2) of v the power nk/(k+3) is the best possible,
= ~ : Iixl ~ IP-Nl(dx)
= 3 N(S ) + i Ixl ~ N(dx)
E S
= (2k)-n(a(s))n
C
where v I =Slxl ~ IQ-Nl(dx). If moreover Q has mean zero and cevariance matrix I
Proof. To prove (5) suppose first that 0 < v = IQ-NI(R k) < 2. Let R k = R+UR - be the
Hahn decomposition of R k with respect to Q-N, i.e. R +, R- are Borel sets such that
We have
13
= f l × l ~ (N-M)(dx) (7)
= N(R +) + 1-Q(R +)
= l-v/2 .
and
M(S r) = (N-M)(S~).
Thus we have
Ss c Ixi 3 (N-M)(dx)
r
and it follows that
qZg~N_j(Rk) 2N(S )
k/(k+3) < r
Vl -- ( ~ ixi3 N(dx))k/(k+3)
S
r
= r(r,k) .
I t is easy to check that f ( r , k ) as a function of r is nondeereasing and f(+O,k)
ck -3/2, which proves (5) when v / 0,2.
in this ease (Q-N)(R +) = l, and thus Q(R +) = i, N(R +) = 0), and we have
To prove (6) consider first the case n=2. Noting that N(.) =Nn(n~) for all
Now since (Q- N)(E) ~ (I/2) IQ- NI(R k) for any Borel set E, we have
Furthermore if g(x) is a bounded Borel measurable function then for any t> 0
~(x) = f g ( x + y ) N t ( d Y ) : f g ( z ) @ t ( x - z)dz
a a
, x
~(x) = Z ~ %~(o) +r all)~3~-- Oa~(ex) (127
llalI~2 a~ ,L
where a are nonnegative integral vectors and I6i s l, For s> 0 and Borel sets
E denoting H(s ) = Q(sE) - N(sE) and observing that H(s ) has the first and second
xa k
IIall~3 ~ Da~(Y) = ~ fg(Z) Cu=lZ(z u - yu)xu/lxI33~t(z y)dz -
where x u (resp. yu,Zu) are the coordinates of x (resp. y,z) and hence, since @
is symmetric
xa
I llall~=3aT.,
Da~(Y)t -< x2J-
~ sup {g(=)l (Tt6 ]'1~113@t(z)dz + t 4 f l z l l * t ( z ) d z ]
-< c sup lg(z) l t 3 l x l 3
Z
From (I0), (II) and (14) with g(x) = × (x) and s,t=l we now deduce (6) for n=2.
~-B
When n>2 let us observe at first that for any Borel set B
Indeed we have seen that this is true for n = 1 ,2 and for n> 2 using induction
on n we have
I(Q-N)n(B)I = If(Q-N)n-l(B-X)(Q-N)(dx) I
sup I(Q- N)n-I(E)IflQ- Nl(dx)
E
(]/2)[IQ- NI(Rk)] n
(the sup here is taken over all Borel sets E). Now, using the above estimates, we
-< { (Q-N)n(n½B)I
n-l
+ If (Q-N) *N(n½B-x)( Z (n] (Q-N) i-I *Nn-i-I (dx) l
i=l
-< (I/2) [IQ " NI (Rk)]n + supl (Q _ N) * N(E)[ -c(n)
E
_< (l/2)[JQ-N1(Rk)]n+c(n)vl,, • D
In the proof of the next |emma we shall need one fact from the theory of convex
sets in R k. Recall that a convex subset of Rk is called regu|ar if all its boundary
16
pmints and all its support hyperplanes are regular. By definition a support hyperplane
of a convex set is regular if it meets its boundary in only one point and a boundary
point of a convex set is regular if it lies on only one support hyperplane. The result
we shall need is: if A is a bounded convex set with non-empty interior then for any
> 0 there exist two bounded closed regular convex sets RI, R 2 such that
~-neighbourhood, i.e. E e = U{S (x),x E E}, SE(x) being the open bali of radius e
with center at x.
Lemma 2. There exist a constant c(k,s),s ~ O, such that for any c > 0 and any convex
A C Rk~
IXl s N(dx) < c(k,s)c.
(~A) ~
Proof. We shall prove the lemma for the case k = 2. With trivial changes the proof
carries over to the case of any k >2. The set (~A) C may be represented as
It is enough to consider only bounded convex sets since for any r > 0
< c(s)c
(here Xl,X 2 are the coordinates of x in a system of coordinates with the ist axis
17
orthogonal to L),
Thus we can suppose that A is bounded and with interior points. Then for any
> 0 there exist regular bounded convex closed sets RI, R 2 satisfying (15). If the
lemma is true for such regular convex sets, then, since obviously A ~ \ A c R ~ + S \ R I ,
we have
for any ~ > O, and hence %he lemma is true for any convex set.
Now let A be a regular closed bounded convex set. Denote by AI(A 2) the set of
those points of ~A at which the cosine of the angle between the external normal to A
and the positive (negative) direction of the 1st axis is >_ 2 -½ . The sets AS, A 4 are
4
defined similarly with respect to the 2nd axis. Clearly 8A = UA.. Denote now
I l
A!z = U{nxc' x ~ Ai},nx~ being the set of points on the external normal to A at x with
distance from x not greater then s. The sets A!l are obviously bounded and closed and
we have
4
As \ A c U A! (17)
1 i
Indeed let y ~ A ~ \ A and let y' be th6 nearest point in 8A to y. We claim that
contains y'. There exists a point z of A belonging to the same open half plane defined
by ~ as y. The segment joining z and y' will contain then a point z' of A which is
Let now L be a straight line which is parallel to the 1st axis and interesects
set L n AJ~ attains its minimum. Let x ° ~ A 1 be the point such that x °' nE and x"
x o
o
be the point on the external normal to A at x with the distance to x equal to s.
o o
Draw the line ~ through the point x" which is parallel to the support line to A at x .
0 0
g\ ,,
18
It follows from the definition of A~ that ~ intersects L in some point x, whose first
co-ordinate is greater than the first coordinate of x' and whose distance from x' does
O 0
not exceed ~ . Since the set A ~ lies on the same side of the line ~ as the point
I
Xo, the set L N A~ is contained in the segment joining X'o and x (of length _< c /~).
In other words, we have shown that the section A](y 2) of the set A~ by any straight
Remark 2. Later we s h a l l need estimates for c(k,O) and c(k~3) i n Lemma 4. From the
c(k,3) < ck 3/Z (see [32]7 Lemma 1). A considerably more complicated proof (see e.g.
Since { x : f e ( x ) > u} = U{S ( y ) : f ( y ) > u}, the funetions ~ ( x ) and } e ( x ) = - (_~)e are
Borel measurable.
Define measure RT by RT(E) = R(TE) and put t = vT-1, f x ( y ) = f ( x + y ) . For any bounded
6 = sup { I f f x ( Y ) H ( d y ) l } , H = P - Q
X
: ^ vt
6T sup { m a x [ I S f ~ ( y ) H * R T ( d Y ) l , l [ f x ( Y ) H . R T ( d Y ) l ] }
X
v
TT = sup { m a X [ l ~ f 2 t ( y ) - f (y))Q(dy),~(f (y)-f2t(y))Q(dy)]}.
X X X X
x
Then f o r any T > 0
2 (2u-l)-i (6T'PYT).
treated in a similar way. For any c > O there exist x such that [fx
. (y)H(dy) > 6 - S,
E
Denoting f l : fx we have
£
>_. 6 - ~ - YT "
>-~-YT'
^±
since, as it easily follows from the definitions, fl(z+y) = fxc+y(Z) and fl(z+y)
ftx +y(Z). Thus
Z ~ ( 2 R T ( S t ) - I ) - Y T - ~"
Since c is arbitrary small and, by (18), RT(S t) = R(S v) ~ u, this proves the lemma.~
Corollary i. Preserving the notations and assumptions of Lemma 3 for a Borel set B
define
20
supIH(B+x) I
X
sup { m a x [ l ( H , R T ) ( B t + x ) l , [ ( H * R T ) ( g -t +x)]}
X
Proof. Take f(x) = XB(X), the indicator function of the set B. Then the definitions
fx(y) = XB_x(y)
vt
~(Y) : XBt_x (Y), fx(Y) : XB_t_ x (Y)
where c(k) is the same as in Lemma 2. According to Remark 2 we may take c(k) ~ ck ½
On the other hand for v we have the relation N(S v) ~ u = 3/4. Take v = v(k) such that
N(S v) = 3/4. To finish the proof of the Corollary it is enough to show now that
v ( k ) k -½ ÷ 1 as k ÷ ~.
N(S v) = w, t h e n
v ( k ) k -1 + ] as k ÷ ~ (19)
21
Let Zl,Z2,... be i.i.d, normal (0,I) random variables. Applying the central limit
+ {~ ,
X=I
k
Since 0 < P(( Z2)
j ½ < v(k))= w < 1, for any XI, X 2 such that 0 < X 1 < 1 < X2 we
1
thus have
f (x) = f(x+y).
Y
If f is Bore1 measurable and M is a Bore1 measure on Rk, denote
~r(c,M) = S ~f(x,c)M(dx)
(21)
m~(x,M) = sup{gf (E,M),y ~ Rk}
Y
Corollary 3. Preserving the notations and assumptions of Lemma 3 we have
ISf(x)(P-Q)(dx)I % (2u-1)-I[wF(Rk)t(P-Q)*RTI(Rk) + ~(2vT-I,Q)].
Proof. We have
= ~)(2t,Q)
The general case is reduced to this one by considering the sequence of random variables
V-½(Xj-p),j=l,2,... which have mean 0 and covariance matrix I. We shall assume also
that v < ~, or, equivalently, that EIX113 < m (otherwise the theorem is trivial). If
< c~ -~
(note that the condition (22) implies n < 4). This proves the theorem in the ease
(22).
^
Let now n > (k+3)k -I. In this case we have v : v. We may also suppose that
v Z cI (23)
(when (23) is violated we have (3) and (1) Follows from (3)).
To prove (i) we shall show now that if it is true for all values of n smaller
than some fixed value, with constant ~(k) large enough to satisfy (4), then it is also
true for this fixed value of n with the same constant c(k).
n-I
Hi •n-i-l, ,
(P(n)-Nn½)*(i! 0 r (n)*mn½ )*m T
n-1
ni ,i , , n-i-i
E ~ ( r , ,-,~ 1,*,~ 1 * NT + n N ~ I * N T ] * ( P ( n ) - N n ½ )
i:l vn) n-~ n~
n-1
( ~ U i + nUo)*H I (24)
i=l
where
23
i i
U i = HI *NT., i =0,n-l, H i = P(n) - N ½, i = l,n-I
I n
HO . ~0'
. ~i
. . ((n i l)/n +T-2) -½ , i = O,n-I (25)
Borel set B
IH~(B)] = IPIn)(B) - Ni~(B) I
n~
= Ipi((n/i)½B) - N((n/i)½B)l
i
c(k)~i -~ (27)
-
c(k)cv2~ i- ½ n - 3 / 2 , (k+3)k -I <isn-2
I
lu i *HI(A) i s (29)
cvn -3/2, O < i < ( k + 3 ) k -1, i <n -2
T. s ( n / ( n - i - l)) ½ s c . (30)
I
Note now that in the same way as we obtained (14) we can prove that (in the
notations of (14))
= ,(0) ]loii , 9
Like (14) implies (26), this implies
Now
flxllP-Nl(dx) = (fSl + J ' S ~ ] J x t l P - N I ( d x )
_< I P - N I ( R k) + v (33)
and by Lemma l
I P - N I (R k) < ck -3/2 vk/(k+3) (34)
Since by (23)
k/(k+3) ~ ci 3/(k+3)v
and, obviously
C-3/(k+3)
l ~ C ~
1
J(Pn " N)*NT(A) I ~ evn ~ + c(k)cv2Tn -1 (37)
T I 1
T = c(k) ~k~)-ln ~
Taking c(k) large enough to make the expression in square brackets in (39) to be
I (which is the case when c(k) ~ck, c being an appropriate constant) we finish
To prove the second part of the theorem again we shall use induction on n and
n Nn½
Pn - N = P(n) -
n
n-2
= ( ~ V i + Hn_ I + nVo] *H] (4o)
i=]
where
V i = H i * N t . ,, i = O , n - 2 , i
H i = P(n) - N ni½' i = I 'n-1
(41)
H0 = 60, t i = ( n / ( n - i - 1)) ½, i = O , n - 2 .
Also
n-I Nn~1(E) i
IHn_I(E) I = I P ( n ) ( E ) - n2
c2v(n -1)-½
and by Lemma I
26
#IHll(dx) = IP - NI(R k)
ck-3/2~ k/(k+3) " (43)
Thus
cn (45)
we obtain
' k/(k+3),-3/2,~
~n ~ c2~n-~[c(c21 + Cl + Cl K JJ . (46)
Taking cI small enough and c2 large enough we can make the expression in square
Theorem 2. Under the same conditions and with the same notations as in Theorem I
I. There exist c(k) such that for any convex Bore] set A for all n=1,2,...
where
2. There exist two constants c3, c3(k) such that if < c3 then for all
n =1,2~...
3. There exist two constants c4, c4(k) such that if ~ < c4 then for any
I. DenoLe
Lemma 6. For any probabiliLy measures P,Q on R k, any s > 0 and any Bore1 seL B C Rk
DenoLe
= f B, if 0 ~ B
I g c, if 0 ~ B
Then
rs(~)IP-QI(B)
< flxl s [ P - Q l ( d x ) .
< c ~ - s - r k s/2
28
5[×1 s P ( d x ) s k s / 2
Proof. The first assertion of the lemma is verified by a direct calculation. The
of Lemma 4.
(1) Suppose Ixl < t. If (a) 0 4 A 2t or if (b) 0~A and r(A t) - r(A) ~ t,
then
r3(At)Q(A t -x) ~ R(A) - ck5/2T-I (5~)
Thus
r3(At)Q(A t-x) ~ r3(At)(p(A) -N(A2t))
By Remark 2,
i), (b). As in the case (a) we have (53). Now 0 6 A implies r(A) ~ r(A t) and
3 + flxI2p(dx) +3t2flxIP(dx) +t 3
2) We have
Proof of Lemma 4. Fix T ~ 1,take v = v(k) such that N(x:ix I < v(k)) = 7/8 and take
e > 0 such that e < k5/2 T -1. Denote t = vT -1. Since R(A) = - R(A c) and
30
6 = sup IR(A) I we can choose A such that R(A ) > 6-c, which is either a convex set
A~ C
or the complement to a convex s e t .
1. Suppose f i r s t that O# A 2 t . Then
E
6 T Z RT(A ~) = ( + ~ )r3(Atc)Q(At-X)NT(dX)
l×l<t ix Izt
= I 1 + 12.
= ~7 (6-c -0k5/2 T - l ) .
2
12 ~ # (R(At-x)-c(klxl+k21xl + IxlS))NT(dX)
Ixizt
- 6 NT(IX I Z t) - ck 3/2 T -1
Thus
2. Let now 0 ~ A2t'c Since R ( A ) = r3(Ac)Q(A)> 6-c and IQ(A)I _< I , we have
r3(A ) > 6-c. We may also suppose 6-~ > (2t) 3 since otherwise 6 < e + (2t) 3 < ek 5/2T-
c
by t h e c h o i c e o f e and ( 1 9 ) , and (50) f o l l o w s , Thus r(A ) > 2 t . Together with
fhus
RT((Ae~Sr_q)t) + RT(S~_ n) = (r-q-t)3(Q.NT)(A~)-RT(Sr_q_t)+R(Sr_q_ t)
At-x~g -x
£ r-q-t
and
r((As\Sr_n)t _ x) ~ min ( r ( A ~ - x ) , r ( S r _ q _ t - x ) ) .
Thus
I R ( ( A ~ Sr_n)t-x) + R(S~_q-x) I
- c ( k l x t + k½ Ixl 2 + Ixt3)NT(dX)
Thas in all cases we have either (61) or (66), and since s can be taken arbitrary
small, this imply (50). D
32
In the proof of Theorem 2 we shall need yet another auxiliary fact contained
in lemma lO. To prove it we first prove
Lemma 9. If Yl'"'' Yn are i.i.d, real random variables with EYi = O, EY~ = I, then
El n
[ Yi 13 2 4nEIY1 t3 + 2n3/2, (67)
1
Proof. We may suppose ElY113 < ~ and denote f o r b r e v i t y sgn u = f ( u ) . We have
n n n n
Ei ~ Yi i3 = E( i Yi)2( i Yi)f(~ Yi) = nE( ~ Yi)2 YIf( ~ Yi )
n n
Denoting
n n n
E = {IY]i > l ~2 Yii }' F1 : { ~ Yi > O}, F2 = { ~ Yi < O}
Obviously
n
J2 = n( I - # )YI( ~ Yi)2dP •
F1 EFI
Moreover denoting XF1 the i n d i c a t o r f u n c t i o n of F 1 we have
Lemma i0. In the notations and conditions of Theorems i-2 with m = O, V =I, for any
Lemma 9
k
ElYni~ < k½E Z IYn313
j=l
< k½ k
(4n-2EJXijl 3 + 2)
j=l
Let us prove first (48). For this we shall use induction on n. When n = i~(48)
is obvious. For n > 1 we shall use representation (40) and notations (41).
= #ff(x+y+u) ix+y+ui3Hi(dY)~t.(u)(du)
I
The function w.
I
is a smooth f u n c t i o n and we can define
a
G i(x) = w i(x) - ]Iall~2x___a!DaWi (0) .
Since the first and the second moments of the signed measure HI = P(n) -N± are
n2
equal to zero, we have
2 a (75)
'2 ~ i c(lW~(x)l + Z I ~ × DaWi(0) F (d×)
S1 j=O IlaEI J a'!-" )lHll "
gi(z) : ff(y+z)[y+zi3Hi(dy) .
Then
w.(x)
I
= fgi(z)*t (z-x)dz
|
-< 9(xi+(lyI3+[z[3)~ i)
35
where ~.
i
= /Isl31Hi[(ds),uiIHil(Rk).From
= the above formulas, Lemma 7 and the
Clearly
/S IxI31Hl l(dx) ~ vn-3/2
1
# cI×l~IHll(dx) ~ #Ixj3lHl (d×) (78)
S1
-3/2
0~~ ~3
Summing now inequalities (80), taking into account (81), (82), (45) and the elementary
n-2 _½ ±
inequality ~ i -<cn 2, we obtain
l
n-2
[ fff(x)l×13vi *Hi(d×) J ~ c,~n-3/2(c3(k)~n+k3/2~,n+k3/2). (831
i=l
By (79) and (5) we also have
+ (I +v)k3/2(c3(k))-l]. (88)
Taking c3 small enough and c~ large enough, and assuming that ~ s c3,
c3(k ) ~c'k
3 3/2, we can make the expression in square brackets in (88) to be less than
Estimate (49) follows easily from (48) and Theorem I. Indeed by Lemma 6 and
c3(k)~n-½ (89)
Assuming that < c4= min(cl,c3), summing (89) and (90) and noting that v<_~
~> c4 (91)
Lemma 6. For n> I we shall use representation (24) with T~ I and notations (25).
vi(x) = r3(A)Ui(A-x)
= Sgi(z)~ (z-x)d~ .
I
Let
X a
~i(x) = vi(x) -llall~2~F. DaVi(O) . (93)
Note that equations (76) remain true if we replace in them wi' ti' gi by vi' ~i'
igi(y+z) i ~ 9 ( r 3 ( A - y - z) + lyl 3 + I z i 3 ) i H i ( A - y - z ) l
9(~i+(lyI3+Jzl3)~I)
where ~ = sup r3(B) IHi(B) I p'. = sup IHi(B) I . Now in the same way as we obtained
C 'J C
(77) we deduce
a
Illall~ .aT× DaV i(Y)I -< clxlJE~(Xi' + IyI3~i')+TJ-3k3/2~!]
i i " (97)
J
By Lemma 5, l) and the inductive hypothesis for any convex Borel set B
From (94) -(99), Lemma 7 and (78) we deduce (using the fact that ~.->-I if T >-I)
I
Finally, by Lemma 6, (86) and the relation 2 "½~T O ~2 ½ valid when T_>l, n~>2,
we have (recall that UO = N 0)
ck3/2 -½ (103)
Relation (24), (101) - (103) and Lemma 4 imply now that for any T~I
+
,3/2 v n -½.+ ck5/2T-i
CK (104)
5/4_4
T (lo5)
(E(k) +k)½~
and with this value of T it is equal to
On the other hand by Lemma 10, r3(A) I(Pn-N)(A) I is always not greater than
Assuming that c(k) has been chosen such that the expressions in square brackets
in (|07) and (109) are ~ | (this is the case when c(k) ~ 6 k 5/2, where c is an
(IO5).
2, define
V|n =
jx I -<n½1xl31~-Nl(dx) +nZlxj>n
- ~ f '~lxI2tp-NI(dx)
g(O) =O,g(l) = 1, g(x) and x/g(x) are non-decreasing, and for a g~F put
I. There exist c'(k) such that for any convex Borel set A
2. There exist two constants c5, c5(k) such that if Vln< c 5 then for any
g~]7
40
-½
[(1 + Ixl2gn([X[))l Pn - Nl(dx) -< c5(k)[Vlnn +~ ( g ( n ' ½ ) ) ' | ]
9
where gn(~) = g ( ~ n = ) I g ( n = ) .
3. There e x i s t two constants c6, c6(k) such that i f Vln < c 6 then f o r any
BoreI set
IPn(B) - N(B) I -< c6(k)(Vln +V2n)n -~a(1 + r 3 (B)) -1
Here we may take c ' ( k ) ~ c ' k 5/2, c5(k ) ~c'5k3/2 , c6(k ) ~c~k 3/2.
speed of convergence in the central limit theorem has already been illustrated in § l.
Now we shall apply this method to obtain an estimate in R k of the type described in
Theorem I. Let X1,X2, ... be a sequence of independent random variables with values
I(this assumption is made only for simplicity of formulas). Denote P the distribut-
n o
ion of the normalized sum n -½ ! (Xj-p), where p = EXI, and let N be the standard
normal distribution on R k. Then there exist c(k) such that for any real bounded
w~(c,N) = sup(N((8(A+y))C),y6 R k}
m~(c,N) ~ c ( k ) c , ~ > 0 ,
The proof of Theorem 1 will be based on several lemmas which we shall prove
Lemma 1. Let r > 0 and ~ e (0,i). There exist a probability measure K on R k such
that
l) K(SI(O)) >
with d e n s i t y
where
sin x )s
c(s) = (S( T dx)-i '
has the characteristic function D(s) which vanishes outside the interval (-s~s). To
show this one may observe that (x-I sin x) s is the Fourier transform of the s-fold
convolution of the uniform distribution on [-s,s] with itself and then consider the
Denote by q the smallest even integer greater than r and let Z1,...~Z k be i.i.d.
Obviously one can find a positive ~ such that P(IZI < y) > ~. Take K to be the
>~
and
[ l x l qK (dx) = Eiy-lZl q
k,
y - q k q/2-1 E # t Z j l q
= T -q k q/2 c ( q ) ~ x - Z ( s i n x) q+2 dx
< co
42
Finally
~(t) : Eei(t,y -IZ)
= ~k Ee i ( T - l t j , Zj)
j=l
and
Eei(T-itj,Zj)
^ (-ltj
: D(q+z ) )
Qb : {x : (Xl,...,x k) E R k, xj ~ 0 if bj : O~
xj < 0 if bj : i } .
k
We have, denoting Ii×II = ~ ixjl for x ~ Rk
Ifll : f lr(x)l dx
k b dx
: ~ ( S -EC ~ )(i + (~(-1) Jxj)k+l)f(x)
b EGQ b n Qb 1 1+t1×il "+1"
O)
Further, .e may suppose Os~(t) ~ L' for a such that II~II = o or k+l since other-
dx < f dx
i + Ilxti k+i - i + IxlI<+l
< co .
X. if IXji < n½
J
Y, = j =l,n
J
0 if Ixjl > n½
The random variables Y, are often called the truncated random variables. Denote also
J
Z. = Y. - EY. and let W be the covariance matrix of Z 1.
J J J
Lemma 3. Denote the density of normal (O,W) distribution on R k and let ~n-n-½EYI.
if
< (Sk) -i n½ (5)
63 = ixi I~ _
< I IX112 dP
IX1 >n 2
-½
2 n 63
and
44
IEYiii = iE(Yii-Xii)i
= I I Xli dPl
Ixii>~½
-½
2 n 83
Thus
Now f a r any y = ( y l , . . . , y k ) E Rk
k
t((W-I)y,Y ) = t }i (wij-~ij)YiYjt
i,j=l
k )2
< 2B3n-½( ~ iYjl
_< 2kB3H-½1y Iz ,
liw-1-111 = i(w-1)w-ill
Iw-zll llw-ljl
(8/3)k63 n-½
< 113.
Using elementary inequality
we may write
l~(x)-~(x)l ~ (2~)-k/2e-lXl2/211-1WI-½exp{(l/2)(Ixl2-(W-]x,x))}l
< (2=)-k/2e-I×12/2(ll-IwI-~l + lwI)el"l (13)
where
w = (i/2)(ixjL(w-ix,x)
((l-w-l)x,x)/2
For w by (12) we have
Now by (12)
Is-Iwl i = lm-lz-w+111
_< c ( k ) r63n 2
--< n-iB3
lani : n~IEYII
2 k½ B3 n 2
<_~(x)(i/Z)llxl2-1X+~n~Xp{(i/Z)llxl2-jX+~nl21}(i7)
and by (16)
k
i=l
-1 -½
= (2k ~ Ixj + 8 )6sn
Lemma #. We preserve the notations of Lemma 3 and denote g the characteristic functior
of Z 1. If
Proof. First consider the case a = O, Note that if U is a real random variable with
Eiu-Oi ~ ! ~ E i U I 3 (19)
47
U = (t,Z1-Z1). We have by (19) of ~l, (19) and the elementary inequality I+~ ~ e~,~
_~ _b -
ig(t n 2) j2 = E exp{i(tn 2,ZI-Z1)}
where
A = - n -i E(t,ZI )2 + (2/3)n -3/2 El(t,Zl) l~.
= I1 12 •
Since
exp {-A} exp{n-l(Et(t,Z1)13)2/3-(2/3)n-3/ZEl(t,Z1 )is}
= e1/3
we have
12 5 exp{(n-s)/3} •
EIZ1 lq = EIY1-EYII q
2qELY1 (q
(2o)
2qEIX1 lq ,
hence
48
and we o b t a i n
Ii 2 exp{-(Wt,t) + It12/3}.
Note now that by (20) we may write for any integer m,i < m < k
= n 21EZlmeXp{i(tn ~,zi)} j
n-½1EZlm[exp{i(t n- 2, ZI) } - i ] I
< kn (q-2)/2
< kn -1 211bll
Thus we always have For any nonnegative integral vector b with llbll z i
c(k) YO
!n!. igYO(tn-½)Dblg(tn-½))Y] "'" (Dbrg(tn-½))Yrl
where r, YO' Y l ' " " Yr are p o s i t i v e integers, bI ...,br, b I # . . . #b are non-
r ~
r r
y j = n, ~ y . b . = a . (24)
0 l J j
f(t) = 1 + o(ltl m)
I~II#II~mp a +
(it) a m)
log f(t) = _<mXa --~---.~+ O([t I (25)
valid for small t, we can express the cumulants in terms of moments. In order to
co (~i) a J
ll!II>l Xa (i~l)--~a= j~1"=(-l)J+l J~ I[!II_>ll~a - - ] (27)
(cf. (25) and (26)) which may be written when the moments and cumulants of all orders
= I~11 Jl Jr
(28)
Xa c(al,..-,ar;Jl,'..,Jr)Pal ... Par
r=l
m Xr(it)
log f(t) = i + ~ r! + °(itim) @
r=l
Define now polynomials Pr(Z), r=l,2,...,z ( C k by the formal relation
P (z) - 1
o
r i__ ~, Xjl+2(z) X"
P (z) = ... Jv +L
r v=l v: Jl ..... Jv (Jl+2)~ (Jv+2)~
Pr(Z) vary between r+2 and 3r and the coefficients of Pr(Z) only involve X~S with
lldlJ I r+2.
The polynomials P (z) arise naturally in the following way. Let f be the chara-
r
cteristic function of a random variable X with values in Rk~ with EX = 0 H eovariance
matrix V and EIXI m < ~, m & 3. Then for the characteristic function fn(tn-½) of the
normalized sum n-½ n Xj, where Xj are i.i.d~with the same distribution as X, we
n -½~
log f (tn j = n l o g f ( t n -½)
m-2
= -~(vt,t)
i
+ X ~Xs+2
i n-S/2
+ °(ltn-½1m)
s=l
as tn -½ + O. It f o l l o w s t h a t f o r any f i x e d t e Rk
m-2
Xs+2 ( i t ) n-S/2 o(n-(m-2)/2)}
fn(tn-½) = exp{-{(Vt,t)}exp{ +
s=l
Lemma 5. Let X be a random variable with values in R k such that EX = O,ElXl m <
for some integer m ~ 3, and let Pr be the polynomials defined by (31) with cumulants
of X. For any nonnegative integer r ~ m-2 and any nonnegative integral vector
a = (a I . . . . ,a k) w i t h j l a j j ! 3r we have
(dl+...+dv)~ dl+...+dv-a
(34)
x (dl+" "+dv-a)! t l )]
52
(terms for which a component of the vector dl+...+d v - a is negative are zero).
To estimate the right side of (34) let us look at the formula (28) relating
cumulants and moments. For any nonnegative integral vector d u : (dul,...,duk) we have
_,dul xdukl
IPd! = ILX 1 "" -k
U
lldull
EIXI
= BLIduII
(here XI,...,X k are the components of X). Thus, using notations of (28), by well
JlLIdlLi/lldli JrItdrii/ildlL
± BtJdll ....... ~tldtt
= 6tldL1'
and it follows that
It is enough to prove (36) when 62 = I since the general case is reduced to this one
- i/2
by replacing X by X/B 2 . The function r ÷ B~/(r-2) is nondecreasing if r÷log B~/(r-2)
log ~i/(r-2)
5r = (r-2) -I (log 6r-log B 2)
and the function r ÷ log 6r is convex on [2,m] (the convexity of the latter function
_rJ ~v-r/(m~2)~r/(m-2)
= C(dl) "''UkUv)~2 ~m ; (37)
obviously
g~-r/(m-2) ~ i + g~ (m-3)/(m-2) (38)
1 + itl 3 r - i l a l j (39)
The 1emma follows now from (34), (37) - (39). 0
In what follows we shall need two auxiliary facts which we state as lemmas 6 and 7.
integer.
where the summation Z* is over all collections of nonzero nonnegative integral vectors
such that ~ djd = b. The product ~' is taken ever all d : 0 < d ~ b.
O<d<b
This 1emma is easily proved by induction on b (if it is true for b then it is
also true for b+ej, i < j < k, where ej is the jth coordinate vector).
Denote by C k the k-fold product of the complex plane C with itself and let l'i
k
be the usual norm in C k, i,e. Izl = ( ~ Izji2)½ for z = (Zl,...,z k) 4 c k,
Lemma 7. Let £ be an absolutely convergent power series on Br = {z:Z ( ck, lzl < r},
54
such that
If(z)! ~ h(Izl), z ~ Br ,
for any nonnegative integral vector b, c > 0 and z E Brl, r I = r( l+k½ e) -1-
< R+r I
= r .
[hus BR(Z o) c Br, f(z) is defined on BR(Z o) and we have for z ~ BR(Z o)
where the summation is over all nonnegative integral vectors d. Consider a set
If z ~ E then
S
IZ-ZoL : k ½ s < R,
Using this representation for Es, multiplying (40) by (2#)-k exp{ -i(d',8)} and
(2~) -k f
i91
e-i(d~s) f(zOl+Se,..,,Zok+SeiOk)del...dgk = Cd,slld'll
[_~,~)k (4i)
(here d' is a nonnegative integral vector). On the other hand taking b-th derivatives
Thus
sup {If(z)l} ± sup h([zl)
Z~Eclzo I z~ Eclzo I
h((l+k ½ c)Izol)
and the lemma follows.
Lemma 8. Let X be a random variable with values in R k such that EX : O,EiXI m <
for some m > 3, and suppose that the eovariance matrix V of X is nonsingular.
m
Denote
then
n -½ -½ m-3 n _ r / 2 ~ _
IDa f (V t n ) - [ P r ( i V - 2 t ) e Itl2/Zl
r=O
< E(t,X)2/2n
= iti2/Zn
-21(m-2)
2 - i Pm
56
< 2 -1 k-2/(m-2)
< 2 -I ,
Thus assuming that ~ < I in (44) we are in the domain where log f(tn -½) is well defined.
Denote now
m-3 -r/~
g(t) = - Lti2/2 + ~ n r+2(it)l(r+2):,
r=l
= Da(exp{h(t)}-l).exp(g(t) }
m-3
g(t) -- ~ n-r/2 Xr+2
(it)/(r+2)!
r=O
m-3
---r=O
~ n-r/2 ir 1]~] --Ir+2(Xd/d!)td
m-3
g(t) + [t12/21 ~ ~ (c(k,m))rc(r)It[ 2 (47)
r=l
Choosing c(k,m) smal] enough the right side of (47) to be less than It12/8 we
shall have
57
eJe b,
and ~' below denoting the summation and the product over all e:O < e ~ b
respectively, then
z'je(z-llell)
H' (Deg) Je e(k,m) ttl
< c(k,m)(Itl2-11bll ÷ Itlllbll),
since
where the summation is over all collections of nonzero nonnegative integral vectors
(as was noted above I f ( t n ~)i ~ if (44) is satisfied with ~ < i). Since
< n elXl
£rll/2 II£rlJ/m
< n p
m
we obtain
Iti2/8 (53)
when c(k,m) is small enough. Thus
D£(exp{h(t)}-l) = D£ exp { h ( t ) }
we have
• Z,(m_llell)je
~'(Deh) Je ~ (c(k,m)P m n-(m-2~2) z'je It I
D~(exp{h(t)}-~ ~ c(k,m)pmn-(m-2)/2(ltlm-il~ll+It[m-2+llZl~exp{Itl2/8}
(55)
give now
e(k,m)Pmn-(m-2)/2(itLm-i[all+jtim+llaiJ)exp{-Itl*/4}(56)
where e(k,m) is the same as in (44) and is small enough (of course (59)coincide with
60
3m-2 3m-2
~ym-2 e(y,z) - ~ym-2 el(Y'Z)~
~m-2
By Lemma 6 the derivative ~ el(Y,Z) is a linear combination of terms of the form
I ss
s e2(Y,z) l =
m-3
I ~
r!
-(T-i--~z
(I ~] z~ )
~-T X Y
r-s
~y r:max(1,s) t~ t=r+2
m-3 r+2 r-s
< c(k,m) [ z Or+ 2 lYl
-- r=max ( l, s)
m-3 Ir+2 r/(m-2) y l r - s
< c(k,m) ~: tz O~, t
-- r=max(1,s)
m-3 1/(m-2) )r-s
c(k,m) ~ (tY s/m-21zt s+2
r=max(1,s) lzf Pm Pm
61
lez(y,z)1% zl2/8
as
I---~ e2(Y,Z) ~ c(k,m) O~/(m-2)
~ Izl s+2 (1 < s < m-2).
ay
Thus
(61)
In Ehe case V=I the iemma follows now from (56) and (62).
If V # I consider the random variable X' = V-½X which has mean zero and covariance
matrix I. Denoting the characteristic function of X' by f' and the polynomials P
r
corresponding to X' by Pr' we have if(t) = f(V -½ t), Pr(it) = Pr(iV-½ t). This
implies the lemma in the general case. E]
since otherwise
62
2 16 kwf(Rk)B3 n-½ .
and
jj'f(x)(Pn-Pn)(dx)l = IE(f(n-½ i Xj)-f(n-½ i Yj) I
Denote now mn = n½ EY1 and let N'(resp. N") be normal (O,W)(resp.(-mn,I)) distribu-
tion. Then
l[fmn(X)(P~-N')(dx)l
+ I~r n(X)(N'-N)(dx) I
+ t~f n(X)i(N-N")(dx) I
= I I + 12 + 13 • (66)
By Lemma 3 we have
12 = i IrOn (x)(~(x)-¢(x)~×l
~fi(Rk)~l~'(x)-¢(x)ldx)
wf(R k) c(k)B 3 n-½ (67)
and
63
where KT(E) = K(TE) for any Borel set E and w~ i s defined by (21), § 2. In what
follows we take
T = n2/(16 cl(k)B3) ,
and denote
3 = I(Pn-N').KTI
O = Slp(x) Idx
A ^ ^
< 2Cl(k) k - | ,
we have
= I f D b e i ( t ' x T - l ) K(dx)l
< c(k).
Thus
64
^ ^
where A
n
= n2/16 63, since
+ f IDaN'ldt
A' n <lti~A n
= Jl + Jz + J3 + J4 ' (71)
Iw 'I = IW-1
llw-lllklz
k-i _~
< ^,
( 4 / 3 ) k/2 []Da(Pn(W- ½ u) - ~ n-r/2 Pr(iW ~ u)e-luJ 2/2) ldu
iu [<_~(k)n½pkl+~k r=O
Thus
J1 2 c(k)B 3 n-½ •
To estimate J2 consider
Lr(t) = Da P r ( i t ) e x p { - ( W t , t ) / 2 )
if ilbi! ! 3r
= o, iF Llbll > 3r. (7~)
Note also t h a t
EIZ112 ! 4EIY112
< 4ElXl t2
4k .
e x p { - ( i / 2 ) ( W t , t ) } = e x p { - i t l 2 / 2 } exp { ( i / 2 ) ( ( l - W ) t , t ) }
! exp { - 3 1 t 1 2 / 8 } . (77)
Furthermore
k
D%(-(I/2)(Wt,t)) = - ~ w.. t For some i if !iZil = i
j:l i J J
66
= 0 if LI~LI > 2
J2 i c(k)63 n-½
J3 ! c(k) I (l+1tlllaLI)exp{-5iti2/24}ltJk(AA)-k dt
AA<Iti<_A n
-k/2
! c(k)Pk+2 n
! c(k)5 3 n-½
34 ! ~ IDa exp{-(Wt,t)/2}ltlk(AA) -k dt
AA<Iti~A n
c(k)63n -½ .
as in Theorem i, ~ 2. Then there exist a constant e(k) such that for any real bounded
~3
~ +frxl3N(d×)
(1 + c c i l k 3 / 2 ) v
Comments on Chapter I
The indicated extensions and improvements I-6 of the Berry-Esseen theorem belong
respectively to M. Katz [17] (more precisely, an equivalent form of it belongs to
M. Katz), S.V. Nagaev [23], V.V. Ulyanov [42] (a weaker version was obtained by
V. Paulauskas F25]), R.N. Bhattacharya [7] (weaker forms were obtained by him earlier~
see [8] for bibliography), C.G. Esseen [14], V.V. Sazonov [32].
Theorems l and 2 of ~ 2 were obtained by V.V. Sazonov and V.V. Ulyanov [37]~
[38]. The estimates of the constants in Theorem l ~ l presented here improve our
previous estimates [38]. The present estimate for c(k) was first obtained by
S. Nagaev lhe used the method of characteristic functions) and then (together with
other estimates) by V. Senator [39]. After H, Bergstrom the method of compositions
was developed by V.V. Sazonov [32, 34-36]~ V. Paulauskas [25-27]~ V.V. Ulyanov
[41-42] and others. A method similar to the m~thod of compositions was used in a
very interesting paper by T.j. Sweeting [40]. However, due to the time limitations
we could not include his work in our lectures. One of his main results states that
in Theorem l of ~ 3 one can replace w~ by ~f. This result also permits to replace
by ~f in Theorem 2 of ~ 3-
In this chapter we shall prove two theorems on the speed of convergence in the
central limit theorem in Hilbert space• The situation here differs substantially
from the finite dimensional case and our knowledge in this area is less advanced.
Let Xl, X2, ... be a sequence of independent random variables with values in
a real separable Hilbert space H and with the same distribution P. Suppose that
under the assumption El×112 < ~ they both exist and coincide and denote V the
(Vx,x) = f(×,z-~)(y,~-~)p(dz)
The central limit theorem in H states that the distributions P of the normalized
n
sums n
(Xr-lJ) converge weakly to the normal distribution G with mean zero
l
and covariance operator V, i.e. the distribution defined by
~exp{i(x,y)}G(dy) = exp{-(l/2)(Vx,x)}
Note that contrary to the finite dimensional case even if E is a rather small
class of "good" subsets of H (e.g. the class of all halfspaces, i.e. sets of the
form {x:(x,y) < u}, y ~ H, u ~ R, or the class of all balls) the convergence
P (E)
n
÷ G(E) E ~ E (1)
implied by the central limit theorem may not be uniform in E ~ E. Indeed let
0 with probability Ps
= (r,s = 1,2 .... )
Xrs ±~ with probability qs
S
where
0 < qs < l , Ps + 2 q s = l , qs ÷ 0 as s ~ ~, ~ ~2 = L 2 < ~ . Put
S= ] s
Xr = ~iXrs e S • Obviously X
r
are i o i.d, EX 1 = O and JXll < L
_ so t h a t EIx~I 2 < ~ .
S
70
Take E to be the class El of all halfspaces. We have for any fixed n~ ~ > 0
Pn(x:(es,X) = O) = P((es,n-½ Zn x .) : o)
r = l r
= P(n-½ n
Xrs : O)
r=l
n
] P(Xrs = O)
r=l
n
= Ps
> 1 -c
G(x:(es,X) = O) = 0
(l/2)Pn(x:(es,X) = O)
I
:~ ( I 1 2 ) ( I - ~)
ft. This follows immediately from (2) if we observe that for any s and u the balls
{ x : I x + (R- U)es] < R} increase with R and their union is the haifspace {x:(es,X) < u}.
However, if we take E to be the class of balls with a fixed center then the con-
vergence (I) is uniform in E~ E and we shall estimate the speed of this convergence.
in a separable Hilbert space H with the same distribution P. Suppose that E[XI2< °°
where
random variable Y = ~ o.Y.e, where {Y.} are i.i.d, normal (O,l) real random
j=l J J J J
variables. In p a r t i c u l a r
Yj = (qjYj - h')2'j Y] +Y2' 3=3~ ~'J' j=]~~'J = [Y- hI2'IY- hl respectively, where h.j = (h,e3).
fj(u)- u
(2~)½oj
exp -
soj
-- +exp f - 2o~ --#JJ
J
"
(4)
Denoting the expression in the square brackets in (4) by dj(u) we may write
l d l(u - v)d2(v)
2~°I°2 f~ (u-v)½v ½ dv
l ~ dw
2~°l~---~f ( 1 - w ) ½ w½
-I -1 (5)
= c~ I o 2 •
a,B ~ 0
(uw) ½+ h. 22-½(u-4h~)½w½+2½(h')½w½±h.
J J
:z2-l(u-4h') ½, j =],2,
72
and we have
u - 4h' I
dj(uw) ~ exp "----T---C~j , J = 1,2 .
J
Thus
gl(U) I ] dl(u(l-w))d2(uw)
dw
2~o]o 2 /o --(1 _w)½ w~
g(u) = f# gl(u-v)g2(v)dv
we have by (5)
g(u)-< coilo22 (7)
On the other hand
= I1 + 12 ,
and by (6)
-< c exp,_4 u _ 4h.)}
'1 ~ L 80 7 ( 5
½ -
v 8ol
for a]] u ~8h' since 2 _ ig2 >- (~_B)2 for ~->B >0.
co
f(u) = 1 if u<O
= 0 if u_>s
for some s>O. Then for any two Borel probability measures QI,Q2 on
XS (h) ~< f ( t x - hl - r)
r
we have
[R(Sr(h)) [ = R(sC(h))
f([x-h[-t) s f(Ix-h[-6)
-< XS (h)
we have
Letting r÷ 0 we see that this inequality is also true for r=O and the lemma
follows. D
f c (u) = l if u ~ 0
= 0 if u > c (12)
If(J)(u)l ~ cg j j = 1,2,3
p(u) = l if u ~ 0
= (l - u4) 4 if O<u<l
= 0 if u ~ I.
gc ( t ) = f c ( { Z l + tzl - s ) ' t ~ R.
Proof. Denote h(t) : Izl + tz I . For t with h(t) = 0 (13) being obvious we may
(z 1 + t z , z )
h'(t) = h(t)
(z~ + t z , z ) 2
(Is)
h"(t) = h ' ~ - h-(t-)3
(z I + t z , z ) I z I 2 (z I + tz,z) 3
h'"(t) = - 3 +3
h(t) 3 h(t) 5
f'(v- s) =
;v f"(s I - s)ds I
E O C
s1
;v I f" (s2 s)ds2ds l
O O
distribution on ff with the same mean and covariance operator as G. For ~ > 0
where B are Borel sets. Then for any h~H, s,~,B,y> 0 and i =0,1,2
76
(Zl,Z)
g~-'(o) = f~(IZT -s)~
(20)
(Zl,Z)2
g~-- (o) = f~'(tz 1 - S) - - +, izi
Iz112 iz113I
Observe that both Py,Gy have mean zero and covariance operator y-2V. Thus for any
zl E~
f(~',z)O(d~) = 0
f(z',z)2Q (dz) = 0
T
fl~j2Q(dz) =o
T
f 9g~x~y
:(J) ,z(O)Qy(dz) = O, j = O , 1 , 2 .
= l f f f G ~ (])- J = O z(J)(o)/j')p(:)(dx)GB(dy
~ ) - x
5
O~(d~)
(2l)
l r
77
where
z~S I
to obtain
5
I r ~ c y - 3 ( l + c -1 + c - 2 ) v ! . (22)
2
I 1
<
- cc
-3fS1 [zl3l%l(dz)ff×s
s , c (h-×-~z) (Y)PZ~(d×)Gs (dy).
(23)
Now by Lemma l
GB(Ss,e(h-x-Bz)) = G(Sl~s,~c(6(h-x-ez))
~< ~ ( V ) ( O l + ~ ( l + l × l + l h l ) B ~ ,
and since
/[×L/l(dx) <-~-l([Ixl2p~(d×))½
-I
=~ (i tr v)½ ,
we have
Thus
Il ~ cc(V)[o| +B(I +~-l(i tr V)½÷lhl]By-3~-2~ (24)
The lemma follows now from (2l) - (24).
where Vl = [IxI3 IQI(dx)" If moreover P has mean zero and covariance operator
IQI(s~) ~ (P+G)(S c)
=_ Q(s C) +2G(S c)
< !QI(s~)+2G(s)
IQI (s~) < ~-3#sC[xl 3[QI (dx)
E
c-3~o
1 "
Moreover by Lemma l
G(S ) ~ COlC(V)~ .
Thus
write
P -G = pn _G n
n (n½) n½
n
= i=1~ I~]Qin½"~Gn-i
½n
n-2
n + ~ c(i , pi ~ Gn-i-l (28)
= Qn ½ i=O 'n) Qn½ (n½~ n½
Indeed (29) is obvious for n = I and for n > 1 using induction on we have
lff(xlQ~½(dxll = l##f(x+y)Qn-ln
½ (dX)Qn½(dY)[
<- (I/2)EIQ[(H)]n-IIQ½T(H)
n
(I/2)EIQI(H)] n •
79
Now i f (I +o I) (I +c(v))v]/3 ~ ] then combining (27) - (30) and (25) and putting
1 - ,, 113 > I the inequality (26) is
= vi/3 we obtain (26). If (I +Ol)(l +c(V))v 1
first that a I = I. For n< 4 the theorem follows from (26). To prove it for n24
i
in = In/2], H i = P(n) - G ni½ ' i = i,n
i : , =
Vi = P(n) * Gt.' i = 0,in , V i Hi*Gt. i in+ l , n - L
I I
Vn_ 1 = Hn_ 1
n _ Gn
Pn - G = P(n) n½
rn-1 i
, * G n-i -7]
L:~0P(n) n½ ~ *Hi
[i~ i n-2
= iP(n) *Gt. + H i *Gt. +Hn_ l + (n- in)Gto ] *H I
i t i=i n + 1 I
80
in)V °
=
C°i'
i I
V
i
+ (n-
]*H l • (32)
For i = in +i ~
n-i putting Gtn_l = 6 O (the distribution concentrated at zero
in H) we have as in (21)
5
Iffc(Ix-al -s)V i *Hl(dx)l ~< Z J (34)
I r
where
Jl = ~ ff; Ig~ '(O)I JHil(dx)Gt.(dy)IHl I(dz)' lOI ~ ]
i
ziES 1
J2 = H I g~(1) iHil(dx)Gt(dy)IH]I(dz)
z~Si
-(J) (t) = gc
gc z(J) (t)X S (h-y-tz) ( x ) , j = 1,2 . . . . . (35)
S~C
and the same argument as in (23) g i v e s
i i
IHil(Ss,~(g)) ~ (P(n) + Gn½)(Ss,c(g))
Hi(B) = ( P i - G)(siB)
81
Gn½
i = Gs. , Gs.(B) = G(s i B)
I I
where s. = (n/i) ½, by (37), Lemma 1 and the inductive hypothesis for all
i
IHi1(Ss,e(g)) ~ 2 sup l ( P i - G ) ( S s ( S l g ) ) I + 2 a ( S s . s , s ~ ( s i g ) )
s>O 1 i
c~(V)((tr V)½+silgllvl/3i-I/6+c~(V)(o1+siJgl)si~
.~ c c ( V ) [ ( t r v)½+Jhl+lyl ] ( ~ v l / 3 n -I/6 +~). (38)
Note also t h a t
[lYlGt.(dy) ~ ((n-i- l)In)½([lyl2G(dy)) ½
i
= ((n-i-1)/n)½(tr V) ½, i ~ n - 1 . (39)
and o b t a i n f o r a l l ~:0<~-<1
5 J ~ cE_2vn_3/2 o
2 r
we may take in (41) to be equal to the left side of (42) and thus obtain
Thus we have proved the Theorem in the case o I = I. If oI# I consider the
sequence X[j = o~Ixj,-, j =1,2 ..... Marking the quantities corresponding to {X[)
J
with prime we have
E(V') = o~E(V)
These formulas imply easily that the general case follows from the case 01 = I.
In this section the speed of convergence in the Central Limit Theorem in Hilbert
space will be improved to n -~. However, this improved estimate is worse than that
of Theorem I, ~l, in some other respects. The difference of the two estimates reflects
EIXII3,, = ~ <
then
In what follows we use the notations of Theorem I, 5J and assume that its con-
g(t,h,V) = fexp{itlx+hI2~G(dx)
= exp~it(Rth,h)~g(t,O,V ) ,
83
where
R t = (I-2itV) -1
Moreover
g(t,h,V) = E e×p{it[Y+hl 2)
ee
= ~ E exp{it(Yj+h.) 2}
j=l J
~ _ f it(x+hj) 2-x2"~
2/zoj.
(2~)-%_ l e dx
j=l J
g(t,h,V) = exp{it(Rth,h)}g(t,O,V).
Finally (2) follows from (3) and the well known formula for the characteristic
Ig(t,h,V) I ~ (] +4Okt2)-k/4 .
with V E F(3,o) then the distribution of IY+h[ 2 has a density which is bounded
IE exp{itlh+X+Yi2~I ~Ig(t,o,v) I .
indeed
IE e x p { i t l h + X + y l 2 } l = IE g ( t , h + X , V ) l
~Ig(t,o,v) I
y = 2 i ( R t h l , h 2) , 6 = 2i(Rth2,h2)
Moreover
IYl ~ 2]hll lh2] , I61 ~ 21h2]2
(5)
Ig(~)I ~ Ig(t,0,v)l
Differentiating this equation we obtain (4). inequalities (5) also follow immediately
Lemma 3. For any ~, 0 < ~ < I, there exists a bounded (two-valued) real random
E~ = O, E~ 2 = ~, E~ 3 = I.
Proof. The indicated properties are enjoyed by a random variable ~ which takes
values
(I *(I +4~3)½)/2~
85
with probabilities
I/2± I/2(I +4~3) ½ . D
Lemma 4. Let X be a random variable with values in H such that EX =0, Elxl 3 <~.
random variable and let ~ =~(e) be a real random variable with properties described
in Lemma 3. Suppose that X,Y,~ are independent. Then X' = ~ X + Y has mean zero,
Moreover
ElX,l m _< c(m,~)ElXl m, m > 2 .
(7)
Proof. Since ¢,X,Y are independent, Eg=O, EX=EY=O in H, E~2=~ and the co-
: 0 )
= (Vhl,h2) ,
i.e. EX' = 0 and the covariance operator of X' is V. Equation (6) is proved in
a similar way: one has only to observe that the symmetry of the d i s t r i b u t i o n of Y
E ( Y , h l ) ( Y , h 2 ) ( Y , h 3) = O.
In order to prove (7) we shall show first that for any positive integer m
have
(9)
(10)
86
If m is even (8) is e a s i l y obtained by comparing (9) and (]0) and observing that
EIYj12~=v c(~), ~ > O. If m is odd we use the moment inequality
we have
E[X'I m ~ 2m-l(El~]mEJXl m + ElY[ m)
= c(m)((]-~)tr V) m/2
= c(m)(]-~)m/2(E[XI2) m/2
c(m)(1- ~)m/2EIXjm.
Together with the obvious inequality EI~I m ~ c(m,~) this proves (7).
Lemma 5. Let Xi,Yi, j = l~...~n be H valued and (j, j = l,...,n be real random
variables. We assume that Xj,Yj,(j, j = l,...,n are independent. Assume also that
Xj, j =l,...~n are identically distributed, IXj! ~B, EXj = O and that the covariance
= ~jXj +Yj and let Pn(resp. Pn ) be the distribution of n-½i Xj (resp. n-½iX'.) .
1 I j
Then for any h ~H
Pn -
P'n = P(n)
n _ (p,(n)) n
n-1 .
= (P(n) " P'(n) ) *jZoP= ~n) *Q(n)n-J-1 *Gn-j-l(n)
87
~exp{itlh+x+zl 2.~G(n
n-j-l(dz)
) = g (t,h+x,Vj,), x ~ H
Hence
6n(t) = fexp{itlh +xI2](Pn- p'n)(dx)
n-l • _n-j-l,
= ~ ffg(t,h+x + y'Vj,~) (m(n/~ -P'(n) )(dx)(P~ n) * QJn)~ * Q(n) ; (dy)
j=0
n-I
= ~ I. .
j=O~ J
Thus expanding gl,j(X) by Taylor's formula up to the term of the fourth order we
obtain • n-j-l)
l'jl ~ (I/4]1## sup I~i,3
(iv)(klJ(P(n) +Pin))(dxl(P}n) *Q(n) (dy). (12)
0~I
Furthermore by Lemma 2
^(iv)
sup Ig I ~ (X) I ~c{t4[(lhl4+lyI4) Ixl4 +IxlSl
O~X~I 'J (13)
+ Itl3rlhl2+lyl2)l×l 2 +[xl4]lxl 2 +t2[xl4]Ig(t,O,Vj, )I.
flx[mm
(n)(dx) = EIn-½XII m
/ I x l m P(n)
, (dx) = EIn-½X~I m
I JyI2P~n) ~Q(n)n.j-l(dy)
' = E In-~ILr=IIX +r r=l+2~rXr] I2
o'filXrI
Lr=l
+r=j+2
i rJXrl ]B2 (15)
88
and similarly
! n )
Ily,mP
• n* Q(n)nil (dy) = Eln-½( l Xr+r +2 rXr14
r
so that n-I
[~n(t)[ <- c((~,B)n-2(l +lhj4) t2(l + t 2) ~ Ig(t,O,Vj,~)J • (17)
j=O
Note now that the distribution of ih+n-½ i XjJ 2 has a bounded density p(~).
1
Applying Esseen inequality [15] and using (17) we may thus write
IT
_< c [I Z ~ n ( t ) . t - l d t + c ( o ) T - 1 (18)
n 0
_< c(c~,o,B)[n-2(l +lhl) 4
;0 t(l
j=O
Jg(t,O,((j + l ) / n ) ( l -~)V)Idt+T-l].
Ig(t,O,u(l-~)VI = Ig(ut,O,(l-~)V)I
-< t -1 l g ( v , O , ( l - e ~ ) V ) I d v = c ( c ~ , c Q t -1.
0
89
If
(1 + I h l ) n -¼ < 1 (21)
the v a l i d i t y of (l]) is o b v i o u s . D
x if Ixl < R
= (22)
o if lxt ~ R .
->0/21 P I
so that
= (1/2)P l + ( ] / 2 ) P 2 (23)
2 2 2 > 2
where a| ~ o 2 ... (resp. alj - a2j ~ ...) are the eigenvalues of the covariance
~2 >- (3/4)a~
Olj , j = ] ,... ,k (24')
^2 ^2
where Olj ; a2j ~ ... are the eigenvalues of the covariance operator of the measure
Moreover
xP(dx) I -.<- BR - 2
x[3P2(dx) ~ 2B .
R1 ~ Cl(k,B) (25)
Proof~ If R > R1, P(SR1) > I/2 we have for any Borel set A
P(A n SR])
(I/2)P(A n SRI)/P(SRI )
= (]/2)PI(A) .
that to prove (24) it is enough to show that for any ~ > 0 we have
Pl = P(SR l )
~1 = fXPl(d×)
- (d×)
= pllfSRlXP
= _ p~l;SR ×~(d×)
we may write
( V l h l , h 2) = f ( h l , x ) ( h 2 ,X)Pl(dX ) - ( h l , ~ l ) ( h 2 , g 2)
Pil#sRl(hl ,x)(h2,x)P(dx)
=
Consequently
+ p[2IfSR~[×IP(dx)]2llhIJlh21 ,
and (26) follows. Relation (24') is proved in a similar way. Finally
R-2fs~I×13p(d×)
6R -2 ,
flxl3P2(d×) ~ 26 .
variable and let ~=~(1/2) be a real random variable described in Lemma 3 and Y
Pl " The random variables X',C,Y are supposed to be independent. Let PI be the
distribution of
flxl3+r#(dx) ~ c(r)6R r .
c(s)(IEX'I s + El~(X'-EX') + YI s)
c(s)E[X'I ~
= o(s)fl×lSp](dx) .
92
flx[3+rp(dx) ~ e(r)flx[3+rp(dx)
< c(r)Rr}IxJr~(dx)
< c(r)t3R r . D
n ] ' n
= n-½ i Xj Pn = P~
n ] ' n
Proof. Denote
P2(A) = P2(A-~2 )
- 1
we have
93
= (2-1(p 1 +P2))n(n½Sr(h))
n
=
nJ ~n
~ (j]2 PI '~' P2 (n~S r (h))
j 0
= ~ °I?
2 nPi '*P2 (n'Sr(h) -j)J l - (n - j)~2)
jo
= 2-nP2(n½Sr(h ) - n~2)
n
+ ~~ r_;_
;n) ~ - n ; / h_) ; / s , ((n/j)½h - j½# - ( n - j ) j - ½ -j-½x)P~-J(dx). (30)
j=l (n/j)~ r 1
Similarly
By Lemma 5 the absolute value of the difference of the integrands in (30) and (31) is
Furthermore by Lemma 6
= (n- j)½(J'txI2P2(dx)) ½
(n-j
.)3i (]1 xl 2P2(dx)) ½
(n-j)½(2B) 1/3
Consequently
94
+ j ~(n- j)½)j-¼
Thus we obtain
n
-
IPn(Sr(b))-Pn(Sr(h))I <c(o,~ ) ~ l{~] 2-n13 -'~+ j-3/4n½1 h !
J
+ j¼R -2 +3"3/41n -2jl +3 -3/4(n- 3)½]
real random variables qi such that P(n i = O) = P(n i = l) = I/2. Note now that by
(I +n/4) -7 + e -3n/32
c(y)n -Y (33)
Moreover
E(I +n)'3/41n -2nl s (E(I +n)-3/2E(n - 2q)2) ½
= n½(E(l +q)-3/2)½
z
= (nl2)2(E(l +n) -3/2)½ (34)
and, obviously
E~ ¼ <_ (E~) ¼
= (n12) ~ (35)
measure and GI is normal (O,(I/2)V l) (see (27) and the definition of PI above)•
• i
Ig(t,O,(l/8)vl)i (37)
Proof of Theorem 2. We may assume that n ~ 2 R 1 =c1(4,6) (see (25)) since otherwise
Let Xj be the truncation of Xj at the level R=n ½ (see (22)) and let Pn
in^
be the distribution of Sn = n-=~Xj " Denoting
n
A. = {w: X. = )(.} A = C~ A.
J J J j=i J
we have
^
IPn(Sr(h))-Pn(Sr(h)) j ~ IP({S n c Sr(h)} hA) +P({Sn ~ Sr(h)} n A) I +P(A c)
n
~j~lP(A~ )
1
= nP(iXll~o~)
_1
_<Bn ~ . (39)
Note that
Elxi[ : IEX1XA1
= IEXI×A~I
Elxllx C
A1
< Bn"1 (4o)
96
E(Xl,f)2
(Vf,f) . (41)
= 2EI×II2×A~
1
2an ~ (42)
variable
y = y - n½~ + y'
Sr(h)ASr(hl ) c S r _ l h _ h l(h)2
;, Ih-hll'
t p y ( S r ( h ) ) - P~(Sr(h)) I = IP~(Sr(h+n~-^ _ y , ) ) _ p ~ ( S r ( h ) ) j
P~(Sr(h+n~ -y,)ASr(h))
~2 ~2 -½
j=3
and 812 _>02^2>-.. . are the eigenvalues of V. Since oj -<o.,j j=l,2,. .. (this follows
Moreover by (40)
];I <- ~n -I
and by (42)
E[¥'I ~< (EI¥'[2) ~
-< (2B)% ¼
Thus
1Py(Sr(h)) - P~f(Sr(h)) I <- c(S,B)(1 +Ihj)n -¼ (43)
NOW (39) and (43) reduce the proof of the theorem to the estimation of
[Pn(Sr(h)) - Pg(Sr(h))l.
Applying Lemma 8 we may write
i n
here Pn is the distribution of n -~ ~ X] , where X. have the same distribution
I J
(27) constructed starting with P=Px. and with R l =ci(4,13) (see (25)) ~ R = n ~.
J
To estimate
Dn
= sup
r_>O
[Pn(Sr(h)) - P~(Sr(h)) [
~
we use the same method as in the proof of Lemma 5. Let P(n) be the distribution
I~ 1
~n
of n ~X l and G(n ) be the normal (n-~p, n-Iv) distribution, so that Pn=P(n),
= ~n
Py (n)" We have
_ p., = ~n ~n
n y P(n) - G(n)
n-I
= (P(n) - G(n) ) "~ Z. o ~" " G(n )
98
and, since
~n-j-I
(n) is normal
'
((n -j -l)n-~, (n-j -l)n-
]~)
we may write
dn(t) = fexp{itjh +xi2}(Pn- P )(dx)
Furthermore by Lemma 2
fl×]S~(n)(d×) = n-S/2flxiS~(d×)
c(s)Bn -3/2. (48)
Denoting Yl a normal (O,V) random variab]e and using (8) we also have
< c(s)n-Sl2(EJYiIS+I~I s)
c(s)n-S/2EIxIIs
Now
99
n-½EIXll 2
1
n ~trV.
Furthermore
El i (xi-~)13 -< (El i ()(i-~)14)3/4
l I
so that
J
El ~ ()(i "~)13 5 c(!3 + (tr V)2)3/4n 3/2
1
Consequently
i
[lh+y+(n-j-1)n =~I3P(J)(dy)
n
< c(S,~)(l+lhl) 3 . (5O)
_< 1 + c(S)nt -1
100
Hence
Idn(t) I ~ c(S,~)(1 + Ihl)3t2(1 + [tl)(1 + n t ' l ) n -3/2 (51)
By Corollary l and Lemma 6 the random variable Ih+YI 2 has a density bounded by
Dn ~ c ( I ~ ! d n ( t ) I t - l d t + c ( S ) T - l )
D n ~ c(S,~)(~ + lhl)3/2n -¼
This proves the Theorem in the case (54). If (54) is not satisfied the Theorem is
obviously true.
Comments on Chapter II
a fixed centre provided by the central limit theorem in Hilbert space was first
considered in a special case by N.P. Kandelaki [16] (see also [44]) then by V.V.
SazOnov [33] (in [33] the logarithmic speed of Kandelaki was improved to n-I/6+e,~>O).
Later a considerable progress was made by J. Kuelbs and T. Kurtz [|8], who obtained
the speed n -I/6+~, e> O, under general moment conditions. The results of [18]
were generalized and made more precise by V. Paulauskas [28]~ V.M. Zolotarev [46],
V.V. Ulyanov [42,43] and others. The proof of Theorem I,§I given in the lectures
101
uses a variant of the method of compositions and is based on the ideas of J. Kueibs
and T. Kurtz with modifications due to V. Paulauskas and V.V. Ulyanov. In the proof
we avoided Frechet derivatives (like J. Kuelbs and T. Kurtz in [18]) to make the
presentation more elementary. Theorem |,~2 belongs to V.V, Yurinskii [45]. Under
bounds with better speeds are also known (see e.g. [24,12]). In 47 F. G~tze obtained
i. Bahr,B. yon, On the central limit theorem in R k, Ark. Mat. 7 (1967), 61-69.
i0. Bikelis, A., On the central limit theorem in R k, Part I, II, Litovsk. Mat.
Sb. 11(1971), 27-58; 12(1972), 53-84.
17. Katz, M., Note on the Berry-Esseen theorem, Ann. Math. Stat. 34(1963),1107-1108.
18. Kuelbs,J. and Kurtz T., Berry-Esseen estimates in Hilbert space and an applica-
tion to the law of the iterated logarithm, Ann. Probab., 2(1974), 387-407.
20. Liapounov,A.N., Sur une proposition de la th6orie des probabilit~s Bull. Acad.
So. St. P~tersbourg (5), 13(1900), 359-386.
23. Nagaev,S.V., Some limit theorems for large deviations, Teor. Verojatnost. i
Primenen, lO (1965), 231-254.
24. Nagaev,S.V. and ~ebotarev V.l., On estimates of the speed of convergence in the
central limit theorem for random vectors with values in ~2' Mathematical
analysis and related topics, Nauka, Novosibirsk (1978), pp. 153-182.
28. - - On the rate of convergence in the central limit theorem in some Banach
spaces, Teor. Verojatnost. i Primenen., 21(1976), 775-791.
29. Rao, Ranga R., On the central limit theorem in Rk, Bull. Amer. Hath. Soc.,
67(1961), 359-361.
32. Sazonov, V.V., On the multi-dimensional central limit theorem, Sankhya Ser. A,
30 (1968), 181-204.
36. , A new general estimate of the rate of converoenee in the central limit
theorem in R k, Proc. Nat. Acad. Sci. USA, 71(1974), Ii8-121.
37. Sazonov, V.V. and Ulyanov, V.V. On the speed o f convergence i n the c e n t r a l l i m i t
theorem, Adv. Appl. Prob. 11(1979), 269-270.
38. - - , On the accuracy of normal approximation, submitted to Multivar. Anal.
39. Senatov, V.V., Several uniform estimates of the rate of convergence in the
multidimensional central limit theorem, Teor. Verojatnost. i Primenen.,
25(1980), 557-770.
104
40. Sweeting, T.J., Speed of convergence for the multidimensional central limit
theorem, Ann. Probab. 5(1977)~ 28-41.
41o Ulyanov, V.V., A non-uniform estimate of the rate of convergence in the central
limit theorem in R k, Teoro Verojatnost, i Primenen., 21(1976), 280-292.
44. Vakhania, N.N. and Kandelaki, N.P., On the estimation of the speed of convergence
in the central limit theorem in Hilbert space, Trans. Comput. Centre
Acad. Sci. Georgian SSR., 10:I(1969), 150-160.
45. Yurinskii, V.V., On the error of the Gaussian approxTmatlon to the probability of
the Gaussian approximation to the probability of hitting a bail, preprlnt.
46. Zolotorev, V.M., ideal metrics in the problem of approximating the distributions
of sums of independent random variables, Teor. Verojatnost. i Primenen.,
22(1977), 449-465.
47, G~tze, F., A~ymptotlc Expansions for Bivariate yon Mises Functional, Z. Wahrschein-
lichkeitstheorie verw. Gebiete 50, (1979), 333-355.
In dex
Berry-Esseen theorem, I
in ~ , I
in ~ , 69
in Hilbert space, 70
nonuniform in ~ , 26
of integral type in R , 67
uniform in ~k , 11
Method of
Polynomials
, 5o
, 5o,51
Pseudomoments, 9, 11
Smoothness lemma
in ~ , 2
in ~& , 18, 27
in Hilbert space, 73