Professional Documents
Culture Documents
-dimensional determin-
istic sequence q with a d
+ d
G,N
(P) = sup
CC
d
1
N
N
i=1
1
C
(p
i
)
G
(C)
,
where 1
C
is the characteristic function of the set C. Furthermore, the G-discrepancy of
P is given by
D
G,N
(P) = sup
RR
d
1
N
N
i=1
1
R
(p
i
)
G
(R)
.
If p is an innite sequence in [0, 1]
d
, then D
G,N
(p) and D
G,N
(p) should be understood
as the discrepancies of its rst N points. Let D
G
(N, d) be the smallest possible G-star
discrepancy of any N-point set in [0, 1]
d
, and dene the inverse of the G-star discrepancy
N
G
(, d) by
N
G
(, d) = minN N[ D
G
(N, d) .
Analogously, we dene D
G
(N, d) and N
G
(, d). Observe that we always have D
G,N
(P)
D
G,N
(P) for all N-point sets P [0, 1]
d
, and consequently D
G
(N, d) D
G
(N, d) and
N
G
(, d) N
G
(, d).
In the case where
G
is the ordinary d-dimensional Lebesgue measure
d
restricted
to [0, 1]
d
, we simply omit any reference to the corresponding distribution function. The
associated discrepancies D
N
(P) and D
N
(P) are the well-known star discrepancy and the
extreme discrepancy (sometimes also called unanchored discrepancy or simply discrep-
ancy). (Notice that our denominations G-star and G-discrepancy dier a bit from the
denominations, e.g., in [8, 11]. We chose these names, because they are consistent with
the classical names star discrepancy and discrepancy.)
G-star and G-discrepancy have, e.g., applications in quasi-Monte Carlo importance
sampling, see [1, 8]. Here we are especially interested in the behavior of the G-star
and G-discrepancy with respect to the dimension d. The following Theorem is a direct
consequence of [4, Theorem 4].
Theorem 1.1. There exists a universal constant C > 0 such that for every d and each
probability measure
G
on [0, 1]
d
we have for all N N
D
G
(N, d) C d
1/2
N
1/2
and D
G
(N, d)
2 C d
1/2
N
1/2
. (1)
This implies
N
G
(, d) C
2
d
2
| and N
G
(, d) 2 C
2
d
2
|. (2)
2
The dependence of the inverses of the G-star and the G-discrepancy on d in (2) can
in general not be improved. This was shown by a lower bound on the inverse of the star
discrepance in [4], which is, of course, also a lower bound of the extreme discrepancy. An
improved lower bound was presented by A. Hinrichs in [5]. To state it in a quite general
form, let us quickly recall the notion of covering numbers.
For a class / of measurable subsets of [0, 1]
d
we dene the pseudo-metric d
G
= d
G
by
d
G
(K, K
) =
G
(KK
) for all K, K
/; here KK
) (K
K) of K and K
/[ d
G
(K, K
) that cover /.
Theorem 4 from [5] gives us in particular:
Theorem 1.2. Let
G
be an arbitrary probability measure on [0, 1]
d
. Assume that there
exists a constant 1 such that
N(G; (
d
, ) ()
d
for all (0, 1]. (3)
Then there exist constants c,
0
> 0 such that
D
G
(N, d) min
0
, cd/N for all N N (4)
and
N
G
(, d) cd/ for all (0,
0
). (5)
We want to stress that the constants c,
0
do not depend on the dimension d. Indeed,
following the proof of [5, Theorem 4] one observes that (4) and (5) hold for the choice
c =
0
= 1/4e.
2 A Result on Covering Numbers
The question that arises here is for which measures
G
do we have a lower bound on the
covering number of the form (3)?
If
G
is a discrete measure of, e.g., the form
G
(A) = (1/M)
M
i=1
1
A
(p
i
) for P =
p
1
, . . . , p
M
[0, 1]
d
, then N(G; (
d
, ) [P C [ C (
d
[ =:
P
2
M
. (Due to Sauers
Lemma, see, e.g., [13, Sect.2.6], one gets actually even the better bound
P
d
j=0
_
M
j
_
.)
Indeed, if one chooses sets C
1
, . . . C
P
(
d
such that [P C
i
[ i = 1, . . . ,
P
[ =
P
, then
the -balls B
C
d
(G; C
i
, ), i = 1, . . . ,
P
, cover (
d
. Hence in this case (3) does obviously
not hold. On the other hand we get in this case D
G
(kM, d) = 0 for all k N, show-
ing that the bound (4) does also not hold. In general, discrete measures do imply less
interesting integration problems, since the integral itself is already a cubature formula.
Thus approximating the integral via a cubature is trivial (at least if we know the discrete
measure explicitly).
So let us conne to another class of distribution functions (, which is interesting for
the mixed sequences we want to study in the next section. A distribution function G
is in ( if we have for i = 1, . . . , d integrable functions g
i
: [0, 1] [0, ) satisfying
3
_
1
0
g
i
(t) dt = 1, and if G
i
(x) =
_
x
0
g
i
(t) dt and G(x) = G
1
(x
1
) . . . G
d
(x
d
) for all x [0, 1]
d
.
In particular, all distribution functions in ( are continuous functions on [0, 1]
d
.
For this restricted class of measures the covering number does not depend on the
particular choice of G:
Proposition 2.1. For all d N and all G ( we have
N(G; (
d
, ) = N((
d
, ) for all (0, 1]. (6)
In particular, we have
N(G; (
d
, ) (8e)
d
for all (0, 1]. (7)
Before starting with the proof, let us dene the generalized inverse functions
G
i
(t) = infs [0, 1] [ G
i
(s) = t and G
i
(t) = sups [0, 1] [ G
i
(s) = t; (8)
these functions are well dened since the mapping G
i
: [0, 1] [0, 1] is surjective. Put
G(x) = (G
1
(x
1
), . . . , G
d
(x
d
)) and G(x) = (G
1
(x
1
), . . . , G
d
(x
d
)) for all x [0, 1]
d
. Let us
consider functions G
i
with G
i
G
i
G
i
. As G
i
and G
i
also G
i
is a right inverse of
G
i
, i.e., we have G
i
(G
i
(t)) = t for all t [0, 1]. (Observe that any right inverse of G
i
has
necessarily to be of the form G
i
.) Notice that if G
i
is injective, then there exists an inverse
G
1
i
and we obviously have G
1
i
= G
i
= G
i
= G
i
. Put
G(x) = (G
1
(x
1
), . . . , G
d
(x
d
)) and
(x) = (G
1
(x
1
), . . . , G
d
(x
d
)). We have
G(
G
([0, )) = G() = G
1
(
1
) . . . G
d
(
d
) = z
1
. . . z
d
=
d
([0, z)). (9)
Now let x
1
, . . . , x
m
[0, 1]
d
be such that the -balls B
C
d
([0, x
i
), ), i = 1, . . . , m, cover
(
d
. Then also the -balls B
C
d
(G; [0,
G
(x
i
)), ), i = 1, . . . , m, cover (
d
. To see this, let
[0, 1]
d
be arbitrary, and let y =
G(). Then we nd an i 1, . . . , m such that
[0, y) B
C
d
([0, x
i
), ), which is equivalent to
d
([0, y)[0, x
i
)) . Now
d
([0, y)[0, x
i
)) =
d
([0, y) [0, x
i
)) +
d
([0, x
i
) [0, y))
=
d
([0, y)) +
d
([0, x
i
)) 2
d
([0, y x
i
)),
where (y x
i
)
j
= miny
j
, x
i
j
. Similarly,
G
([0, )[0,
G
(x
i
))) =
G
([0, )) +
G
([0,
G
(x
i
))) 2
G
([0,
G
(x
i
))).
Since y =
G(), x
i
=
G(
(x
i
)), and y x
i
=
G(
G
(x
i
)), we get from (9)
G
([0, )[0,
G
(x
i
))) =
d
([0, y)[0, x
i
)) ,
implying [0, ) B
C
d
(G; [0,
G
(x
i
)), ). Thus we have shown that N((
d
, ) N(G; (
d
, ).
Similarly, if we have
1
, . . . ,
k
[0, 1]
d
such that the -balls B
C
d
(G; [0,
i
), ), i = 1, . . . , k,
cover (
d
, then also the -balls B
C
d
([0,
G(
i
)), ), i = 1, . . . , k, cover (
d
. This gives
N((
d
, ) N(G; (
d
, ).
The bound (7) was established by A. Hinrichs for N((
d
, ) in the course of the proof
of [5, Theorem 2]. Identity (6) shows that it is true for general G (.
4
Corollary 2.2. Let d N, and let G (. Then the bounds (4) und (5) hold for
c =
0
= 1/32e
2
.
3 Mixed Sequences
Let d, d
, d
N with d = d
+ d
, G
and [0, 1]
d
(x
)G
(x
)
for all x = (x
, x
) [0, 1]
d
[0, 1]
d
. Furthermore, let q = (q
k
) be a deterministic sequence
in [0, 1)
d
, and let X = (X
k
) be a sequence of independent G
,N
(q) O(log(N)
d
,N
(q)
_
1
1
2
1
4N
(D
G
,N
(q) + 1). (10)
We will show now that this result is unfortunately incorrect, regardless of the distri-
bution function G (: Assume (10) is true. Let (0, 1/2] be given. Choose d
= 1.
There exists a C > 0 such that D
G
,N
(q) C log(N)/N for N 2. Let us consider N
large enough to satisfy C log(N)/N < /2. Recall that D
G
,N
(q) 1. Then (10) implies
for the mixed sequence m
P
_
D
G,N
(m)
_
P
_
D
G,N
(m) ( C log(N)/N) + D
G
,N
(q)
_
1
1
( C log(N)/N)
2
1
2N
1
2
N
2
.
It is easy to see that there exists a constant K independent of d and such that for all
N K
2
the inequalities C log(N)/N < /2 and P(D
G,N
(m) ) > 0 hold. Hence, for
each d N and each N K
2
there exists in particular an N-point set P [0, 1)
d
such
that D
G,N
(P) . Thus N
G
(, d) K
2
| for all d N, contradicting Corollary 2.2,
which says that for < 1/32e
2
we have N
G
(, d) N
G
(, d) d/32e
2
.
5
Since the statement of [11, Theorem 8] is incorrect, the proof has to be incorrect,
too. The source of the incorrectness is [11, Lemma 7] whose statement is ambigious (an
all-quantier J is missing in some place in the if-condition) and which is used for an
inadmissable conclusion in the proof of [11, Theorem 8].
Remark 3.1. In [11, Corollary 6] it was shown for a mixed sequence m that
Var
_
1
N
N
k=1
1
J
(m
k
)
G
(J)
_
1
4N
(D
G
,N
(q) + 1),
where J 1
d
and the variance is taken with respect to the random variables X
1
, X
2
, . . ..
In fact a better result holds, namely
Var
_
1
N
N
k=1
1
J
(m
k
)
G
(J)
_
1
4N
.
To realize this, one has to estimate the sum
N
k=1
1
J
(q
k
)/N appearing in the course of
the proof of [11, Corollary 6] by 1. The same is true for the analogous estimate in [7,
Lemma 1].
3.2 New Probabilistic Bounds
Now we derive correct probabilistic bounds for the G-star and G-discrepancy of a mixed
sequence, which additionally exhibit a better asymptotical behavior in N than the incor-
rect bound (10). First we dene G--covers and prove two helpful results about them.
Denition 3.2. Let (0, 1], and let /
d
(
d
, 1
d
. A nite set /
d
is called a
G--cover of /
d
if for all B /
d
there are A, C such that A B C and
G
(C)
G
(A) . We put ^(G; /
d
, ) := min[[ [ is a G--cover of /
d
.
In the case where /
d
= (
d
, it is convenient to identify the elements [0, x) of (
d
with
their upper right edge points x. Following this convention, we view -covers of (
d
as nite
subsets of [0, 1]
d
.
Proposition 3.3. For all d N and all G ( we have
^(G; (
d
, ) = ^((
d
, ) and ^(G; 1
d
, ) = ^(1
d
, ) for all (0, 1]. (11)
Proof. Let be a -cover of (
d
. Put
G
:= G(x) [ x . Then [
G
[ = [[ (since G
is injective). Let [0, 1)
d
be given and y =
G(). There exist x ,z 0 with
x y z and
d
([0, z))
d
([0, x)) . If y
i
,= 0 for all i, then it is easy to see that
we can choose x and z in such a way that x
i
< y
i
for all i. (Consider for suciently
small vectors y
:= (y
1
, . . . , y
d
). Since is nite, we nd x, z 0 with
d
([0, z))
d
([0, x)) such that for innitely many > 0 we have y
[x, y].
Hence y [x, y] and x
i
< y
i
for all i.) Since
G is non-decreasing with respect to each
6
component, we have G(x) G(z). Furthermore,
G
([0, G(z)))
G
([0, G(x))) =
d
([0, z))
d
([0, x)) . If y
i
= 0 for some i, then
d
([0, x)) = 0. Thus we get
0 G(z) and
G
([0, G(z)))
G
([0, 0)) =
G
([0, G(z))) =
d
([0, z)) . Hence
G
is a G--cover of (
d
. Thus ^(G; (
d
, ) ^((
d
, ).
Let now
G
be an arbitrary G--cover of (
d
. Then it is easily veried that :=
G() [
G
is a -cover of (
d
with [[ [
G
[. Hence ^(G; (
d
, ) ^((
d
, ).
Let
be a -cover of 1
d
. We may assume /
. Put
G
:= [G(x), G(z)) [ [x, z)
.
We have [
G
[ = [
[. Let B := [, ) 1
d
be given, and let
B := [
G(),
G()). Then
there exist
A,
C
with
A
B
C and
d
(
C)
d
(
A) . If
B ,= , we
may assume that
A = [x, x) with
G
i
(
i
) < x
i
x
i
<
G
i
(
i
) for all i. (Consider for
suciently small boxes
B
= [
G()
+
,
G()
). Since
is nite, we nd
A,
C
with
d
(
C)
d
(
A) such that
A
B
C for innitely many > 0. These
A and
C
do the job.) But then [G(x), G(x)) [, ), since
G is non-decreasing with respect to each
component. In any case, we have for
C = [z, z) that [, ) [G(z), G(z)). Furthermore,
G
([G(z), G(z)))
G
([G(x), G(x)))
=
d
i=1
(G
i
(G
i
(z
i
)) G
i
(G
i
(z
i
)))
d
i=1
(G
i
(G
i
(x
i
)) G
i
(G
i
(x
i
)))
=
d
i=1
(z
i
z
i
)
d
i=1
(x
i
x
i
) =
d
(
C)
d
(
A) .
If
B = , then
d
(
A) = 0. If
C = [z, z), we thus have B [G(z), G(z))
and
G
([G(z), G(z)))
G
() . This shows that
G
is a G--cover of 1
d
. Thus
^(G; 1
d
, ) ^(1
d
, ).
Again, the second inequality is easier to show; for any G--cover
G
of 1
d
the set
:= [
G(),
G()) [ [, )
G
is a -cover of 1
d
with [
[ [
G
[.
From [2, Thm. 1.15] we know that ^((
d
, ) (2d)
d
(
1
+ 1)
d
/d! (2e)
d
(
1
+ 1)
d
.
In [2, Lemma 1.17] the inequality ^(1
d
, ) ^((
d
, /2)
2
was proved. Thus we get from
Proposition 3.3
^(G; (
d
, ) (2e)
d
(
1
+ 1)
d
and ^(G; 1
d
, ) (2e)
2d
(2
1
+ 1)
2d
. (12)
One may use -covers to discretize the G-star or G-discrepancy at the cost of a dis-
cretization error at most .
Lemma 3.4. Let P = p
1
, . . . , p
N
[0, 1]
d
. Let be a G--cover of (
d
, then
D
G,N
(P) D
G,N,
(P) + , where D
G,N,
(P) := max
x
G
([0, x))
1
N
N
j=1
1
[0,x)
(p
j
)
.
Let
be a G--cover of 1
d
, then
D
G,N
(P) D
G,N,
e
(P) + , where D
G,N,
e
(P) := max
R
e
G
(R)
1
N
N
j=1
1
R
(p
j
)
.
7
Proof. Let B 1
d
. Then we nd A, C
such that A B C and
G
(C)
G
(A) . Hence we get
G
(A) +
1
N
N
j=1
1
A
(p
j
)
G
(B)
1
N
N
j=1
1
B
(p
j
)
G
(C)
1
N
N
j=1
1
C
(p
j
).
From this the statement for the G-discrepancy follows. Similarly, one shows the statement
for the G-star discrepancy.
Theorem 3.5. Let q = (q
k
) be a deterministic sequence in [0, 1)
d
, X = (X
k
) be a sequence
of independent and G-distributed random vectors in [0, 1)
d
, and let m = (m
k
) = (q
k
, X
k
)
be the resulting d-dimensional mixed sequence. Then we have for all (0, 1]
P
_
D
G,N
(m) D
,N
(q) <
_
> 1 2 ^(G; (
d
, /2) exp
_
2
N
2
_
(13)
and
P
_
D
G,N
(m) D
G
,N
(q) <
_
> 1 2 ^(G; 1
d
, /2) exp
_
2
N
2
_
. (14)
Let [0, 1). We have with probability strictly larger than
D
G,N
(m) < D
,N
(q) +
_
2
N
_
d ln() + ln
_
2
1
__
1/2
, (15)
where = (N, d) := 6e(max1 , N/(2 ln(6e)d))
1/2
, and
D
G,N
(m) < D
G
,N
(q) +
_
2
N
_
2d ln( ) + ln
_
2
1
__
1/2
, (16)
where = (N, d) := 5e(max1 , N/(ln(5e)d))
1/2
.
Remark 3.6. In front of the exponential function in the bounds (13) and (14) there
necessarily has to appear a function depending exponentially on the dimension d: If we
have a bound of the form P
_
D
G,N
(m)D
,N
(q)
_
1f(q; d, ) exp
_
2
N
2
_
for all
in some interval (0,
], all d > d
1
d
, R
1
d
such that R = R
.
Put
k
=
k
(R) :=
G
(R) 1
R
(m
k
) for k = 1, 2, . . .. The
k
, k = 1, 2, . . ., are independent
random variables with E(
k
) =
G
(R
)(
G
(R
) 1
R
(q
k
)). Thus we have
E
_
1
N
N
k=1
k
_
=
G
(R
G
(R
)
1
N
N
k=1
1
R
(q
k
)
D
G
,N
(q).
8
Let := /2. Then Hoedings inequality (see, e.g., [10, p.191]) gives us
P
_
1
N
N
k=1
D
G
,N
(q) +
_
P
_
1
N
N
k=1
(
k
E(
k
))
_
2 exp(2
2
N). (17)
Now let
be a minimal G--cover of 1
d
. Then Lemma 3.4 and (17) gives us
P
_
D
G,N
(m) D
G
,N
(q) <
_
P
_
D
G,N,
e
(m) D
G
,N
(q) <
_
=1 P
_
D
G,N,
e
(m) D
G
,N
(q) +
_
1
R
e
P
_
1
N
N
k=1
k
(R)
D
G
,N
(q) +
_
>1 2[
[ exp
_
2
2
N
_
.
(In the last estimate we obtained >, since we have necessarily I = [0, 1)
d
and
k
(I) = 0 for all k.) This proves (14).
Now 1 2[
[ exp(2
2
N) i
1
2N
_
ln [
[ + ln
_
2
1
__
. (18)
Due to (12) we have [
[ (2e)
2d
(2
1
+1)
2d
. Therefore it is easily veried that (18) holds
if we choose to be
=
_
1
2N
_
2d ln( ) + ln
_
2
1
__
1/2
.
This proves that (16) holds with probability > .
The verication of (13) and (15) can be done with the same proof sheme and was done
in [3, Theorem 3.3] for the classical star discrepancy.
3.3 An Alternative Approach
One could prove a version of Theorem 3.5 without proving the results on G--covers in
Proposition 3.3 and Lemma 3.4, and proving instead the next theorem on the G-star
discrepancy and a similar one for the G-discrepancy:
Theorem 3.7. Let d N, and let G (. Let P = p
1
, . . . , p
N
[0, 1)
d
. Then
D
N
(
G(P)) D
G,N
(P). (19)
Under the additional assumption
G(P) [0, 1)
d
we get
D
N
(
G(P)) = D
G,N
(P). (20)
In particular, for all P = p
1
, . . . , p
N
[0, 1)
d
and a right inverse
G
of
G as dened
above we get
D
N
(P) = D
G,N
(
(P)). (21)
9
Remark 3.8. Without the condition
G(P) [0, 1)
d
the bound D
G,N
(P) D
N
(
G(P))
does not necessarily hold as the following simple example reveals: Let d = 2, G
1
(x) =
min1, 2x, and G
2
(x) = x. Furthermore, let (0, 1/2) and P = p
1
, p
2
with p
1
=
(1/4, ), p
2
= (1/2, ). Then
G(P) = (1/2, ), (1, ) and D
N
(
G,N
(P).
Using Theorem 3.7 and [3, Theorem 3.3] one can conclude (13) and (15) under the
additional assumption that
G
(q
k
) [0, 1)
d
d
([0, x)) =
G
([0, G(x))). Let p [0, 1)
d
. Then
G(p) [0, x) = [0,
G(G(x))) i p
[0, G(x)), since
G is monotone increasing with respect to every component. This gives
1
N
N
i=1
1
[0,x)
(
G(p
i
))
d
([0, x))
1
N
N
i=1
1
[0,G(x))
(p
i
)
G
([0, G(x)))
.
This shows that D
N
(
G(P)) D
G,N
(P).
Let us now additionally assume that
G(P) [0, 1)
d
and prove D
G,N
(P) D
N
(
G(P)):
Let [0, 1]
d
be given and z =
G(). Then
G
([0, )) =
d
([0, z)). Let p [0, 1)
d
with
G(p) [0, 1)
d
. Since
G(p) [0, z) implies p [0, ), we obtain
G
([0, ))
1
N
N
i=1
1
[0,)
(p
i
)
d
([0, z))
1
N
N
i=1
1
[0,z)
(
G(p
i
)).
Since
G is not necessarily injective, we cannot conclude from p [0, ) that
G(p) [0, z)
holds. So let us dene z
i
() = minz
i
+ , 1 for > 0 and i = 1, . . . , d. For all
i 1, . . . , d we get the following: If z
i
= 1, then G
i
(p
i
) [0, z
i
) = [0, z
i
()). If z
i
< 1,
then we get from p
i
<
i
that G
i
(p
i
) G
i
(
i
) = z
i
< z
i
(). Thus p [0, ) always implies
i=1
1
[0,)
(p
i
)
G
([0, ))
1
N
N
i=1
1
[0,z())
(
G(p
i
))
d
([0, z))
=
1
N
N
i=1
1
[0,z())
(
G(p
i
))
d
([0, z())) + ().
10
Since we have lim
0
() = 0, we obtain
1
N
N
i=1
1
[0,)
(p
i
)
G
([0, )) sup
>0
1
N
N
i=1
1
[0,z())
(
G(p
i
))
d
([0, z()))
.
These arguments establish D
G,N
(P) D
N
(
N
(P) = D
N
(
G(
(P))) = D
G,N
(
(P)).
References
[1] Chelson, P.: Quasi-random techniques for Monte Carlo methods. Ph.D. Dissertation, The Claremont
Graduate School, Claremont, California, 1976.
[2] Gnewuch, M.: Bracketing numbers for d-dimensional boxes and applications to geometric discrep-
ancy. J. Complexity 24 (2008), 154172.
[3] Gnewuch, M.: On probabilistic results for the discrepancy of a hybrid-Monte Carlo sequence.
Preprint 2008 (submitted).
[4] Heinrich, S., Novak, E., Wasilkowski, G. W., Wozniakowski, H.: The inverse of the star-discrepancy
depends linearly on the dimension. Acta Arith. 96 (2001), 279302.
[5] Hinrichs, A.: Covering numbers, Vapnik-