Professional Documents
Culture Documents
MSU-STT-465: Summer-20B
since each of the three sets has one 1’s and two 0’s. This property is
called exchangeability. Loosely speaking, the subscripts of the
observations convey no information about the outcomes.
Definition 0.1
Exchangeability. Let Y1 , . . . , Yn be n random variables. If the joint density
p (y1 , . . . , yn ) = p (yπ1 , . . . , yπn ) for all permutations π = (π1 , . . . , πn ) of
{1, . . . , n}, then Y1 , . . . , Yn are exchangeable
Definition 0.2
Two random variables X and Y are said to have the identical
d
distribution, denoted by X = Y , if Fx (x ) = Fy (x ) , ∀x ∈ R.
d d
Note: Let X and Y be random vectors. If X = Y , then f (X ) = f (Y ), where
f is a real-valued (deterministic) function.
Assume the population has 1300 units and also we know the proportion θ
of units having a particular property. We observe the property on 8 units
and suppose it is reasonable to to assume
Pr (Y8 = 1 | θ) = θ
Pr (Y8 = 1 | Y1 = y1 , . . . , Y7 = y7 , θ) = θ
Pr (Y4 = 1 | Y1 = y1 , . . . , Y3 = y3 , Y5 = y5 , . . . , Y8 = y8 , θ) = θ
and for similar conditional distributions. Since the population size is quite
large the compared to the sample size n = 8, the sampling without
replacement (WOR) is approximately same as sampling with replacement
(WR). That is, we may consider Yi ’s to be conditionally independent.
Proof.
Let (π1 , . . . , πn ) be a permutation of (1, . . . , n). Then
Z
p (y1 , . . . , yn ) = p (y1 , . . . , yn | θ)p (θ) d θ
n
Z Y
p (yi | θ) p (θ) d θ (i.i.d.)
=
i =1
Z Y n
p (yπi | θ) p (θ) d θ (order of product doesn’t matter)
=
i =1
= p (yπ1 , . . . , yπn ).
Remark 0.1
Frequentists assume the Bernoulli variables X1 , X2 , . . . , Xn are
independent outcomes of the same experiment (e.g. a coin flip), that is,
independence.
Theorem 0.1
de Finetti’s theorem. Let Y1 , . . . , Yn be a finite subset of an infinitely
exchangeable, but not necessarily i.i.d., rvs so that
We next give de Finetti’s theorem for Bernoulli rvs. We need the following.
(−1)r ∆r cn ≥ 0, r ≥ 0,
Theorem 0.2
A sequence {ck } of reals with c0 = 1 corresponds to the moments of a
distribution function F on [0,1] if and only if it is completely monotone.
−∆ck = ck − ck +1 = E(X k (1 − X )) ≥ 0.
Inductively, it follows
Theorem 0.3
de Finetti’s theorem. Let {Xn }n≥1 be an exchangeable sequence of
Bernoulli random variables and Sn = ni=1 Xi . Then there exists a
P
probability distribution F on [0, 1] such that
P (X1 = 1, . . . , Xk = 1, Xk +1 = 0, . . . , Xn = 0)
Z 1
= θk (1 − θ)n−k dF (θ);
0
!Z 1
n
P (Sn = k ) = θk (1 − θ)n−k dF (θ)
k 0
pk ,n = P (X1 = 1, . . . , Xk = 1, Xk +1 = 0, . . . , Xn = 0)
For n = 1, 2, . . .,
Thus, in general
and