Professional Documents
Culture Documents
1/31/10
1 Preliminaries
2 Great Expectations
7 Limit Theorems
8 Statistics Tidbits
Preliminaries
Will assume that you know about sample spaces, events, and the
definition of probability.
Example: Toss a fair die. Let A = {1, 2, 3} and B = {3, 4, 5, 6}. Then
P (A ∩ B) 1/6
P (A|B) = = = 1/4. 2
P (B) 4/6
Example: Toss two dice. Let A = “Sum is 7” and B = “First die is 4”.
Then
P (A) = 1/6, P (B) = 1/6, and
P (A ∩ B) = P ((4, 3)) = 1/36 = P (A)P (B).
So A and B are independent. 2
Example: Let X be the sum of two dice rolls. Then X((4, 6)) = 10. In
addition,
1/36 if x = 2
2/36 if x = 3
..
P (X = x) = . 2
1/36 if x = 12
0 otherwise
Examples: Here are some well-known discrete RV’s that you may
know: Bernoulli(p), Binomial(n, p), Geometric(p), Negative Binomial,
Poisson(λ), etc.
Outline
1 Preliminaries
2 Great Expectations
7 Limit Theorems
8 Statistics Tidbits
Great Expectations
x 2 3 4
f (x) 0.3 0.6 0.1
Then E[X 3 ] = 3
P
x x f (x) = 8(0.3) + 27(0.6) + 64(0.1) = 25. 2
Outline
1 Preliminaries
2 Great Expectations
7 Limit Theorems
8 Statistics Tidbits
Discrete Example: Let X denote the number of H’s from two coin
tosses. We want the pmf for Y = X 2 − X.
x 0 1 2
f (x) 1/4 1/2 1/4
y = x2 − x 0 0 2
G(y) = P (Y ≤ y)
= P (X 2 ≤ y)
√ √
= P (− y ≤ X ≤ y)
Z √y
= √
|x| dx = y, 0 < y < 1.
− y
P (Y ≤ y) = P (F (X) ≤ y)
= P (X ≤ F −1 (y))
= F (F −1 (y)) = y,
Outline
1 Preliminaries
2 Great Expectations
7 Limit Theorems
8 Statistics Tidbits
Remark: The marginal cdf of X is FX (x) = F (x, ∞). (We use the X
subscript to remind us that it’s just the cdf of X all by itself.) Similarly,
the marginal cdf of Y is FY (y) = F (∞, y).
Example: The following table gives the joint pmf f (x, y), along with the
accompanying marginals.
X=2 X=3 X=4 fY (y)
Y =4 0.3 0.2 0.1 0.6
Y =6 0.1 0.2 0.1 0.4
fX (x) 0.4 0.4 0.2 1
Alexopoulos and Goldsman 1/31/10 20 / 45
Jointly Distributed Random Variables
Theorem: X and Y are indep if you can write their joint pdf as
f (x, y) = a(x)b(y) for some functions a(x) and b(y), and x and y don’t
have funny limits (their domains do not depend on each other).
Examples: If f (x, y) = cxy for 0 ≤ x ≤ 2, 0 ≤ y ≤ 3, then X and Y are
independent.
21 2
If f (x, y) = 4 x y for x2 ≤ y ≤ 1, then X and Y are not independent.
If f (x, y) = c/(x + y) for 1 ≤ x ≤ 2, 1 ≤ y ≤ 3, then X and Y are not
independent. 2
iid
Notation: X1 , . . . , Xn ∼ f (x). (The term “iid” reads independent and
identically distributed.)
iid
P If X1 , . . . , Xn ∼ f (x) and the sample mean
Example:
X̄n ≡ ni=1 Xi /n, then E[X̄n ] = E[Xi ] and Var(X̄n ) = Var(Xi )/n.
Thus, the variance decreases as n increases. 2
Outline
1 Preliminaries
2 Great Expectations
7 Limit Theorems
8 Statistics Tidbits
and
Var(X − Y ) = Var(X) + Var(Y ) − 2Cov(X, Y ).
Theorem: −1 ≤ ρ ≤ 1.
and
E[XY ] − E[X]E[Y ]
ρ = p = −0.415. 2
Var(X)Var(Y )
Outline
1 Preliminaries
2 Great Expectations
7 Limit Theorems
8 Statistics Tidbits
X ∼ Bernoulli(p).
(
p if x = 1
f (x) =
1 − p (= q) if x = 0
iid
Y ∼ Binomial(n,Pp). If X1 , X2 , . . . , Xn ∼ Bern(p) (i.e., Bernoulli(p)
n
trials), then Y = i=1 Xi
∼ Bin(n, p).
!
n y n−y
f (y) = p q , y = 0, 1, . . . , n.
k
f (x) = q x−1 p, x = 1, 2, . . . .
Y ∼ NegBin(r, p) is the sum of r iid Geom(p) RV’s, i.e., the time until
the rth success occurs. For example, “FFFSSFS” implies that
NegBin(3, p) = 7.
!
y − 1 y−r r
f (y) = q p , y = r, r + 1, . . . .
r−1
X ∼ Poisson(λ).
e−λ λx
f (x) = , x = 0, 1, . . . .
x!
E[X] = λ = Var(X).
Alexopoulos and Goldsman 1/31/10 33 / 45
Some Probability Distributions
X ∼ Gamma(α, λ).
λα xα−1 e−λx
f (x) = , x ≥ 0,
Γ(α)
iid
If X1 , X2 , . . . , Xn ∼ Exp(λ), then Y ≡ ni=1 Xi ∼ Gamma(n, λ). The
P
Gamma(n, λ) is also called the Erlangn (λ). It has cdf
n−1
X (λy)j
FY (y) = 1 − e−λy , y ≥ 0.
j!
j=0
a ab
E[X] = and Var(X) = .
a+b (a + b)2 (a + b + 1)
−(x − µ)2
1
f (x) = √ exp , x ∈ R.
2πσ 2 2σ 2
E[X] = µ, Var(X) = σ 2 .
If Z ∼ Nor(0,
p 1), Y ∼ χ2 (k), and Z and Y are independent, then
T = Z/ Y /k has the Student t distribution with k df. Notation:
T ∼ t(k). Note that the t(1) is the Cauchy distribution.
Outline
1 Preliminaries
2 Great Expectations
7 Limit Theorems
8 Statistics Tidbits
Limit Theorems
This is a special case of the Law of Large Numbers, which says that
X̄n approximates µ well as n becomes large.
iid
Central Limit Theorem: If X1 , X2 , . . . , Xn ∼ f (x) with mean µ and
variance σ 2 , then
Pn √
i=1√Xi − nµ n(X̄n − µ) d
Zn ≡ = −→ Nor(0, 1).
nσ σ
iid
Example: Suppose X1 , X2 , . . . , X100 ∼ Exp(1) (so µ = σ 2 = 1).
100
X 90 − 100 110 − 100
P 90 ≤ Xi ≤ 110 = P √ ≤ Z100 ≤ √
i=1
100 100
≈ P (−1 ≤ Nor(0, 1) ≤ 1) = 0.683. 2
Outline
1 Preliminaries
2 Great Expectations
7 Limit Theorems
8 Statistics Tidbits
Statistics Tidbits
For this iid case, we have already seen that E[X̄n ] = µ, i.e., X̄n is
unbiased for µ.
1 Pn
Definition: The sample variance is S 2 = n−1 i=1 (Xi − X̄n )2 .
(n − 1)S 2 2 (n − 1)S 2
≤ σ ≤ ,
χ2α ,n−1 χ21− α ,n−1
2 2