Professional Documents
Culture Documents
IE 212 STATISTICS
Week 4: Sampling Distributions
SAMPLING DISTRIBUTIONS
❖ The field of statistical inference is basically concerned with
generalizations and predictions.
❖ We computed a statistic from a sample selected from the
population, and from this statistic we made various statements
concerning the values of population parameters that may or may not
be true.
❖ Since a statistic is a random variable that depends only on the
observed sample, it must have a probability distribution.
1 (𝑥 − 𝜇)2
𝑓𝑋 (𝑥) = exp {− 2
} , 𝑥 ∈ 𝑅;
√2𝜋𝜎 2𝜎
𝑥
𝐹𝑋 (𝑥) ≡ ∫ 𝑓𝑋 (𝑢)𝑑𝑢 , 𝑥 ∈ 𝑅.
−∞
𝑌 = 𝑎1 𝑋1 + 𝑎2 𝑋2 + ⋯ + 𝑎𝑛 𝑋𝑛
∑𝑛𝑖=1 𝑎𝑖 𝜇𝑖 and variance 𝜎𝑌2 = 𝑎12 𝜎12 + 𝑎22 𝜎22 + ⋯ + 𝑎𝑛2 𝜎𝑛2 = ∑𝑛𝑖=1 𝑎𝑖 𝜎𝑖2 .□
𝑋̅ − 𝜇
𝑍= 𝜎
√𝑛
as 𝑛 → ∞, is the standard normal distribution with parameters 0 and 1,
respectively.□
1
𝜇𝑋̅ = 𝐸 (𝑋̅) = 𝐸 ( ∑𝑛𝑖=1 𝑋𝑖 ) = 𝜇 and
𝑛
1 𝜎2
𝜎𝑋2̅ = 𝑉𝑎𝑟(𝑋̅) = 𝑉𝑎𝑟 ( ∑𝑛𝑖=1 𝑋𝑖 ) = .□
𝑛 𝑛
DR. BAŞAK GEVER 6
IE 212 STATISTICS WEEK 1
775−800
Corresponding to 𝑥̅ = 775, we find that 𝑧 = = −2.5
10
−2.5
and therefore, 𝑃(𝑋̅ < 775) = 𝑃(𝑍 < −2.5) = ∫−∞ 𝜑𝑍 (𝑢)𝑑𝑢
−2.5 1 −𝑢 2 /2 1 𝑥
= ∫−∞ √2𝜋 𝑒 𝑑𝑢 = {erf ( ) + 1} = 0.0062. (Go to Z Table)
2 √2
𝑛
1
𝑆2 = ∑(𝑋𝑖 − 𝑋̅)2 .
𝑛−1
𝑖=1
➢ Include 𝑋𝑖 − 𝑋̅ = 𝑋𝑖 − 𝜇 + 𝜇 − 𝑋̅:
𝑛
1
⇒ 𝐸 {∑[(𝑋𝑖 − 𝜇) + (𝜇 − 𝑋̅)]2 }
𝑛−1
𝑖=1
DR. BAŞAK GEVER 12
IE 212 STATISTICS WEEK 1
𝑛
1
= 𝐸 {∑[(𝑋𝑖 − 𝜇)2 + (𝜇 − 𝑋̅)2 + 2(𝑋𝑖 − 𝜇)(𝜇 − 𝑋̅)]}
𝑛−1
𝑖=1
𝑛 𝑛 𝑛
1
= 𝐸 {∑(𝑋𝑖 − 𝜇)2 + ∑(𝜇 − 𝑋̅)2 + ∑ 2(𝑋𝑖 − 𝜇)(𝜇 − 𝑋̅)}
𝑛−1
𝑖=1 𝑖=1 𝑖=1
𝑛 𝑛
1
= 𝐸 {∑(𝑋𝑖 − 𝜇)2 + 𝑛(𝜇 − 𝑋̅)2 + 2(𝜇 − 𝑋̅) ∑(𝑋𝑖 − 𝜇)}
𝑛−1
𝑖=1 𝑖=1
𝑛
1
= 𝐸 {∑(𝑋𝑖 − 𝜇)2 + 𝑛(𝑋̅ − 𝜇)2 − 2𝑛(𝑋̅ − 𝜇)2 }
𝑛−1
𝑖=1
𝑛
1
= {∑ 𝐸 (𝑋𝑖 − 𝜇)2 − 𝑛𝐸 (𝑋̅ − 𝜇)2 }
𝑛−1
𝑖=1
𝑛 2
1 1 𝑛𝜎𝑋
= {∑ 𝑉𝑎𝑟(𝑋𝑖 ) − 𝑛𝑉𝑎𝑟(𝑋̅)} = {𝑛𝜎𝑋2 − }
𝑛−1 𝑛−1 𝑛
𝑖=1
1
= 𝜎𝑋2 (𝑛 − 1) = 𝜎𝑋2 . □
𝑛−1
1 𝑛
𝑓𝑋 (𝑥) = 𝑛 𝑥 2 −1 𝑒 −𝑥/2 , 𝑥 > 0. □
𝑛
2 Γ( )
2
2
❖ If you recall, Chi-Squared distribution is a speacial case of Gamma
distribution.
❖ When the parameters of Gamma distribution 𝛼 = 𝑛/2 and 𝜆 = 1/
2, then we obtain Chi-Squared distribution with degrees of freedom
𝜈 = 𝑛.
𝑈 = ∑ 𝑍𝑖2 .
𝑖=1
3. T – Student Distribution
(𝜈 + 1 ) (𝜈+1)
Γ[ ] 𝑡2
−
2
2
ℎ(𝑡) = 𝜈 (1 + ) , 𝑡 ∈ 𝑅.
Γ ( ) √𝜋𝜈 𝜈
2
This is known as the t-distribution with 𝜈 degrees of freedom.□
➢ From the foregoing and the theorem above we have the following
corollary.
Example: Find k such that 𝑃{𝑘 < 𝑇 < −1.761} = 0.045 for a random
𝑋̅−𝜇
sample of size 15 selected from a normal distribution and .
𝑆/√𝑛
Solution:
𝑃{𝑘 < 𝑇 < −1.761} = 0.045 ⇒ 𝑃{𝑇 < −1.761} − 𝑃{𝑇 < 𝑘} = 0.045
Because of the symmetry, 𝑃{𝑇 < −1.761} = 𝑃{𝑇 > 1.761} can be
written. From t-table
𝑃{𝑇 > 1.761} = 0.05.
Then 𝑃{𝑇 < −1.761} − 𝑃{𝑇 < 𝑘} = 0.045 ⇒ 0.05 − 𝑃{𝑇 < 𝑘} =
0.045 ⇒
𝑃{𝑇 < 𝑘} = 0.005 = 𝑃{𝑇 > −𝑘} = 0.005 ⇒ 𝑘 = −𝑡14;0.005 = −2.977
4. F Distribution
❖ We have motivated the t-distribution in part by its application to
problems in which there is comparative sampling (i.e., a comparison
between two sample means).
❖ While it is of interest to let sample information shed light on two
population means, it is often the case that a comparison of variability
is equally important, if not more so.
❖ The F-distribution finds enormous application in comparing sample
variances.
❖ Applications of the F-distribution are found in problems involving
two or more samples.