Professional Documents
Culture Documents
CHAPTER 4
PROBABILITY AND SAMPLING
DISTRIBUTIONS
Outline of the Subject
2
Chapter 1: Introduction-to-Applied-Probability-Statistics
Chapter 2: Numerical_Summary_Measure
Chapter 3: Bivariate_Multivariate_Data_Distribution
Chapter 4: Probability_Sampling_Distributions
q For instance, the chance experiment of conducting a series of stress tests on three
metal parts has the eight possible outcomes PPP, PPF, PFP, FPP, PFF, FPF, FFP,
and FFF, where P and F denote the test results “pass” and “fail,” and the order in
which the letters appear corresponds to the part number tested.
• When two events A and B have no simple events in common, we say that
they are mutually exclusive or disjoint ( cannot occur simultaneously).
• Several of the previous definitions can be extended to include events
formed from more than two events.
DEFINITIONS
Given a chance experiment and any events A1, A2, A3, . . . , Ak:
1. The event A1 or A2 or A3 or . . . or Ak consists of all the simple events that are
contained in at least one of the events A1, A2, A3, . . . , or Ak. It can also be described
as the event that at least one of the events A1, A2, A3, . . . , or Ak occurs.
2. The event A1 and A2 and A3 and . . . and Ak consists of all simple events
common to all the events A1, A2, A3, . . . , and Ak. This event can be described as
the event that all of the events A1, A2, A3, . . . , and Ak occur.
3. Several events A1, A2, A3, . . . , and Ak are said to be mutually exclusive or
disjoint if no two of them have any simple events in common.
4.1 Chance Experiments and Probability Concepts
à Probability Concepts
Probability Axioms
1. The probability of any event must lie between 0 and 1. That is,
0 £ P(A) £ 1 for any event.
2. The total probability assigned to the sample space of an
experiment must be 1.
• More generally, for any collection of disjoint events A1, A2, A3, . . . ,
Ak,
4.1 Chance Experiments and Probability Concepts
à Complementary Events
Two events, A and B, are independent events if the probability that either one occurs is
not affected by the occurrence of the other. In this case,
Several events, A1, A2, A3, . . . , Ak, are independent if the probability of each
event is unaltered by the occurrence of any subset of the remaining events.
In this case, the product rule can be applied to any subset of the k events. That
is, the probability that all the events in any subset occur equals the product of
their individual probabilities of occurring. In particular, for all k events,
4.3 Random variables
à Introduction
Example 4.1: A coin is tossed three times and the sequence of heads
and tails is noted. The sample space for this experiment is
u(x)
1
X and Y are two random variables defined on the same sample space S.
1. 𝑓 𝑥, 𝑦 ≥ 0
2. ∑!""($,&) 𝑓 𝑥, 𝑦 = 1
¨ Example 4.2:
4.3 Joint Distributions
Distributions for two discrete variables
à The marginal probability mass functions
26
¨ Def:
¨ The second condition indicates that the total volume under this density surface
is 1.
à Figure :
Volume representing the proportion
of ( x, y) in the region A
4.3 Joint Distributions
Distributions for two continuous variables
28
¨ Example 4.3:
1.
2.
4.3 Joint Distributions
Distributions for two continuous variables
à The marginal probability mass functions
30
¨ Def:
What is the probability distribution of busy time from for the walk-up
window for the period between one-quarter and third-quarter of the time ?
à The marginal pdf of Y, which gives the probability distribution of busy
time for the walk-up window without reference to the drive-up facility is
Leads to
4.3 Joint Distributions
The Bivariate Normal Distribution
32
Note that once independence is assumed, one has only to select appropriate
distributions for 𝑥 and 𝑦 separately and then use (1) to yield the joint
distribution.
4.3 Joint Distributions
Correlation and the Bivariate Normal Distribution
33
¨ Example (cont’):
àBecause x and y are
independent, we can write the
joint density as the product:
Suppose that we draw 1000 random samples, each of size n = 25, from a
normal population with a mean of 50 and a standard deviation of 2. If we
calculate the mean 𝑥̅ of each sample, then the distribution of all 1000 𝑥̅ values
gives a good approximation to the sampling distribution of 𝑥.̅
• Notice that the 1000 sample means stack up around the population mean (µ = 50).
• Variation among the sample means is smaller than variation in the population.
• In particular, none of the sample means fall outside the range of 48.5 to 51.5 (i.e.,
none are more than 1.5 units away from µ).
• In fact, it also appears that very few sample means fall outside the interval 49 to 51;
that is, they are generally within 1 unit of µ.
• From the shape and location of the sampling distribution, we can begin to see
which values of the sample statistic are more likely to occur than others.
• The sampling distribution of 𝑥,̅ also called the sampling distribution of the mean,
is the probability distribution that describes the behavior of 𝑥̅ in repeated random
samples from a population or process.
• Like any distribution, the sampling distribution of 𝑥̅ has its own unique mean and
standard deviation, which we denote by µ𝑥 ̅ and s𝑥 ̅ , respectively.
4.5 Describing Sampling Distributions
!
à Sampling Distribution of 𝒙
• These equations hold regardless of the particular form of the population distribution.
• One of the key features of the standard error of the mean s𝑥 ̅ is that it decreases as
the sample size increases. In fact, many statistics have this.
• This makes intuitive sense, since we expect that more information ought to provide
better estimates (i.e., smaller standard errors). As a result, increasing the size of a
random sample has the desirable effect of increasing the probability that the estimate
𝑥̅ will lie close to the population mean µ.
4.5 Describing Sampling Distributions
à Sampling Distribution of 𝒙!
à Sampling from a Normal Population
• The normality of 𝑥̅ , along with the fact that its mean µ𝑥 ̅ and standard
error s𝑥 ̅ can be determined from µ and s, is enough to completely
characterize the sampling distribution of x in this case.
When a population
distribution is normal,
the sampling distribution
# is also normal,
of 𝒙
regardless of the size of
the sample.
4.5 Describing Sampling Distributions
à The Central Limit Theorem
Every random sample drawn from such a population will consist entirely of 0s and
1s. Suppose, for instance, that a particular sample of size 10 contains the observations
{0, 0, 1, 1, 0, 1, 0, 0, 1, 0}. Then the sample mean is (0+0 +1+1+0+1+0 +0+1+0)/10 =
.40. That is, the sample mean is simply the proportion of 1s in the sample.
à We use the notation p to denote the proportion of successes, also called the sample
proportion, in a random sample of size n.