Professional Documents
Culture Documents
Lecturer :
Dr. Iqbal Kharisudin S.Pd., M.Sc.
Arranged by :
Adelia Salsabila (4101418127)
FebriyantiErviMu’azzizah (4101418117)
Rembulan Permata Octalia (4101418067)
COVER .................................................................................................................................. i
REFRENCES ......................................................................................................................... 31
ii
CHAPTER I
INTRODUCTION
1.1 Background
The concept of a random variable then was introduced so that events could be
associated with sets of real numbers in the range space of the random variable. This makes
it possible to mathematically express the probability model for the population or
characteristic of interest in the form of a pdf or a CDF for the associated random variable,
say X. In this case, X represents the initial characteristic of interest, and the pdf,f(x), may
be referred to as the population pdf. It often maybe the case that some function of this
variable also is of interest. Thus, if X represents the age in days of some component,
another experimenter may be expressing the age, Y, in hours, so that Y = 24X. Similarly,
W = ln X or some other function of X may be of interest. Any function of a random
variable X is itself a random variable, and the probability distribution of a function of X is
determined by the probability distribution of X. For example, for Y above, ,
- , -, and so on. Clearly, probabilities concerningfunctions of a random
variable may be of interest, and it is useful to be able toexpress the pdf or CDF of a
function of a random variable in terms of the pdf orCDF of the original variable. Such pdf's
sometimes are referred to as "derived"distributions. Of course, a çertain pdf may represent
a population pdf in oneapplication, but correspond to a derived distribution in a different
application. In this paper we will disscuss about Fuctions of Random Variable such as
Transformation Methods, Sums of Random Variables, and Order Statistics.
1.2 Problems
1
1.3 Objectives
2
CHAPTER II
DISCUSSION
( ) ( ( )) (6.3.1)
where * ( ) +
Proof :
This follows because ( ) , - , ( ) - , ( )- ( ( ))
Example 2.1.1
Let ( ), so that
( )
Another frequently encountered random variable that also is called geometric is
of the form so that ( ) , ( ) , and
( ) ( )
( )
which is nothing more than the pdf of the number of failures before the first success.
Example 2.1.2
3
A distribution that is related to the normal distribution, but for which the random
variable assumes only positive values, is the lognormal distribution, whichis defined
by the pdf
( ) ( )
(6.3.3)
√
and in this notation teta becomes a scale parameter. It is clear that cumulative
lognormal probabilities can be expressed in terms of normal probabilities, because if
relationship ( ) , then
( ) , - , -
( ) , - , -
( ) , - . / (6.3.5)
Example 2.1.3
If ( ), then the pdf of is
( )
( )
and the CDF is
( )
for all real z. 1f we introduce location and scale parameters and 0, respectively, by
the transformation ( ) , then the pdf of ( ) is
0 1
( ) (6.3.6)
2 0 13
4
for all real y. The distribution of Y is known as the logistic distribution, denoted by
( ). This is another example of a symmetric distribution, which follows by
noting that
. /
( ) ( )
( ) ( )
The transformation provides a general approach to introducing location
and scale parameters into a model.
Consider the case in which u(x) is an increasing function so that the inverse
transformation, x = w(y), also is increasing :
, ( )- ∫ ( ) ( )
, ( )- ∫ , ( )- , ( )- ( )
, ( )- ∫ ( )
, ( )- ( )
The case in which u(x) is decreasing is similar. A very useful special transformation
is given by the following theorem.
5
Example 2.1.4 Let X be a continuous random variable with pdf
( ) {
The CDF of X, whose graph is shown in Figure 6.2, is not one-to-one, because it
assumes the value 1/2 for all .
( ) {
The function G(u) has another important application, which follows from the next
theorem.
Theorem 2.1.4 Let F(x) be a CDF and let G(u) be the function defined by (6.3.7). If
( ) then ( ) ( )
Example 2.1.5
If ( ), then
( ) {
and
( ) 2
6
2.1.2 Transformations That Are Not One-To-One
Theorem 2.1.5 Theorem 2.1.1 can be extended to functions that are not one-to-one
by replacing equation (6.3.1) with
( ) ∑ . ( )/ (6.3.9)
Example 2.1.6
and
( ) ( )
( ) ( ) ( )
( ) ( ) ( )
( ) 0. / . / 1
An expression analogous to equation (6.3.9) for the discrete case is obtained for
continuous functions that are not one-to-one by extending equation (6.3.2) to
( ) ∑ ( ( )) | ( )| (6.3.10)
If Y = X2 , then
( ) (√ ) ( √ )
and by taking derivatives
( ) [ (√ ) ( √ )] (6.3.11)
√
7
Theorem 2.1.6 If X is a vector of discrete random variables with joint pdf f x(x) and
Y = u(X) defines a one-to-one transformation, then the joint pdf of Y is
( ) ( ) (6.3.14)
where x1, x2 , ..., xk, are the solutions of y = u(x), and consequently depend on y 1 ,
Y2,,Yk
If the transformation is not one-to-one, and ifa partition exists, say A1 , A2 ,
such that the equation y = u(x) has a unique solution x = xj or
( ) (6.3.15)
over Aj, then the pdf of Y is
( ) ∑ ( ) (6.3.16)
| | (6.3.17)
Example 2.1.7
It is desired to transform x1 and x2 into x1 and the product x1 x2 . Specifically, let
y1 = x1 and y2 = x1x2. The solution is x1 = y1 and x2 = y2 /y1, and the Jacobian is
| |
|| ||
8
Theorem 2.1.7 Suppose that ( )is a vector of continuous random
variables with joint pdf ( ) on A, and ( ) is
defined by the one-to-one transformation
( ) i=1,2,...,k
If the Jacobian is continuous and nonzero over the range of the transformation, then
the joint pdf of Y is
( ) ( ) (6.3.19)
Where ( ) is the solution of ( )
Proof :
As noted earlier, the problem of finding the pdf of a function of a random vari able is related
to a change of variables in an integral. This approach extends readily to transformations of
k variables. Denote by B the range of a transformation y = u(x) with inverse x = w(y).
Assume D B , and let C be the set of all points x = (x1, ..., xk) that map into D under the
transformation. We have
, - ∫ ∫ ( )
, - ∫ ∫ ( )
, - ∫ ∫ , ( ) ( )-
| |
And thus
9
( ) ( )
( ) ,( )
and zero otherwise. The set B is obtained by transforming the set A, and this
corresponds to and . Thus
*( ) +, which is a triangular region in the plane with
boundaries and .
The marginal pdf’s of Y1 and Y2 are given as follows :
( ) ∫
( )
( ) ∫
( )
Note that ( )
We will discuss about special methods for dealing with the important special case of
sums of random variables.
( ) ∫ ( ) (2.2.1)
( ) ∫ ( ) ( ) (2.2.2)
Example 2.2.1
Let and X2 be independent and uniform, X iUNIF(0, 1), and let
. The region corresponding to the transformation and
10
is *( ) + and this is shown in
Figure 2.2.1.
Regions corresponding to the transformation and
Figure 2.2.1
( ) ∫
Example 2.2.2
( )
( ) ( )
| | ( ) ( )
and
11
( )
( ) [ ( )]
( ) ( )
( )
( ) ( )
12
( ) ( )
( ) ( )
▄
( ) , ( )-
The MGF of a random variable uniquely determines its distribution. It is clear that
the MGF can be used as a technique for determining the distribution of a function of
a random variable, and it is undoubtedly more important for this purpose than for
computing moments. The MGF approach is particularly useful for determining the
distribution of a sum of independent random variables, and it often will be much
more convenient than trying to carry out a joint transformation. If the MGF of a
variable is ascertained, then it is necessary to recognize what distribution has that
MGF.
Example 2.2.3
( ) ( ) ( )
( ) ( )
( )
We recognize that this is the binomial MGF with parameters and , and
thus, ( ).
Example 2.2.4
13
( ) , ( )- , ( )-
,( )( )-
( ) ( ) ( ) ( ) (6.5.2)
if , and zero otherwise.
▄
Example 2.3.1
Suppose that and represent a random sample of size 3 from a population
with pdf
( )
14
It follows that the joint pdf of the order statistics and is
( ) ( ) ( ) ( )
( ) ∫ ∫
( )
If we want to know the probability that the smallest observation is below some value, say
0.1, it follows that
, - ∫ ( )
It is possible to derive an explicit general formula for the distribution of the k-th order
statistic in terms of the pdf, f(x), and CDF, F(x), of the population random variable X. If X
is a continuous random variable with ( ) on (a may be and b may
be ), then, for example, for n = 3,
( ) ∫ ∫ ( ) ( ) ( )
( )∫ ( ), ( ) ( )-
, ( )-
( ) |
( ), ( )-
Similarly,
( ) ∫ ∫ ( ) ( ) ( )
( ), ( ) ( )- ∫ ( )
( ), ( )- ( )
where ( ) and ( ) .
15
These results may be generalized to the n-dimensional case to obtain the following
theorem.
Theorem 2.3.2 Suppose that denotes a random sample of size n from a
continuous pdf, f(x), where ( ) for . Then the pdf of the kth order
statistic , is given by
( ) , ( )- , ( )- ( ) (6.5.3)
( )( )
Figure 6.6
The multinomial form gives the joint pdf for and as
( ) , ( )- ( )
( ) ( ) ( )
[ ( ) ( )] [ ( )] ( ) (6.5.4)
if , and zero otherwise. This is illustrated by Figure 6.7.
16
Figure 6.7
The smallest and largest order statistics are of special importance, as are certain functions
of order statistics known as the sample median and range.
The sample median is the middle observation of
where ( ) , if n = odd
and where , if n =even
The sample range is the difference of the smallest from the largest,
.
For continuous random variables, the pdf's of the minimum and maximum, and ,
which are special cases of equation (6.5.3) are
( ) , ( )- ( ) (6.5.5)
and
( ) , ( )- ( ) (6.5.6)
For discrete and continuous random variables, the CDF of the minimum or maximum of
the sample can be derived directly by following the CDF technique.
For the minimum.
( ) , -
, -
, -
, ( )- (6.5.7)
17
, ( )- (6.5.8)
Theorem 2.3.3 For a random sample of size n from a discrete or continuous CDF, F(x), the
marginal CDF of the k-th order statistic is given by
( ) ∑ . /, ( )- , ( )- (6.5.9)
Example 2.3.2
Consider a random sample of size n from a distribution with pdf and CDF given by
( ) and ( ) ; . From equations (6.5.5) and (6.5.6), we
have that
( ) ( )
And
( ) ( )
Example 2.3.3
Suppose that in Example 2.3.2 we are interested in the density of the range of the sample,
. From expression (6.5.4), we have
( ) , - ( )
( )
( ) ( )
( ), -
18
Figure 6.8
( ) ∫ ( )
∫ ∫ ( ), ( ) ( )- ( )
( )
∫ ( ), ( ) ( )- (6.5.11)
Note, however, that great care must be taken in applying this formula to an example where
the region with ( ) has finite limits.
19
Theorem 2.3.4 Type II Censored Sampling The joint marginal density function
of the first r order statistics from a random sample of size n from a continuous
pdf, f(x), is given by
( ) , ( )- ∏ ( ) (6.5.12)
( )
and
( )
( ) ( )
( ) , ( )- ∏ ( ) (6.5.17)
( )
20
which simplifies to equation (6.5.17).
Example 2.3.4
We will assume that failure times of airplane air conditioners follow an
exponential model EXP( ). We will study properties of random variables in the
next chapter that will help us characterize a distribution and interpret the physical
meaning of parameters such as . However, for illustration purposes, suppose the
manufacturer claims that an exponential distribution with = 200 provides a good
model for the failure times of such air conditioners, but the mechanics feel =
150 provides a better model. Thirteen airplanes were placed in service, and the
first 10 air conditioner failure times were as follows (Proschan, 1963):
23, 50, 50, 55, 74, 90, 97, 102, 130, 194
For Type II censored sampling, the likelihood function for the exponential
distribution is given by equation (6.512) as
( )
g( ) ( )
0 1 0 ∑ 1
[ (∑ ( ) ) ]
( )
∑ ( )
21
Thus, if one wished to choose a value of based on these data, the value =
144.7 seems reasonable.
For illustration purposes, suppose that Type I censoring had been used and that
the experiment had been conducted for 200 flying hours for each plane to obtain
the preceding data. The likelihood function now is given by equation (6.5.17):
( ) [ (∑ ( ) ) ]
( )
(∑ ( ) )
As a final illustration, suppose that a large fleet of planes is placed in service and a
repair depot decides to record the failure times that occur before 200 hours.
However, some units in service may be taken to a different depot for repair, so it
is unknown how many units have not failed after 200 hours. That is, the sample
size n is unknown. Given that r ordered observations have been recorded, the
conditional likelihood is given by equation (6.5.16):
( ∑ )
( )
, ( )-
where r = 10 and = 200.
The value of that maximizes this joint pdf cannot be expressed in closed form;
however, the approximate value for this case based on the given data is = 245.
This value is not too close to the other values obtained, but of course the data were
not actually obtained under this mode of sampling. If two different assumptions
are made about the same data, then one cannot expect to always get similar results
(although the Type I and Type II censoring formulas are quite similar).
22
2.4 EXERCISE
1. Let be independentnormally distributed random variables, ( ),
MGF of Y!
Solution :
( ) 4 5 4 5
4 5
( )
,( ) -
Then we get ( ).
( ) ( )
Solution:
( ) ( ) ( )
( ) ( )
So consequently, ( ).
( ) {
23
First, we find the inverse transformation:
( )
( ) √ (by inverse transformation) and its
derivative:
( )
Thus, for
( ) , ( )- | ( )|
( )
( )
( ) {
( ) {
The domain (*domain is the region where the pdf is non-zero) for ( ) is
* +; thus, because , it follows that the domain for U is
* +.
The cdf of U is:
( ) (𝑈 ) ( )
( ) (𝑈 ) ( )
( ) (𝑈 ) ( )
( ) (𝑈 ) ( )
( ) (𝑈 ) ( )
Because
( ) ( )
24
Taking derrivatives, we get for u > 0,
( ) ( ) ( )
( ) {
( ) { ( )
( ) , ( )- | ( )|
( ) ( ) , ( )-
( ) ( )
So, we get the PDF of U is,
( ) ( )
{
25
c. Find the pdf of the sample range R = Y2 − Y1
Solution :
a. The marginal pdf of the smallest and largest order statistics from a sample of size
n = 2 can be derived by using the marginal pdf formula, but we first need the
CDF of X :
( ) ∫
Now, we use the marginal pdf of formula :
( ) , ( )- , ( )- ( )
( ) ( )
, ( ) -
, ( ) -
and
( ) , ( )- , ( )- ( )
( ) ( )
,( ) -
( )
b. The joint pdf of Y1 and Y2 can be obtained from the pdf f (x) (the joint pdf can be
written as ( ) ( ) ( ) for n independently sampled
observations from the distribution defined by f(x)) taking into account the
permutations of and :
( )
( ) ( )
( )( )
c. The pdf of the sample range is just a transformation of the joint PDF, similar to all
of the convolutions that we've been doing. Consider the inverse transforms: Y1 =
Y1 and Y2 = R +Y1. Then we can use the transform methods:
( ) ( )
Where | | So, applying the formula for the joint from point b. So we
get:
26
( ) ( )
which we now use to integrate out yl to get the marginal of R. We proceed, paying
careful attention to the change in bounds (draw a picture to convince yourself-first
with Y1 and Y2 and then with R and Y1)
( ) ∫ ( )
0 . /1
. ( ) ( ) /
( ( ) ( ))
; zero otherwise.
a. Give the joint pdf of the order statistics.
b. Give the pdf of the smallest order statistic, Y 1.
c. Give the pdf of the largest order statistic, Y n.
d. Derive the pdf of the sample range, R = Yn−Y1, for n= 2.
e. Give the pdf of the sample median, Yr, assuming that n is odd so that (
) .
Solution:
a. The joint pdf can be represented as a relabeling of the original joint pdf of X1
through Xn , the random variables representing the sample.
In particular, we can write:
( ) ( ) ( )
b. The pdf of the smallest order statistic is the PDF when we integrate out all of the
other order statistics. We already have a formula for this, so we'll use it, rather
than deriving it all over again. We use k = 1 and n:
27
( ) , ( )- , ( )- ( )
( ) ( )
( ) , ( )- ( )
[ ∫ (̃ ) ̃ ] ( )
0 . /1
0 1
c. The pdf for the largest order statistic is computed from the same formula as for
the smallest order statistic:
( ) , ( )- , ( )- ( )
( ) ( )
( ) , ( )- ( )
[∫ (̃ ) ̃ ] ( )
0. /1
0 1
d. We again define two random variables. Call them S = Y1 and R = Yn –Y1 and the
inverse transforms of S = Y1 and Y2 = R +S. The joint pdf from part a can now be
used:
( )
We use the transformation formula, noting that the Jacobian has a determinant of
1: | | .
( ) ( )
( )
( ) ∫ ( )
28
This is the partial fraction decomposition:
( ) ( ) ( )
which we can now integrate. The solution we obtain for the pdf is
( ) ∫ 0 1
( ) ( )
( )
6 7
( )
[ . / ]
( )
. /
( )
[ ( )]
where r > 0. The PDF of the sample range and the first and second order statistics
in a sample of size n = 2 are plotted below, and compared to the kernel estimates
from 1 million samples of size n = 2 to benchmark the theory.
e. The sample median is the order statistic with r = k = (n+ 1)/2. We can compute
this in the same way as before, using the formula:
( ) , ( )- , ( )- ( )
( ) ( )
( )
[ ] [ ]
( ) ( )
( ) ( )
,( )-
29
CHAPTER III
CLOSING
3.1 CONCLUSION
The main purpose of this chapter was to develop methods for deriving the distribution
of a function of one or more random variables. The CDF technique is a general method
that involves expressing the CDF of the "new" random variable in terms of the
distribution of the "old random variable (or variables). When one k-dimensional vector of
random variables (new variables) is defined as a function of another k-dimensional vector
of random variables (old variables) by means of a set of equations, transformation
methods make it possible to express the joint pdf of the new random variables in terms of
the joint pdf of the old random variables. The continuous case also involves multiplying
by a function called the Jacobian of the transformation. A special transformation, called
the probability integral transformation, and its inverse are useful in applications such as
computer simulation of data.
The transformation that orders the values in a random sample from smallest to largest
can be used to define the order statistics. A set of order statistics in which a specified
subset is not observed is termed a censored sample. This concept is useful in applications
such as life-testing of manufactured components, where it is not feasible to wait for all
components to fail before analyzing the data.
3.2 SUGGESTION
The reader advised to look more examples and do more exercises to enrich of
knowledge about functions of random variable.
30
REFRENCES
Bain, L. J., & Engelhardt, M. (1993). Introduction to Probability and Mathematical Statistics. In
Biometrics (Vol. 49, Issue 2). https://doi.org/10.2307/2532587
31