ajd

Attribution Non-Commercial (BY-NC)

5 views

ajd

Attribution Non-Commercial (BY-NC)

- G R Dodagoudar_reliability Pap_2011
- Error Function Table PDF
- cs1538
- Answer MSE1sttest
- Statistics for Business and Economics: bab 5
- Mth514NotesCh4,5,6,10
- Ma2b12rec1-12
- I02620020220115066Random Variables and Probability Distributions
- Lecture713-03-4
- Lecture 1
- Cats Guide
- Cont Examp
- Dip Image Inhancment
- MonteCarlo
- Chapter 7 (1)
- Solución 2.3
- W3INSE6220.pdf
- Digital Image Processing
- ES9-Chap06
- Random Variables Tarea Teoría

You are on page 1of 38

Monotonic Transformations

Expectations

Random Variables

Monotonic Transformations

Expectations

Outline

Random Variables Discrete Random Variables Continuous Random Variables Monotonic Transformations Expectations Properties of Expectations Measures of Dispersion

Monotonic Transformations

Expectations

Random Variables

1

Denition: A discrete random variable is a random variable that can take on only a nite (or countably innite) number of values. Denition: A continuous random variable is a random variable that can take on a continuum of values (uncountable number of values).

Let F (x ) be the CDF for a continuous random variable X . Properties of F (same as for discrete random variables):

limx F (x ) = 0, limx + F (x ) = 1 F (x ) is increasing so x1 < x2 F (x1 ) F (x2 )

The CDF is a smooth increasing function over some interval of real numbers.

Monotonic Transformations

Expectations

In dealing with continuous random variables, we talk about the distribution of the random variable over the sample space in terms of the probability density function [PDF] not probability mass functions. Denition: The CDF of a continuous random variable is x F (x ) = P (X x ) = f (x ) x . If f is continuous at x , then the PDF is f (x ) = F (x )/ y . In other words, the PDF is the derivative of the CDF. A PDF f (x ) is a function such that: f (x ) 0 for all x , < x < + f is piecewise continuous, and f (x ) x = 1. These conditions imply that: P (a X b) =

b a

f (X ) X .

Monotonic Transformations

Expectations

Continuous Random Variables EXAMPLE: The Uniform distribution on the interval [0, 1] chooses any number between 0 and 1. The PDF of the Uniform is given by: 1 if 0 x 1 f (x ) = 0 else f(x)

Monotonic Transformations

Expectations

EXAMPLE: Uniform Distribution

A consequence is that P (X = a) =

a a

f (x ) x = 0.

The probability that a given value for a continuous random variable is realized is 0. Why is this the case? If the probability of observing the realization a was any positive probability, it would be the same for any number (by denition of the Uniform). But then the sum of any countably innity subset of [0,1] (e.g., the set of rationals) would be > , which is not possible. If P (X = x ) = p > 0, then F (x ) would have a discontinuity (jump) of size p at point x, violating the assumption of continuity. Practically speaking, what is the likelihood of seeing rainfall measurement of exactly 3.435 inches?

Monotonic Transformations

Expectations

EXAMPLE: Uniform Distribution The CDF of the Uniform Distribution described above is (note the pdf is just the derivative of F (x ): given by: 0 if x < 0 x if 0 x 1 F (x ) = 1 if x > 1 f (x ) = 1 0 if 0 x 1 else

Pr(X <= x)

0.0 2

0.4

0.8

0 x

Monotonic Transformations

Expectations

Normal Probability Distributions

For > 0 and , Y is a normal random variable if the probability distribution is dened as: 2 2 1 e(x ) /(2 ) , x p(X = x ) = p(x ) = 2 Standard Normal: = 0 and = 1

Monotonic Transformations

Expectations

Normal Probability Distributions

Monotonic Transformations

Expectations

Standard Normal Probability Distributions

Monotonic Transformations

Expectations

Other Probability Distributions

Random Variables

Monotonic Transformations

Expectations

Monotonic Transformation

Suppose g(x) is a strictly monotone function with inverse g 1 (). Let Y be {y |g (x ) = y for some x X )}. Suppose that fX (x ) is continuous. Then

fY (y ) = fX (g 1 (y ))

g 1 y (y )

, on Y .

Random Variables

Monotonic Transformations

Expectations

Monotonic Transformation

EXAMPLE: Say pdf fX (x ) = exp(x ), x > 0 and 0 elsewhere Find the pdf for Y = 1/X

fY (y ) = e(1/y )

y

1 y2

fY (y ) = ee |ey |

Random Variables

Monotonic Transformations

Expectations

Events in a sample space can be used to model experiments. Probability measure on the events enables us to evaluate the frequency of events. (Discrete and Continuous) random variables can be dened on the sample space to quantify events. Random variables can be described/characterized using pmf/pdf and cdf. However, is there a summary measure of a quantity of interest that can be used to characterize the outcome of the experiment? Of course there is. We refer to such quantities as moments.

Random Variables

Monotonic Transformations

Expectations

Discrete Random Variable: Expected Value E[X] One characteristic that we might desire is the expected realization of the random variable X. For a discrete random variable X with values xi , 1 = 1...N , we calculate the expected value (or expectation or mean or the mean of the probability distribution) as follows:

N

E [X ] =

i =1

xi f (xi ) = = x

EXAMPLE 1 Let X be the number of heads in 2 fair coin tosses. Then E [X ] = 0 1/4 + 1 1/2 + 2 1/4 = 1.

.50 Pr(X=x) 0 .25

Random Variables

Monotonic Transformations

Expectations

Discrete Random Variable: Median(X)

An alternative measure of central tendency is given by the median of a random variable. The median xM of random variable X is dened as the choice xM such that P (X xM ) 1/2.

Pr(X < x) 0.8 0.0 2 0.4

For the experiment described above, note that xM = 1. x Also realize that, by denition, whereas the median depends only on the rank ordering of the realizations of X, the mean depends on the actual values of X. In other words, since the realizations xi i 1...N enter into our calculations, the mean is sensitive to extreme values (sometimes called outliers").

Random Variables

Monotonic Transformations

Expectations

Example: Mean vs. Median

EXAMPLE Consider the random variables X and Y with the following PMF. x f(x) y f(y) 0 .5 0 .5 1 .49 1 .5 1000000 .01

What is E[X] and E[Y] = ? What is the median of X and Y = ? E[X] = .5 E[Y] = 10000.49 Median of X and Y are both .5

Random Variables

Monotonic Transformations

Expectations

Discrete Random Variable: Mode(X)

We can also dene the mode of a random variable, which is the element of X that occurs most frequently. mode(X ) = value of xi which maximizes p(xi )

Random Variables

Monotonic Transformations

Expectations

N = E [X ] = =x i =1 xi f (xi ) median(X ) = xM such that P (X xM ) 1/2 mode(X ) = value of xi which maximizes p(xi )

Continuous Random Variable = E [X ] = xf (x ) x =x xM median(X ) = xM such that f (x ) x 1/2 mode(X ) = Value of xi which maximizes f (xi )

Random Variables

Monotonic Transformations

Expectations

Measures of Central Tendency EXAMPLE Suppose f (x ) = 2x 3 x > 1. Calculate the expected value, median and mode.

2.0

3.0

1 x

Random Variables

Monotonic Transformations

Expectations

EXAMPLE Suppose f (x ) = 2x 3 x > 1. Calculate the expected value, median and mode. E [X ] = = = = = = = x 2x 3 x b limb 1 x 2x 3 x 1 b limb 1 2x 2 x b limb 2 1 x 2 x 2 limb 2 2

1 x 1 1 2 2 limb b 1 1

b 1

1 2

f (x ) x limb a f (x ) x by standard improper integral results. Since kf (x ) x = k f (x ) x . 3 By Power Rule": x n x = 1/(n + 1)x n+1 for n = 1.

Random Variables

Monotonic Transformations

Expectations

EXAMPLE Suppose f (x ) = 2x 3 x > 1. Calculate the expected value, median and mode. The median of X is solved for as follows.

1 2 1 2 1 2 1 m2

= =

m 1

2x 3 x

m 1

= = m=

1 x 2 2 1 1 1 m2 1 2

Inspection of the PDF reveals that the mode of X is obviously at 1. Note that as the above example indicates, there is no reason to expect that all measures of central tendency will be identical. Consequently, depending on the measure you choose, your characterization of the random variable may differ. Furthermore, there is no necessary reason for the median and mode to be unique although it is the case that the mean (E[X]) is unique.

Random Variables

Monotonic Transformations

Expectations

EXAMPLE Consider the random variable X with the following PMF. x f(x) 0 .2 1 .3 2 .1 3 .3 4 .1

E [X ] = 0(.2) + 1(.3) + 2(.1) + 3(.3) + 4(.1) = 1.8.4 mode(X ) = {1, 3} 1 5 median(X ) [1, 2] since P (X < xm ) 1 2 and P (X > xm ) 2 .

4 5

Note that there is no reason why E[X] has to be an actual value of X. Sometimes expresses as 1.5.

Monotonic Transformations

Expectations

Properties of Expectations

For any function of the random variable g(x) we can still calculate the expected value of the transformed" random variable using the following result:

E [g (x )] =

x

E [g (x )] =

g (x )f (x ) x if X continuous

Monotonic Transformations

Expectations

Properties of Expectations

EXAMPLE Consider the random variable X with the following PMF. x f (x ) 0 .5 1 .5 x2 f (x 2 ) 0 .5 1 .5

E [X ] = 0(.5) + 1(.5) = .5 implying that (E [X ])2 = .52 = .25. E [X 2 ] = 0(.5) + 1(.5) = .5. EXAMPLE Let g(x) = a + b(x). E [a + bX ] = = = = =

(a + bx )f (x ) x af (x ) x + bxf (x ) x a f (x ) x + b xf (x ) x 6 a 1 + b E [X ]

a + bE [X ]

Monotonic Transformations

Expectations

Measures of Dispersion

Just as there are several measures of central tendency, so too are there several measures of dispersion. The two most commonly used are the range and the variance. The range is dened as: range(X) = max(X) - min(X). Note that it is usually very uninformative. A more useful description of a random variables dispersion is given by the variance. Intuitively, the variance of a random variable measures how much the random variable X typically" deviates from its typical" value (or distance from the population mean). Mathematically: Var [X ] = E [(X E [X ])2 ]. 2 = E [(X E [X ])2 ] = E [X ])2 f (x ) if X is discrete (x E [X ])2 f (x ) x if X is continuous x

x (x

Monotonic Transformations

Expectations

Measures of Dispersion

i (xi

E [X ])f (x ) = = = =

i xi f (x ) i E [X ]f (x ) E [X ] E [X ] i f (x )7 E [X ] E [X ]8 0.

In other words, the average deviation of X from its average is always 0; values above the mean cancel out values below the mean by denition of the mean. Consequently, we need to square the difference. Note that we could also take the absolute difference but that is harder" to work with mathematically.

7 8

By denition of E[X] and since E[X] is a constant respectively. Since i f (x ) = 1 by denition of probability.

Monotonic Transformations

Expectations

Measures of Dispersion

EXAMPLE x f (x ) 1 .2 2 .2 3 .2 4 .2 5 .2

E [X ] = 1(.2) + 2(.2) + 3(.2) + 4(.2) + 5(.2) = .2 + .4 + .6 + .8 + 1 = 3 Var [X ] = E [(X E [X ])2 ] = (1 3)2 (.2) + (2 3)2 (.2) + (3 3)2 (.2) + (4 3)2 (.2) + (5 3)2 (.2) = 2.5

Monotonic Transformations

Expectations

Measures of Dispersion

We can also derive another expression for the variance. E [(X E [X ])2 ] = = = = = E [X 2 2XE [X ] + (E [X ])2 ] E [X 2 ] 2E [XE [X ]] + E [(E [X ])2 ] E [X 2 ] 2E [X ]E [X ] + E [(E [X ])2 ] E [X 2 ] 2E [X ]2 + E [X ]2 E [X 2 ] E [X ]2

Monotonic Transformations

Expectations

Measures of Dispersion

EXAMPLE Consider the continuous random variable with the PDF. Calculate Var[X]. f (x ) = 2(1 x ) 0

1 0

= E [X ] = = = = = = E [X ]2 =

x 2(1 x ) x

1 2 0 x x x 1 1 2 0 x x 0 x x 1 2 1 1 3 1 x ]0 3 x ]0 2 1 1 3 2 1 1 = 3 6 2 1 9

2 2 2 2 2

1 3

Monotonic Transformations

Expectations

Measures of Dispersion

E [X 2 ] = = = = = = 2 = Var [X ]

1 0

x 2 2(1 x ) x

1 2 3 0 x x x 1 2 1 3 0 x x 0 x x 1 3 1 x ]0 1 x 4 ]1 0 3 4 1 1 3 4 1 1 = 6 12

2 2 2 2 2

= E [X 2 ] E [X ] 2 = 1 1 6 9 1 = 18

Monotonic Transformations

Expectations

Measures of Dispersion

The interpretation of the variance may not be that intuitive, as it tells us the average squared deviation of a random variable X. Consequently, if X is measured in dollars, Var[X] is measures in squared dollars. To make interpretation easier, we often times refer to the standard deviation of a random variable. The standard deviation is simply the square root of the variance. More generally, we can characterize a random variable using a moment generating function. The moment generating function allows us to dene a sequence of moments which can completely characterize the probability distribution.

The kth moment around zero is dened as E [0 E [X ]]k or E [X ]k . Note that the rst moment about zero is the mean: E[X]. The kth moment around the mean is dened as: E [(X E [X ])k ]. The second moment about the mean is the variance.

Monotonic Transformations

Expectations

Measures of Dispersion

Why is this terminology useful? Because it provides a common framework for talking about our measures of central tendency and dispersion. Higher moments about the mean also have special terms associated with them. The third moment around the mean E [(X E [X ])3 ] is called the skew of the distribution. The skew tells us whether the dispersion about the mean is symmetric (if skew = 0 ), or if it is negatively skewed (if skew < 0; implying that E[X] < median(X)) or positively skewed (if skew > 0; implying that E[X] > median(X)).

Monotonic Transformations

Expectations

Measures of Dispersion

Skewness

Positive Skew

0.04 EX med[X] Density 0.02 0.00

50

100 x

150

Symmetric

0.015 0.010

Density

EX

0.005

med[X]

0.000

50

100 x

150

Negative Skew

0.04

EX med[X]

Density

0.00

0.02

50

100 x

150

Monotonic Transformations

Expectations

Measures of Dispersion

Kurtosis

The fourth moment about the mean E [(X E [X ])4 ] is known as the kurtosis and it measures how thick the tails" of the distribution are.9

Monotonic Transformations

Expectations

Measures of Dispersion

Kurtosis

0.020

0.015

Density

Density 0 50 x 150

0.010

0.005

0.000

0.000 0 50 x

0.005

0.010

0.015

0.020

150

Monotonic Transformations

Expectations

Measures of Dispersion

Example: Normal Distribution

EXAMPLE The Normal Distribution (also sometimes called the Gaussian Distribution) is completely characterized by two moments. The PDF is: 2 1 (x ) 1 f (x |, 2 ) = e 2 2 2 where E [X ] = and Var [X ] = 2 . Since two parameters dene a normal distribution, we denote this by: X N (, 2 ). In doing empirical work, it is always important to understand what the characteristics are of the data you are working with. A good rst step is to calculate the summary statistics of variables you work with as well as plot them out so you can visualize what you are working with. Doing so will also (hopefully) help you catch errors and identify outliers for further investigation.

Monotonic Transformations

Expectations

Measures of Dispersion

If X is a random variable with nite variance, then for any constants a and b, Var [aX + b] = = = = = E [(aX + b) E [(aX + b)]2 E [aX + b aE [X ] b]2 E [aX aE [X ]]2 a2 E [X E [X ]]2 a2 Var [X ]

- G R Dodagoudar_reliability Pap_2011Uploaded bydodagoudar@gmail.com
- Error Function Table PDFUploaded byTamika
- cs1538Uploaded byAchmad Hidayat
- Answer MSE1sttestUploaded bySneha Kumar B Positive
- Statistics for Business and Economics: bab 5Uploaded bybalo
- Mth514NotesCh4,5,6,10Uploaded byt_man222
- Ma2b12rec1-12Uploaded byMichael Yurko
- I02620020220115066Random Variables and Probability DistributionsUploaded bySara Zalanderi
- Lecture713-03-4Uploaded byrao253
- Lecture 1Uploaded byhyperpearl
- Cats GuideUploaded byRodney Pereira
- Cont ExampUploaded bySumedh Kakde
- Dip Image InhancmentUploaded byMatthew Wagner
- MonteCarloUploaded byPutri Suardani
- Chapter 7 (1)Uploaded bymaustro
- Solución 2.3Uploaded byLina Maria Jimenez
- W3INSE6220.pdfUploaded bypicala
- Digital Image ProcessingUploaded byAnushka Bajpai
- ES9-Chap06Uploaded bynikowawa
- Random Variables Tarea TeoríaUploaded byMarco Espinosa
- Computational Geosciences With Mathematica [William C. Haneberg, 2004] @Geo PediaUploaded byPhạm Nam
- ma2110_rv_assgn_2Uploaded bytarak
- MATH30-8 ME 1st Q 2015-16Uploaded byRon Inagan
- Maths QuestionsUploaded byabhishek kumar
- psheet7Uploaded byMichael Corleone
- Continuous 1Uploaded byHanif Mohammad
- CHAPTER 1Uploaded bySalihah Abdullah
- 4 Discrete Time Random ProcessesUploaded byBomber Killer
- MathExpectUploaded byOput Magfirah Yanuar Putra
- Final ExamUploaded byPeter Hua

- The Game FrameUploaded byjxkjxk
- Global CorporateUploaded bymhyigit
- Convergent and Divergent ThinkingUploaded bydons
- 2015 Hate Crime Report PDFUploaded byJD
- Vindhyavasini Cynthia HumeUploaded byphil_hine
- nullUploaded byapi-27788847
- LUploaded byMariaGrazia Pirruccio
- Evolutionary Psychology. an Emerging Integrative Perspective Within the Science ... Psychology - Kennair (2002)Uploaded byEduardo Aguirre Dávila
- Devi Khadgamala With MeaningUploaded byAanandkrishna
- Howard Barker InterviewUploaded byMike Pugsley
- aUploaded byMohammad Rameez
- LinUploaded byEpic Win
- Full Compilation Qawaid FiqhiyyahUploaded byanis suraya mohamed said
- Exp6 Superposition and Reciprocity (1)Uploaded byDefer
- Testing Theories of American PoliticsUploaded byAnonymous gJ7KQOv6NE
- Cyber Shaman Manual 1Uploaded byClóvis Charlanti
- Seth in the Magickal TextsUploaded byIkenna Ohiaeri
- 971 Horses + 4 Zebras by Lilly HusbandsUploaded byAnimate Projects
- Limits of the law Human Rights and the Emancipation of SexualUploaded bySilvino González Morales
- Baha'I- Inspired Perspectives on Human RightsUploaded byMassovia
- nop landscapedUploaded byapi-232087103
- A Companion to Federico Garc a Lorca Monograf as aUploaded byUlises Bravo
- Torts 2 Outline.docxUploaded byMarinéFigueroa
- HIlbro SurgicalUploaded byshazay_77
- a separate peace essayUploaded byapi-317509912
- Empowerments of the Kagyu Ngak DzodUploaded byJustin Von Bujdoss
- Brassier-Badiou's Materialist Epistemology of Mathematics (1)Uploaded byDaniel Sacilotto
- 488329Uploaded byAlex Luppe
- A History of the Jews and Judaism in the Second Temple Period the Early Tic Period 335 175 BCE Lester GrabbeUploaded byDavid Bailey
- Session break down.pdfUploaded byDiego Baldi

## Much more than documents.

Discover everything Scribd has to offer, including books and audiobooks from major publishers.

Cancel anytime.