7 views

Uploaded by modparadigm

foo

save

- Convenience Sample to Support Pronvenience Sample to Support Probability Sample
- Lecture 2
- Basic Statistics
- Estimation
- Chapter2_solutions.doc
- sluo2006lec4
- CT3_QP_1111.pdf
- npsDB65
- A11n39_2
- Almost Unbiased Ratio and Product Type Estimator of Finite Population Variance Using the Knowledge of Kurtosis of an Auxiliary Variable in Sample Surveys
- Unit 1
- Interval Estimation
- JLarge Slides MT123 2011
- 2paperandmcqs
- A Weight Smoothing Method for Dealing With
- Constant Expected Return
- Probability
- Properties
- 01.estimation.pdf
- RANDOMNESS&OPTIMAL ESTIMATION, by M. Khoshnevisan, S. Saxena, H. P. Singh, S. Singh, F. Smarandache
- KM 1.1 Vrije Universiteit Amsterdam Summary
- tema1
- Chapter Two - Estimators.2
- Estimating Entropy of Data Streams Using Compressed Counting
- Bayes’ estimators of generalized entropies
- DIS_ch_9
- Applied Statistics and Probability for Engineers, 5th edition
- Estimation of Weibull Shape Parameter by Shrinkage Towards An Interval Under Failure Censored Sampling
- 407-2267-2-PB
- A Study on the application of 6-Sigma on the enhancement of service quality of fitness club
- Notes
- SPN1130 OralExam1 Preparation
- DCalc.pdf
- Foo
- TODO Today
- Math 4413 Syl
- foo bar.pdf
- Ch4
- Math 4413 Syl
- Finding the Equation of a Tangent Line.pdf

You are on page 1of 4

**1. Let X1 and X2 be independent random variables each Bernoulli(θ) where
**

θ ∈ {1/4, 3/4}. Let D = Θ and assume squared error loss. Let ∆ be all non-

randomized decision rules base on (X1 , X2 ). Let ∆T be all non-randomized

decision rules base on T = X1 + X2 . Show that there is a member of ∆ − ∆T

which is not dominated by any member of ∆T .

2. Let X1 , . . . , Xn be iid each Poisson(λ) where λ ∈ (0, +∞) = Θ. Let

D = [0, +∞) and assume convex loss. Show that if γ(λ) has an unbiased

estimator then it can be expressed as a power series in λ. Find the best unbiased

estimator of γ(λ) = λi where i is some poiitive integer. Show that the best

unbiased esitmator of

γ(λ) = Pλ (X1 = k) = exp(−λ)λk /k!

is

1 t

δ(t) = k (1 − 1/n)t−k for 0≤k≤t

n k (1)

=0 otherwise

Pn

where T = i=1 Xi

3. Let X be Poisson(λ) where λ ∈ {0, 1, . . . }. Show that X is not complete.

4. Let X1 , . . . , Xn be iid each uniform on {1, 2, . . . , N } where N ∈ {1, 2, . . .} =

Θ. Show that T (X1 , . . . , Xn ) = max Xi is complete and sufficient. Let the de-

cision space D be the set of real numbers. Let Γ be the class of all functions

defined on Θ having unbiased estimators. Show that Γ is the class of all real-

valued functions defined on Θ. Given a γ ∈ Γ find an explicit representation for

its unbiased estimator. Apply this to the case where γ(N ) = N .

5. Let X be a random variable with

θ

fθ = (1/2)θ for x = 0, 1, . . . , θ

x (2)

=0 otherwise

**We wish to estimate γ(θ) = θ with squared error loss and with the decision
**

space D equal to the set of real numbers.

i) If Θ = {0, 1, 2, . . . } show that X is complete and sufficient and find the

best unbiased estimator of θ.

ii) If Θ = {1, 2, . . .} show that X is no longer complete and that there does

not exist a best unbiased estimator for θ.

iii) If Θ = {1, 2, . . .} and D = {1, 2, . . .} show that a best unbiased estimator

of θ exists. Find this estimator.

6. Consider a coin with unknown probability θ ∈ [0, 1] of coming up heads.

Let a and n be known integers satisfying 1 < a < n. Consider the experiment

which consists of the tossing the coin independently until a heads are observed

or n tosses are completed (i.e. the coin is tossed at most n times).

1

. When no change point exist we take as our prior for λ the exponential distribution (i. x2 ) up to the normalizing constant. 1) it again moves a unit to the right or up and continues in this way until it reaches a stopping point. 1) such that on [0. A fraction θ of the coins is of Type I where each has probability λ of coming up heads on a given toss while the remaining fraction of the coins is of a second type. say Type II. b) and λ ∼ beta(α. Consider the following idealized situation.1) independent of the priors for λ1 and λ2 . The walk starts at (0. For 0 < x < 1 let k(x) be the number of events that occurred in [0. An urn contains an infinite number of coins. assumed to be known. ii) Consider the statistic T which is the total number of tails observed when the the random walk stops. With prob- ability θ. Suppose events are happening in time over the unit interval. From the resulting point. Given that a change point does exist our prior distribution for where it occurs is uniform(0. 1] the events follow a Poisson process with parameter λ2 . Suppose we have have observed the process over the unit interval and the data consists of k events occurring at times 0 < t1 < t2 < · · · < tk < 1. i) Show that under this model the probability of observing these data is θ/2k+1 + (1 − θ)φ(data) 2 . g(λ) = exp(−λ) for λ > 0). ii) Now suppose that all of the coins that turned up heads on the first set of tosses are all tossed again and this time there were x2 tails observed where 0 < x2 < N − x1 . Then suppose that each of the selected coins are tossed and it is observed that x1 of the tosses resulted in a tail where (0 < x1 < N ). where each has probability 1 of coming up heads when tossed. iv) Find the best unbiased estimator of θ. 8. iii) Prove that T is complete for this model. Assume that the prior distributions for θ and λ are independent beta distributions where θ ∼ beta(a. When a change point does exist our prior for λ1 and λ2 are independent exponentials. x). Find the probability distribution of T (as a function of θ) and show that T is sufficient for this model. the distribution of the events follows a Poisson process with intensity parameter λ or with probability 1−θ there is some change point x ∈ (0. Suppose N coins are selected at random from the urn and let N1 be the number of coins that are of Type I. β) i) Find the posterior distribution (up to the normalizing constant) of p(N1 | x1 ). (1.e. 7. 0) and moves one unit to the right or up according to whether the first toss is a head or a tail. Now find p(N1 | x1 . The two types of coins are indistinguishable however. 0) or (0. x) the events follow a Poisson process with pa- rameter λ1 and on [x. i) Give the probability of an outcome which contained h heads and t tails in some specified order. The outcomes of such an experiment can be represented as a random walk in the plane.

Find the Bayes estimator for the data problem for each of the two loss functions. d) = exp[a(d − θ)] − a(d − θ) − 1 where a 6= 0 is a fixed. d) = (θ − d)2 + C 2 φ(d) where C > 0 is a fixed known real number and 0 if d = 0 φ(d) = 1 if d 6= 0 i) Let g be a prior density for θ with mean µ. variance σ 2 and moment generating function ψ. Show that there exist θ0 such that Eθ ψ(X) ≤ 0 for θ < θ0 and Eθ ψ(X) > 0 for θ > θ0 unless Eθ ψ(X) is either positive for all θ or negative for all θ. σ 2 ). for large values of x. Let g(θ) = exp −α(θ) be a prior distribution for −∞ < θ < ∞. 9. there exists a value x0 such that ψ(x) ≤ 0 for x < x0 and ψ(x) ≥ 0 for x ≥ x0 . 11. 3 . That is. Assume Θ is an open interval of real numbers. p(θ | x).where 1 xk(x) (1 − x)k−k(x) Z φ(data) = k(x)+1 dx 0 (1 + x) (2 − x)k−k(x)+1 ii) Find the posterior probability that a change point exists. Assume that this family of densities has the monotone likelihood ratio property in X. We are interested in studying the behavior of the posterior distribution of θ given x. iii) Briefly describe scenarios where the two loss functions would be more reasonable than squared error loss. Let X be a real valued random variable with {fθ : θ ∈ Θ} a family of everywhere positive densities. ii)Now consider the data problem where the distribution of X given θ is Normal(θ. known real number and L2 (θ.1) and the prior g is Normal(µ. Let X be a normal random variable with mean θ and variance equal to one. Hint: let θ1 < θ2 and show that if Eθ1 (X) > 0 then Eθ2 (X) > 0. For each of the loss functions find the Bayes estimator for θ for the no data problem for the prior g. 10. Hint: let θ1 < θ2 and consider the sets A = {x : fθ2 (x) < fθ1 (x)} and B = {x : fθ2 (x) > fθ1 (x)} ii) Now let ψ be a function with a single change of sign. We will consider two possible loss functions L1 (θ. Consider a decision problem where both the parameter space Θ and the decision space D are the set of real numbers. i) Show that if ψ is a non-decreasing function of x then Eθ ψ(X) is a non- decreasing function of θ.

i) Show that for x sufficiently large p(θ | x). What do the above results imply about the behavior of δg (x) for large values of x. For such a y find the posterior distribution of θ given y. λ0 (θ) ≤ M and for θ > K. exists and that there exist constants K and M such that for θ ≤ K. ii) For x > K let y = (λ0 )−1 (x). c2 and c3 such that for y > c3 fy (z) ≤ c1 exp(−c2 z 2 ) for z ∈ (−∞. iv) Let δg (x) be the Bayes estimate of θ against the prior g when the loss is squared error. 4 . Let fy be the posterior density of θ − y given y. ∞) Hint: It may be useful to expand α(y + z) in a Taylor series about y. Show that for every real number z r 1+c z2 lim fy (z) = exp {− (1 + c)} y→∞ 2π 2 In addition show that there exist constants c1 . the derivative of λ. λ0 is strictly increasing and everywhere greater than M . as a function of θ is maximized at θ = (λ0 )−1 (x) where (λ0 )−1 is the inverse of λ0 on the set θ > K. Let λ(θ) = θ2 /2 + α(θ). iii) Suppose that α00 (·) exists for all θ and that infθ α00 (θ) ≥ a > −1 and limθ→∞ α00 (θ) = c where −1 < c < ∞. We suppose that λ0 .

- Convenience Sample to Support Pronvenience Sample to Support Probability SampleUploaded bydiana.yanti
- Lecture 2Uploaded byDingzhou Li
- Basic StatisticsUploaded byzulkfal
- EstimationUploaded byfaisal gazi
- Chapter2_solutions.docUploaded byMiguel Gonzalez
- sluo2006lec4Uploaded byBoda Kishan
- CT3_QP_1111.pdfUploaded byNitai Chandra Ganguly
- npsDB65Uploaded byAUCOM
- A11n39_2Uploaded byabdelaziz_ismail685662
- Almost Unbiased Ratio and Product Type Estimator of Finite Population Variance Using the Knowledge of Kurtosis of an Auxiliary Variable in Sample SurveysUploaded byscience2010
- Unit 1Uploaded byMarian Nelson
- Interval EstimationUploaded byAntariksha Ganguly
- JLarge Slides MT123 2011Uploaded byleafko
- 2paperandmcqsUploaded byUsman Tahir
- A Weight Smoothing Method for Dealing WithUploaded byraychell_santos2801
- Constant Expected ReturnUploaded byCristian Núñez Clausen
- ProbabilityUploaded byKanchana Randall
- PropertiesUploaded byali_mudassar
- 01.estimation.pdfUploaded byShafiullaShaik
- RANDOMNESS&OPTIMAL ESTIMATION, by M. Khoshnevisan, S. Saxena, H. P. Singh, S. Singh, F. SmarandacheUploaded bymarinescu
- KM 1.1 Vrije Universiteit Amsterdam SummaryUploaded bymaensh
- tema1Uploaded byalex
- Chapter Two - Estimators.2Uploaded bybshankar481
- Estimating Entropy of Data Streams Using Compressed CountingUploaded byterminatory808
- Bayes’ estimators of generalized entropiesUploaded bysleepanon4362
- DIS_ch_9Uploaded byVictoria Liendo
- Applied Statistics and Probability for Engineers, 5th editionUploaded byMuhammadHazmiMokhtar
- Estimation of Weibull Shape Parameter by Shrinkage Towards An Interval Under Failure Censored SamplingUploaded byscience2010
- 407-2267-2-PBUploaded byCibi Kulandaisamy
- A Study on the application of 6-Sigma on the enhancement of service quality of fitness clubUploaded byWong Chee Haur

- NotesUploaded bymodparadigm
- SPN1130 OralExam1 PreparationUploaded bymodparadigm
- DCalc.pdfUploaded bymodparadigm
- FooUploaded bymodparadigm
- TODO TodayUploaded bymodparadigm
- Math 4413 SylUploaded bymodparadigm
- foo bar.pdfUploaded bymodparadigm
- Ch4Uploaded bymodparadigm
- Math 4413 SylUploaded bymodparadigm
- Finding the Equation of a Tangent Line.pdfUploaded bymodparadigm