Professional Documents
Culture Documents
Multiple Choice Question Bank For BUEC 333
Multiple Choice Question Bank For BUEC 333
This set of multiple choice questions has been prepared to supply you with a means of
checking your command of the course material. For each week of the course I have tried
to provide questions that cover the full range of material discussed during that week. I
have also tried not to include “duplicate” questions, namely questions that are basically
the same as other questions but, for example, just use different numbers. The questions
range quite widely in character. Some check your knowledge of definitions, some address
implications of concepts, some require finding numbers in tables, some ask you to do
calculations, some are very easy, and some are very difficult. Because of this, not all
questions are such that they are likely to appear on an exam, so how you score on these
questions is not as important as how well you understand the logic of the answers.
Spending quality time on these questions should help your efforts to learn this course
material. My advice is to work through each chapter’s questions after you believe you
have a good command of that chapter’s material, and then if there are any questions for
which you don’t understand why the answer provided is the best answer, make sure you
find out why.
These questions have not been tested, so there may be problems with them. Some
questions may be vague or defective, and some answers may be incorrect. Please bring to
my attention any questions that are messed up.
x 1 2 3
Prob(x) .1 .2 k
1. The value of k is
a) .3 b) .5 c) .7 d) indeterminate
4. The variance of x is
a) 0.44 b) 0.66 c) 4.6 d) indeterminate
5. If all the x values were increased by 5 in this table, then the answer to question 2
would be
a) unchanged b) increased by 5 c) multiplied by 5 d) indeterminate
2
6. If all the x values were increased by 5 in this table, then the answer to question 3
would be
a) unchanged b) increased by 25 c) multiplied by 25 d) none of the above
7. If all the x values were increased by 5 in this table, then the answer to question 4
would be
a) unchanged b) increased by 25 c) multiplied by 25 d) none of the above
8. If all the x values were multiplied by 5 in this table, then the answer to question 2
would be
b) unchanged b) increased by 5 c) multiplied by 5 d) indeterminate
9. If all the x values were multiplied by 5 in this table, then the answer to question 3
would be
b) unchanged b) increased by 25 c) multiplied by 25 d) none of the above
10. If all the x values were multiplied by 5 in this table, then the answer to question 4
would be
a) unchanged b) increased by 25 c) multiplied by 25 d) none of the above
The next 17 questions refer to variables X and Y with the following joint distribution
prob(X,Y)
13. If I don’t know anything about the value of Y, then the probability that X=3 is
a) .1 b) .2 c) .4 d) .6
20. If all the X values in the table above were increased by 8, then the answer to question
18 would be
a) unchanged b) increased by 8 c) multiplied by 8 d) multiplied by 64
21. If all the X values in the table above were increased by 8, then the answer to question
19 would be
a) unchanged b) increased by 8 c) multiplied by 8 d) multiplied by 64
22. If all the X values and all the Y values in the table above were increased by 8, then
the answer to question 18 would be
a) unchanged b) increased by 8 c) multiplied by 8 d) multiplied by 64
23. If all the X values and all the Y values in the table above were increased by 8, then
the answer to question 19 would be
a) unchanged b) increased by 8 c) multiplied by 8 d) multiplied by 64
24. If all the X values in the table above were multiplied by 8, then the answer to question
18 would be
a) unchanged b) increased by 8 c) multiplied by 8 d) multiplied by 64
25. If all the X values in the table above were multiplied by 8, then the answer to question
19 would be
a) unchanged b) increased by 8 c) multiplied by 8 d) multiplied by 64
26. If all the X values and all the Y values in the table above were multiplied by 8, then
the answer to question 18 would be
a) unchanged b) increased by 8 c) multiplied by 8 d) multiplied by 64
27. If all the X values and all the Y values in the table above were multiplied by 8, then
the answer to question 19 would be
a) unchanged b) increased by 8 c) multiplied by 8 d) multiplied by 64
28. The distribution of X when Y is known is called the ________ distribution of X, and
is written as ________. These blanks are best filled with
a) conditional, p(X) b) conditional, p(X |Y)
c) marginal, p(X) d) marginal, p(X |Y)
4
29. The distribution of X when Y is not known is called the ________ distribution of X,
and is written as ________. These blanks are best filled with
a) conditional, p(X) b) conditional, p(X |Y)
c) marginal, p(X) d) marginal, p(X |Y)
The next 5 questions refer to the following information. You have estimated the equation
wage = alphahat + betahat*experience to predict a person’s wage using years of
experience as an explanatory variable. Your results are that alphahat is 5.0 with standard
error 0.8, betahat is 1.2 with standard error 0.1, and the estimated covariance between
alphahat and betahat is –0.005. What this means is that 5.0 is a realization of a random
variable with unknown mean and standard error 1.0, and 1.2 is a realization of another
random variable which has unknown mean and standard error 0.01.
30. The estimated variance of your forecast of the wage of a person with no experience is
a) 0.64 b) 0.8 c) 0.81 d) none of these
31. The estimated variance of your forecast of the wage of a person with one year of
experience is
a) 0.01 b) 0.64 c) 0.65 d) none of these
32. The estimated variance of your forecast of the wage of a person with two years of
experience is
a) 0.64 b) 0.65 c) 0.66 d) 0.67
33. The estimate of the increase in wage enjoyed by a person with three additional years
of experience is
a) 3.6 b) 8.6 c) 15 d) none of these
34. The estimated variance of the estimate of the increase in wage enjoyed by a person
with three additional years of experience is
a) 0.01 b) 0.03 c) 0.09 d) none of these
The next 9 questions refer to the following information. The percentage returns from
stocks A, B, and C are random variables with means 0.05, 0.08, and 0.12 respectively,
and variances 0.04, 0.09, and 0.16, respectively. The covariance between A and B returns
is minus 0.01; the return from stock C is independent of the other two. A GIC is available
with a guaranteed return of 0.03.
35. If you buy a thousand dollars each of A and B, your expected percentage return for
this portfolio is
a) 0.05 b) 0.065 c) 0.08 d) none of these
36. If you buy a thousand dollars each of A and B, the variance of your percentage return
for this portfolio is
a) 0.11 b) 0.12 c) 0.13 d) none of these
5
37. If you buy a thousand dollars of A and two thousand dollars of B, your expected
percentage return for this portfolio is
a) 0.05 b) 0.07 c) 0.08 d) none of these
38. If you buy a thousand dollars of A and two thousand dollars of B, the variance of
your percentage return for this portfolio is
a) 0.04 b) 0.044 c) 0.73 d) none of these
39. If you were to supplement either of the above portfolios with some of stock C your
expected return should go ____ and if you were to supplement with some GIC your
expected return should go ____. The best ways to fill these blanks are
a) up, up b) up, down c) down, up d) down, down
40. If you were to supplement either of the above portfolios with some GIC the variance
of your return should
a) increase
b) decrease
c) remain unchanged
d) can’t tell what will happen
41. If you were to supplement either of the above portfolios with some of stock C, the
variance of your return should
a) increase
b) decrease
c) remain unchanged
d) can’t tell what will happen
42. Suppose you bought a thousand dollars of each of A, B, C and GIC. The expected
return of this portfolio is
a) .0625 b) .07 c) .087 d) none of these
43. Suppose you bought a thousand dollars of each of A, B, C and GIC. The variance of
the return of this portfolio is
a) .017 b) .068 c) .075 d) .27
44. Suppose we have a sample of size 100 from a random variable x with mean 3 and
variance 4. The standard deviation of xbar, the average of our sample values, is
a) 0.04 b) 0.2 c) 2 d) 4
45. You have obtained the following data on the wages of randomly-obtained
observationally-identical teenagers: 7, 8, 8, 7, 9, 8, 10, 8, 7, 8, 8. You calculate the
average as 8 and intend to report this figure; you also want to provide a confidence
interval but to do this you have to estimate the standard error of this average. The
estimated standard error you should use is approximately the square root of
a) 0.073 b) 0.08 c) 0.8 d) none of these
6
46. From a sample of size 300 you have estimated the percentage of workers who have
experienced an injury on the job last year to be six percent. You wish to report this
figure but you also want to provide a confidence interval. To do this you need to
estimate the standard error of this estimate. The estimated standard error you should
use is approximately
a) 0.0002 b) 0.014 c) 0.056 d) none of these
47. A negative covariance between x and y means that whenever we obtain an x value
that is greater than the mean of x
a) we will obtain a corresponding y value smaller than the mean of y
b) we will obtain a corresponding y value greater than the mean of y
c) we have a greater than fifty percent chance of obtaining a corresponding y value
smaller than the mean of y
d) we have a greater than fifty percent chance of obtaining a corresponding y value
greater than the mean of y
48. The central limit theorem assures us that the sampling distribution of the mean
a) is always normal
b) is always normal for large sample sizes
c) approaches normality as the sample size increases
d) appears normal only when the sample size exceeds 1,000
49. For a variable x the standard error of the sample mean is calculated as 20 when
samples of size 25 are taken and as 10 when samples of size 100 are taken. A
quadrupling of sample size has only halved the standard error. We can conclude that
increasing sample size is
a) always cost effective b) sometimes cost effective c) never cost effective
50. In the preceding question, what must be the value of the standard error of x?
a) 1000 b) 500 c) 377.5 d) 100
51. Suppose a random variable x has distribution given by f(x) = 2x, for 0 ≤ x ≤ 1 and
zero elsewhere. The expected value of x is
a) less than 0.5 b) equal to 0.5 c) greater than 0.5 d) indeterminate
52. Suppose a random variable x has distribution given by f(x) = kx, for 0 ≤ x ≤ 2 and
zero elsewhere. The value of k is
a) 0.5 b) 1.0 c) 2.0 d) indeterminate
7
1. Suppose that if the null that beta equals one is true a test statistic you have calculated
is distributed as a t statistic with 17 degrees of freedom. What critical value cuts off
5% of the upper tail of this distribution?
a) 1.65 b) 1.74 c) 1.96 d) 2.11
2. Suppose that in the previous question beta is equal to 1.2. Then the critical value from
the previous question will cut off ______ of the upper tail of the distribution of your
test statistic. The blank is best filled with
a) less than 5% b) 5% c) more than 5%
3. Suppose that if the null that alpha and beta both equal one is true a test statistic you
have calculated is distributed as a chi-square statistic with 2 degrees of freedom.
What critical value cuts off 5% of the upper tail of this distribution?
a) 3.84 b) 5.02 c) 5.99 d) 7.38
4. Suppose that if the null that alpha and beta both equal one is true a test statistic you
have calculated is distributed as an F statistic with 2 and 22 degrees of freedom for
the numerator and denominator respectively. What critical value cuts off 5% of the
upper tail of this distribution?
b) 3.00 b) 3.44 c) 4.30 d) 5.72
5. Suppose that if the null that beta equals one is true a test statistic you have calculated
is distributed as a z (standard normal) statistic. What critical value cuts off 5% of the
upper tail of this distribution?
a) 0.31 b) 0.48 c) 1.65 b) 2.57
6. Suppose that if the null that beta equals one is true a test statistic you have calculated
is distributed as a z (standard normal) statistic. If you choose 1.75 as your critical
value, what is your (one-sided) type I error probability?
a) 4% b) 5% c) 6% d) 7%
7. Suppose that if the null that beta equals one is true a test statistic you have calculated
is distributed as a z (standard normal) statistic. If you choose 1.28 as your critical
value, what is your (two-sided) type I error probability?
a) 5% b) 10% c) 15% d) 20%
8. A type I error is
a) failing to reject the null when it is false
b) rejecting the null when it is true
8
13. A power curve graphs the degree of falseness of the null against
a) the type I error probability
b) the type II error probability
c) one minus the type I error probability
d) one minus the type II error probability
15. Other things equal, when the sample size increases the power curve
a) flattens out
b) becomes steeper
c) is unaffected
16. Other things equal, when the type I error probability is increased the power curve
a) shifts up b) shifts down c) is unaffected
9
18. A manufacturer has had to recall several models due to problems not discovered with
its random final inspection procedures. This is an example of
a) a type I error b) a type II error c) both types of error d) neither type of error
19. As the sample size becomes larger, the type I error probability
a) increases b) decreases c) does not change d) can’t tell
20. Consider the following two statements: a) If you reject a null using a one-tailed test,
then you will also reject it using a two-tailed test at the same significance level; b) For a
given level of significance, the critical value of t gets closer to zero as the sample size
increases.
a) both statements are true b) neither statement is true
c) only the first statement is true d) only the second statement is true
24. In the Eviews software, in the OLS output the intercept estimate by default is
a) printed last and called “I” for “intercept”
b) printed first and called “I”
c) printed last and called “C” (for “constant”)
d) printed first and called “C”
10
25. A newspaper reports a poll estimating the proportion u of the adult population in
favor of a proposition as 65%, but qualifies this result by saying that “this result is
accurate within plus or minus 3 percentage points, 19 times out of twenty.” What does
this mean?
a) the probablilty is 95% that u lies between 62% and 68%
b) the probability is 95% that u is equal to 65%
c) 95% of estimates calculated from samples of this size will lie between 62% and 68%
d) none of the above
26. In the Eviews software, when you run an OLS regression by clicking on buttons, the
parameter estimates are put in a vector called
a) c (for “coefficient vector”) with the first element in this vector the intercept estimate
b) c (for “coefficient vector”) with the last element in this vector the intercept estimate
c) b (for “beta vector”) with the first element in this vector the intercept estimate
d) b (for “beta vector”) with the last element in this vector the intercept estimate
27. A newspaper reports a poll of 400 people estimating the proportion u of the adult
population in favor of a proposition as 60%, but qualifies this result by saying that “this
result is accurate within plus or minus x percentage points, 19 times out of twenty.” The
value of x in this case is about
a) 2 b) 3 c) 4 d) 5
28. In the Eviews software, in the OLS output the far right column reports
a) the coefficient estimate b) the standard error c) the t value d) none of these
30. When you calculate a 95% confidence interval for an unknown parameter beta, the
interpretation of this interval is that
a) the probability that the true value of beta lies in this interval is 95%
b) 95% of repeated calculations of estimates of beta from different samples will lie in
this interval
c) 95% of intervals computed in this way will cover the true value of beta
d) none of the above
31. Suppose from a very large sample you have estimated a parameter beta as 2.80 with
estimated variance 0.25. Your 90% confidence interval for beta is 2.80 plus or minus
approximately
a) 0.41 b) 0.49 c) 0. 82 d) 0.98
11
The next 8 questions refer to the following information. You have an estimate 1.75 of a
slope coefficient which you know is distributed normally with unknown mean beta and
known variance 0.25. You wish to test the null that beta = 1 against the alternative that
beta > 1 at the 10% significance level.
33. You should _____ the null. If you had used a 5% significance level you would
______ the null. The blanks are best filled with
a) accept; accept b) accept; reject c) reject; accept d) reject; reject
35. If the true value of beta is 1.01, the power of your test is approximately
a) 1% b) 5% c) 10% d) nowhere near these values
36. If the true value of beta is 10.01, the power of your test is approximately
a) 1% b) 5% c) 10% d) nowhere near these values
37. If the true value of beta is 1.75, the power of your test is approximately
a) 10% b) 40% c) 60% d) 90%
38. If the true value of beta is 1.65, the power of your test is approximately
a) 10% b) 50% c) 70% d) 90%
39. If the true value of beta is 1.25, the power of your test is approximately
a) 22% b) 40% c) 60% d) 78%
12
The next 3 questions refer to the following information. Suppose the regression
specification y = α + βx + δz + ε was estimated as y = 2 + 3x + 4z. We have a new
observation for which x = 5 and z = -2. For this new observation
14. For the regression specification y = α + βx + ε the OLS estimates result from
minimizing the sum of
a) (α + βx)2
b) (α + βx + ε)2
c) (y - α + βx)2
d) none of these
15. For the regression specification y = α + βx + ε a computer search to find the OLS
estimates would search over all values of
a) x b) α and β c) α, β, and x d) α, β, x, and y
19. When there are more explanatory variables the adjustment of R-square to create
adjusted R-square is
a) bigger b) smaller c) unaffected
20. Compared to estimates obtained by minimizing the sum of absolute errors, OLS
estimates are _______ to outliers. The blank is best filled with
a) more sensitive b) equally sensitive c) less sensitive
22. R-squared is
a) The minimized sum of squared errors as a fraction of the total sum of squared errors.
b) The sum of squared errors as a fraction of the total variation in the dependent
variable.
c) One minus the answer in a).
d) One minus the answer in b).
23. You have 46 observations on y (average value 15) and on x (average value 8) and
from an OLS regression have estimated the slope of x to be 2.0. Your estimate of the
mean of y conditional on x is
a) 15 b) 16 c) 17 d) none of the above
The following relates to the next two questions. Suppose we have obtained the following
regression results using observations on 87 individuals: yhat = 3 + 5x where the standard
errors of the intercept and slope are 1 and 2, respectively.
24. If an individual increases her x value by 4, what impact do you predict this will have
on her y value? Up by
a) 4 b) 5 c) 20 d) 23
15
26. Suppose wage = α + βage + ε and we have 100 observations on wage and age, with
average values 70 and 30, respectively. We have run a regression to estimate the slope of
x as 2.0. Consider now a new individual whose age is 20. For this individual the
predicted wage from this regression is
a) 40 b) 50 c) 60 d) impossible to predict without knowing the intercept estimate
28. You have regressed y on x to obtain yhat = 3 + 4x. If x increases from 7 to 10, what is
your forecast of y?
a) 12 b) 31 c) 40 d) 43
29. Suppose wage = α + βexp + ε and we have 50 observations on wage and exp, with
average values 10 and 8, respectively. We have run a regression to estimate the intercept
as 6.0. Consider now a new individual whose exp is 10. For this individual the predicted
wage from this regression is
a) 6 b) 10 c) 11 d) impossible to predict without knowing the slope estimate
30. If the expected value of the error term is 5, then after running an OLS regression
a) the average of the residuals should be approximately 5
b) the average of the residuals should be exactly zero
c) the average of the residuals should be exactly five
d) nothing can be said about the average of the residuals
31. Suppose we run a regression of y on x and save the residuals as e. If we now regress e
on x the slope estimate should be
a) zero b) one c) minus one d) nothing can be said about this estimate
32. Suppose your data produce the regression result y = 10 + 3x. Consider scaling the
data to express them in a different base year dollar, by multiplying observations by 0.9.
If both y and x are scaled, the new intercept and slope estimates will be
a) 10 and 3 b) 9 and 3 c) 10 and 2.7 d) 9 and 2.7
34. Suppose your data produce the regression result y = 10 + 3x. Consider scaling the
data to express them in a different base year dollar, by multiplying observations by 0.9. If
y is scaled but x is not (because y is measured in dollars and x is measured in physical
units, for example), the new intercept and slope estimates will be
a) 10 and 3 b) 9 and 3 c) 10 and 2.7 d) 9 and 2.7
39. Your data produce the regression result y = 8 + 5x. If the x values were scaled by
multiplying them by 0.5 the new intercept and slope estimates will be
a) 4 and 2.5 b) 8 and 2.5 c) 8 and 10 d) 16 and 10
17
1. Whenever the dependent variable is a fraction we should use as our functional form
the
a) double log b) semi-log c) logarithmic d) none of these
10. The most common functional form for estimating wage equations is
a) Linear
b) Double log
c) semilogarithmic with the dependent variable logged
d) semilogarithmic with the explanatory variables logged
15. Whenever the dependent variable is a fraction, using a linear functional form is OK if
a) most of the dependent variable values are close to one
b) most of the dependent variable values are close to zero
c) most of the dependent variable values are close to either zero or one
d) none of the dependent variable values are close to either zero or one
19
16. Violation of the CLR assumption that the expected value of the error is zero is a
problem only if this expected value is
a) negative
b) constant
c) correlated with an explanatory variable
d) uncorrelated with all explanatory variables
21. Correlation between the error term and an explanatory variable can arise because
a) of error in measuring the dependent variable
b) of a constant non-zero expected error
c) the equation we are estimating is part of a system of simultaneous equations
d) of multicollinearity
25. Suppose you are using the specification wage = α + βEducation + δMale +
θExperience + ε. In your data the variables Education and Experience happen to be
highly correlated because the observations with a lot of education happen not to have
much experience. As a consequence of this negative correlation the OLS estimates
a) are likely to be better because the movement of one explanatory variable offsets the
other, allowing the computer more easily to isolate the impact of each on the
dependent variable
b) are likely to be better because the negative correlation reduces variance making
estimates more reliable
c) are likely to be worse because the computer can’t tell which variable is causing
changes in the dependent variable
d) are likely to be worse because compared to positive correlation the negative
correlation increases variance, making estimates less reliable
21
3. An example of a statistic is
a) a parameter estimate but not a t value or a forecast
b) a parameter estimate or a t value, but not a forecast
c) a parameter estimate, a t value, or a forecast
d) a t value but not a parameter estimate or a forecast
5. Suppose we know that the CLR model applies to y = βx + ε, and that we estimate
using β* = Σy/Σx = β + Σε/Σx. This appears to be a good estimator because the
second term is
a) zero because Eε = 0
b) small because Σx is large
c) small because Σε is small
d) is likely to be small because because Σε is likely to be small
12. Suppose we estimate an unknown parameter with the value 6.5, ignoring the data.
This estimator
a) has minimum variance
b) has zero variance
c) is biased
d) all of the above
19. The OLS estimator is not used for all estimating situations because
a) it is sometimes difficult to calculate
b) it doesn’t always minimize R-squared
c) it doesn’t always have a good-looking sampling distribution
d) sometimes other estimators have better looking sampling distributions
21. To create a random variable that is normally distributed with mean 6 and variance 9
we should have the computer draw a value from a standard normal and then we
should
a) add 6 to it and multiply the result by 3
b) add 6 to it and multiply the result by 9
c) multiply it by 3 and add 6 to the result
d) multiply it by 9 and add 6 to the result
22. Suppose we have performed a Monte Carlo study to evaluate the sampling
distribution properties of an estimator betahat in a context in which we have chosen
the true parameter value beta to be 1.0. We have calculated 2000 values of betahat
and found their average to be 1.3, and their sample standard error to be 0.5. The
estimated MSE of betahat is
a) 0.34 b) 0.59 c) 0.8 d) none of these
23. Suppose we have performed a Monte Carlo study to evaluate the sampling
distribution properties of a test statistic that is supposed to be distributed as a t
statistic with 17 degrees of freedom if the null hypothesis is true. Forcing the null
hypothesis to be true we have calculated 3000 values of this statistic. Approximately
___ of these values should be greater than 1.333 and when ordered from smallest to
largest the 2850th value should be approximately ____. These blanks are best filled
with
a) 300, 1.74 b) 300, 2.11 c) 600, 1.74 d) 600, 2.11
For the next two questions, suppose you have programmed a computer as follows:
i. Draw 50 x values from a distribution uniform between 10 and 20.
ii. Count the number g of x values greater than 18.
iii. Divide g by 50 to get h1.
iv. Repeat this procedure to get 1000 h values h1 to h1000.
v. Calculate the average hav and the variance hvar of the h values.
26. Suppose the CNLR model applies and you have used OLS to estimate a slope as 2.4.
If the true value of this slope is 3.0, then the OLS estimator
a) has bias of 0.6
b) has bias of –0.6
c) is unbiased
d) we cannot say anything about bias here
25
For the next two questions, suppose you have programmed a computer as follows:
i. Draw randomly 25 values from a standard normal distribution.
ii. Multiply each of these values by 8 and add 5.
iii. Take their average and call it A1.
iv. Repeat this procedure to obtain 400 averages A1 through A400.
v. Compute the average of these 400 A values. Call it Abar.
vi. Compute the standard error of these 400 A values. Call it Asterr.
29. Four econometricians have proposed four different estimates for an unknown slope.
The estimators that have produced these estimates have bias 1, 2, 3, and 4,
respectively, and variances 18, 14, 10, and 6, respectively. From what you have learned
in this course, which of these four should be preferred?
a) first b) second c) third d) fourth
30. Suppose the CNLR model applies and you have used OLS to estimate beta as 1.3 and
the variance of this estimate as 0.25. The sampling distribution of the OLS estimator
a) has mean 1.3 and variance 0.25.
b) has a normal distribution shape
c) has a smaller variance than any other estimator
d) has bias equal to the difference between 1.3 and the true value of beta
For the next three questions, suppose you have programmed a computer as follows:
i. Draw 12 x values from a distribution uniform between 5 and 15.
ii. Draw randomly 12 e values from a standard normal distribution.
iii. Create 12 y values as y = 3*x + 2*e.
iv. Calculate bhat1 as the sum of the y values divided by the sum of the x
values.
v. Calculate bstar1 as the sum of the xy values divided by the sum of the x
squared values.
vi. Repeat this procedure from ii above to obtain 4000 bhat values bhat1
through bhat4000 and 4000 bstar values bstar1 through bstar4000.
vii. Compute the averages of these 4000 values. Call them bhatbar and
bstarbar.
viii. Compute the variances of these 4000 values. Call them bhatv and bstarv.
33. In the previous question suppose you had subtracted three from each of the bhat
values to get new numbers called q1 through q4000 and then ordered these numbers
from smallest to largest. The 3600th of these q values should be
a) approximately equal to 1.29
b) approximately equal to 1.36
c) approximately equal to 1.80
d) not very close to any of these values
35. Suppose you have a random sample of 100 observations on a variable x which is
distributed normally with mean 14 and variance 8. The sample average, xbar, is 15, and
the sample variance is 7. Then the mean of the sampling distribution of xbar is
a) 15 and its variance is 7
b) 15 and its variance is 0.07
c) 14 and its variance is 8
d) 14 and its variance is 0.08
27
The next 13 questions are based on the following information. Suppose we specify that y
= α + βx + δ1Male + δ2Female + θ1Left + θ2Center + θ3Right + ε where Left, Center,
and Right refer to the three possible political orientations. A variable Fringe is created as
the sum of Left and Right, and a variable x*Male is created as the product of x and Male.
5. Using Fringe instead of Left and Right separately in this specification is done to force
the slopes of Left and Right to be
a) the same
b) half the slope of Center
c) twice the slope of Center
d) the same as the slope of Center
9. Suppose we regress y on an intercept, x, Male, Left and Center and then do another
regression, regressing y on an intercept, x, and Center and Right. The interpretation of
the slope estimate on Center should be
a) the intercept for those from the political center in both regressions
b) the difference between the Center and Right intercepts in the first regression, and the
difference between the Center and Left intercepts in the second regression
c) the difference between the Center and Left intercepts in the first regression, and the
difference between the Center and Right intercepts in the second regression
d) none of these
10. Suppose we regress y on an intercept, x, Male, Left and Center and then do another
regression, regressing y on an intercept, x, and Center and Right. The slope estimate
on Center in the second regression should be
a) the same as the slope estimate on Center in the first regression
b) equal to the difference between the original Center coefficient and the Left coefficient
c) equal to the difference between the original Center coefficient and the Right
coefficient
d) unrelated to the first regression results
11. Suppose we regress y on an intercept, Male, Left, and Center. The base category is
a) a male on the left
b) a female on the left
c) a male on the right
d) a female on the right
12. Suppose we regress y on an intercept, Male, Left, and Center. The intercept is
interpreted as the intercept of a
a) male
b) male on the right
c) female
d) female on the right
29
14. In the preceding question, the base categories for specifications A and B are,
respectively,
a) male on the right and female on the right
b) male on the right and female on the left
c) female on the right and female on the right
d) female on the right and male on the right
16. Suppose you have estimated wage = 5 + 3education + 2gender, where gender is one
for male and zero for female. If gender had been one for female and zero for male, this
result would have been
a) Unchanged
b) wage = 5 + 3education - 2gender
c) wage = 7 + 3education + 2gender
d) wage = 7 + 3education - 2gender
The following relates to the next three questions. In a study investigating the effect of a
new computer instructional technology for economics principles, a researcher taught a
control class in the normal way and an experimental class using the new technology. She
regressed student final exam numerical grade (out of 100) on GPA, Male, Age, Tech (a
dummy equaling unity for the experimental class), and interaction variables Tech*GPA,
Tech*Male, and Tech*Age. Age and Tech*GPA had coefficients jointly insignificantly
different from zero, so she dropped them and ended up with
grade = 45 + 9*GPA + 5*Male + 10*Tech - 6*Tech*Male - 0.2*Tech*Age
with all coefficients significant. She concludes that a) age makes no difference in the
control group, but older students do not seem to benefit as much from the computer
technology, and that b) the effect of GPA is the same regardless of what group a student
is in.
21. These point estimates measure the impact of the new technology on male and female
scores, respectively, to be
a) 5 and zero b) 4 and 10 c) –1 and 10 d) 9 and 10
27. In the logit model the predicted value of the dependent variable is interpreted as
a) the probability that the dependent variable is one
b) the probability that the dependent variable is zero
c) the fraction of the observations in the sample that are ones
d) the fraction of the observations in the sample that are zeroes.
28. To find the maximum likelihood estimates the computer searches over all possible
values of the
a) dependent variable
b) independent variables
c) coefficients
d) all of the above
31. To predict the value of a dependent dummy variable for a new observation we should
predict it as a one if
a) the estimated probability of this observation’s dependent variable being a one is
greater than fifty percent
b) more than half of the observations are ones
c) the expected payoff of doing so is greater than the expected payoff of predicting it as
a zero
d) none of these
32. Which of the following is the best way to measure the prediction success of a logit
specification?
a) the percentage of correct predictions across all the data
b) the average of the percent correct predictions in each category
c) a weighted average of the percent correct predictions in each category, where the
weights are the fractions of the observations in each category
d) the sum across all the observations of the net benefits from each observation’s
prediction
34. You have estimated a logit model and found for a new individual that the estimated
probability of her being a one (as opposed to a zero) is 40%. The benefit of correctly
classifying this person is $1,000, regardless of whether she is a one or a zero. The
cost of classifying this person as a one when she is actually a zero is $500. You
should classify this person as a one when the other misclassification cost exceeds
what value?
a) $750 b) $1,000 c) $1250 d) $1500
35. You have estimated a logit model and found for a new individual that the estimated
probability of her being a one (as opposed to a zero) is 40%. The benefit of correctly
classifying this person is $2,000, regardless of whether she is a one or a zero. The
cost of classifying this person as a zero when she is actually a one is $1600. You
should be indifferent to classifying this person as a one or a zero when the other
misclassification cost equals what value?
a) $100 b) $200 c) $300 d) $400
33
36. You have estimated a logit model to determine the probability that an individual is
earning more than ten dollars an hour, with observations earning more than ten
dollars an hour coded as ones; your estimated logit index function is
-22 + 2*Ed – 6*Female + 4*Exp
where Ed is years of education, Female is a dummy with value one for females, and Exp
is years of experience. You have been asked to classify a new observation with 10 years
of education and 2 years of experience. You should classify her as
a) a one b) a zero c) too close to call
d) not enough information to make a classification
37. In the preceding question, suppose you believe that the influence of experience
depends on gender. To incorporate this into your logit estimation procedure you
should
a) add an interaction variable defined as the product of Ed and Female
b) estimate using only the female observations and again using only the male
observations
c) add a new explanatory variable coded as zero for the male observations and whatever
is the value of the experience variable for the female observations
d) none of the above
38. From estimating a logit model you have produced a slope estimate of 0.3 on the
explanatory variable x. This means that a unit increase in x will cause
a) an increase in the probability of being a y=1 observation of 0.3
b) an increase in the probability of being a y=0 observation of 0.3
c) an increase in the ratio of these two probabilities of 0.3
d) none of the above
39. You have obtained the following regression results using data on law students from
the class of 1980 at your university:
Income = 11 + .24GPA - .15Female + .14Married - .02Married*Female
where the variables are self-explanatory. Consider married individuals with equal GPAs.
Your results suggest that compared to female income, male income is higher by
a) 0.01 b) 0.02 c) 0.15 c) 0.17
42. The coefficient δ is interpreted as the ceteris paribus difference in y between an urban
immigrant and
a) a rural native
b) a rural immigrant
c) an urban native
d) none of these
43. You have estimated a logit model to determine the success of an advertising program
in a town, with successes coded as ones; your estimated logit index function is -70 +
2*PerCap + 3*South where PerCap is the per capita income in the town (measured in
thousands of dollars), and South is a dummy with value one for towns in the south
and zero for towns in the north, the only other region. If the advertising program is a
success, you will make $5000; if it is a failure you will lose $3000. You are
considering two towns, one in the south and one in the north, both with per capita
incomes of $35,000. You should undertake the advertising program
a) in both towns
b) in neither town
c) in only the south town
d) in only the north town
35
3. If a null hypothesis is true, when we impose the restrictions of this null the minimized
sum of squared errors
a) becomes smaller b) does not change c) becomes bigger
d) changes in an indeterminate fashion
4. If a null hypothesis is false, when we impose the restrictions of this null the
minimized sum of squared errors
a) becomes smaller b) does not change c) becomes bigger
e) changes in an indeterminate fashion
5. Suppose you have 25 years of quarterly data and specify that demand for your
product is a linear function of price, income, and quarter of the year, where quarter of
the year affects only the intercept. You wish to test the null that ceteris paribus
demand is the same in spring, summer, and fall, against the alternative that demand is
different in all quarters. The degrees of freedom for your F test are
a) 2 and 19 b) 2 and 94 c) 3 and 19 d) 3 and 94
6. In the preceding question, suppose you wish to test the hypothesis that the entire
relationship (i.e., that the two slopes and the intercept) is the same for all quarters,
versus the alternative that the relationship is completely different in all quarters. The
degrees of freedom for your F test are
a) 3 and 94 b) 6 and 88 c) 9 and 82 d) none of these
7. In the preceding question, suppose you are certain that the intercepts are different
across the quarters, and wish to test the hypothesis that both slopes are unchanged
across the quarters, against the alternative that the slopes are different in each quarter.
The degrees of freedom for your F test are
a) 3 and 94 b) 6 and 88 c) 9 and 82 d) none of these
10. After running a regression, to find the covariance between the first and second slope
coefficient estimates we
a) calculate the square root of the product of their variances
b) look at the first off-diagonal element of the correlation matrix
c) look at the first diagonal element of the variance-covariance matrix
d) none of these
11. Suppose you have used Eviews to regress output on capital, labor, and a time trend by
clicking on these variables in the order above, or, equivalently, using the command ls
y cap lab time c. To test for constant returns to scale using the Wald – Coefficient
Restrictions button you need to provide the software with the following information
a) cap+lab =1
b) c(1)+c(2)=1
c) c(2)+c(3) = 1
d) none of these
12. When testing a joint null, an F test is used instead of several separate t tests because
a) the t tests may not agree with each other
b) the F test is easier to calculate
c) the collective results of the t test could mislead
d) the t tests are impossible to calculate in this case
13. The rationale behind the F test is that if the null hypothesis is true, by imposing the
null hypothesis restrictions on the OLS estimation the per restriction sum of squared
errors
a) falls by a significant amount
b) rises by a significant amount
c) falls by an insignificant amount
d) rises by an insignificant amount
14. Suppose we are regressing wage on an intercept, education, experience, gender, and
dummies for black and hispanic (the base being white). To find the restricted SSE to
calculate an F test to test the null hypothesis that the black and hispanic coefficients
are equal we should regress wage on an intercept, education, experience, gender, and
a new variable constructed as the
a) sum of the black and hispanic dummies
b) difference between the black and hispanic dummies
c) product of the black and hispanic dummies
d) none of these
37
15. In the preceding question, if the null hypothesis is true then, compared to the
unrestricted SSE, the restricted SSE should be
a) smaller b) the same c) larger d) unpredictable
16. In question 14, if we regress wage on an intercept, education, experience, gender, and
a dummy for white, compared to the restricted SSE in that question, the resulting sum
of squared errors should be
a) smaller b) the same c) larger d) unpredictable
17. Suppose you have specified the demand for beer (measured in liters) as
LnBeer = β0 + β1lnBeerprice + β2lnOthergoodsprice + β3lnIncome + ε
where the notation should be obvious. Economists will tell you that in theory this
relationship should be homogeneous of degree zero, meaning that if income and prices all
increase by the same percent, demand should not change. Testing homogeneity of degree
zero means testing the null that
a) β1 = β2 = β3 = 0 b) β1 + β2 + β3 = 0 c) β1 + β2 + β3 = 1 d) none of these
Suppose you have run a logit regression in which defaulting on a credit card payment is
related to people’s income, gender, education, and age, with the coefficients on income
and age, but not education, allowed to be different for males versus females. The next 4
questions relate to this information.
18. The degrees of freedom for the LR test of the null hypothesis that gender does not
matter is
a) 1 b) 2 c) 3 d) 4
19. To calculate the LR test statistic for this null we need to compute twice the difference
between the
a) restricted and unrestricted maximized likelihoods
b) restricted and unrestricted maximized loglikelihoods
c) unrestricted and restricted maximized likelihoods
d) unrestricted and restricted maximized loglikelihoods
20. Suppose the null that the slopes on income and age are the same for males and
females is true. Then compared to the unrestricted maximized likelihood, the
restricted maximized likelihood should be
a) smaller b) the same c) bigger d) unpredictable
21. The coefficient on income can be interpreted as ceteris paribus the change in the
______ resulting from a unit increase in income.
a) probability of defaulting
b) odds ratio of defaulting versus not defaulting
c) log odds ratio of defaulting versus not defaulting
d) none of these
38
Week 9: Specification
8. Suppose you know for sure that a variable does not belong in a regression as an
explanatory variable. If someone includes this variable in their regression, in general
this will create
a) bias and increase variance
b) bias and decrease variance
c) no bias and increase variance
d) no bias and decrease variance
13. A sensitivity analysis is conducted by varying the specification to see what happens
to
a) Bias
b) MSE
c) R-square
d) the coefficient estimates
15. To perform the RESET test we rerun the regression adding as regressors the squares
and cubes of the
a) dependent variable
b) suspect explanatory variable
c) forecasts of the dependent variable
d) none of these
18. Selecting the lag length in a distributed lag model is usually done by
a) minimizing the MSE
b) maximizing R-square
c) maximizing the t values
d) minimizing an information criterion
21. In the Koyck distributed lag model, as the lag lengthens the coefficients on the lagged
explanatory variable
a) increase and then decrease b) decrease forever
c) decrease for awhile and then become zero d) none of these
22. Using the lagged value of the dependent variable as an explanatory variable is often
done to
a) avoid bias
b) reduce MSE
c) improve the fit of a specification
d) facilitate estimation of some complicated models
41
4. Multicollinearity causes
a) low R-squares
b) biased coefficient estimates
c) biased coefficient variance estimates
d) none of these
5. A symptom of multicollinearity is
a) estimates don’t change much when a regressor is omitted
b) t values on important variables are quite big
c) the variance-covariance matrix contains small numbers
d) none of these
8. A friend has told you that his multiple regression has a high R2 but all the estimates of
the regression slopes are insignificantly different from zero on the basis of t tests of
significance. This has probably happened because the
a) intercept has been omitted
b) explanatory variables are highly collinear
c) explanatory variables are highly orthogonal
d) dependent variable doesn’t vary by much
13. Suppose you are estimating y = α + βx + δz + θw + ε for which the CLR assumptions
hold and x, z, and w are not orthogonal to one another. You estimate incorporating
the information that β = δ. To do this you will regress
a) y on an intercept, 2x, and w
b) y on an intercept, (x+z), and w
c) y-x on an intercept, z, and w
d) none of these
14. In the preceding question, suppose that in fact β is not equal to δ. Then in general,
compared to regressing without this extra information, your estimate of θ
a) is unaffected
b) is still unbiased
c) has a smaller variance
d) nothing can be said about what will happen to this estimate
43
15. Economic theory tells us that when estimating the real demand for exports we should
use the ______ exchange rate and when estimating the real demand for money we
should use the _______ interest rate. The blanks should be filled with
a) real; real
b) real; nominal
c) nominal; real
d) nominal; nominal
16. You have run a regression of the change in inflation on unemployment. Economic
theory tells us that our estimate of the natural rate of unemployment is
a) the intercept estimate
b) the slope estimate
c) minus the intercept estimate divided by the slope estimate
d) minus the slope estimate divided by the intercept estimate
17. You have thirty observations from a major golf tournament in which the percentage
of putts made was recorded for distances ranging from one foot to thirty feet, in
increments of one foot (i.e., you have 30 observations). You propose estimating
success as a function of distance. What functional form should you use?
a) linear
b) logistic
c) quadratic
d) exponential
18. Starting with a comprehensive model and testing down to find the best specification
has the advantage that
a) complicated models are inherently better
b) testing down is guaranteed to find the best specification
c) testing should be unbiased
d) pretest bias is eliminated
21. When the sample size is quite large, a researcher needs to pay special attention to
a) coefficient magnitudes
b) t statistic magnitudes
c) statistical significance
d) type I errors
44
22. Your only measure of a key economic variable is unsatisfactory but you use it
anyway. This is an example of
a) knowing the context
b) asking the right questions
c) compromising
d) a sensitivity analysis
25. When testing if a coefficient is zero it is traditional to use a type I error rate of 5%.
When testing if a variable should remain in a specification we should
a) continue to use a type I error rate of 5%
b) use a smaller type I error rate
c) use a larger type I error rate
d) forget about the type I error rate and instead choose a type II error rate
28. The adage that begins with “Graphs force you to notice ….” is completed with
a) outliers
b) incorrrect functional forms
c) what you never expected to see
d) the real relationships among data
45
34. The adage that begins with “All models are wrong, ….” is completed with
a) especially those with low R-squares
b) but some are useful
c) so it is impossible to find a correct specification
d) but that should not concern us
35. Those claiming that statistical significance is being misused are referring to the
problem that
a) there may be a type I error
b) there may be a type II error
c) the coefficient magnitude may not be of consequence
d) there may be too much multicollinearity
36. Those worried that researchers are “using statistical significance to sanctify a result”
suggest that statistical analysis be supplemented by
a) looking for corroborating evidence
b) looking for disconfirming evidence
c) assessing the magnitude of coefficients
d) all of the above
46
37. To deal with results tainted by subjective specification decisions undertaken during
the heat of econometric battle it is suggested that researchers
a) eliminate multicollinearity
b) report a senstitivity analysis
c) use F tests instead of t tests
d) use larger type I error rates
38. You have regressed yt on xt and xt-1, obtaining a positive coefficient estimate on xt, as
expected, but a negative coefficient estimate on lagged x. This
a) indicates that something is wrong with the regression
b) implies that the short-run effect of x is smaller than its long-run effect
c) implies that the short-run effect of x is larger than its long-run effect
d) is due to high collinearity
41. Suppose you are estimating the returns to education and so regress wage on years of
education and some other explanatory variables. One problem with this is that people
with higher general ability levels, for which you have no measure, tend to opt for
more years of education, creating bias in your estimation. This bias is referred to as
a) multicollinearity bias
b) pretest bias
c) self-selection bias
d) omitted variable bias
3. Upon discovering via a test that you have nonspherical errors you should
a) use generalized least squares
b) find the appropriate transformation of the variables
c) double-check your specification
d) use an autocorrelation- or heteroskedasticity-consistent variance covariance matrix
estimate
4. GLS can be performed by running OLS on variables transformed so that the error
term in the transformed relationship is
a) homoskedastic
b) spherical
c) serially uncorrelated
d) eliminated
5. Second-order autocorrelated errors means that the current error εt is a linear function
of
a) εt-1 b) εt-1 squared c) εt-2 d) εt-1 and εt-2
6. Suppose you have an autocorrelated error with rho equal to 0.4. You should transform
each variable xt to become
a) .4xt b) .6xt c) xt - .4xt-1 d) .6xt - .4xt-1
12. To use the Breusch-Godfrey statistic to test the null of no autocorrelation against the
alternative of second order autocorrelated errors, we need to regress the OLS
residuals on ________ and use _____ degrees of freedom for our test statistic. The
blanks are best filled with
a) two lags of the OLS residuals; 2
b) the original explanatory variables and one lag of the OLS residuals; 1
c) the original explanatory variables and two lags of the OLS residuals; 2
d) the original explanatory variables, their lags, and one lag of the OLS residuals; 1
16. Suppose your dependent variable is aggregate household demand for electricity for
various cities. To correct for heteroskedasticity you should
a) multiply observations by the city size
b) divide observations by the city size
c) multiply observations by the square root of the city size
d) divide observations by the square root of the city size
e) none of these
17. Suppose your dependent variable is crime rates for various cities. To correct for
heteroskedasticity you should
a) multiply observations by the city size
b) divide observations by the city size
c) multiply observations by the square root of the city size
d) divide observations by the square root of the city size
e) none of these
18. When using the eyeball test for heteroskedasticity, under the null we would expect the
relationship between the squared residuals and the explanatory variable to be such
that
a) as the explanatory variable gets bigger the squared residual gets bigger
b) as the explanatory variable gets bigger the squared residual gets smaller
c) when the explanatory variable is quite small or quite large the squared residual will
be large relative to its value otherwise
d) there is no evident relationship
19. Suppose you are estimating the relationship y = α + βx + δz + ε but you suspect that
the 50 male observations have a different error variance than the 40 female
observations. The degrees of freedom for the Goldfeld-Quandt test are
a) 50 and 40 b) 49 and 39 c) 48 and 38 d) 47 and 37
20. In the previous question, suppose you had chosen to use the studentized BP test. The
degrees of freedom would then have been
a) 1 b) 2 c) 3 d) 4
21. In the previous question, to conduct the studentized BP test you would have regressed
the squared residuals on an intercept and
a) x b) z c) x and z d) a dummy for gender
50
22. Suppose you are estimating demand for electricity using aggregated data on
household income and on electricity demand across 30 cities of differing sizes Ni.
Your specification is that household demand is a linear function of household income
and city price. To estimate using GLS you should regress
a) per capita demand on an intercept, price and per capita income
b) aggregate demand on an intercept, price, and aggregate income
c) per capita demand on the inverse of Ni, price divided by Ni, and per capita income
d) none of these
23. Suppose you are estimating student performance on an economics exam, regressing
exam score on an intercept, GPA, and a dummy MALE. The CLR model assumptions
apply except that you have determined that the error variance for the male
observations is eight but for females it is only two. To estimate using GLS you should
transform by
a) dividing the male observations by 8 and the female observations by 2
b) multiplying the male observations by 2 and the female observations by 8
c) dividing the male observations by 2
d) multiplying the female observations by 8
24. Suppose the CLR model applies except that the errors are nonspherical of known
form so that you can calculate the GLS estimator. Then
a) the R-square calculated using the GLS estimates is smaller than the OLS R-square
b) the R-square calculated using the GLS estimates is equal to the OLS R-square
c) the R-square calculated using the GLS estimates is larger than the OLS R-square
d) nothing can be said about the relative magnitudes of R-square
Consider a case in which there is a nonspherical error of known form so that you can
calculate the GLS estimator. You have conducted a Monte Carlo study to investigate the
difference between OLS and GLS, using the computer to generate 2000 samples with
nonspherical errors, from which you calculate the following.
a) 2000 OLS estimates and their average betaolsbar
b) 2000 estimated variances of these OLS estimates and their average betaolsvarbar
c) the estimated variance of the 2000 OLS estimates, varbetaols.
d) 2000 corresponding GLS estimates and their average betaglsbar
e) 2000 estimated variances of these GLS estimates and their average betaglsvarbar
f) the estimated variance of the 2000 GLS estimates, varbetagls
The following six questions refer to this information.
31. Suppose the CLR model holds but the presence of nonspherical errors causes the
variance estimates of the OLS estimator to be an underestimate. Because of this,
when testing for the significance of a slope coefficient using for our large sample the
critical t value 1.96, the type I error rate
a) is higher than 5%
b) is lower than 5%
c) remains fixed at 5%
d) not possible to tell what happens to the type I error rate
32. Suppose you want to undertake a Monte Carlo study to examine the impact of
heteroskedastic errors of the form V(ε) = 4 + 9x2 where x is one of the explanatory
variables in your specification. After getting the computer to draw errors from a
standard normal, to create the desired heteroskedasticity you need to multiply the ith
error by
a) 3xi b) 2 + 3xi c) 4 + 9xi2 d) none of these
52
33. Suppose the CLR model assumptions apply to y = α + βx + θz + ε except that the
variance of the error is proportional to x squared. To produce the GLS estimator you
should regress y/x on
a) an intercept, 1/x, and z/x
b) an intercept and z/x
c) 1/x and z/x
d) not possible to produce GLS because the factor of proportionality is not known
35. Suppose income is the dependent variable in a regression and contains errors of
measurement (i) caused by people rounding their income to the nearest $100, or (ii)
caused by people not knowing their exact income but always guessing within 5% of
the true value. In case (i) there is
a) heteroskedasticity and the same for case (ii)
b) heteroskedasticity but not for case (ii)
c) no heteroskedasticity but heteroskedasiticity for case (ii)
d) no heteroskedasticity and the same for case (ii)
36. Suppose you have regressed score on an economics exam on GPA for 50 individuals,
ordered from smallest to largest GPA. The DW statistic is 1.5; you should conclude
that
a) the errors are autocorrelated
b) there is heteroskedasticity
c) there is multicollinearity
d) there is a functional form misspecification
2. Suppose a classical statistician estimates via OLS an unknown parameter beta and
because the CLR model assumptions hold declares the resulting estimate’s sampling
distribution to be such that it is unbiased and has minimum variance among all linear
unbiased estimators. For the Bayesian the sampling distribution
a) is also unbiased
b) is biased because of the prior
c) has a smaller variance
d) does not exist
Suppose the CNLR model applies and with a very large sample size the classical
statistician produces an estimate betahat = 6, with variance 4. With the same data, using
an ignorance prior, a Bayesian produces a normal posterior distribution with mean 6 and
variance 4. The next ten questions refer to this information.
5. In this example the Bayesian estimate of beta would be the same as the classical
estimate if the loss function were
a) all-or-nothing b) absolute c) quadratic d) all of the above
6. If the Bayesian had used an informative prior instead of an ignorance prior the
posterior would have had
a) the same mean but a smaller variance
b) the same mean but a larger variance
c) a different mean and a smaller variance
d) a different mean and a larger variance
54
8. For the classical statistician, the probability that beta is greater than 7 is
a) 40% b) 31% c) 16% d) not a meaningful question
9. Suppose we want to test the null hypothesis that beta is equal to 4, against the
alternative that beta is greater than 4. The classical statistician’s p value is
approximately
a) .16 b) .31 c) .32 d) none of these
10. Suppose we want to test the null hypothesis that beta is less than or equal to 4, against
the alternative that beta is greater than 4. The Bayesian statistician’s probability that
the null is true is approximately
a) .16 b) .69 c) .84 d) none of these
11. The Bayesian would interpret the interval from 2.7 to 9.3 as
a) an interval which if calculated in repeated samples would cover the true value of beta
90% of the time
b) a range containing the true value of beta with 90% probability
c) an interval that the Bayesian would bet contains the true value of beta
12. Consider the interval from 2.7 to 9.3. For the Bayesian the probability that the true
value of beta is not in this interval is
a) approximately equal to the probability that beta is less than 3.4
b) a lot greater than the probability that beta is less than 3.4
c) a lot less than the probability that beta is less than 3.4
d) not a meaningful question
14. The subjective element in a Bayesian analysis comes about through use of
a) an ignorance prior
b) an informative prior
c) the likelihood
d) the posterior
16. The usual “Bayesian point estimate” is the mean of the posterior distribution. This
assumes
a) a quadratic loss function
b) an absolute loss function
c) an all-or-nothing loss function
d) no particular loss function
18. From the Bayesian perspective a sensitivity analysis checks to see by how much the
results change when a different
a) loss function is used
b) prior is used
c) posterior is used
d) data set is used
22. Suppose you are a Bayesian and your posterior distribution for next month’s
unemployment rate is a normal distribution with mean 8.0 and variance 0.25. If this
month’s unemployment rate is 8.1 percent, what would you say is the probability that
unemployment will increase from this month to next month?
a) 50% b) 42% c) 5% d) 2.3%
56
23. If a Bayesian has a quadratic loss function, his/her preferred point estimate is
a) the mean of the posterior distribution
b) the median of the posterior distribution
c) the mode of the posterior distribution
d) cannot be determined unless the specific quadratic loss function is known
24. Suppose the net cost to a firm of undertaking a venture is $1800 if beta is less than or
equal to one and its net profit is $Q if beta is greater than one. Your posterior
distribution for beta is normal with mean 2.28 and variance unity. Any value of Q
bigger than what number entices you to undertake this venture?
a) 100 b) 200 c) 300 d) 450
25. A Bayesian has a client with a loss function equal to the absolute value of the
difference between the true value of beta and the point estimate of beta. The posterior
distribution is f(beta) = 2*beta for beta between zero and one, with f(beta) zero
elsewhere. (This distribution has mean two-thirds and variance one-eighteenth.)
Approximately what point estimate should be given to this client?
a) 0.50 b) 0.66 c) 0.71 d) 0.75
26. A Bayesian has a client with a quadratic loss function. The posterior distribution is
beta = 1, 2, and 3 with probabilities 0.1, 0.3 and 0.6, respectively. What point
estimate should be given to this client?
a) 1 b) 2 c) 3 d) none of these
27. A Bayesian has a client with an all-or-nothing loss function. The posterior
distribution is beta = 1, 2, and 3 with probabilities 0.1, 0.3 and 0.6, respectively. What
point estimate should be given to this client?
a) 1 b) 2 c) 3 d) none of these
57
Answers
Week 1: Statistical Foundations I 1c, 2c, 3c, 4a, 5b, 6d, 7a, 8c, 9c, 10c, 11a, 12c, 13d,
14c, 15c, 16a, 17a, 18d, 19c, 20a, 21a, 22a, 23a, 24c, 25a, 26d, 27a, 28b, 29c, 30a, 31b,
32c, 33a, 34c, 35b, 36d, 37b, 38a, 39b, 40b, 41d, 42b, 43a, 44b, 45a, 46b, 47c, 48c, 49b,
50d, 51c, 52a
Week 2: Statistical Foundations II 1b, 2c, 3c, 4b, 5c, 6a, 7d, 8b, 9a, 10a, 11d, 12d, 13d,
14a, 15b, 16a, 17a, 18b, 19c, 20d, 21b, 22d, 23b, 24c, 25d, 26b, 27d, 28d, 29d, 30c, 31c,
32a, 33c, 34b, 35c, 36d, 37c, 38b, 39a
Week 3: What is Regression Analysis? 1a, 2c, 3b, 4d, 5c, 6d, 7a, 8d, 9a, 10d, 11b, 12b,
13b, 14d, 15b, 16c, 17c, 18b, 19a, 20a, 21d, 22d, 23d, 24c, 25d, 26b, 27a, 28d, 29c, 30b,
31a, 32b, 33c, 34d, 35b, 36b, 37a, 38c, 39c
Week 4: The CLR Model 1d, 2c, 3b, 4c, 5d, 6c, 7d, 8b, 9b, 10c, 11d, 12c, 13c, 14d, 15d,
16c, 17c, 18a, 19a, 20a, 21c, 22b, 23d, 24a, 25c
Week 5: Sampling Distributions 1d, 2d, 3c, 4d, 5d, 6c, 7d, 8c, 9b, 10c, 11a, 12d, 13c,
14d, 15d, 16d, 17d, 18d, 19d, 20c, 21c, 22a, 23a, 24b, 25b, 26c, 27b, 28c, 29b, 30b, 31b,
32c, 33d, 34b, 35d
Week 6: Dummy Variables 1d, 2c, 3b, 4a, 5a, 6a, 7c, 8b, 9b, 10b, 11d, 12d, 13c, 14c,
15c, 16d, 17c, 18c, 19a, 20c, 21b, 22c, 23d, 24d, 25c, 26b, 27a, 28c, 29d, 30d, 31c, 32d,
33b, 34c, 35d, 36d, 37c, 38d, 39d, 40b, 41b, 42d, 43a
Week 7: Hypothesis Testing 1b, 2c, 3c, 4c, 5b, 6d, 7b, 8d, 9b, 10d, 11b, 12c, 13d, 14a,
15c, 16b, 17b, 18c, 19d, 20a, 21c
Week 9: Specification 1c, 2b, 3b, 4c, 5b, 6c, 7b, 8c, 9c, 10d, 11b, 12d, 13d, 14c, 15d,
16d, 17c, 18d, 19c, 20d, 21b, 22d
Week 10: Multicollinearity; Applied Econometrics 1c, 2d, 3c, 4d, 5d, 6c, 7d, 8b, 9d,
10c, 11d, 12d, 13b, 14c, 15b, 16c, 17b, 18c, 19c, 20d, 21a, 22c, 23d, 24d, 25c, 26d, 27c,
28c, 29c, 30d, 31a, 32b, 33b, 34b, 35c, 36d, 37b, 38c, 39c, 40a, 41c, 42d
Week 11: Nonspherical Errors 1d, 2b, 3c, 4b, 5d, 6c, 7b, 8d, 9d, 10d, 11d, 12c, 13c,
14c, 15b, 16d, 17c, 18d, 19d, 20a, 21d, 22d, 23c, 24a, 25c, 26a, 27b, 28b, 29d, 30a, 31a,
32d, 33a, 34c, 35c, 36d, 37c
Week 12: Bayesian Statistics 1b, 2d, 3b, 4a, 5d, 6c, 7b, 8d, 9a, 10a, 11b, 12a, 13d, 14b,
15a, 16a, 17b, 18b, 19c, 20d, 21d, 22b, 23a, 24b, 25c, 26d, 27c