Professional Documents
Culture Documents
Econometria Basica
Econometria Basica
Appendix A Page 1
The Summation
Operator and
Descriptive Statistics
The summation operator is a useful shorthand for manipulating expressions involving the sums of many numbers, and it plays a key role in statistics and
econometric analysis.
Definition
If denotes a sequence of numbers, then we write the sum of these numbers as
Property Sum.1 Property Sum.2 Property Sum.3 These are not true
For any constant , For any constant , If is a set of pairs of
numbers, and and are constants, then
. .
,
.
Notice that
Notice that
Appendix A Page 2
Properties of Linear
Functions
Univariate Linear Function
Example:
Definition
If and are two variables related by Suppose that the relationship between monthly housing expenditure and
monthly income is
,
.
then we say that is a linear function of , and
and are two parameters (numbers) describing
this relationship.
The intercept is .
The slope is .
Linear functions are easily defined for more than two Example:
variables. Suppose that is related to two variables,
and , in the general form Suppose that the monthly quantity demanded of compact discs is related to the
price of compact discs and monthly discretionary income by
.
.
The change in for given changes in and is
Partial Effect of on :
Appendix A Page 3
.
Partial Effect of on :
if .
Partial Effect of on :
if .
The slope of the demand curve, -9.8, is the partial effect of price on
quantity: holding income fixed, if the price of compact discs increases by one
dollar, then the quantity demanded falls by 9.8.
Appendix A Page 4
Proportions and
Percentages
Definition: Example:
The proportionate change in in moving from to If an individual’s income goes from per year to per year,
, sometimes called the relative change, is simply then the proportionate change is
Definition: Example:
The percentage change in in going from to is When income goes from to , income has increased by
simply 100 times the proportionate change, i.e., ; to get this, we simply multiply the proportionate change, , by
.
.
⚠ When is a percentage itself, and is moving from to it's advisable to report the percentage point change, i.e.,
instead.
Example:
Let denote the percentage of adults in a particular city having a college education. Suppose the initial value is (24% have a college
education), and the new value is . The change in , . The percentage of people with a college education has
increased by six percentage points.
Appendix A Page 5
Special Functions &
Their Properties
Definition:
A nonlinear function between and is
characterized by the fact that the change in for a
given change in depends on the starting value of .
1. Quadratic Functions
2. Natural Logarithm
3. Exponential Function
Appendix A Page 6
Quadratic Functions
, Example:
⚠ The function is
nonlinear in , but it is
where , , and are parameters. , linear in , , and .
,
.
Optima: ➡
In this case .
Example:
Suppose the relationship between hourly wages and years in the workforce ( ) is
given by
.
The first year of experience is worth approximately .48 (48 cents). Each additional year
of experience increases wage by less than the previous year (diminishing marginal
return to experience)
Appendix A Page 7
Natural Logarithm
The logarithm of a number, , is the exponent to which
another fixed value, the base, must be raised to produce
that number, i.e.,
• .
• .
Properties:
An Important Approximation:
. proportionate
change
. percentage
change
Definition (Elasticity)
The elasticity of with respect to is defined as The elasticity of with respect
to is the percentage change in
when increases by 1%.
.
Appendix A Page 8
.
Then , so
Then,
Semi-elasticity of with
respect to : The percentage
.
change of when increases
by one unit.
It follows that one more year of education increases hourly wage by about
.
Then , so ,
is the unit change in when increases by In other words, a increase in wage increases the weekly hours worked
. [assuming that ]. by about , or slightly less than one-half hour.
Appendix A Page 9
Exponential Function
We write the exponential function as
Properties:
Appendix A Page 10
Differential Calculus
Function Derivative
Appendix A Page 11
Chapter 1: The Nature of
Econometrics and
Economic Data
1. What Is Econometrics?
2. Steps in Empirical Economic Analysis
3. The Structure of Economic Data
4. Causality & Ceteris Paribus
Ch01 Page 1
What Is Econometrics?
The term “econometrics” is believed to have been crafted by Ragnar Frisch (1895-1973) of Norway, one of the three
principal founders of the Econometric Society, first editor of the journal Econometrica, and co-winner of the first
Nobel Memorial Prize in Economic Sciences in 1969. It is therefore fitting that we turn to Frisch’s own words in the
introduction to the first issue of Econometrica to describe the discipline.
A word of explanation regarding the term econometrics may be in order. Its definition is implied in the statement of
the scope of the [Econometric] Society, in Section I of the Constitution, which reads: “The Econometric Society is an
international society for the advancement of economic theory in its relation to statistics and mathematics.... Its main
object shall be to promote studies that aim at a unification of the theoretical-quantitative and the empirical-
quantitative approach to economic problems....”a
Ch01 Page 2
Steps in Empirical
Economic Analysis
A model in ML
An empirical analysis uses data to test a theory or to estimate a relationship. is called the
Machine or the
⚠ When testing a theory, a formal economic model is often required: Hypothesis.
Example: Economic Model of Crime Example: Job Training & Worker Productivity A deterministic
relationship!
Becker, G. S. (1968) "Crime and Punishment: An What is the effect of additional training on worker Once we know the
Economic Approach," Journal of Political Economy 76, productivity? values of we
169-217. automatically know the
Simple reasoning leads to a model such as value of y if the
functional form f is
An equation for criminal activity based on utility known!
maximization is derived: ,
, where
where
hourly wage, A deterministic
hours spent in criminal activities, relationship!
years of formal education, Once we know the
"wage" for an hour spent in criminal activity, values of educ, exper
hourly wage in legal employment, years of workforce experience, and
and training we
income other than from crime or employment, weeks spent in job training. automatically know
the value of wage if
probability of getting caught, the functional form f
probability of being convicted if caught, is known!
expected sentence if convicted, and
age.
⚠ One needs to transformed the economic model into an econometric model. This often requires that
where
hourly wage,
where
years of formal education,
some measure of the frequency of criminal activity, years of workforce experience, and
the wage that can be earned in legal employment, weeks spent in job training.
the income from other sources (assets, inheritance, and so
on), unobservables characteristics such as “innate ability,”
the frequency of arrests for prior infractions (to approximate quality of education, family background, etc.
the probability of arrest), the frequency of conviction, and
the average sentence length after conviction.
unobservables characteristics such as moral character, "wages" in
criminal activities, family socio-economic background, etc.
Ch01 Page 3
⭐ Econometrics will allow us to estimate (guess) the values of the parameters , , , with economic data on
, , , and (let's say) while accounting for the presence of .
Ch01 Page 4
The Structure of
Economic Data
Cross-Sectional Data
Ch01 Page 5
Pooled Cross Sections
Ch01 Page 6
Panel (longitudinal) Data
Ch01 Page 7
Ch01 Page 8
Causality & Ceteris
Paribus
Definition ✏ The notion of 'ceteris paribus' (which A model used for prediction can solely use
means "other (relevant) factors being correlations, but it is understood that causal
The causal effect of on can be defined as equal") plays an important role in the models should in principle predict better!
"How does the variable change if the variable causal analysis. The notion of causation is implicit when talking
changes but all other relevant factors are held about 'model interpretability' in ML.
constant?" Correlation does not mean causation
Ch01 Page 9
Chapter 2: The Simple
Regression Model
1. Definition
2. OLS: Derivation
3. OLS: Algebraic Properties
4. Nonlinearities
5. OLS: Expected Values & Variance
Ch02 Page 1
Definition
An example of a model that will "explain in terms of " is the simple linear regression model:
, Parameters:
if
Intercept
The change in is simply
Slope multiplied by the change in
holding fixed.
• The error term contains factors such as land • The error term contains factors such as labor
quality, rainfall, and so on. force experience, innate ability, tenure with current
employer, work ethic, and so on.
• The coefficient (parameter) measures the
effect of fertilizer on yield, holding other factors • The coefficient (parameter) measures the change in
fixed, i.e., . hourly wage given another year of education, holding
other factors fixed, i.e., .
⚠ This is a very strong assumption and unlikely to hold in practice for non-experimental data.
Choose several one-acre plots of land; randomly Choose a group of people; randomly assign different
assign different amounts of fertilizer to the different amounts of education to them (infeasable!); compare
plots; compare yields. wage outcomes.
Because of random assignment, the random Because of random assignment, the random variables
variables and are statistically independent and and are statistically independent and therefore mean-
therefore mean-independent. independent.
Without loss of generality the mean independence assumption, , is further strengthen to:
Why?
Firstly, notice that putting the mean independence assumption together with the zero conditional mean assumption
one has .
Now assume that you're certain that the unobservable, , cannot have mean zero but instead . In this case
Ch02 Page 2
= , where is a new intercept that relates to the
original one by , and .
Therefore, even if you believe that the original unobservable, , cannot have mean zero, we have just shown that the
original model can be re-written in terms of an alternative model with the same slope parameters, , (the ones we
usually care for), a different intercept and a new mean-zero error term, .
.
"Simple Linear Regression Model"
The goal is to approximate the mapping function so well ❗ Also notice that the zero conditional mean assumption has two
that when you have new input data ( ) that you can other implications:
predict the output variables ( ) for that data.
,
From <https://machinelearningmastery.com/supervised-and-unsupervised-machine-learning-
algorithms/>
.
Ch02 Page 3
OLS: Derivation
Recall that the zero conditional mean assumption implied
"random sample" means that any two pairs and for are
(random sampling assumption) statistically independent and came from the same joint distribution.
We have a random sample of size , , "following the population model" means that each pair, say and for
following the population model in equation . , in the sample are such that and .
How can we 'guess' (estimate) the numerical values of and based on a random sample fulfilling the random sampling assumption?
.
Further simplification using the results in Appendix A yield Since the estimators and are found by solving an ordinary
system of linear equations:
.
They are called the Ordinary Least Squares (OLS).
The "Least Squares" part of the name comes from the fact
that this ordinary system of linear equations are the first-order
conditions (FOC) from minimizing (least) the sum of the
squared errors (squares), i.e.,
Further Estimators
Ch02 Page 4
Example: CEO Salary & Return on Equity
Regression Model: .
Fitted Regression: .
Regression Model: .
Fitted Regression:
.
Ch02 Page 5
Residuals:
Fitted Values:
Ch02 Page 6
OLS: Algebraic
Properties
(1) The sum (therefore the sample average) of the OLS residuals,
, is zero.
Goodness-of-Fit
Definition
Regression Model: .
Fitted Regression:
.
Regression Model: .
Fitted Regression:
.
Ch02 Page 7
The share of campaign expenditures explains over 85% of the
variation in the election outcomes for this sample.
Ch02 Page 8
Nonlinearities
Regression Model: .
Fitted Regression:
.
Regression Model:
Fitted Regression:
.
Ch02 Page 9
OLS: Expected Values &
Variance
Recall that
❗ Therefore, the OLS estimators and are functions of random variables, so they are
random variables themselves, i.e., we can calculate expected values and variances.
Unbiasedness of OLS
where and are the population intercept and slope parameters, respectively.
Theorem (Unbiasedness of OLS) • The estimated coefficients may be smaller or larger, depending on the sample that
Let Assumptions SLR.1-SLR.4 hold, then is the result of a random draw.
• However, on average, they will be equal to the values that characterize the true
, relationship between and in the population.
• “On average” means if sampling was repeated, i.e. if drawing the random sample
. and doing the estimation was repeated many times.
• In a given sample, estimates may differ considerably from true values.
Depending on the sample, the estimates will be nearer or farther away from the true population values.
How far can we expect our estimates to be away from the true population values on average (= sampling variability)?
Sampling variability is measured by the estimator‘s variances
,
.
,
.
Ch02 Page 10
Assumption SLR.5 (Homoskedasticity)
The error has the same variance given any value of the explanatory variable, i.e.,
.
Theorem (Sampling Variance of OLS) • The sampling variability of the estimated regression coefficients will be the higher,
Let Assumptions SLR.1-SLR.5 hold, then the larger the variability of the unobserved factors, and the lower, the higher the
variation in the explanatory variable.
,
Although we can calculate and from the sample,
we do not observe .
.
⚠ Therefore, in order to estimate and , we need
to estimate the generally unknown error variance, .
.
0
Infeasible Estimator of Feasible Estimators of
Biased:
.
Unbiased:
Standard Error of :
Ch02 Page 11
Chapter 3: Multiple Linear
Regression (Estimation)
1. Definition
2. OLS: Derivation
3. OLS: Interpretation & Partialling Out
4. OLS: Algebraic Properties
5. OLS: Expected Values
6. OLS: Variance
7. OLS: Efficiency (Gauss-Markov Theorem)
Ch03 Page 1
Definition
The general multiple linear regression model (also called the multiple regression model) can be written in the population as
Parameters: Variables:
Intercept , , ,
Slope parameter associated with . Dependent variable Independent variables Error variable
Slope parameter associated with . Explained variable Explanatory variables Disturbance variable
Response variable Control variables Unobservable variable
Slope parameter associated with . Predicted variable Predictor variables
Regressand Regressors
There are parameters in the model.
❗ Also notice that the zero conditional mean assumption has the following implications:
Ch03 Page 2
Ch03 Page 3
OLS: Derivation
Population Estimator
Moment
with respect to , , , , , .
Ch03 Page 4
OLS: Interpretation &
Partialling Out
The SRF (OLS Regression line) is
or in term of changes,
Definition
The coefficient on measures the change in due to a one-unit
increase in , holding all other independent variables fixed. That is,
,
holding fixed. Thus, we have controlled for the variables
when estimating the effect of on . The other coefficients
have a similar interpretation.
. . . .
. . . .
. . . . . .
or 2.61%.
Ch03 Page 5
One can show that the estimated coefficient, , of an explanatory
variable, , in a multiple regression
Ch03 Page 6
OLS: Algebraic Properties
(1) The sum (therefore the sample average) of the OLS residuals,
, is zero.
Fitted Regression 1:
. . . .
.
Fitted Regression 2:
. . . . .
.
Ch03 Page 7
Ch03 Page 8
OLS: Expected Values
Unbiasedness of OLS
Assumption MLR.1 (Linear in Parameters) Recall the model must be linear in the parameters of interest (no on
In the population model, the dependent variable, the explanatory variables).
, is related to the independent variable, , ,
, , and the error (or disturbance), , as
,
,
,
,
,
where .
Ch03 Page 9
Ch03 Page 10
Including Irrelevant
Regressors
Suppose we specify the model ( )
Example of
, a misspecified
model.
However, has no effect on after and have been controlled for, which means that .
If we were to regress on , , and and obtain the OLS estimates, , , , and by virtue of the Assumptions
MLR.1 - MLR.4 we have
Including one or more irrelevant variables in a multiple regression model, or overspecifying the model, does not affect the
unbiasedness of the OLS estimators.
Ch03 Page 11
Excluding Relevant
Regressors
Suppose the true model ( ) is
Therefore, if one runs a regression of on , then if Assumptions SLR.1-SLR.4 hold for this regression, then
Example of
a misspecified , ⚠ Unless or
model.
, .
.
bias
bias
There are not similar results for the general (any) case.
Ch03 Page 12
Ch03 Page 13
OLS: Variance
As in the simple linear regression case we have that for the population model
,
,
.
Error Variance Total Sample Variation in the Explanatory Variable Linear Relationships Among Regressors
• A high error variance increases the sampling variance • More sample variation leads to more precise estimates. • The -squared of this regression will be the
because there is more “noise” in the equation. • Total sample variation automatically increases with the sample higher when can be better explained by
• A large error variance doesn‘t necessarily make size. the other independent variables.
estimates imprecise. • Increasing the sample size is thus a way to get more precise • The sampling variance of the slope
• The error variance does not decrease with sample size. estimates. estimator for will be higher when can
be better explained by the other
⚠ This term is generally unknown and will be independent variables.
estimated (below). Under perfect multicollinearity, the
variance of the slope estimator is not
defined (infinity).
❓ Is Multicollinearity a Problem?
so that .
Ch03 Page 14
Ch03 Page 15
Misspecified
Models
Suppose the true model ( ) is
Regress on and : ,
Regress on : ,
,
, ,& are unbiased because .
,& are biased
Ch03 Page 16
Estimating : Standard
Errors
Firstly, recall that by the Law of Total Variance: . Therefore, by Assumptions MLR.4 and MLR,5 we have that
We can figure out why the degrees of freedom adjustment is necessary by returning to the first order conditions for the OLS estimators.
These can be written and where . Thus, in obtaining the OLS estimates, restrictions are imposed
on the OLS residuals. This means that, given of the residuals, the remaining residuals are known: there are only
degrees of freedom in the residuals. (This can be contrasted with the errors , which have degrees of freedom in the sample.)
Ch03 Page 17
Example: Explaining Arrest Records
. . . . .
. . . . .
.
Ch03 Page 18
OLS: Efficiency (Gauss-
Markov Theorem)
Simple Linear Regression:
Consider the simple linear regression model ( ) again but without an intercept ( ):
NO
It is also unbiased: .
It turns out that, under Assumptions SLR.1-SLR.5, the OLS estimator has the smallest possible variance among any other
estimator that can be written as a linear function of the response variable , i.e., , and is also
unbiased. In this sense the OLS estimator is the best.
Ch03 Page 19
Derivacion alternativa de MCO
Saturday, June 26, 2021 8:54 PM
Ch03 Page 20
Ch03 Page 21
Ch03 Page 22
Ch03 Page 23