CH 10
CH 10
CH 10
Walter R. Paczkowski
Rutgers University
Chapter 10: Random Regressors and
Principles of Econometrics, 4th Edition Page 1
Moment-Based Estimation
Chapter Contents
10.1.1
The Small Sample
Properties of the
Least Squares
Estimators
10.1.1
The Small Sample
Properties of the
Under assumptions A10.1–A10.6:
Least Squares
Estimators 1. The least squares estimator is unbiased
2. The least squares estimator is the best linear
unbiased estimator of the regression
parameters, and the usual estimator of σ2 is
unbiased
3. The distributions of the least squares
estimators, conditional upon the x’s, are normal,
and their variances are estimated in the usual
way
• The usual interval estimation and hypothesis
testing procedures are valid
Chapter 10: Random Regressors and
Principles of Econometrics, 4th Edition Page 10
Moment-Based Estimation
10.1
Linear Regression
with Random x’s
10.1.1
The Small Sample
Properties of the
Least Squares
Estimators
10.1.2
Large Sample
Properties of the
Least Squares
Estimators
10.1.2
Large Sample
Properties of the
Now we can say:
Least Squares
Estimators
– Under assumptions A10.1, A10.2, A10.3*, A10.4, and A10.5, the least
squares estimators:
1. Are consistent.
– They converge in probability to the true parameter values as
N→∞.
2. Have approximate normal distributions in large samples, whether
the errors are normally distributed or not.
– Our usual interval estimators and test statistics are valid, if the
sample is large.
3. If assumption A10.3* is not true, and in particular if cov(x,e) ≠ 0
so that x and e are correlated, then the least squares estimators are
inconsistent.
– They do not converge to the true parameter values even in very
large samples.
– None of our usual hypothesis testing or interval estimation
procedures are valid.
Chapter 10: Random Regressors and
Principles of Econometrics, 4th Edition Page 13
Moment-Based Estimation
10.1
Linear Regression
with Random x’s
FIGURE 10.1 (a) Correlated x and e
10.1.3
Why Least Squares
Estimation Fails
10.1.3
Why Least Squares
Estimation Fails
10.1.3
Why Least Squares
Estimation Fails
10.2.1
Measurement Error
10.2.1
Measurement Error Let y = annual savings and x* = the permanent
annual income of a person
– A simple regression model is:
Eq. 10.1 yi 1 2 xi* vi
– Current income is a measure of permanent
income, but it does not measure permanent
income exactly.
• It is sometimes called a proxy variable
• To capture this feature, specify that:
Eq. 10.2 xi xi* ui
10.2.1
Measurement Error
Substituting:
y 1 2 x* vi
1 2 x u v
Eq. 10.3
1 2 x v 2u
1 2 x e
10.2.1
Measurement Error
10.2.1
Measurement Error
10.2.2
Simultaneous
Equations Bias
10.2.2
Simultaneous
Equations Bias
10.2.3
Omitted Variables
10.2.3
Omitted Variables
10.2.3
Omitted Variables
10.2.4
Least Squares
Estimation of a
Estimating our wage equation, we have:
Wage Equation
10.3.1
Method of Moments
Estimation of a
Population Mean
and Variance The kth moment of a random variable Y is the
expected value of the random variable raised to
the kth power:
Eq. 10.7 E Y k k k th moment of Y
– The kth population moment in Eq. 10.7 can be
estimated consistently using the sample (of size
N) analog:
E Y ˆ k k th sample moment of Y yik N
Eq. 10.8
k
10.3.1
Method of Moments
Estimation of a
Population Mean
and Variance
10.3.1
Method of Moments
Estimation of a
Population Mean
and Variance
10.3.1
Method of Moments
Estimation of a
Population Mean
and Variance
10.3.2
Method of Moments
Estimation in the
Simple Linear
Regression Model
10.3.2
Method of Moments
Estimation in the
Simple Linear
Regression Model
10.3.2
Method of Moments
Estimation in the
These are equivalent to the least squares normal
Simple Linear
Regression Model equations and their solution is:
b2
xi x yi y
xi x
2
Eq. 10.16
b1 y b2 x
– Under "nice" assumptions, the method of
moments principle of estimation leads us to the
same estimators for the simple linear regression
model as the least squares principle
10.3.3
Instrumental
Variables Estimation
Suppose that there is another variable, z, such that:
1. z does not have a direct effect on y, and thus it
in the Simple Linear
Regression Model
10.3.3
Instrumental
Variables Estimation
in the Simple Linear
Regression Model If such a variable z exists, then it can be used to
form the moment condition:
Eq. 10.16 E ze 0 E z y 1 2 x 0
– Use Eqs. 10.13 and 10.16, the sample moment
conditions are:
yi ˆ 1 ˆ 2 xi 0
1
N
Eq. 10.17
zi yi ˆ 1 ˆ 2 xi 0
1
N
Chapter 10: Random Regressors and
Principles of Econometrics, 4th Edition Page 40
Moment-Based Estimation
10.3
Estimators Based on
the Method of
Moments
10.3.3
Instrumental
Variables Estimation
in the Simple Linear
Regression Model Solving these equations leads us to method of
moments estimators, which are usually called the
instrumental variable (IV) estimators:
ˆ N zi yi zi yi zi z yi y
N zi xi zi xi zi z xi x
2
Eq. 10.18
ˆ 1 y ˆ 2 x
10.3.3
Instrumental
Variables Estimation
These new estimators have the following
in the Simple Linear
Regression Model properties:
– They are consistent, if z is exogenous, with
E(ze) = 0
– In large samples the instrumental variable
estimators have approximate normal
distributions
• In the simple regression model:
2
Eq. 10.19 ˆ 2 ~ N 2 , 2
r x x 2
zx i
Chapter 10: Random Regressors and
Principles of Econometrics, 4th Edition Page 42
Moment-Based Estimation
10.3
Estimators Based on
the Method of
Moments
10.3.3
Instrumental
Variables Estimation
These new estimators have the following
in the Simple Linear
Regression Model properties (Continued):
– The error variance is estimated using the
estimator:
yi ˆ 1 ˆ 2 xi
2
ˆ 2IV
N 2
10.3.3a
The Importance of
Using Strong
Instruments
Note that we can write the variance of the
instrumental variables estimator of β2 as:
2 var b2
var ˆ 2
xi x
2
r 2
zx
rzx2
– Because rzx2 1 the variance of the instrumental
variables estimator will always be larger than
the variance of the least squares estimator, and
thus it is said to be less efficient
10.3.4
Instrumental
Variables Estimation
in the Multiple
Regression Model To extend our analysis to a more general setting,
consider the multiple regression model:
y β1 β 2 x2 β K xK e
– Let xK be an endogenous variable correlated
with the error term
– The first K - 1 variables are exogenous
variables that are uncorrelated with the error
term e - they are ‘‘included’’ instruments
10.3.4
Instrumental
Variables Estimation
in the Multiple
Regression Model
10.3.4
Instrumental
Variables Estimation
in the Multiple
Regression Model The first stage regression has the endogenous
variable xK on the left-hand side, and all
exogenous and instrumental variables on the
right-hand side
– The first stage regression is:
Eq. 10.20 xK 1 2 x2 K 1 xK 1 1 z1 L z L vK
10.3.4
Instrumental
Variables Estimation
The second stage regression is based on the original
in the Multiple
Regression Model specification:
Eq. 10.22 y β1 β 2 x2 β K xK e
ˆ *
10.3.4
Instrumental
Variables Estimation
in the Multiple
Regression Model
2
Eq. 10.23
σ̂ 2
yi βˆ 1 βˆ 2 x2i βˆ K xKi
IV
N K
10.3.4a
Using Surplus
Instruments in
Simple Regression
In the simple regression, if x is endogenous and
we have L instruments:
xˆ ˆ 1 ˆ 1 z1 ˆ L z L
– The two sample moment conditions are:
y βˆ
1
i 1 βˆ 2 xi 0
N
xˆ y βˆ
1
i i 1 βˆ 2 xi 0
N
10.3.4a
Using Surplus
Instruments in
Simple Regression
β̂ 2
xˆ xˆ y y xˆ x y y
i i
i i
xˆ xˆ x x xˆ x x x
i i i i
βˆ 1 y βˆ 2 x
10.3.4b
Surplus Moment
Conditions
E z2e E z2 y β1 β 2 x 0
10.3.4b
Surplus Moment
Conditions
1
yi βˆ 1 βˆ 2 xi mˆ i 0
N
1
N
z y βˆ βˆ x mˆ 0
i1 i 1 2 i 2
z y βˆ βˆ x mˆ 0
1
i2 i 1 2 i 3
N
10.3.5
Assessing
Instrument Strength
Using the First Stage
Model
10.3.5a
One Instrumental
Variable
10.3.5b
More Than One
Instrumental
Variable
Eq. 10.25 xK 1 2 x2 K 1 xK 1 1 z1 L zL vK
10.3.6
Instrumental
Variables Estimation
of the Wage
Equation
10.3.6
Instrumental
Variables Estimation
of the Wage
Equation
10.3.6
Instrumental
Variables Estimation
of the Wage
Equation
10.3.6
Instrumental
Variables Estimation
of the Wage
Equation
10.3.6
Instrumental
Variables Estimation
of the Wage
Equation
10.3.6
Instrumental
Variables Estimation
of the Wage
Equation
10.3.7
Partial Correlation
10.3.7
Partial Correlation
10.3.8
Instrumental
Variables Estimation
in a General Model
10.3.8
Instrumental Think of G = Good explanatory variables, B = Bad
Variables Estimation
in a General Model explanatory variables and L = Lucky instrumental variables
– It is a necessary condition for IV estimation that L ≥ B
– If L = B then there are just enough instrumental
variables to carry out IV estimation
• The model parameters are said to just identified or
exactly identified in this case
• The term identified is used to indicate that the model
parameters can be consistently estimated
– If L > B then we have more instruments than are
necessary for IV estimation, and the model is said to be
overidentified
10.3.8
Instrumental
Variables Estimation
Consider the B first-stage equations:
in a General Model
xG j 1 j 2 j x2 Gj xG 1 j z1 Lj zL v j ,
Eq. 10.29
j 1, , B
The predicted values are:
xˆG j ˆ 1 j ˆ 2 j x2 ˆ Gj xG ˆ 1 j z1 ˆ Lj z L ,
j 1, , B
In the second stage of estimation we apply least
squares to:
Eq. 10.30 y 1 2 x2 G xG G 1 xˆG 1 K xˆK e*
10.3.8a
Assessing
Instrument Strength
in a General Model
xG 1 γ11 γ 21 x2 γ G1 xG θ11 z1 θ 21 z2 v1
xG 2 γ12 γ 22 x2 γ G 2 xG θ12 z1 θ 22 z2 v2
10.3.8b
Hypothesis Testing
with Instrumental
Variables Estimates
When testing the null hypothesis H0: βk = c, use of
the test statistic t ˆ k c se ˆ k is valid in large
samples
– It is common, but not universal, practice to use
critical values, and p-values, based on the
distribution rather than the more strictly
appropriate N(0,1) distribution
– The reason is that tests based on the t-
distribution tend to work better in samples of
data that are not large
Chapter 10: Random Regressors and
Principles of Econometrics, 4th Edition Page 69
Moment-Based Estimation
10.3
Estimators Based on
the Method of
Moments
10.3.8b
Hypothesis Testing
with Instrumental
Variables Estimates
When testing a joint hypothesis, such as
H0: β2 = c2, β3 = c3, the test may be based on the
chi-square distribution with the number of degrees
of freedom equal to the number of hypotheses (J)
being tested
– The test itself may be called a “Wald” test, or a
likelihood ratio (LR) test, or a Lagrange
multiplier (LM) test
– These testing procedures are all asymptotically
equivalent
Chapter 10: Random Regressors and
Principles of Econometrics, 4th Edition Page 70
Moment-Based Estimation
10.3
Estimators Based on
the Method of
Moments
10.3.8c
Goodness-of-Fit
with Instrumental
Variables Estimates
10.4.1
The Hausman Test
for Endogeneity
10.4.1
The Hausman Test
for Endogeneity
If null hypothesis is true, both the least squares
estimator and the instrumental variables estimator are
consistent
– Naturally if the null hypothesis is true, use the more
efficient estimator, which is the least squares
estimator
If the null hypothesis is false, the least squares estimator
is not consistent, and the instrumental variables
estimator is consistent
– If the null hypothesis is not true, use the instrumental
variables estimator, which is consistent
Chapter 10: Random Regressors and
Principles of Econometrics, 4th Edition Page 75
Moment-Based Estimation
10.4
Specification Tests
10.4.1
The Hausman Test
for Endogeneity
10.4.1
The Hausman Test
for Endogeneity
Consider the model: y 1 2 x e
– Let z1 and z2 be instrumental variables for x.
1. Estimate the model x 1 1 z1 2 z2 vby least
squares, and obtain the residuals
v̂ x ˆ ˆ z ˆ z
.
1 1 1 2 2
10.4.1
The Hausman Test
for Endogeneity
Consider the model (Continued): y 1 2 x e
2. Include the residuals computed in step 1 as an
explanatory variable in the original regression,
y 1 2 x vˆ e
– Estimate this "artificial regression" by least
squares, and employ the usual t-test for the
hypothesis of significance
10.4.1
The Hausman Test
for Endogeneity
Consider the model (Continued): y 1 2 x e
3. If more than one variable is being tested for
endogeneity, the test will be an F-test of joint
significance of the coefficients on the
included residuals
10.4.2
Testing Instrument
Validity
A test of the validity of the surplus moment
conditions is:
1. Compute the IV estimates ˆ k using all
available instruments, including the G
variables x1=1, x2, …, xG that are presumed to
be exogenous, and the L instruments
2. Obtain the residuals eˆ y ˆ 1 ˆ 2 x2 ˆ K xK .
3. Regress ê on all the available instruments
described in step 1
10.4.2
Testing Instrument
Validity
A test of the validity of the surplus moment conditions
is (Continued):
4. Compute NR2 from this regression, where N is the
sample size and R2 is the usual goodness-of-fit
measure
5. If all of the surplus moment conditions are valid,
NR ~ ( L B ) .
2 2
then
• If the value of the test statistic exceeds
(2L B ) the
100(1−α)-percentile from the distribution,
then we conclude that at least one of the surplus
moment conditions restrictions is not valid
Chapter 10: Random Regressors and
Principles of Econometrics, 4th Edition Page 81
Moment-Based Estimation
10.4
Specification Tests Table 10.2 Hausman Test Auxiliary Regression
10.4.3
Specification Tests
for the Wage
Equation
10A.1
Conditional
Expectations
var Y | X x y E Y | X x f y | x
2
10A.2
Iterated
Expectations
10A.2
Iterated
Expectations
EX E Y | X
10A.2
Iterated
Expectations
Eq. 10A.3 E XY E X XE Y | X
10A.3
Regression Model
Application
Eq. 10A.5 E ei Ex E ei | xi Ex 0 0
Eq. 10A.6 E xi ei Ex xi E ei | xi Ex xi 0 0
10A.3
Regression Model
Application
xi E xi yi E yi 2 xi E xi xi E xi ei
2
E xi E xi yi E yi 2 E xi E xi E xi E xi ei
2
cov x, y 2 var x cov x, e
or
Chapter 10: Random Regressors and
Principles of Econometrics, 4th Edition Page 93
Moment-Based Estimation
10B
The Inconsistency
of the Least
Squares Estimator
– If cov(x, e) ≠ 0, then:
cov x, y cov x, e
2
var x var x
cov x, y cov x, e
Eq. 10B.4 b2 2 2
var x var x
ˆ zi z yi y z, y
N 1
cov
Eq. 10C.1
zi z xi x N 1 cov
z, x
2
Eq. 10C.2
ˆ cov z , y
cov z , x
2
cov z, y cov z, e
Eq. 10C.3 2
cov z , x cov z , x
If cov(x, e) = 0, then:
Eq. 10C.4
ˆ cov z , y
cov z , x
2 2
Eq. 10D.2 x 0 1 z v
– Substituting:
y 1 2 x e 1 2 E x v e
1 2 E x 2v e
Eq. 10D.4
Eq. 10D.5 x xˆ vˆ
y 1 2 x e 1 2 xˆ vˆ e
Eq. 10D.6
1 2 xˆ 2vˆ e
Eq. 10D.7
y 1 2 xˆ vˆ e
– If we omit vˆ
Eq. 10D.8 y 1 2 xˆ e
Eq. 10D.9 1 2 xˆ vˆ 2 vˆ e
1 2 x vˆ e
10E.1
A Test for Weak
Identification
10E.1
A Test for Weak
Identification
10E.1
A Test for Weak
Identification
10E.1
A Test for Weak
Identification
10E.2
Examples of
Testing for Weak
Identification
where
NWIFEINC FAMINC WAGE HOURS 1000
10E.2
Examples of
Testing for Weak
Identification
Weak IV Example 1: Endogenous: MTR;
Instrument: EXPER
– The estimated first-stage equation for MTR is
Model (1) of Table 10E.3
– The estimated coefficient of MTR in the
estimated HOURS supply equation in Model
(1) of Table 10E.4 is negative and significant at
the 5% level
10E.2
Examples of
Testing for Weak
Identification
10E.2
Examples of
Testing for Weak
Identification
10E.2
Examples of
Testing for Weak
Weak IV Example 2: Endogenous: MTR;
Identification
Instruments: EXPER, EXPER2, LARGECITY
– The first-stage equation estimates are reported
in Model (2) of Table 10E.3
– The estimated coefficient of MTR in the
estimated HOURS supply equation in Model
(2) of Table 10E.4 is negative and significant at
the 5% level, although the magnitudes of all the
coefficients are smaller in absolute value for
this estimation than the model in Model (1)
10E.2
Examples of
Testing for Weak
Weak IV Example 3 Endogenous: MTR, EDUC;
Identification
Instruments: MOTHEREDUC, FATHEREDUC
– The first-stage equations for MTR and EDUC
are Model (3) and Model (4) of Table 10E.3
– The estimates of the HOURS supply equation,
Model (3) of Table 10E.4, shows parameter
estimates that are wildly different from those in
Model (1) and Model (2), and the very small t-
statistic values imply very large standard errors,
another consequence for instrumental variables
estimation in the presence of weak instruments
Chapter 10: Random Regressors and
Principles of Econometrics, 4th Edition Page 114
Moment-Based Estimation
10E
Testing for Weak
Instruments
10E.3
Testing for Weak
Identification:
If instrumental variables are ‘‘weak,’’ then the
Conclusions
instrumental variables, or two-stage least squares,
estimator is unreliable
When there is a single endogenous variable, the
first-stage F-test of the joint significance of the
external instruments is an indicator of instrument
strength
If there is more than one endogenous variable on
the right-hand side of an equation, then the F-test
statistics from the first stage equations do not
provide reliable information about instrument
strength
Chapter 10: Random Regressors and
Principles of Econometrics, 4th Edition Page 115
Moment-Based Estimation
10F
Monte Carlo
Simulation
10F.1
Illustrations Using
Simulated Data
10F.1
Illustrations Using
Simulated Data
10F.1
Illustrations Using
Simulated Data
10F.1
Illustrations Using
Simulated Data
The outcome of two-stage least squares estimation
using the two instruments z1 and z2 where we first
obtain the first-stage regression of x on the two
instruments z1 and z2:
xˆ 0.1947 0.5700 z1 0.2068 z2
Eq. 10F.1
se 0.079 0.089 0.077
– The instrumental variables estimates are:
yˆ IV _ z1 , z2 1.1376 1.0399 x
Eq. 10F.2
se 0.116 0.194
10F.1.1
The Hausman Test To implement the Hausman test, estimate the first-
stage equation shown in Eq. 10F.1 using the
instruments z1 and z2
– Compute the residuals:
vˆ x xˆ x 0.1947 0.5700 z1 0.2068 z2
– Include the residuals as an extra variable in the
regression equation and apply least squares:
yˆ 1.1376 1.0399 x 0.9957vˆ
se 0.080 0.133 0.163
10F.1.2
Test for Weak
Instruments
10F.1.3
Testing the Validity
of Surplus
If we use z1, z2, and z3 as instruments, there are two
Instruments
surplus moment conditions.
– The IV estimates using these three instruments
are:
yˆ IV _ z1 , z2 , z3 1.0626 1.3535 x
10F.2
The Repeated
Sampling
Properties of
IV/2SLS
10F.2
The Repeated
Sampling
Properties of
IV/2SLS
The IV estimator it is
β̂
2
10000
m1 2m β2 10000 1.0088