Econometrics I 2
Econometrics I 2
Econometrics I 2
William Greene
Department of Economics
Stern School of Business
Applied Econometrics
1.50
1.00
P
.50
.00
1.50 2.00 2.50 3.00 3.50 4.00 4.50 5.00 5.50
Q
Variation Around E[P|Q]
(Conditioning Reduces Variation)
1.50
1.00
P
.50
.00
1.50 2.00 2.50 3.00 3.50 4.00 4.50 5.00 5.50
Q
Means of P for Given Group Means of Q
1.50
1.00
P
.50
.00
1.50 2.00 2.50 3.00 3.50 4.00 4.50 5.00 5.50
Q
Another Conditioning Variable
1.50
E[P|Q,Y=1]
1.00
P
E[P|Q,Y=0]
.50
.00
1.50 2.00 2.50 3.00 3.50 4.00 4.50 5.00 5.50
Q
HIGHIN=1 HIGHIN=2
Conditional Mean Functions
840
560
280
0
0 1 2 3 4 5 6 7 8 9101112
1314
1516
1718
1920
2122
2324
2526
2728
2930
3132
3334
3536
3738
3940
4142
4344
4546
4748
4950
5152
5354
5556
5758
59
DOCVIS
Conditional Mean and Projection
Doctor Visits: Conditional Mean and Linear Projection
4.22
3.32
This area is
2.43
Most of the outside the range
data are in
D o cV isit
of the data
here
1.53
.63
-.26
0 4 8 12 16 20
INCOME
CONDMEAN PROJECTN
= X +
The assumption means that the rank of the matrix X is K.
No linear dependencies => FULL COLUMN RANK of the matrix X.
Expected Values of Deviations from
the Conditional Mean
Observed y will equal E[y|x] + random variation.
y = E[y|x] + (disturbance)
Is there any information about in x? That is,
does movement in x provide useful information
about movement in ? If so, then we have not fully
specified the conditional mean, and this function we
are calling ‘E[y|x]’ is not the conditional mean
(regression)
There may be information about in other variables.
But, not in x. If E[|x] 0 then it follows that
Cov[,x] 0. This violates the (as yet still not fully
defined) ‘independence’ assumption
Zero Conditional Mean of ε
E[|all data in X] = 0
E[|X] = 0 is stronger than E[i | xi] = 0
The second says that knowledge of xi provides no
information about the mean of i. The first says that
no xj provides information about the expected value
of i, not the ith observation and not any other
observation either.
“No information” is the same as no correlation.
Proof: Cov[X,] = Cov[X,E[|X]] = 0
The Difference Between E[ε |X]=0 and E[ε]=0
Conditional Homoscedasticity and
Nonautocorrelation
Disturbances provide no information about each
other, whether in the presence of X or not.
Var[|X] = 2I.
Does this imply that Var[] = 2I? Yes: Proof:
Var[] = E[Var[|X]] + Var[E[|X]].
Insert the pieces above. What does this mean? It is
an additional assumption, part of the model. We’ll
change it later. For now, it is a useful simplification
Nonrandom (Fixed) Regressors