ARE107 L3 Detailed
ARE107 L3 Detailed
ARE107 L3 Detailed
Lecture 3
1
RECAP
2
Lecture 3 - Roadmap
So far...
I Predictive vs. Causal Modeling
I Randomized Control Trials
Causal Effect: E [Y1i − Y0i |Di = 1]
The conditional expectation given Di = 1 means we are taking the average over
individuals that received the treatment.
Note: (1) Recall that cross-sectional data means we have a random sample of observations. (2) By
when here we mean under what assumptions.
3
Linear Regression Model: Causal Effects and Selection Bias
We will start with a binary regressor, Di
Yi = α + βDi + ui
Using the definitions of Y1i and Y0i , define the causal effect.
Note: Can you see the key problem in identifying the causal effect in the linear model?
4
Linear Regression Model: Causal Effects and Selection Bias
In the linear model, β1 is the causal effect for an individual.
Exercise 2. Now decompose the difference in group means,
E [Yi |Di = 1] − E [Yi |Di = 0], into the average causal effect and selection bias
for the linear regression model.
5
Linear Regression Model: Causal Effects and Selection Bias
In the linear model Yi = α + βDi + ui , the difference in group means can be
decomposed into the following
REMARKS:
I β is the Causal Effect: In the linear regression model, β is the causal
effect we are after.
I β is the be-all end-all : Note that β is not only the average causal
effect, but also the causal effect of D on Y for each individual i, i.e. the
causal effect is the same for everyone. This is a byproduct of the linear
model.
6
Linear Regression Model: Causal Effects and Selection Bias
In the linear model Yi = α + βDi + ui , the difference in group means can be
decomposed into the following
REMARKS:
I Selection Bias: Consider the example where Y is income and D is
whether one has a college degree or not. We can think of ui as innate
ability, which is unobservable. Hence, selection bias here is the
difference in average innate ability between individuals that earn a
college degree and those who do not.
I Linear Regression of Y on D: Instead of comparing group means as
we did before, we can just run a regression of the outcome variable Y
on treatment status D to estimate the causal effect from an RCT.
7
Linear Regression Model: Causal Effects and Selection Bias
E [ui |Di ] = 0.
8
Linear Regression Model: Causal Effects and Selection Bias
Now we are ready to move beyond a binary regressor.
Consider the example of earnings (Y ) and years of schooling (X )
Yi = α + βXi + ui
9
Linear Regression Model: Causal Effects and Selection Bias
10
Linear Regression Model: Causal Effects and Selection Bias
E [Yi |Xi = 12] − E [Yi |Xi = 11] = β + E [ui |Xi = 12] − E [ui |Xi = 11]
|{z} | {z }
Causal Effect Selection Bias
11
Linear Regression Model: Causal Effects and Selection Bias
Exogeneity, E [ui |Xi ] = 0, implies that E [ui |Xi ] does not change as Xi
changes, then
E [Yi |Xi ] = α + βXi + E [ui |Xi ] = α + βXi
| {z }
=0 held constant!
12
Linear Regression Model: Causal Effects and Selection Bias
13
ARE 107 Lecture Notes
Lecture 4
14
Lecture 4 - Roadmap
Last Lecture, we talked about the linear regression model with
cross-sectional data and asked when the coefficients in the linear
regression model have a causal interpretation.
I Causal Effects and Selection Bias X
I Causality and Randomized Control Trials (RCTs) X
I Observational Data and Endogeneity Bias
Question: How does this reflect on the OLS estimates?
- Formula for the OLS Estimator
- Omitted Variable Bias
- Application: Sales and Advertising
Reading: Chapter 2 in Mastering Metrics (especially 2.3 onwards)
15
Linear Regression Model: Observational Data
16
Linear Regression Model: Observational Data
E [Yi |Xi = 12] − E [Yi |Xi = 11] = β + E [ui |Xi = 12] − E [ui |Xi = 11]
| {z } |{z} | {z }
Difference in Group Means Av. Causal Effect Selection Bias
17
Linear Regression Model: Observational Data
1. β
Yi = α + βXi + ui
2. βOLS
The β that minimized sum of squared residuals. From ARE 106, you
learned that
Cov (Xi , Yi )
βOLS =
Var (Xi )
βOLS is a measure of correlation between Y and X !
Note: Cov (Yi , Xi ) is the covariance of Yi and Xi . The mathematical definition for covariance is
Cov (Yi , Xi ) = E [(Yi − E [Yi ])(Xi − E [Xi ])].
18
Linear Regression Model: Observational Data
Exercise 5. Decompose βOLS into β, the causal effect, and selection bias.
19
Linear Regression Model: Observational Data
Now we have shown that the OLS coefficient on X consists of the average
causal effect as well as a selection bias term as follows
Cov (Xi , ui )
βOLS = β +
|{z} Var (Xi )
Average Causal Effect | {z }
Selection Bias
Hence, βOLS = β. The OLS coefficient equals the average causal effect.
Cov (Xi , ui ) 6= 0.
20
20
Linear Regression Model: Observational Data
Example
A manager has data on a product’s sales (1,000 units) and spending on TV,
newspaper, and radio advertising (1,000 dollars) in 200 markets where the
product is sold. The manager wants to decide whether to increase spending on
newspaper advertising or not. He/she regresses sales on newspaper advertising
and finds the following relationship.
21
Linear Regression Model: Observational Data
22
Linear Regression Model: Observational Data
Yi = α + βXi + γWi + i
23
Linear Regression Model: Observational Data
Let us take a closer look at the omitted variable bias (OVB) formula
Cov (Wi , Xi )
OVB = γ
|{z} Var (Xi )
Coefficient on W in Y Eq. | {z }
?
24
Linear Regression Model: Observational Data
Cov (Wi , Xi )
OVB = γ ×
|{z} Var (Xi )
Coefficient on W in Y Eq. | {z }
OLS coefficient of regressing W on X
(1)
(2)
I negative if:
(1)
(2)
I zero if:
25
Linear Regression Model: Observational Data
Now back to our example, recall the short regression of sales on newspaper
advertising
Question:
I Is the coefficient on newspaper significant in the long regression?
26
Linear Regression Model: Observational Data
Exercise 8. Using the OVB formula, can you calculate the bias due to
omitting Radio from the short regression?
What is the difference between the coefficient on Newspaper from the short
vs. the long regression?
27
Linear Regression Model: Observational Data
Yi = α + βXi + γWi + i
Cov (Wi , Xi )
OVB = βSR − βLR = γ
Var (Xi )
βSR is the OLS coefficient on X in the short regression of Y on X , βLR is the
OLS coefficient on X from the long regression of Y on W and X ,
REMARKS:
I The above shows that the omitted variable bias exactly measures the
bias due to omitting a variable.
28
Table: Sales and Advertising: Short and Long Regressions
Outcome variable:
Sales
(1) (2) (3) (4)
Newspaper 0.055∗∗∗ 0.044∗∗∗ 0.007 −0.001
(0.017) (0.010) (0.015) (0.006)
(0.001)
TV 0.047∗∗∗ 0.046∗∗∗
(0.003) (0.001)
29