Econometrie
Econometrie
Econometrie
to Econometrics,
Third Edition
Update
Chapters 1, 2 and 3
Variables:
• 5PthP grade test scores (Stanford-9 achievement test,
combined math and reading), district average
• Student-teacher ratio (STR) = no. of students in the
district divided by no. full-time equivalent teachers
= –
= 657.4 – 650.0
= 7.4
Is this a large difference in a real-world sense?
– Standard deviation across districts = 19.1
– Difference between 60PthP and 75PthP percentiles of test
score distribution is 667.6 – 659.4 = 8.2
– This is a big enough difference to be important for school
reform discussions, for parents, or for a school
committee?
Size sBYB n
small 657.4 19.4 238
large 650.0 17.9 182
= 4.05
( – ) ± 1.96×SE( – )
= 7.4 ± 1.96×1.83 = (3.8, 11.0)
Two equivalent statements:
1. The 95% confidence interval for Δ doesn’t include 0;
2. The hypothesis that Δ = 0 is rejected at the 5% level.
Population
• The group or collection of all possible entities of interest
(school districts)
• We will think of populations as infinitely large (∞ is an
approximation to “very big”)
Random variable Y
• Numerical summary of a random outcome (district average
test score, district STR)
skewness =
= measure of asymmetry of a distribution
• skewness = 0: distribution is symmetric
• skewness > (<) 0: distribution has long right (left) tail
kurtosis =
= measure of mass in tails
= measure of probability of large values
• kurtosis = 3: normal distribution
• skewness > 3: heavy tails (“leptokurtotic”)
So is the correlation…
corr(X,Z) = = rBXZB
• –1 ≤ corr(X,Z) ≤ 1
• corr(X,Z) = 1 mean perfect positive linear association
• corr(X,Z) = –1 means perfect negative linear association
• corr(X,Z) = 0 means no linear association
Conditional distributions
• The distribution of Y, given value(s) of some other random
variable, X
• Ex: the distribution of test scores, given that STR < 20
Conditional expectations and conditional moments
• conditional mean = mean of conditional distribution
– = E(Y|X = x) (important concept and notation)
• conditional variance = variance of conditional distribution
• Example: E(Test scores|STR < 20) = the mean of test
scores among districts with small class sizes
The difference in means is the difference between the
means of two conditional distributions:
Estimation
is the natural estimator of the mean. But:
a) What are the properties of ?
b) Why should we use rather than some other estimator?
• YB1B (the first observation)
• maybe unequal weights – not simple average
• median(YB1B,…, YBnB)
The starting point is the sampling distribution of …
• General case – that is, for Yi i.i.d. from any distribution, not
just Bernoulli:
• mean: E( ) = E( )= = = μY
= E[ – μY]2
=E
=E
=
© Pearson Education Limited 2015
1-36
Mean and variance of sampling
distribution of , ctd.
E( ) = μY
var( )=
Implications:
1. is an unbiased estimator of μY (that is, E( ) = μY)
2. var( ) is inversely proportional to n
1. the spread of the sampling distribution is
proportional to 1/
2. Thus the sampling uncertainty associated with
is proportional to 1/ (larger samples, less
uncertainty, but square-root law)
• is unbiased: E( ) = μY
• is consistent: μY
• is the “least squares” estimator of μY; solves,
= =
Set derivative to zero and denote optimal value of m
by m̂ :
n
= mˆ = nmˆ or m̂ =
i 1
=
© Pearson Education Limited 2015
1-44
Why Use To Estimate μY, ctd.
p-value = ,
=
~ probability under left+right N(0,1) tails
=
where = std. dev. of the distribution of = σY/ .
= = “sample variance of Y”
Fact:
If (Y1,…,Yn) are i.i.d. and E(Y4) < ∞ , then
p-value = ,
~
= (large n)
so
p-value = ( estimated)