Econometrics Chap 3
Econometrics Chap 3
Econometrics Chap 3
(Eq. 2.17)
• Consistency implies:
✓ As the sample size (T) tends to infinity, the probability of ˆβ being δ away
from its true value tends to zero.
✓ In other words, as the number of observations increases, the estimator
approaches its true value.
• Consistency is an asymptotic property:
✓ It holds for large sample sizes, approaching infinity.
• Assumptions for consistency:
(2.22)
✓ Assumption 1 of the CLRM was that the expected or average value of the
errors is zero. Under this assumption, (2.22) above reduces to
✓ s is also known as the standard error of the regression or the standard error
of the estimate.
✓ It is sometimes used as a broad measure of the fit of the regression equation.
✓ Everything else being equal, the smaller this quantity is, the closer is the fit of
the line to the actual data
ii) Some comments on the standard error estimators
(1) Larger sample size (T) → Smaller coefficient standard errors.
✓ T explicitly affects SE(ˆα) and implicitly affects SE(ˆβ).
✓ More information from a larger sample lead to increased confidence in the
estimates.
(2) Both SE(ˆα) and SE(ˆβ) depend on s^2 (or s) - the estimate of error variance.
✓ Larger s^2 → More dispersed residuals → Greater uncertainty in the model.
(3) Sum of squares of (xt - ¯x) affects both formulae:
✓ Larger sum of squares → Smaller coefficient variances.
✓ Figure 2.7: Small (xt - ¯x)^2 → Difficult to determine the line's position.
✓ Figure 2.8: Large (xt - ¯x)^2 → More confidence in the estimates.
(4) Term x^2_t affects only the intercept standard error, not the slope standard
error.
✓ x^2_t measures how far points are from the y-axis.
✓ Figure 2.9: Points far from the y-axis → Difficult to estimate the intercept
accurately.
✓ Figure 2.10: Points closer to the y-axis → Easier to determine where the line
crosses the y-axis.
_________end_________
✓ β=0.5091isasingle(point)estimate of the unknown population parameter,β.
✓ As stated above, the reliability of the point estimate is measured by the
coefficient’s standard error.
✓ Sample coefficients and their standard errors are used to make inferences
about population parameters.
✓ Example: The estimate of the slope coefficient is ˆβ = 0.5091, but it is
expected to vary from one sample to another.
✓ Hypothesis Testing:
✓ Hypothesis testing helps answer questions about the plausibility of
population parameters based on sample estimates.
________--end________________
❖ *********Hypothesis testing: some concepts
Hypothesis Testing Framework:
✓ Hypothesis testing involves two hypotheses that go together.
✓ Null Hypothesis (H0 or HN):
-The null hypothesis is the statement or statistical hypothesis being tested.
✓ Alternative Hypothesis (H1 or HA):
-The alternative hypothesis represents the remaining outcomes of interest.
✓ Both hypotheses are essential for hypothesis testing to compare and assess
the evidence in favor of one over the other.
✓ Example of Hypothesis Testing:
• Hypothesis: Testing the true value of β = 0.5 using the regression results
above.
• Null Hypothesis (H0):
H0: β = 0.5
The null hypothesis states that the true, but unknown value of β is 0.5.
• Alternative Hypothesis (H1):
H1: β ≠ 0.5
The alternative hypothesis represents the remaining outcomes where β is not
equal to 0.5.
• Two-Sided Test:
This is a two-sided test since the alternative hypothesis includes both
possibilities: β < 0.5 and β > 0.5.
• Hypothesis: Testing the true value of β = 0.5 using the regression results
above, with prior information suggesting β > 0.5.
• Null Hypothesis (H0):
H0: β = 0.5
The null hypothesis states that the true, but unknown value of β is 0.5.
• Alternative Hypothesis (H1):
H1: β > 0.5
The one-sided alternative hypothesis suggests that β is more than 0.5.
• One-Sided Test:
• This is a one-sided test because the alternative hypothesis only considers the
possibility of β being greater than 0.5, and β < 0.5 is no longer of interest in
this context.
✓ Two Ways to Conduct a Hypothesis Test:
1. Test of Significance Approach:
• The test of significance approach involves statistical comparison of the
estimated coefficient value and its value under the null hypothesis.
• If the estimated value is significantly different from the hypothesized value,
the null hypothesis is likely to be rejected.
2.Confidence Interval Approach:
• The confidence interval approach also compares the estimated coefficient
value with its value under the null hypothesis.
• If the value under the null hypothesis falls within the confidence interval, the
null hypothesis is less likely to be rejected.
✓ Comparison of Estimated and Hypothesized Values:
• In general terms, if the estimated value is far from the hypothesized value,
the null hypothesis is more likely to be rejected.
• Conversely, if the value under the null hypothesis and the estimated value
are close to each other, the null hypothesis is less likely to be rejected.
Example:
• Consider the estimated value ˆβ = 0.5091 from above.
• A hypothesis that the true value of β is 5 is more likely to be rejected than a
null hypothesis.
- that the true value of β is 0.5
- because the estimated value is far from 5 but relatively close to 0.5.
What is required now is a statistical decision rule that will permit the formal
testing of such hypotheses.
▪ The probability distribution of the least squares estimators
✓ In order to test hypotheses, assumption 5 of the CLRM must be used,
-namely that ut ∼ N(0,σ2)-- i.e. that the error term is normally distributed.
✓ The normal distribution is a convenient one to use for it involves only two
parameters (its mean and variance).
✓ This makes the algebra involved in statistical inference considerably simpler
than it otherwise would have been.
✓ Since yt depends partially on ut,
- it can be stated that if ut is normally distributed, yt will also be normally
distributed.
✓ Least squares estimators (ˆβ) are linear combinations of random variables
(yt).
✓ ˆβ = wtyt, where wt are effectively weights in the regression equation.
✓ The weighted sum of normal random variables (yt) is also normally
distributed.
✓ As a result, the coefficient estimates (ˆβ) will also follow a normal
distribution. Thus ,
✓ Will the coefficient estimates still follow a normal distribution if the errors do
not follow a normal distribution?
- Well, briefly, the answer is usually ‘yes’,
-provided that the other assumptions of the CLRM hold, and the sample
size is sufficiently large.
✓ Standard normal variables can be constructed from ˆα and ˆβ by subtracting
the mean and dividing by the square root of the variance
The square roots of the coefficient variances are the standard errors.
✓ Unfortunately, the standard errors of the true coefficient values under the
PRF are never known,
-- all that is available are their sample counterparts,
-the calculated standard errors of the coefficient estimates, SE(ˆα) and SE(
ˆβ).
✓ Replacing the true values of the standard errors with the sample estimated
versions induces another source of uncertainty,
- and also means that the standardised statistics follow a t-distribution with T
− 2 degrees of freedom (defined below) rather than a normal distribution, so
✓ A normal variate can be scaled to have zero mean and unit variance by
subtracting its mean and dividing by its standard deviation.
✓ There is a specific relationship between the t- and the standard normal
distribution, and the t-distribution has another parameter, its degrees of
freedom.
✓ There are broadly two approaches to testing hypotheses under regression
analysis:
a) the test of significance approach
b) and the confidence interval approach.
(5) Given a significance level, a rejection region and non-rejection region can be
determined.
-A 5% significance level means that 5% of the total distribution (area under the
curve) will be in the rejection region.
✓ Two-Sided Test Rejection Region:
-In a two-sided test, the 5% rejection region is split equally between the two
tails of the distribution.
✓ One-Sided Test Rejection Region:
-In a one-sided test, the 5% rejection region is located solely in one tail of the
distribution.
-Figure 2.14 shows the rejection region for a test with the alternative of the 'less
than' form.
-Figure 2.15 shows the rejection region for a test with the alternative of the
'greater than' form.
(6) Use the t-tables to obtain a critical value or values with which to compare the
test statistic.
-The critical value will be that value of x that puts 5% into the rejection region.
(7) Finally perform the test.
-If the test statistic lies in the rejection region then reject the null hypothesis
(H0), else do not reject H0.
✓ In Step 2, the estimated value of β is compared to the value under the null
hypothesis.
✓ The difference is "normalised" or scaled by the standard error of the
coefficient estimate.
✓ The standard error measures the confidence in the coefficient estimate from
the first stage.
✓ A small standard error leads to a larger test statistic relative to a large
standard error.
✓ A small standard error implies that even a small difference between the
estimated and hypothesised values can lead to rejecting the null hypothesis.
✓ Dividing by the standard error ensures that, under the five CLRM
assumptions, the test statistic follows a tabulated distribution.
✓ This distribution is used to determine the critical values and p-values for
hypothesis testing.
b)The confidence interval approach to hypothesis testing
✓ But this is just the rule for non-rejection under the confidence interval
approach.
✓ So it will always be the case that, for a given significance level, the test of
significance and confidence interval approaches will provide the same
conclusion by construction.
✓ One testing approach is simply an algebraic rearrangement of the other.