Linear Regression Analysis: Module - Vii
Linear Regression Analysis: Module - Vii
Linear Regression Analysis: Module - Vii
MODULE – VII
Lecture – 25
Generalized and Weighted
Least Squares Estimation
Dr. Shalabh
Department of Mathematics and Statistics
Indian Institute of Technology Kanpur
2
The usual linear regression model assumes that all the random error components are identically and independently
distributed with constant variance. When this assumption is violated, then ordinary least squares estimator of regression
coefficient looses its property of minimum variance in the class of linear and unbiased estimators. The violation of such
assumption can arise in anyone of the following situations:
1. The variance of random error components is not constant.
2. The random error components are not independent.
3. The random error components do not have constant variance as well as they are not independent.
In such cases, the covariance matrix of random error components does not remain in the form of an identity matrix but can
be considered as any positive definite matrix. Under such assumption, the OLSE does not remain efficient as in the case of
identity covariance matrix. The generalized or weighted least squares method is used in such situations to estimate the
parameters of the model.
In this method, the deviation between the observed and expected values of yi is multiplied by a weight ωi where ωi
is chosen to be inversely proportional to the variance of yi.
n
For simple linear regression model, the weighted least squares function is S ( β 0=
, β1 ) ∑ω ( y − β − β1 xi ) .
2
i i 0
The least squares normal equations are obtained by differentiating S ( β 0 , β1 ) with respect to β 0 and β1 and equating
them to zero as n n n
βˆ0 ∑ ωi +βˆ1 ∑ ωi xi =
∑ ωi yi
=i 1 =i 1 =i 1
n n n
0 ∑ ωi xi +β1 ∑ ωi xi =
βˆ ∑ ωi xi yi .
ˆ 2
=i 1 =i 1 =i 1
Solution of these two normal equations give the weighted least squares estimate of β 0 and β1 .
3
The OLSE of β is
b = ( X ' X ) −1 X ' y.
In such cases OLSE gives unbiased estimate but has more variability as
KK ' = Ω .
K −1 y K −1 X β + K −1ε
=
or
z Bβ + g
=
=
where
−1
z K= −1
y, B K= X , g K −1ε . Now observe that
= −1)
E ( g ) K= E (ε ) 0
E { g − E ( g )}{ g − E ( g )} '
V (g) =
= E ( gg ')
So either minimize
S (β ) = g ' g
= ε ' Ω −1ε
=( y − X β ) ' Ω −1 ( y − X β )
and get normal equations as
(X'Ω-1 X ) βˆ =
X ' Ω −1 y
or βˆ =Ω
( X ' −1 X ) −1 X ' Ω −1 y.
βˆ = ( B ' B) −1 B ' z
= ( X ' K '−1 K −1 X ) −1 X ' K '−1 K −1 y
( X ' −1 X ) −1 X ' Ω −1 y.
=Ω
or βˆ − β =
( B ' B ) −1 B ' g .
6
Then
E ( βˆ − β ) ( B ' B=
= ) −1 B ' E ( g ) 0
which shows that GLSE is an unbiased estimator of β . The covariance matrix of GLSE is given by
{ }{
E βˆ − E ( βˆ ) βˆ − E ( βˆ ) '
V ( βˆ ) =
}
= E ( B ' B ) −1 B ' gg ' B '( B ' B ) −1
( X ' −1 X ) −1 X ' Ω −1 E ( y )
=Ω
( X ' −1 X ) −1 X ' Ω −1 X β
=Ω
= β.
= ( X ' Ω −1 X ) −1.
Thus,
Var ( ' βˆ ) = 'Var ( βˆ )
Let β be another unbiased estimator of β that is a linear combination of the data. Our goal, then, is to show that
Var ( ' β ) ≥ '( X ' Ω X ) with at least one such that Var ( ' β ) ≥ '( X ' Ω −1 X ) −1 .
−1 −1
We first note that we can write any other estimator of β that is a linear combination of the data as
( X ' Ω −1 X ) −1 X ' Ω −1 + B y + b*
β= 0
where B is an p x n matrix and bo* is a p x 1 vector of constants that appropriately adjusts the GLS estimator to form
the alternative estimate. Then
(
( β ) E ( X ' Ω −1 X ) −1 X ' Ω −1 + B y + bo*
E= )
= ( X ' Ω −1 X ) −1 X ' Ω −1 + B E ( y ) + b0*
(
V ( β ) Var ( X ' Ω −1 X ) −1 X ' Ω −1 + B y
= )
= ( X ' Ω −1 X ) −1 X ' Ω −1 + B V ( y ) ( X ' Ω −1 X ) −1 X ' Ω −1 + B '
(
= ' ( X ' Ω −1 X ) −1 + BΩB ' )
= '( X ' Ω −1 X ) −1 + ' BΩB '
We note that Ω is a positive definite matrix. Consequently, there exists some nonsingular matrix K such that Ω =K ' K As
a result, BΩB ' =
BK ' KB ' is at least positive semidefinite matrix; hence, ' BΩB ' ≥ 0. Next note that we can define
* = KB ' .
As a result,
p
' BΩB ' = =*' *
∑
i =1
*2
i
all i = 1, 2, …, k. Thus, the GLS estimate of β is the best linear unbiased estimator.
10
1
ω 0 0 0
1
1
2
0 0 0
V (ε )= σ Ω= σ
2
ω2 .
1
0 0 0
ωn
The estimation procedure is usually called as weighted least squares.
Let W = Ω −1 then the weighted least squares estimator of β is obtained by solving normal equation ( X 'WX ) βˆ = X 'Wy
which gives βˆ = ( X 'WX ) X 'Wy where
−1
ω1 , ω2 ,..., ωn are called the weights.
The observations with large variances usual have smaller weights than observations with small variance.