Académique Documents
Professionnel Documents
Culture Documents
HETEROSKEDASTICITY
HETEROSKEDASTICITY
Heteroskedasticity arises when the error term of a regression
equation does not have a constant variance:
Var(ei) ≠ σ2 .
That is, the error term can VARY with a change in one of
the explanatory variables.
For example, household expenditure on food will likely be
heteroskedastic because expenditure on food will rise with
income.
It is most likely – but not limited to – to appearing in
regressions taken from CROSS-SECTIONAL DATA.
CONSEQUENCES:
The OLS estimator is NO LONGER BLUE:
It is still LINEAR and UNBIASED but is no longer BEST.
The Estimated Standard Errors are no longer correct.
Therefore hypothesis testing may give misleading results.
MATHEMATICALLY
Var ˆ
2
The maths behind the consequences
are shown to the side.
i 1 i
N
x x 2
wrong, rather than the bottom.
N
x x i
2
i2
i 1
N
i 1
xi x 2
2
CORRECTING FOR HETEROSKED. I
To still be able to use OLS estimation
and retrieve reliable estimated standard
errors, we can calculate WHITE’S
HETEROSKEDASTICITY-
2
Then retrieve the residual errors (the ˆ
var ˆ
distance from the data points to the
FITTED line):
2
N x x 2
Residual ‘i’ = yi – b1 – b2xi
i 1 i
However:
This method allows us to use hypothesis
testing...
BUT is still not the BEST method of
estimation.
CORRECTING FOR HETEROSKED II
GENERALIZED LEAST SQUARES:
If there is Heterosked, we may impose an
assumption to make it easier to deal with.
For example, assume that in the model:
Y = B1 + B2 X + E
Var ei X i 2
That Y is expenditure on food and X is income.
1 X 2
2 Var ei i 2
We suspect that X is the explanatory variable
driving heterosked in the model.
Therefore, we can model the variance of the
Xi Xi
model as being dependent on X – shown in the
top equation.
Hence, if we work backwards – as shown – we
can calculate that we have to divide all the
terms by the SQUARE ROOT of X. Yi 1 2 X i ei
We then have a constant variance.
HOWEVER: Xi Xi Xi Xi
NO LONGER A CONSTANT TERM.