CHAPTER 8
TEACHING NOTES
This is a good place to remind students that homoskedasticity played no role in showing that
OLS is unbiased for the parameters in the regression equation.
In addition, you probably should
mention that there is nothing wrong with the
R
squared or adjusted
R
squared as goodnessoffit
measures.
The key is that these are estimates of the population
R
squared, 1 – [Var(
u
)/Var(
y
)],
where the variances are the
unconditional
variances in the population.
The usual
R
squared, and
the adjusted version, consistently estimate the population
R
squared whether or not Var(
u

x
) =
Var(
y

x
) depends on
x
.
Of course, heteroskedasticity causes the usual standard errors,
t
statistics,
and
F
statistics to be invalid, even in large samples, with or without normality.
By explicitly stating the homoskedasticity assumption as conditional on the explanatory
variables that appear in the conditional mean, it is clear that only heteroskedasticity that depends
on the explanatory variables in the model affects the validity of standard errors and test statistics.
The version of the BreuschPagan test in the text, and the White test, are ideally suited for
detecting forms of heteroskedasticity that invalidate inference obtained under homoskedasticity.
If heteroskedasticity depends on an exogenous variable that does not also appear in the mean
equation, this can be exploited in weighted least squares for efficiency, but only rarely is such a
variable available.
One case where such a variable is available is when an individuallevel
equation has been aggregated.
I discuss this case in the text but I rarely have time to teach it.
As I mention in the text, other traditional tests for heteroskedasticity, such as the Park and
Glejser tests, do not directly test what we want, or add too many assumptions under the null.
The GoldfeldQuandt test only works when there is a natural way to order the data based on one
independent variable.
This is rare in practice, especially for crosssectional applications.
Some argue that weighted least squares estimation is a relic, and is no longer necessary given the
availability of heteroskedasticityrobust standard errors and test statistics.
While I am
sympathetic to this argument, it presumes that we do not care much about efficiency.
Even in
large samples, the OLS estimates may not be precise enough to learn much about the population
parameters.
With substantial heteroskedasticity we might do better with weighted least squares,
even if the weighting function is misspecified.
As discussed in the text on pages 287288, one
can, and probably should, compute robust standard errors after weighted least squares.
For
asymptotic efficiency comparisons, these would be directly comparable to the
heteroskedasiticityrobust standard errors for OLS.
This preview has intentionally blurred sections. Sign up to view the full version.
View Full Document
This is the end of the preview.
Sign up
to
access the rest of the document.
 Spring '10
 Jack
 Econometrics, Regression Analysis, Standard Error, standard errors, OLS

Click to edit the document details