Lecture 3
Lecture 3
Misspecification
• Model specification refers to the set of variables included in the regression and
the regression equation’s functional form.
• When estimating a regression, we assume it has the correct functional form, an
assumption that can fail in different ways, as shown in Exhibit 2.
Violations of regression
assumptions: Heteroskedasticity
• An important assumption underlying linear regression is that the variance of
errors is constant across observations (errors are homoskedastic).
• Residuals in financial model estimations, however, are often heteroskedastic,
meaning the variance of the residuals differs across observations.
• Heteroskedasticity may arise from model misspecification, including omitted
variables, incorrect functional form, and incorrect data transformations,
as well as from extreme values of independent variables.
Consequences of Heteroskedasticity
• Positive serial correlation also means a negative residual for one observation
increases the chance of a negative residual for another observation.
• We examine positive serial correlation because it is the most common type and
assume first-order serial correlation, or correlation between adjacent
observations. In a time series, this means the sign of the residual tends to
persist from one period to the next.
• Positive serial correlation does not affect the consistency of regression
coefficients, but it does affect statistical tests.
• First, the F-statistic may be inflated because the MSE will tend to
underestimate the population error variance.
• In the case of asset prices, if these prices were to exhibit a pattern, investors
would likely discern this pattern and exploit it to capture alpha, thereby
eliminating such a pattern.
• However, the DW test is limiting because it applies only to testing for first-
order serial correlation.
• Methods for adjusting standard errors are standard in many software packages.
The corrections are known by various names, including serial-correlation
consistent standard errors, serial correlation and heteroskedasticity adjusted
standard errors, Newey–West standard errors, and robust standard errors.
• With more than two independent variables, high pairwise correlations are
not a necessary condition for multicollinearity.
• VIF increases as the correlation increases; the higher the VIF, the more likely a
given independent variable can be accurately predicted from the remaining
independent variables, making it increasingly redundant.