02 Simple Regression
02 Simple Regression
Topics
Simple regression
– For simplicity, say k=1. This is the situation where y depends on
only one dependent variable (x).
Terminology for Simple Regression
Simple Regression: An Example
We have some intuition that the beta on this fund is positive, and we
therefore want to find whether there appears to be a relationship
between x and y given the data that we have. The first stage would
be to form a scatter plot of the two variables.
Graph (Scatter Diagram)
45
Excess return on fund XXX
4045
Excess return on fund XXX
3540
3035
2530
2025
20
15
15
10
10
5
5
0
0
00 55 10
10 1515 20 20 25 25
Excess returnon
Excess return onmarket
marketportfolio
portfolio
Finding a Line of Best Fit
Choose intercept and slope so that the (vertical) distances from the
data points to the fitted lines are minimised (so that the line fits the data
as closely as possible):
y
yi y
ût
û i
yt denote the actual data point t
ŷt
ŷi
ŷt denote the fitted value from
the regression line
ût denote the residual, yt - ŷt
xi x
Finding a Line of Best Fit
ordinary least square (OLS)
5
5, or minimise t .
ˆ
So min. 1
u 2
ˆ
u 2
2
ˆ
u 2
3
ˆ
u 2
4
ˆ
u 2
ˆ
u 2
This is
t 1
known as the residual sum of squares.
y t = + x t + u t
So minimising t t
y ˆ
y 2
is equivalent to minimising
t with respect to
ˆ
u 2
$ and $ .
Deriving the OLS Estimator
ˆ t ˆ ˆxt , so let L
But y t t
( y ˆ
y ) 2
t
( y ˆ ˆ
x ) 2
t
t i
But y t Ty and xt Tx .
Deriving the OLS Estimator (cont’d)
t t t
x y y x ˆ
x t
x ˆ
t 0
x
2
xt yt Tyx ˆTx 2 ˆ xt 0
2
t
Deriving the OLS Estimator (cont’d)
Rearranging for $ ,
So overall we have
ˆ x y Tx y
t t
and ˆ y ˆx
x Tx2
t
2
The PRF is yt b1 b2 xt ut
We observe data for xt, but since yt also depends on ut, we must be
specific about how the ut are generated.
We usually make the following set of assumptions about the ut’s (the
unobservable error terms):
Technical Notation Interpretation
1. E(ut) = 0 The errors have zero mean
2. Var (ut) = 2 The variance of the errors is constant and finite
over all values of xt (Homoskedasticity)
3. Cov (ui,uj)=0 The errors are statistically independent of
one another (Zero Autocorrelation)
4. Cov (ut,xt)=0 No relationship between the error and
corresponding x variate
Homoskedasticity vs. Heteroskedasticity
Homoskedasticity Heteroskedasticity
Properties of the OLS Estimator
The least squares estimators $ and $ are consistent. That is, the
estimates will converge to their true values as the sample size
increases to infinity. Need the assumptions E(xtut)=0 and
Var(ut)=2 < to prove this. Consistency implies that
lim Pr ˆ 0 0
T
Estimator or Estimate?
ˆ t 2 t
) are given by x y Nx y
and ˆ y ˆx
xt Nx 2
N
t
u 2
N
uˆ t
But this estimator is a biased estimator of 2.
N 2
where uˆ 2
t is the residual sum of squares and N is the sample
size.
Questions?