Chapter Three
Chapter Three
1
i) Classical Assumptions
For the validity of a regression function and its attributes the data we use or the terms related to
our regression function should fulfill the following conditions known as classical assumptions.
1. The error terms „Ui‟ are randomly distributed or the disturbance terms are not correlated.
This means that there is no systematic variation or relation among the value of the error
terms (Ui and Uj); Where i = 1, 2, 3, …….., j = 1, 2, 3, ……. and i j .
Cov (Ui , Uj) = 0 for i j . Note that the same argument holds for residual terms when
we use sample data or sample regression function. Thus, Cov (ei, ej) = 0 for i j .
Otherwise, the error terms do not serve an adjustment purpose rather it causes an
autocorrelation problem.
2. The disturbance terms „Ui‟ have zero mean. This implies the sum of the individual
disturbance terms is zero. The deviations of the values of some of the disturbance terms are
negative; some are zero and some are positive and the sum or the average is zero. This is
given by the following identities.
U i
E(Ui) = 0 . Multiplying both sides by (sample size „n‟) we obtain the following.
n
E(Ui) = U i 0 . The same argument is true for sample regression function and so
for residual terms given as follows ei 0
If this condition is not met, then the position of the regression function (or curve) will not
be the same as where it is supposed to be. This results in an upward (if the mean of the
error term or residual term is positive) or down ward (if the mean of the error term or
residual term is negative) shift in the regression function. For instance, suppose we have
the following regression function.
i 0 1 i Ui
i E (Yi / X i ) 0 1 i U i
i E (Yi / X i ) 0 1 X i if E (U i ) 0
Otherwise the estimated models will be biased and cause the regression function to shift. For
instance, if E (U i ) 0 (or positive) it is going to shift the estimation upward from the true
representative model. Similar argument is true for residual term of sample regression
function. This is demonstrated by the following figure.
2
Figure 1: Regression Function/curve if the mean of error term is not zero
3. The disturbance terms have constant variance in each period. This is given as follows:
Var U i E ((U i E (U i )) 2 = u . This assumption is known as the
2 2
3
7. Normality assumption-The disturbance term Ui is assumed to have a normal distribution
with zero mean and a constant variance. This assumption is given as follows:
Ui ˜
N 0, u
2
. This assumption is a combination of zero mean of error term
assumption and homoscedasticity assumption. This assumption or combination of
assumptions is used in testing hypotheses about significance of parameters. It is also useful
in both estimating parameters and testing their significance in maximum likelihood
method.
8. Explanatory variables should not be perfectly, linearly and/or highly correlated. Using
explanatory variables which are highly or perfectly correlated in a regression function
causes a biased function or model. It also results in multicollinearity problem.
9. The relationship between variables (or the model) is correctly specified. For instance all the
necessary variables are included in the model. The variables are in the form that best
describes the functional relationship. For instance, “Y = f (X 2)” may better reflect the
relationship between Y and X than “Y = f (X )”.
10. The explanatory variables do not have identical value. This assumption is very important
for improving the precision of estimators.
ii) OLS method of Estimation
Estimating a linear regression function using the Ordinary Least Square (OLS) method is simply
about calculating the parameters of the regression function for which the sum of square of the
error terms is minimized. The procedure is given as follows. Suppose we want to estimate the
following equation
Yi 0 1 X i U i
Since most of the time we use sample (or it is difficult to get population data) the corresponding
sample regression function is given as follows.
The method of OLS involves finding the estimates of the intercept and the slope for which the
sum squares given by the Equation is minimized. To minimize the residual sum of squares we
take the first order partial derivatives of Equation 2.8 and equate them to zero.
4
That is, the partial derivative with respect to ˆ 0 :
ei
2
2 Yi ˆ0 ˆX i (1) 0 2.9
ˆ
0
X i Yi ˆ0 X i ˆ1 X i
2
2.15
^ ^
Note that the equation (Yi 0 1 X i ) 2 is a composite function and we should apply a chain
rule in finding the partial derivatives with respect to the parameter estimates.
Equations 2.12 and 2.15 are together called the system of normal equations. Solving the system
of normal equations simultaneously we obtain:
n XY ( X )( Y )
ˆ1
Or
n X 2 ( X ) 2
_ _
XY n Y X and we have ˆ 0 Y ˆ1 X
1 _
from above
X i2 n X 2
Now we have the formula to estimate the simple linear regression function. Let us illustrate with
example.
Example 2.4: Given the following sample data of three pairs of „Y‟ (dependent variable) and „X‟
(independent variable), find a simple linear regression function; Y = f(X).
5
Yi Xi
10 30
20 50
30 60
n XY ( X )( Y )
3(3100) (140)(60)
ˆ1 0.64
3(7000) (140) 2
n X 2 ( X ) 2
6
x y
i i
Then, the coefficients can be represented by alternative formula given below ˆ1
x i
2
Then
x i yi
300 , and ˆ0 Y ˆ1 X =20-(0.64) (46.67) = -10 with results
ˆ1 0.64
466.67
x i
2
7
3. The mean of 0 E ( 0 ) 0
U2 X i2
4. The variance of 0 E (( 0 E ( 0 )) 2
n xi2
ei2
5. The estimated value of the variance of the error term U
2
n2
8
The total variation of the dependent variable is given in the following form; TSS=ESS + RSS,
which means total sum of square of the dependent variable is split into explained sum of square
and residual sum of square.
ei yi y i
yi yi ei
2
yi2 yi ei2 2 yi ei
2
y i2 y i ei2 2 y i ei
But y i ei 0
2
Therefore, y y i ei2
2
i
(Yˆ Y )2 yˆ
2
i i
R2
Explained Variation in Y 2.16
Total Variation in Y
(Y i Y )2 y i
2
2
Since y i 1 x i y i the coefficient of determination can also be given as
1 xi y i
R2
y i2
Or
e e t t 1 ˆ
P 2.17
e
2
t 1
The higher the coefficient of determination is the better the fit. Conversely, the smaller the
coefficient of determination is the poorer the fit. That is why the coefficient of determination is
used to compare two or more models. One minus the coefficient of determination is called the
coefficient of non-determination, and it gives the proportion of the variation in the dependent
variable that remained undetermined or unexplained by the model.
ii) Testing the significance of a given regression coefficient
Since the sample values of the intercept and the coefficient are estimates of the true population
parameters, we have to test them for their statistical reliability.
The significance of a model can be seen in terms of the amount of variation in the dependent
variable that it explains and the significance of the regression coefficients.
There are different tests that are available to test the statistical reliability of the parameter
estimates. The following are the common ones;
9
A) The standard error test
B) The standard normal test
C) The students t-test
Now, let us discuss them one by one.
A) The Standard Error Test
This test first establishes the two hypotheses that are going to be tested which are commonly
known as the null and alternative hypotheses. The null hypothesis addresses that the sample is
coming from the population whose parameter is not significantly different from zero while the
alternative hypothesis addresses that the sample is coming from the population whose parameter
is significantly different from zero. The two hypotheses are given as follows:
H0: βi=0
H1: βi≠0
The standard error test is outlined as follows:
1. Compute the standard deviations of the parameter estimates using the above formula for
variances of parameter estimates. This is because standard deviation is the positive square root of
the variance.
U2
se( 1 )
xi2
U2 X i2
se( 0 )
n xi2
2. Compare the standard errors of the estimates with the numerical values of the estimates and
make decision.
A) If the standard error of the estimate is less than half of the numerical value of the estimate, we
1
can conclude that the estimate is statistically significant. That is, if se( i ) ( i ) , reject the
2
null hypothesis and we can conclude that the estimate is statistically significant.
B) If the standard error of the estimate is greater than half of the numerical value of the estimate,
1
the parameter estimate is not statistically reliable. That is, if se( i ) ( i ) , conclude to accept
2
the null hypothesis and conclude that the estimate is not statistically significant.
B) The Standard Normal Test
This test is based on the normal distribution. The test is applicable if:
The standard deviation of the population is known irrespective of the sample size
The standard deviation of the population is unknown provided that the sample size is
sufficiently large (n>30).
The standard normal test or Z-test is outline as follows;
10
1. Test the null hypothesis H 0 : i 0 against the alternative hypothesis H1 : i 0
2. Determine the level of significant ( ) in which the test is carried out. It is the probability
of committing type I error, i.e. the probability of rejecting the null hypothesis while it is
true. It is common in applied econometrics to use 5% level of significance.
3. Determine the theoretical or tabulated value of Z from the table. That is, find the value of
Z 2 from the standard normal table. Z 0.025 1.96 from the table.
4. Make decision. The decision of statistical hypothesis testing consists of two decisions;
either accepting the null hypothesis or rejecting it.
If Z cal Z tab , accept the null hypothesis while if Z cal Z tab , reject the null hypothesis. It
is true that most of the times the null and alternative hypotheses are mutually exclusive.
Accepting the null hypothesis means that rejecting the alternative hypothesis and
rejecting the null hypothesis means accepting the alternative hypothesis.
Example: If the regression has a value of 1 =29.48 and the standard error of 1 is 36. Test the
hypothesis that the value of 1 25 at 5% level of significance using standard normal test.
Solution: We have to follow the procedures of the test.
H 0 : 1 25
H1 : 1 25
After setting up the hypotheses to be tested, the next step is to determine the level of significance
in which the test is carried out. In the above example the significance level is given as 5%.
The third step is to find the theoretical value of Z at specified level of significance. From the
standard normal table we can get that Z 0.025 1.96 .
The fourth step in hypothesis testing is computing the observed or calculated value of the
standard normal distribution using the following formula.
ˆ1 1 29.48 25
Z cal 0.12 . Since the calculated value of the test statistic is less than the
se( ˆ1 ) 36
tabulated value, the decision is to accept the null hypothesis and conclude that the value of the
parameter is 25.
C) The Student t-Test
In conditions where Z-test is not applied (in small samples), t-test can be used to test the
statistical reliability of the parameter estimates. The test depends on the degrees of freedom that
the sample has. The test procedures of t-test are similar with that of the z-test. The procedures are
outlined as follows;
1. Set up the hypothesis. The hypotheses for testing a given regression coefficient is given by:
11
H 0 : i 0
H1 : i 0
2. Determine the level of significance for carrying out the test. We usually use a 5% level
significance in applied econometric research.
3. Determine the tabulated value of t from the table with n-k degrees of freedom, where k is the
number of parameters estimated.
4. Determine the calculated value of t. The test statistic (using the t- test) is given by:
ˆi
tcal
se( ˆi )
The test rule or decision is given as follows:
Reject H0 if | t cal | t / 2,nk
iii) Confidence Interval Estimation of the regression Coefficients
We have discussed the important tests that that can be conducted to check model and parameters
validity. But one thing that must be clear is that rejecting the null hypothesis does not mean that
the parameter estimates are correct estimates of the true population parameters. It means that the
estimate comes from the sample drawn from the population whose population parameter is
significantly different from zero. In order to define the range within which the true parameter
lies, we must construct a confidence interval for the parameter. Like we constructed confidence
interval estimates for a given population mean, using the sample mean (in Introduction to
Statistics), we can construct 100(1- ) % confidence intervals for the sample regression
coefficients. To do so we need to have the standard errors of the sample regression coefficients.
The standard error of a given coefficient is the positive square root of the variance of the
coefficient. Thus, we have discussed that the formulae for finding the variances of the regression
coefficients are given as.
X
2
i
1
Variance of the slope ( ˆ1 ) is given by var( ˆ1 ) u
2
2.19
x
2
i
e
2
i
Where, u2 (2.20) is the estimate of the variance of the random term and k is
nk
the number of parameters to be estimated in the model. The standard errors are the positive
square root of the variances and the 100 (1- ) % confidence interval for the slope is given by:
12
1 t (n k )(se( 1 )) 1 1 t (n k )(se( 1 ))
2 2
13
Table 2: Data for computation of different parameters
Time Y X XY X2 Y2 x Y Xy x2 y2 Yˆ ei ei2
1 69 9 621 81 4761 0 6 0 0 36 63.00 6.00 36.00
2 76 12 912 144 5776 3 13 39 9 169 72.75 3.25 10.56
3 52 6 312 36 2704 -3 -11 33 9 121 53.25 -1.25 1.56
4 56 10 560 100 3136 1 -7 -7 1 49 66.25 -10.25 105.06
5 57 9 513 81 3249 0 -6 0 0 36 63.00 -6.00 36.00
6 77 10 770 100 5929 1 14 14 1 196 66.25 10.75 115.56
7 58 7 406 49 3364 -2 -5 10 4 25 56.50 1.50 2.25
8 55 8 440 64 3025 -1 -8 8 1 64 59.75 -4.75 22.56
9 67 12 804 144 4489 3 4 12 9 16 72.75 -5.75 33.06
10 53 6 318 36 2809 -3 -10 30 9 100 53.25 -0.25 0.06
11 72 11 792 121 5184 2 9 18 4 81 69.50 2.50 6.25
12 64 8 512 64 4096 -1 1 -1 1 1 59.75 4.25 18.06
Sum 756 108 6960 1020 48522 0 0 156 48 894 756.00 0.00 387.00
e
2
i
387
R2 1 1 1 0.43 0.57
894
y i
2
This result shows that 57% of the variation in the quantity supplied of the commodity under
consideration is explained by the variation in the price of the commodity; and the rest 37%
remain unexplained by the price of the commodity. In other word, there may be other important
explanatory variables left out that could contribute to the variation in the quantity supplied of the
commodity, under consideration.
14
2. Run significance test of regression coefficients using the following test methods
Fitted regression line for the data given is:
Yi 33.75 3.25 X i , where the numbers in parenthesis are standard errors of the respective
(8.3) (0.9)
coefficients.
A. Standard Error test
In testing the statistical significance of the estimates using standard error test, the following
information needed for decision.
Since there are two parameter estimates in the model, we have to test them separately.
Testing for 1
We have the following information about 1 i.e 1 =3.25 and se( 1 ) 0.9
The following are the null and alternative hypotheses to be tested.
H 0 : 1 0
H 1 : 1 0
Since the standard error of 1 is less than half of the value of 1 , we have to reject the null
hypothesis and conclude that the parameter estimate 1 is statistically significant.
Testing for 0
Again we have the following information about 0
0 33.75 and se( 0 ) 8.3
The hypotheses to be tested are given as follows;
H0 : 0 0
H1 : 0 0
Since the standard error of 0 is less than half of the numerical value of 0 , we have to reject the
null hypothesis and conclude that 0 is statistically significant.
B. The students t-test
In the illustrative example, we can apply t-test to see whether price of the commodity is
significant in determining the quantity supplied of the commodity under consideration? Use
=0.05.
The hypothesis to be tested is:
15
H 0 : 1 0
H 1 : 1 0
Fitted regression model is indicated Yi 33.75 3.25 X i ,where the numbers in parenthesis are
(8.3) (0.9)
standard errors of the respective coefficients. To estimate confidence interval we need standard
error which is determined as follows
e
2
i
387 387
u2 38.7
nk 12 2 10
1 1
var( ˆ1 ) u 38.7( ) 0.80625
2
48
x 2
The standard error of the slope is se( ˆ1 ) var( ˆ1 ) 0.80625 0.8979
16
The tabulated value of t for degrees of freedom 12-2=10 and /2=0.025 is 2.228.
Hence the 95% confidence interval for the slope is given by:
ˆ1 3.25 (2.228)(0.8979) 3.25 2 3.25 2, 3.25 2 1.25, 5.25 . The result tells us that
at the error probability 0.05, the true value of the slope coefficient lies between 1.25 and 5.25
17
Y
i i=1,2,...K
X
Y
X => Y
K
X
Output
Input
18
Thus, given the assumption of a constant elasticity, the proper form is the exponential (log-
linear) form.
Given: Yi 0 X i i eU i
The log-linear functional form for the above equation can be obtained by a logarithmic
transformation of the equation.
ln Yi ln 0 i ln X i U i
The model can be estimated by OLS if the basic assumptions are fulfilled.
demand gd(log f)
ln Yi ln 0 1 ln X i
1
Yi 0 X i
The model is also called a constant elasticity model because the coefficient of elasticity between
Y and X (1) remains constant.
Y X d ln Y
1
X Y d ln X
This functional form is used in the estimation of demand and production functions.
Note: We should make sure that there are no negative or zero observations in the data set before
we decide to use the log-linear model. Thus log-linear models should be run only if all the
variables take on positive values.
c) Semi-log Form
The semi-log functional form is a variant of the log-linear equation in which some but not all of
the variables (dependent and independent) are expressed in terms of their logs. Such models
expressed as:
( i ) Yi 0 1 ln X 1i U i ( lin-log model ) and ( ii ) ln Yi 0 1 X 1i U i ( log-lin
model ) are called semi-log models. The semi-log functional form, in the case of taking the log
of one of the independent variables, can be used to depict a situation in which the impact of X on
Y is expected to „tail off‟ as X gets bigger as long as 1 is greater than zero.
19
y
1<0
Y=0+1Xi
1>0
Example: The Engel‟s curve tends to flatten out, because as incomes get higher, a smaller
percentage of income goes to consumption and a greater percentage goes to saving.
Consumption thus increases at a decreasing rate.
Growth models are examples of semi-log forms
d) Polynomial Form
Polynomial functional forms express Y as a function of independent variables some of which are
raised to powers others than one. For example in a second degree polynomial (quadratic)
equation, at least one independent variable is squared.
Y 0 1 X 1i 2 X 1i 3 X 2i U i
2
Such models produce slopes that change as the independent variables change. Thus the slopes of
Y with respect to the Xs are
Y Y
1 2 2 X 1 , and 3
X 1 X 2
In most cost functions, the slope of the cost curve changes as output changes.
20
Y Y
A) B)
X
Xi Impact of age on earnings
a typical cost curve
Simple transformation of the polynomial could enable us to use the OLS method to estimate the
parameters of the model
X1 X 3
2
Setting
Y 0 1 X 1i 2 X 3 3 X 2i U i
e) Reciprocal Transformation (Inverse Functional Forms)
The inverse functional form expresses Y as a function of the reciprocal (or inverse) of one or
more of the independent variables (in this case X1):
1
Yi 0 1 ( ) 2 X 2i U i
X 1i
Or
1
Yi 0 1 ( ) 2 X 2i U i
X 1i
The reciprocal form should be used when the impact of a particular independent variable is
expected to approach zero as that independent variable increases and eventually approaches
infinity. Thus as X1 gets larger, its impact on Y decreases.
1 0 0
Y 0
X 1i 1 0
0
1 0 0
Y 0
X 1i 1 0
21
An asymptote or limit value is set that the dependent variable will take if the value of the X-
variable increases indefinitely i.e. 0 provides the value in the above case. The function
approaches the asymptote from the top or bottom depending on the sign of 1.
Example: Phillips curve, a non-linear relationship between the rate of unemployment and the
percentage wage change.
1
Wt 0 1 ( ) Ut
Ut
22