3A Demand Estimationl
3A Demand Estimationl
3A Demand Estimationl
&
Forecasting
Definition of Elasticity of demand
Price Elasticity of demand:
Income Elasticity:
Cross Price Elasticity:
|
|
.
|
\
|
A
A
=
q
p
p
q
e
p
|
|
.
|
\
|
A
A
=
q
I
I
q
e
I
|
|
.
|
\
|
A
A
=
q
p
p
q
e
r
r
p
r
Interpreting the Price Elasticity of
Demand: How Elastic Is Elastic?
Demand is elastic if the price elasticity of
demand is greater than 1
Inelastic if the price elasticity of demand is
less than 1, and
Unit-elastic if the price elasticity of
demand is exactly 1.
Highway department
charges for crossing a
bridge
Nature of goods according to Income
elasticity
e
I
>0 => Normal Goods
e
I
< 0 => Inferior Goods
e
I
<1 => Necessities
e
I
>1 => Luxury Goods
Cross-Price Elasticity
Goods are substitutes when the cross-price
elasticity of demand is positive
e.g. Coke & Pepsi, Zen & Santro
Goods are complements when the cross-price
elasticity of demand is negative
e.g. tea & sugar, petrol & petrol-driven car
Alcoholic Beverages elasticities (e)
Many public policy issues are related to the
consumption of alcoholic beverages
Spirits refer to all beverages that contain
alcohol other than beer & wine
Price elasticity (e
pb
) of dd for beer -0.23
Cross-price (e
pb,
pw
) 0.31
Cross-price (e
pb,
ps
) 0.15
Income elasticity (e
Ib
) -0.09
Income elasticity (e
Iw
) 5.03
Income elasticity (e
Is
) 1.21
Alcoholic Beverages elasticities (e)
Demand for beer inelastic
10% increase in beer price will result in 2.3% decrease in
beer demand
Wine & spirit are substitutes for beer
A 10% increase in wine price will result in 3.1% increase in
the quantity of beer demanded
Similarly for spirit, a 10% increase will increase 1.5%
increase in quantity of beer demand
Beer is an inferior good
10% increase in income will result in 0.9% decline in
quantity of beer demanded
Both wine & spirit are luxury goods as income
elasticities are >1
Information about demand is essential for
making pricing and production decisions
A knowledge of future demand conditions can
also be extremely useful for managers due to
huge financial implications
Large business firms pioneered the use of
empirical demand functions and forecasts for
future production-pricing decisions
Determinants of Demand
Consumer Income (more purchasing power)
Price of the product
The prices of related goods
Substitute Goods (e.g. petrol vs. diesel)
Complementary Goods (diesel car & diesel sale)
Consumer expectations of future price & income
Population & growth of the economy
Consumer tastes and preferences
Demand=f(Y, Pr, Po, ..)
Methods of Demand Estimation
Interview and Experimental Methods
Expert Opinion
Consumer Interviews/ surveys
Interviews can solicit useful information when market data
is scarce.
Sample selection to represent consumer population & skill
of surveyors are important
Market Experiments
Controlled experiments in test markets can generate useful
insights
Advantage over surveys as it reflect actual consumer
behavior
Experiments can be expensive
General Empirical Demand Specification
Q = f(P, M, P
R
N)
Where,
Q= quantity demanded
P = Price of the good
M = Consumers income
P
R
= Price(s) of the related product(s)
N = Number of buyers
Linear form of the demand function is
Q = a + bP + cM + dP
R
+ eN
We need to know the value of a, b, e..how ??
There are many ways but most common one is
through Regression Analysis
Regression Analysis
Regression analysis is concerned with
the study of the relationship between
one variable called explained or
dependent variable (y) and one or more
other variables called independent or
explanatory variables (x
1,
x
2
x
n
)
Y = f (x
1,
x
2
x
n
)
Methodology for Regression Analysis
Theory
Mathematical model of theory
Econometric model of theory
Data collection
Estimation of econometric model
Hypothesis testing
Forecasting
Using the model for policy purpose
Specification of Mathematical & Econometric Model
Y = B
1
+ B
2
X; Mathematical model (Deterministic)
Y = B
1
+ B
2
X + u Econometric model (Example of
linear regression model)
Y Dependent Variable; X Independent Variable; u Error term
B
1
& B
2
are parameters to be estimated
X
Y
B
2
* * *
* * *
X
Y
B
1
Econometric Model
Actual = systemic part+ random error
Say, Consumption (C) = Function (f) of income
(I) with error (u)
C = f(I) + u
u represents the combined influence on
dependent variable of a large number of
independent variables that are not explicitly
introduced in the regression model
We hope that influence of those omitted or
neglected variables is small and at best
random
Assumptions
The relationship between X & Y is linear
The Xs are non-stochastic variables whose
values are fixed
The error has zero expected value; E(u)=0
The error term has constant variance; E(u
2
) =
2
homoscedastic
Errors are statistically independent.
Thus, E(u
i
u
j
)=0 for all i j no auto
correlation
The error term is normally distributed;
u ~ N (0,
2
)
u
i
X
i
= 0 u & X are uncorrelated
Y~ N (B
1
+ B
2
X,
2
)
Linearity Assumption
The term linear in a simple regression model does not mean a
linear relationship between variables, but a model in which the
parameters enter the model in a linear way
A function is said to be linear in parameter if it
appears with a power of one and is not multiplied or
divided by any other parameters
Useful Functional Form
Linear:
Reciprocal
Log-Log
Useful Functional Form
Log-linear
Linear-log
Log-inverse
Population Regression Function
Let Y represents weekly expenditure on
lottery &
X represents weekly personal disposable
income
For simplicity, we assume a hypothetical
population of 100 players, which has been
divided into 10 PDI classes in increments
of $25 starting with $150 and ending
with $375
Weekly exp on Lotto and weekly PDI
150 175 200
PDI,X
Y, Weekly exp on Lotto
PRL
E(Y/X
i
) = B
1
+ B
2
X
(mathematical)
Y
i
= B
1
+ B
2
X
i
+u
i
(stochastic,
individual values
different from mean
values)
B
1
B
2
parameters
225
u
i
u
i
PRF
For any X value, there are 10 Y values
Also, there is a general tendency for Y
to increase as X increases people with
higher PDI likely to spend more on
lottery.
This will be more clear if we take mean
value of Y corresponding to various Xs
If we connect various mean values of Y,
the resulting line is called PRL
Sample Regression Function (SRF)
In practice, we rarely have population data but only sample from
the population
Suppose we have randomly selected sample of Y values
corresponding to X values
Now we have only one Y corresponding to each X
We cannot say which SRL represent PRL
Can we estimate PRF from sample data?
Y X Y X
18 150 23 150
24 175 18 175
26 200 24 200
23 225 25 225
30 250 28 250
27 275 27 275
34 300 31 300
35 325 29 325
* * * *
* * * * * *
* * * * *
* * * * *
SRL 1
SRL 2
X
Y
SRF
Here, SRL: =b
1
+b
2
X
i
Where , b
1
,b
2
are estimator of E(Y/X
i
), B
1
and B
2
An estimator is a formula that suggests how
we can estimate population parameter
A particular numerical value obtained by the
estimator in an application is an estimate
Stochastic SRF: Y
i
=b
1
+b
2
X
i
+e
i,
e
i
=estimator of
u
i
SRF
Thus, e
i
= Y
i
Granted that SRF is only
approximation of PRF, can we find a
method that will make this
approximation as close as possible?
Or, how should we construct SRF so
that b
1
& b
2
are as close as B
1
& B
2
?
Population & Sample Regression Line
Suppose we would like to estimate demand of
rice in Gurgaon and the demand =f(income)
One way to estimate this is to go each person
in Gurgaon to collect data on income and rice
consumption to estimate the equation
C = B
1
+ B
2
M, where B
1
& B
2
are parameters
to be estimated
Other way is to collect data from a sample of
say 100 people and estimate C = b
1
+ b
2
M
Population & Sample Regression Line
However, for another sample we may get C =
c
1
+ c
2
M and so on
We cannot say which SRL represent PRL
Can we estimate PRF from sample data?
Granted that SRF is only approximation of PRF,
can we find a method that will make this
approximation as close as possible?
Or, how should we construct SRF so that b
1
&
b
2
are as close as B
1
& B
2
?
Estimation of parameters:
Method of Ordinary Least Squares
We have, e
i
= Y
i
= Y
i
- b
1
- b
2
X
i
Objective is to choose b
1
& b
2
so that e
i
are as
small as possible
OLS states that b
1
& b
2
should be chosen in such a
way that RSS in minimum
Thus, minimise e
i
2
= (Y
i
- b
1
- b
2
X
i
)
2
b
2
= x
i
y
i
/ x
i
2
=
b
1
= - b
2
(
t
X -
X
_
) (
t
Y -
Y
_
)/ (
t
X -
X
_
)
2
Estimating coefficients
Consider a firm with a fixed capital stock that has
been rented under a long-term lease for Rs 100 per
production period. Other input of the firms
production process is labor, which can be increased or
decreased depending on the firms needs. So, cost of
the capital input is fixed and cost of labor is variable.
The manager of the firm wants to know the
relationship between output and cost. This will allow
the manager to predict the cost of any specified rate
of output for the next production period
The manager is interested to estimate the
coefficients b
1
and b
2
of the function
Y = b
1
+ b
2
X, where Y is total cost and X
is total output
Estimates
Cost
(Y
t
)
Output
(X
t
)
t
Y -
Y
_
t
X -
X
_
(
t
X -
X
_
)
2
(
t
X -
X
_
) (
t
Y -
Y
_
)
100 0 -137 -12.29 151.04 1645.45
150 5 -87.14 -7.29 53.14 635.25
160 8 -77.14 -4.29 18.4 330.93
240 10 -2.86 -2.29 5.24 -6.55
230 15 -7.14 2.71 7.34 -19.35
370 23 132.86 10.71 114.7 1422.93
410 25 172.86 12.71 161.54 2197.05
Y
_
=
237.14
X
_
=
12.29
(
t
X -
X
_
)
2
= 511.4
(
t
X -
X
_
) (
t
Y -
Y
_
)
=6245.71
Estimates
Y
= 87.08 + 12.21 X
One unit change in X results in 12.21 units change in Y
b
2
= (
t
X - X ) (
t
Y - Y )/ (
t
X - X )
2
= 12.21
b
1
= Y - b
2 X
= 237.14 12.21 (12.29) = 87.08
EVIEWS
Estimates
So far we have estimated b
1
& b
2
using OLS
It is evident that least square estimates are
a function of sample data
Since the data are likely to change from
sample to sample, the estimates will also
change
Therefore, what is needed is some measure of
reliability or precision of the estimators b
1
&
b
2,
which can be measured by standard error
Variances (& SEs) of OLS estimators
(T-2) is called dof, number of independent observations, as we loose 2 dof
to compute b
1
& b
2
in estimating Y(cap)
Computing sources of variation
Y
t
Total
Variation
(
t
Y - Y )
2
t
Y
=
1
b +
2
b X
t
Explained
Variation
(
t
Y
- Y )
2
Unexplained
Variation
(
t
Y -
t
Y
)
2
100 18,807.38 87.08 22,518 166.93
150 7593.38 148.13 7922.78 3.5
160 5950.58 148.76 2743.66 613.06
240 8.18 209.18 781.76 949.87
230 50.98 270.23 1094.95 1618.45
370 17,651.78 357.91 17,100.79 4.37
410 29,880.58 392.33 24,083.94 312.23
Y = 237.14 (
t
Y - Y )
2
=79,942.86
(
t
Y
- Y )
2
=76,245.88
(
t
Y -
t
Y
)
2
=3668.41
Standard error of estimate
Var (b
2
) = [ (
t
Y -
t
Y
)
2
/(T 2)]/ (
t
X - X )
2
= [3668.41/ (7 -2)]/511.4 = 1.4161
se (b
2
) = 4161 . 1 = 1.19
= 87.08 + 12.21 X
(***) (1.19)
- where figures in parentheses are estimated std. errors, which measures
variability of estimates from sample to sample
- t-test is used to determine if there is a significant relationship between
dependent variable and each independent variable
- The test requires that s.e. of the estimated regression coefficient be computed
Hypothesis testing
Say, prior knowledge or expert opinion tells us that true
average price to earning (p/e) ratio in the population of
BSC is 20
Suppose a particular random sample of 30 stocks gives
this estimate as 23
Is the value of 23 statistically different from 20?
Due to sample fluctuations it is possible that 23 may not
statistically different from 20
In this case we may not reject the hypothesis that true
value of p/e is 20
This can be done through hypothesis testing
Hypothesis testing
Suppose someone suggests that X has no effect
in Y
Null hypothesis: H
0
: B
2
= 0
If H
0
is accepted, there is no point in including X
in the model
If X really belongs to the model then one would
expect that H
0
must be rejected against
alternate hypothesis H
1,
which says that
B
2
0
It could be positive or negative
Though in our analysis b
2
0, we should not look
at numerical results alone because of sampling
fluctuations
Statistical evaluation of regression results
This can be done through ttest
t-test: test of statistical significance of each
estimated regression coefficient
b: estimated coefficient
SE
b
: standard error of the estimated coefficient
Rule of 2: if absolute value of t is greater than 2,
estimated coefficient is significant at the 5% level
If coefficient passes t-test, the variable has a true
impact on demand
b
SE
b
t =
CI Vs TOS
In CI approach, we specify a plausible range
of values for the true parameter and find out
if CI includes the hypothesized value of the
parameter
If it does, we do not reject H
o
but if it lies
outside CI, we can reject H
o
In test of significance approach, instead of
specifying a range of values, we pick a specific
value of the parameter suggested by H
o
In practice, whether we use CI approach or
TOS approach of hypothesis testing is a
matter of personal choice and convenience
Test of significance
One property of normal distribution is that
any linear function of normally distributed
variables is itself normally distributed
Since b
1
and b
2
are linear function of u, which
is normally distributed
Therefore, b
1
and b
2
should also be normally
distributed
Test of significance
b
1
~ N (B
1
,
2
1 b
)
b
2
~N (B
2
,
2
2 b
)
Z = (b
2
B
2
)/ se(b
2
) = (b
2
B
2
)/ / (
2
i
x ) ~ N ( o . 1 )
Where x
i
= (X
i
- X )
- Since we dont know , w e h a v e t o u s e t h e e s t i m a t e o f
.
- In that case, (b
2
B
2
)/
/ (
2
i
x ) ~ t
n-2
= estimator (b
2
) hypothesized value (B
2
*
)/se of estimator (b
2
)
- If the absolute value of this ratio is equal to greater than
the table value of t for (n-2) dof, b
2
is said to be
statistically significant
- In our case, t = b
2
/ [se(b
2
)] = 12.21/1.19 = 10.26 > table
value of t stat at 95% confidence interval and at 5 dof,
which is 2.015
- So H
0
: B
2
= 0 is rejected
|
2
|
2
+sd |
2
+2.58sd |
2
-sd |
2
-2.58sd
0.5% 0.5%
5% level
1% level
hypothetical distribution under
0
2 2 0
: | | = H
acceptance region for b
2
0 0 0 0 0
5
b
2
t-statistic
The diagram show the acceptance region and the
rejection regions for a 5% and 1% significance
test.
2.5% 2.5%
Explanatory power of a model
Y X
Y
X
Y
X
Breakdown of total variation
X
X
t
(X
t
,Y
t
)
SRF
Total Variation
(Y
t
- )
(
t
- )=
variation in Y
t
explained
by regression
e
t
=(Y
t
-
t
)
Decomposition of Sum of Squares
(Y
t
- ) = (
t
- ) + (Y
t
-
t
)
After squaring both sides and algebraic
manipulations, we get
TSS = ESS + RSS
2 2 2
( ) ( ) ( )
t t t
Y Y Y Y Y Y = +
2
2
2
( )
( )
t
Y Y
ExplainedVariation
R
Total Variation Y Y
= =
Goodness of fit: R
2
Value of R
2
ranges from 0 to 1
If the regression equation explains none of
the variation of Y
i
(i.e. no relationship
between X & Y), R
2
will be zero
If the equation explains all the variation, R
2
will be one
In general, higher the R
2
value
,
the better the
regression equation
A low R
2
would be indicative of a rather poor
fit
2V Ex
Three Variable Regression
Model
Y
i
= B
1
+B
2
X
2i
+B
3
X
3i_
Nonstochastic form,
PRF
Y
i
= B
1
+B
2
X
2i
+B
3
X
3i
+u
i
stochastic
B
2,
B
3
called partial regression or partial
slope coefficients
B
2
measures the change in mean value of Y,
per unit change in X
2
holding the value of
X
3
constant
Y
i
= b
1
+b
2
X
2i
+b
3
X
3i
+e
i
SRF
Assumptions
Linear relationship
Xs are non-stochastic variables.
No linear relationship exists between two or
more independent variables (no multi-
collinearaity). Ex:X
2i
= 3 +2X
3
Error has zero expected value, constant
variance and normally distributed
RSS = e
2
= (Y
i
i
)
2
= (Y
i
b
1
-b
2
X
2i
-b
3
X
3i
)
2
Testing of hypothesis, t-test
Say,
i
= -1336.09 + 12.7413X
2i
+85.7640X
3i
(175.2725) (0.9123) (8.8019)
p=0.000 0.000 0.000
R
2
= 0.89, n =32
H
0
: B
1
=0, b
1
/se(b
1
)~ t
(n-3)
H
0
: B
2
=0, b
2
/se(b
2
)~ t
(n-3)
H
0
: B
3
=, (b
3
- )
/se(b
3
)~ t
(n-3)
Testing Joint Hypothesis, F Test
H
0
: B
2
= B
3
= 0
Or, H
0
: R
2
= 0
X
2
& X
3
explain zero percent of the
variation of Y
H
1
: At least one B 0
A test of either hypothesis is called a test
of overall significance of the estimated
multiple regression
We know, TSS = ESS + RSS
F test
If computed F value exceeds critical F value, we
reject the null hypothesis that the impact of
explanatory variables is simultaneously equal to zero
Otherwise we cannot reject the null hypothesis
It may happen that not all the explanatory
variables individually have much impact on dependent
variable (i.e., some of the t values may be
statically insignificant) yet all of them collectively
influence dependent variable (H
0
is rejected in F
test)
This happen only we have the problem of
multicollinearity
Specification error
In this example we have seen that
both the explanatory variables are
individually and collectively different
from zero
If we omit any one of these
explanatory variable from our model,
then there would be specification
error
What would be b
1
, b
2
& R
2
in 2-
variable model?
Specification error
i
= -1336.09 + 12.7413X
2i
+85.7640X
3i
(175.2725) (0.9123) (8.8019)
p=0.000 0.000 0.000
R
2
= 0.89, n =32
i
= -191.66 + 10.48X
2
(264.43) (1.79)
R
2
= 0.53
i
= 807.95 + 54.57X
3i
(231.95) (23.57)
R
2
= 0.15
R
2
versus Adjusted R
2
Larger the number of explanatory variables
in the model, the higher the R
2
will be
However, R
2
does not take into account dof
Therefore, comparing R
2
values of the two
models with same dependent variable but
different numbers of explanatory variables
is essentially like comparing apples and
bananas
We need a measure of fit that is adjusted
for the no. of explanatory variables in the
model
R
2
versus Adjusted R
2
Such a measure is called Adj R
2
If k > 1, Adj R
2
R
2,
as the no of explanatory
variables increases in the model, Adj R
2
becomes increasingly smaller than R
2
It enable us to compare two models that have
same dependent variable but different
numbers of independent variables
In our example, it can be shown that
Adj R
2
=0.88 < 0.89 (R
2
)
2 2
( 1)
1 (1 )
( )
n
R R
n k