ExamFinal Topics
ExamFinal Topics
Listed below are the major topics covered in class that are likely to be in
the Final Exam. Good Luck!
P (Y = y, X = x) = P (Y = y | X = x)P (X = x)
X
P (Y = y) = P (Y = y, X = x)
x
1
P (2 < Z < 2) 0.95; P ( 2 < X < + 2) 0.95.
Pn
x)2
i=1 (xi
p
s2x = , sx = s2x
n1
Pn
(yi y)2 q
s2y = i=1 , sy = s2y
n1
Least squares estimation:
sy
b0 = y b1 x, b1 = rxy
sx
2
Interpreting covariance, correlation and regression coefficients.
SST, SSR, SSE
n
X
SST = (yi y)2
i=1
n
X n
X
SSR = (yi y)2 = [(b0 + b1 xi ) y]2
i=1 i=1
n
X n
X n
X
SSE = e2i = (yi yi )2 = [yi (b0 + b1 xi )]2
i=1 i=1 i=1
Y = 0 + 1 X + , N (0, 2 )
yi = 0 + 1 xi + i , i N (0, 2 )
yi N (0 + 1 xi , 2 )
(0 + 1 xi ) 2.
b1 2sb1 , b0 2sb0 .
Hypothesis testing:
We test the null hypothesis H0 : 1 = 10 versus the alternative H1 : 1 6= 10 .
3
b1 10
The t-stat t = s b1 measures the number of standard errors the estimate
b1 is from the proposed value 10 .
The p-value provides a measure of how weird your estimate b1 is if the null
hypothesis is true.
We usually reject the null hypothesis if |t| > 2, p < 0.05, or 10 is not within
the 95% confidence interval (b1 2sb1 , b1 + 2sb1 ).
Significance level and type I error.
Forecasting:
Given Xf , the 95% plug-in prediction interval of Yf is (b0 + b1 Xf ) 2s.
A large predictive error variance (high uncertainty) comes from a large s, a
small n, a small sx and a large difference between Xf and X.
Statistical model:
Y = 0 + 1 X1 + 2 X2 + + p Xp + , N (0, 2 )
Y |X1 . . . Xp N (0 + 1 X1 + + p Xp , 2 )
e = 0, Corr(Xj , e) = 0, Corr(Y , e) = 0
2
R2 = Corr(Y, Y ) = SSR SST = 1 SST
SSE
4
If f is large (p-value is small), we reject H0 .
Understanding multiple linear regression
Correlation is not causation
Multiple linear regression allows us to control all important variables by
including them into the regression model
Dependencies between the explanatory variables (Xs) will affect our inter-
pretation of regression coefficients
Dependencies between the explanatory variables (Xs) will inflate the stan-
dard errors of regression coefficients
s2
s2bj =
variation in Xj not associated with other Xs
Dummy variables
Gender: Male, Female; Education level: High-school, Bachelor, Master, Doc-
tor; Month: Jan, Feb, , Dec
A variable of n categories can be included into multiple linear regression
using C dummy variables, where 1 C n 1
Representing a variable of n categories with n dummy variables will lead to
the problem of perfect multicollinearity
Interpretation: the same slope but different intercepts
Interactions
Interpretation: different intercepts and slopes
Diagnostics
Model assumptions:
Statistical model:
Y = 0 + 1 X1 + 2 X2 + + p Xp + , N (0, 2 )
Y = 0 + 1 X + 2 X 2 + + m X m + , N (0, 2 )
5
We can always increase m if necessary, but m = 2 is usually enough.
Be very careful about over-fitting and doing prediction outside the data
range, especially if m is large.
For Y = 0 + 1 X + 2 X 2 + , the marginal effect of X on Y is
E[Y |X]
= 1 + 22 X,
X
which means the slope is a function of X (no longer a constant).
Handing non-constant variance with Log-Log transformation
Statistical model:
Y = e0 X 1 e , N (0, 2 )
Interpretation: about 1 % change in Y per 1% change in X.
Example: price elasticity
95% plug-in prediction interval of log(Y )
(0 + 1 log(X)) 2s
Log transformation of Y
Statistical model:
log(Y ) = 0 + 1 X + , N (0, 2 )
Y = e0 e1 X e , N (0, 2 )
Interpretation: about (1001 )% change in Y per unit change in X (if 1 is
small).
Example: exponential growth
Time Series
6
Modeling non-linearity by adding t2 into the regression model: the slope
changes as time changes.
95% plug-in prediction interval
Autoregressive models
Random walk model: Yt = 0 + Yt1 + t , t N (0, 2 )
Autoregressive model of order 1 (AR(1)):
Yt = 0 + 1 Yt1 + t , t N (0, 2 )
Yt = 0 + 1 Yt1 + 2 t + t , t N (0, 2 )
Modeling seasonality
Using no more than 11 dummy variables for 12 months; using no more than
3 dummy variables for 4 quarters
Seasonal model:
Yt = 0 + 1 Jan + + 11 N ov + t
Yt = 0 + 1 Jan + + 11 N ov + 12 Yt1 + 13 t + t
Model Selection
7
Validate a model using out-of-sample prediction
Model selection criteria (AIC, BIC, Adjusted R2 )
Forward regression, backward regression, stepwise regression
Decision tree
Represent a payoff table with a decision tree
Time proceeds from left to right
Folding back procedure
Risk profile
Sensitivity analysis
Decision making and Bayes theorem
The value of information
Value of perfect information
Expected value of perfect information (EVPI)
Value of sample information
Expected value of sample information (EVSI)
Bayes theorem and the value of information
8
Flip a coin, toss a die, flip two coins, toss tow dice
Normal random numbers, Students t random numbers
Understand how to simulate from a discrete distribution
Understand how to use simulation to estimate P (X < x), E[X] and Var[X],
where X is a random variable following some distribution.
Understand how to use simulation to demonstrate Law of Large Numbers
Understand how to use simulation to demonstrate the sampling distribution of
sample mean
Understand how to use simulation to demonstrate the Central Limit Theorem
Simulation and decision making
Simulate an AR(1)+Trend+Logtransformation time series model
Using simulation to estimate the prediction intervals
Understand how to construct a random experiment and find relevant answers by
simulating the same experiment repeatedly under identical conditions