+part 04 - AMEFA - 2024 - Introduction and Repetition
+part 04 - AMEFA - 2024 - Introduction and Repetition
Pär Sjölander
+Part 04 - AMEFA - 2024 - Introduction and Repetition.pptx
Simple Linear Regression vs. Multiple Linear Regression
Thus, Multiple Linear OLS regression is rather similar to Single Linear OLS regression,
but with more variables – more assumptions need to be satisfied too.
Multiple Linear Regression
• Simple Linear Regression can be extended (generalized) to models with more than
one independent variable, for example, a model with p variables, like:
• The formulae for OLS estimators of multiple linear regression become complicated
when more than 2 independent variables are involved. One needs to use matrix
algebra. However, instead, we use statistical software packages to solve this problem
in the course.
• A slope 𝛽𝛽𝑘𝑘 , for 𝑘𝑘 = 1, … , 𝑝𝑝, is interpreted as the amount of change in the dependent
variable 𝑌𝑌, per unit change of the independent variable 𝑋𝑋𝑘𝑘 , holding the values of all
other independent variables fixed/constant (ceteris paribus).
OLS Classical Linear Regression Model (CLRM) Assumptions
Classical Linear Regression Model (CLRM) assumptions must be satisfied (except 10) for an estimator to be considered the Best Linear
Unbiased Estimator (BLUE). These assumptions are relevant to the Gauss-Markov theorem, which proves the BLUE property.
1) Linearity (linear model in parameters, not necessarily in variables since e.g. log transformations can linearize variables).
2) Independent 𝑋𝑋’s from the error terms, i.e., 𝑐𝑐𝑐𝑐𝑐𝑐 𝑋𝑋𝑘𝑘𝑘𝑘 , 𝑢𝑢𝑗𝑗 = 0, for all 𝑘𝑘 = 1, … , 𝑝𝑝, and 𝑖𝑖, 𝑗𝑗 = 1, … , 𝑛𝑛.
3) Zero mean error terms, i.e., 𝐸𝐸 𝑢𝑢𝑖𝑖 = 0, for all 𝑖𝑖 = 1, … , 𝑛𝑛.
4) Homoscedasticity (constant variance of the error terms), i.e., 𝑣𝑣𝑣𝑣𝑣𝑣 𝑢𝑢𝑖𝑖 = 𝜎𝜎 2 , for all 𝑖𝑖 = 1, … , 𝑛𝑛.
5) No autocorrelation between the error terms, i.e., 𝑐𝑐𝑐𝑐𝑐𝑐 𝑢𝑢𝑖𝑖 , 𝑢𝑢𝑗𝑗 = 0, for all 𝑖𝑖, 𝑗𝑗 = 1, … , 𝑛𝑛, where 𝑖𝑖 ≠ 𝑗𝑗.
6) The number of observations must be larger than the number of parameters to be estimated (slopes and intercept),
i.e., 𝑛𝑛 > 𝑝𝑝 + 1.
7) There must be variation in the values of 𝑋𝑋’s (other than intercept), i.e., there is no 𝑘𝑘, such that 𝑋𝑋𝑘𝑘𝑘𝑘 = 𝑐𝑐, for all 𝑖𝑖 = 1, … , 𝑛𝑛.
8) No perfect multicollinearity between 𝑋𝑋’s.
9) No specification bias, i.e., the model should be specified correctly (e.g. not overspecified/underspecified).
10) Assumption of error normality, i.e., 𝑢𝑢𝑖𝑖 ~𝑁𝑁 0, 𝜎𝜎 2 . Normality is not required for OLS to be BLUE (minimum variance
coefficient estimates for linear unbiased models). Error normality is necessary for small sample hypothesis testing (e.g.
hypothesis tests as t-tests) but less important for large samples due to the Central Limit Theorem (CLT).
(For MLE, however, normality is necessary for defining the likelihood function). More clarifications on the next slide
While the normal distribution of errors is a common assumption, it is not necessary for OLS to provide BLUE (Best Linear Unbiased Estimator) estimates. In simple terms, even if the errors are not normally distributed, OLS can still give us good estimates of our
parameters. However, this assumption becomes important when we want to make precise predictions about how often our estimates might be off (hypothesis testing). When we have a lot of data, the Central Limit Theorem helps us because it says that our
estimates will tend to follow a normal distribution even if the errors do not. But for small samples, if we want to use common statistical tests, we rely on this normality assumption. In contrast, for methods like Maximum Likelihood Estimation (MLE), which is
a different estimation approach, assuming normality is crucial for the technique to work properly. For MLE: If the assumption of error normality is violated, it does not necessarily mean that MLE estimators lose their consistency, but they may not be efficient
or unbiased, particularly in small samples. In large samples, however, MLE estimators often retain their consistency and asymptotic efficiency, even if the error terms are not normally distributed.
What is BLUE? * Note: To be correct, though not crucial to memomrize, the Gauss-Markov theorem asserts in a linear regression context: if errors exhibit zero
mean, constant variance (homoscedasticity), and no correlation, then OLS estimators are distinguished as the Best Linear Unbiased Estimators (BLUE).
However, roughly speaking we say that the CLRM assumptions (except error normality) should hold – then we can use OLS – and it will be BLUE.
The term "BLUE" in econometrics stands for "Best Linear Unbiased Estimator." The estimator (that
gives the coef.) is:
- Best: The estimator has the smallest variance among all linear unbiased estimators.
- Linear: A linear function of the observed data.
- Unbiased: Has an expected value equal to the true parameter value (not systematically over or
underestimating it).
Under the Classical Linear Regression Model (CLRM) assumptions*, the Ordinary Least Squares (OLS)
estimators are the Best Linear Unbiased Estimators (BLUE). This implies that, within the class of all
linear unbiased estimators, OLS estimators have the minimum variance, giving them the most efficient
estimators for the parameters.
For OLS regression, the property of being BLUE is related to the estimators of the regression
parameters (e.g. slope parameters/coefficient), rather than to the residuals/errors and the hypothesis
testing (by e.g. t-tests, p-values, S.E.) of these parameters. Hypothesis testing is not related to BLUE.
TO CONCLUDE: BLUE OLS parameters get a minimum variance (among linear unbiased parameters)!
However, we simplify, to be correct note that there is a distinction: An estimator is a rule or a formula that tells you how to calculate the estimate of a parameter from your data. For example, the sample mean (X̄) is an estimator of the population mean (μ), and
the sample variance (S²) is an estimator of the population variance (σ²). It is actually this estimator we refer to, not the parameter.
• Note: The assumption of error normality is not a requirement for an estimator to be BLUE. However,
a requirement for reliable hypothesis testing (reject/not reject H0) and interval estimation
(confidence intervals)…
…thus error nomality is a separate concern from the estimator's properties as defined by the Gauss-
Markov theorem).
TO CONCLUDE: Error Normality Hypothesis testing works (e.g. “rejecting/not rejecting” H0)!
BUT
CLT and OLS Assumptions / BLUE MLE = Maximum Likelihood Estimation
Two-tailed test:
𝐻𝐻0 : 𝛽𝛽1 = 𝛽𝛽
𝐻𝐻1 : 𝛽𝛽1 ≠ 𝛽𝛽
Left-tailed test:
𝐻𝐻0 : 𝛽𝛽1 = 𝛽𝛽
𝐻𝐻1 : 𝛽𝛽1 < 𝛽𝛽
Right-tailed test:
𝐻𝐻0 : 𝛽𝛽1 = 𝛽𝛽
𝐻𝐻1 : 𝛽𝛽1 > 𝛽𝛽
• F test is a test of overall effect of the explanatory variables on the dependent variable,
i.e., it is a test for model significance. F test statistics follows 𝐹𝐹~𝐹𝐹(𝑝𝑝, 𝑛𝑛 − 𝑝𝑝 − 1).
• The MSE from the table is the same as the regression error variance estimator 𝜎𝜎� 2 .
S.O.V (Source of Variation), df (Degrees of Freedom). SS (Sum of Squares), MS (Mean Squared), F (F Test Statistic), ESS (Explained
Sum of Squares, also known as SSR Sum of Squares (due to) Regression), RSS (Residual Sum of Squares) a.k.a. SSE (Sum of Squares
Error), TSS (Total Sum of Squares) a.k.a. Sum of Square Total, SST, MSR Mean Square (due to) Regression), MSE (Mean Square Error)
Example 1: Cobb-Douglas.wf1/dta
Using the data in the table, Cobb-Douglas production function is estimated.
Year GDP Employment Fixed Capital
ln 𝐺𝐺𝐺𝐺𝐺𝐺𝑡𝑡 = 𝛽𝛽0 + 𝛽𝛽1 ln 𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝑡𝑡 + 𝛽𝛽2 ln 𝐹𝐹𝐹𝐹𝐹𝐹𝐹𝐹𝐹𝐹_𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝑡𝑡 + 𝑢𝑢𝑡𝑡 (T) (Y) (L) (K)
* In Stata (must log first): 1955 114043 8310 182113
1956 120410 8529 193749
gen log_y = log(y)
1957 129187 8738 205192
gen log_l = log(l)
1958 134705 8952 215130
' In EViews: gen log_k = log(k) • The coefficient of employment is ≈0.34, 1959 139960 9171 225021
ls log(y) c log(l) log(k) reg log_y log_l log_k
which represents GDP elasticity for 1960 150511 9569 237026
• The estimate for employment is not significant with α=0.05, since its p-value (0.0849) is larger than 0.05.
• The F test statistic is 1719.231, with p-value (0.000), meaning that the overall estimated model is significant.
• The coefficient of determination R2=0.995. It means 99.5% of the variation in ln(GDP) (not GDP itself) is explained by
the estimated model ln 𝐺𝐺𝐺𝐺𝐺𝐺𝑡𝑡 = −1.65242 + 0.33973 × ln 𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝐸𝑡𝑡 + 0.845997 × ln 𝐹𝐹𝐹𝐹𝐹𝐹𝐹𝐹𝐹𝐹_𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝐶𝑡𝑡 .
Example 6: Regression hypothesis testing and prediction
Example 5: (F Test & ANOVA Table)
Example 5: (F Test & ANOVA Table)
Example 5: (F Test & ANOVA Table)
Example 5: (F Test & ANOVA Table)
Example 5: (F Test & ANOVA Table)
Two-tailed
test since we
test H0: bi=0
and H1: bi≠0
Example 5: (F Test & ANOVA Table)
Example 5: (F Test & ANOVA Table)
RSS = Residual SS
In Eviews: File/New/Workfile. Dated/Annual. Start data: 1955 End date: 1974. Quick/Empty Group (Edit Series). On the first row write at Column1: T, Col2: Y, Col3: L, Col4: K. Copy the data and paste into the cells.
Then in the command window: ls log(y) c log(l) log(k) AND THEN ls log(y) c log(l) log(k) l k For this last regression Choose: View/Coefficient Diagnostics/Wald Test Coefficient Restrictions. Type in c(4)=c(5)=0
Year GDP Employment Fixed Capital
(since these are the two last estimated coefficinets when log(y) c log(l) log(k) l k is estimated. Thus, we test the coefficients of L and K, that is this test hypothesis 𝐻𝐻0:𝛽𝛽3=𝛽𝛽4=0
(T) (Y) (L) (K)
Using the data in the following table, Cobb-Douglas production function is estimated. 1955 114043 8310 182113
1956 120410 8529 193749
ln 𝑌𝑌𝑡𝑡 = 𝛽𝛽0 + 𝛽𝛽1 ln 𝐿𝐿𝑡𝑡 + 𝛽𝛽2 ln 𝐾𝐾𝑡𝑡 + 𝑢𝑢𝑡𝑡 Is there any value added of these 2 extra variables in this unrestricted model? 1957 129187 8738 205192
The transcendental production function (a generalization of Cobb-Douglas function) is 1958 134705 8952 215130
as follows. ln 𝑌𝑌𝑡𝑡 = 𝛽𝛽0 + 𝛽𝛽1 ln 𝐿𝐿𝑡𝑡 + 𝛽𝛽2 ln 𝐾𝐾𝑡𝑡 + 𝛽𝛽3 𝐿𝐿𝑡𝑡 + 𝛽𝛽4 𝐾𝐾𝑡𝑡 + 𝑢𝑢𝑡𝑡 1959 139960 9171 225021
1960 150511 9569 237026
Restricted model Unrestricted model (with all variables=no restrictions)
1961 157897 9527 248897
1962 165286 9662 260661
1963 178491 10334 275466
1964 199457 10981 295378
1965 212323 11746 315715
1966 226977 11521 337642
1967 241194 11540 363599
1968 260881 12066 391847
1969 277498 12297 422382
1970 296530 12955 455049
1971 306712 13338 484677
1972 329030 13738 520553
1973 354057 15924 561531
1974 374977 14154 609825
To check if the Cobb-Douglas The critical value is (Fobs=31.01923)>(3.68=Fcrit,2,15.5%).
production function is the Therefore, the null hypothesis is rejected. At 5%
correct model, we can test the significance level. The second unrestricted model
hypothesis significantly fits better. OR
𝐻𝐻0 : 𝛽𝛽3 = 𝛽𝛽4 = 0. The p-value for the F-statistic in the Wald test is:
(which are the 4th c(4) and the (p-value=0.0000)<(0.05=sign.level). Therefore, the null
5th c(5) coefficients since we hypothesis is rejected. At 5% significance level. The second
have an intercept as c(1)) unrestricted model significantly fits better.
* In Stata (must log first): Cobb-Douglas.wf1/dta
reg log_y log_l log_k
reg log_y log_l log_k l k Year GDP Employment Fixed Capital
testparm l k (T) (Y) (L) (K)
Using the data in the following table, Cobb-Douglas production function is estimated. 1955 114043 8310 182113
1956 120410 8529 193749
ln 𝑌𝑌𝑡𝑡 = 𝛽𝛽0 + 𝛽𝛽1 ln 𝐿𝐿𝑡𝑡 + 𝛽𝛽2 ln 𝐾𝐾𝑡𝑡 + 𝑢𝑢𝑡𝑡 Is there any value added of these 2 extra variables in this unrestricted model? 1957 129187 8738 205192
The transcendental production function (a generalization of Cobb-Douglas function) is 1958 134705 8952 215130
as follows. ln 𝑌𝑌𝑡𝑡 = 𝛽𝛽0 + 𝛽𝛽1 ln 𝐿𝐿𝑡𝑡 + 𝛽𝛽2 ln 𝐾𝐾𝑡𝑡 + 𝛽𝛽3 𝐿𝐿𝑡𝑡 + 𝛽𝛽4 𝐾𝐾𝑡𝑡 + 𝑢𝑢𝑡𝑡 1959 139960 9171 225021
1960 150511 9569 237026
Restricted model Unrestricted model (with all variables=no restrictions)
1961 157897 9527 248897
1962 165286 9662 260661
1963 178491 10334 275466
1964 199457 10981 295378
1965 212323 11746 315715
1966 226977 11521 337642
1967 241194 11540 363599
1968 260881 12066 391847
1969 277498 12297 422382
1970 296530 12955 455049
1971 306712 13338 484677
1972 329030 13738 520553
1973 354057 15924 561531
1974 374977 14154 609825
To check if the Cobb-Douglas The critical value is (Fobs=31.02)>(3.68=Fcrit,2,15.5%).
production function is the Therefore, the null hypothesis is rejected. At 5%
correct model, we can test the significance level. The second unrestricted model
hypothesis significantly fits better. OR
𝐻𝐻0 : 𝛽𝛽3 = 𝛽𝛽4 = 0 The p-value for the F-statistic in the Wald test is:
(p-value=0.0000)<(0.05=sign.level). Therefore, the null
hypothesis is rejected. At 5% significance level. The second
unrestricted model significantly fits better.
Or equivalently using one of these formulas:
USING R2: To check if the Cobb-Douglas production function is the correct model, we can test the hypothesis 𝐻𝐻0 : 𝛽𝛽3 = 𝛽𝛽4 = 0.
(0.999042−0.995080)/((20−2−1)−(20−4−1))
𝐹𝐹 = = 31.0177453~𝐹𝐹 2,15
(1−0.999042)/(20−4−1)
where ~𝐹𝐹 𝑑𝑑𝑑𝑑𝑅𝑅𝑅𝑅𝑅𝑅𝑅𝑅𝑅𝑅𝑅𝑅𝑅𝑅𝑅𝑅 − 𝑑𝑑𝑑𝑑𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈 , 𝑑𝑑𝑑𝑑𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈𝑈 or ~𝐹𝐹 (20 − 2 − 1) − (20 − 4 − 1), (20 − 4 − 1)
The critical value is 3.68. (Fobs=31.0177453)>(3.68=Fcrit). Therefore, the null hypothesis is rejected. The second unrestricted
model significantly fits better.
USING RSS (Residual SS): To check if the Cobb-Douglas production function is the correct model, we can test the hypothesis
𝐻𝐻0 : 𝛽𝛽3 = 𝛽𝛽4 = 0.
(0.013604 − 0.002649)/(20 − 2 − 1 − (20 − 4 − 1))
𝐹𝐹 = = 31.0177453~𝐹𝐹 2,15
(0.002649)/(20 − 4 − 1)
The critical value is 3.68. (Fobs=31.01642129)>(3.68=Fcrit). Therefore, the null hypothesis is rejected. The second unrestricted
model significantly fits better.
F Test & Nested Models
Thus, easiest to use these formulas
(or most convenient to use a software and give the command for a Wald test)
• Using R2:
RSS = Residual SS
The unrestricted model with all variables – we save RSS (Residual SS)
WITH STANDARDIZATION: Investment Returns = β0 + β1∗ × (Std. Market Risk Premium) + β∗2 × (Std. Book−to−Market Ratio) + ε
• Standardized Coefficients:
• β1∗ (Standardized Market Risk Premium): 0.5
• β∗2 (Standardized Book-to-Market Ratio): 0.6
• Interpretation:
• A 1 SD increase in market risk premium is associated with a 0.5 SD increase in investment returns.
• A 1 SD increase in the book-to-market ratio is associated with a 0.6 SD increase in investment returns.
• Corrected Conclusion:
• Book-to-market ratio has a greater standardized impact on investment returns than market risk premium (0.6 vs. 0.5).
Through standardization, the coefficients are rescaled so that they can be compared directly, revealing that the book-to-market
ratio has a slightly greater impact on investment returns than market risk premium when comparing the effects in terms of
standard deviations. This contrasts with the initial, potentially misleading conclusion drawn from non-standardized coefficients.
OLS with
Savings
Standardized
Savings Credit
Variables Creditscore standardized coef.wf1/dta
'In EViews TO STANDARDIZE ALL VARIABLES: series y_star = (y - @mean(y)) / @stdev(y)
Age (SEK) (USD) Score *with mean=0, and stdev=1, using @mean(y) and @stdev(y)
AGE SAVINGS_SEK SAVINGS_USD CREDITSCORE
'Standardizing the variables - and adding the suffix "_STAR" to the new standardized variables
76.46 207202.18 20720.22 664.79 series AGE_star = (AGE - @mean(AGE)) / @stdev(AGE)
56.00 272713.68 27271.37 651.31 series SAVINGS_SEK_star = (SAVINGS_SEK - @mean(SAVINGS_SEK)) /
@stdev(SAVINGS_SEK)
64.68 238051.89 23805.19 645.25 series SAVINGS_USD_star = (SAVINGS_USD - @mean(SAVINGS_USD)) /
@stdev(SAVINGS_USD)
83.61 206083.75 20608.38 649.14 series CREDITSCORE_star = (CREDITSCORE - @mean(CREDITSCORE)) /
78.01 222193.16 22219.32 661.15 @stdev(CREDITSCORE)
𝟎𝟎. 𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎 > 𝟎𝟎. 𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎 = 𝟗𝟗. 𝟎𝟎𝟎𝟎𝟎𝟎 − 𝟎𝟎𝟎𝟎 = 𝟗𝟗. 𝟎𝟎𝟎𝟎 × 𝟏𝟏𝟏𝟏−𝟓𝟓 𝟏𝟏𝟏𝟏 𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕 𝒉𝒉𝒉𝒉𝒉𝒉𝒉𝒉𝒉𝒉𝒉𝒉 𝒄𝒄𝒄𝒄𝒄𝒄𝒄𝒄. 𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕 𝒇𝒇𝒇𝒇𝒇𝒇 𝑺𝑺𝑺𝑺𝑺𝑺 𝟗𝟗. 𝟎𝟎𝟎𝟎𝟎𝟎 − 𝟎𝟎𝟎𝟎 = 𝟎𝟎. 𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎 < 𝟎𝟎. 𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎 𝟏𝟏𝟏𝟏 𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕 𝒍𝒍𝒍𝒍𝒍𝒍𝒍𝒍𝒍𝒍 𝒄𝒄𝒄𝒄𝒄𝒄𝒄𝒄. 𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕 𝒇𝒇𝒇𝒇𝒇𝒇 𝑼𝑼𝑼𝑼𝑼𝑼
Despite different scales (currencies), standardization makes coefficients for SAVINGS_USD = SAVINGS_SEK_STAR identical and comparable with AGE_STAR
Standardization solves the problem, but note that SE, t-stat, F-stat, R2 etc. are the same in the models
OLS with
Savings
Standardized
Savings Credit
Variables Creditscore standardized coef.wf1/dta
𝟎𝟎. 𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎 > 𝟎𝟎. 𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎 𝟏𝟏𝟏𝟏 𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕 𝒉𝒉𝒉𝒉𝒉𝒉𝒉𝒉𝒉𝒉𝒉𝒉 𝒄𝒄𝒄𝒄𝒄𝒄𝒄𝒄. 𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕 𝒇𝒇𝒇𝒇𝒇𝒇 𝑺𝑺𝑺𝑺𝑺𝑺 𝟎𝟎. 𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎 < 𝟎𝟎. 𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎𝟎 𝟏𝟏𝟏𝟏 𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕 𝒍𝒍𝒍𝒍𝒍𝒍𝒍𝒍𝒍𝒍 𝒄𝒄𝒄𝒄𝒄𝒄𝒄𝒄. 𝒕𝒕𝒕𝒕𝒕𝒕𝒕𝒕 𝒇𝒇𝒇𝒇𝒇𝒇 𝑼𝑼𝑼𝑼𝑼𝑼
Despite different scales (currencies), standardization makes coefficients for SAVINGS_USD = SAVINGS_SEK_STAR identical and comparable with AGE_STAR
Standardization solves the problem, but note that SE, t-stat, F-stat, R2 etc. are the same in the models
OLS with Standardized Independent Variables
• Standardization of an independent variable 𝑋𝑋𝑗𝑗 (excluding intercept) is a linear
transformation, based on itself and the intercept, as shown below.
𝑋𝑋𝑗𝑗 − �𝑗𝑗
𝑋𝑋
𝑋𝑋𝑗𝑗∗ = 𝑠𝑠𝑋𝑋𝑗𝑗 is the standard deviation of the independent variable X in a dataset.
j
𝑠𝑠𝑋𝑋𝑗𝑗
• The consequences of standardizing of the independent variable 𝑿𝑿𝒋𝒋 are as follows:
It does not affect the estimators of other slopes.
The slope estimator of 𝑋𝑋𝑗𝑗 becomes 𝛽𝛽̂𝑗𝑗∗ = 𝑠𝑠𝑋𝑋 𝛽𝛽̂𝑗𝑗 . 𝑗𝑗
* Standardizing variables
egen mean_X_SEK = mean(X_SEK)
egen sd_X_SEK = sd(X_SEK)
egen mean_Y = mean(Y)
egen sd_Y = sd(Y)
See many practical examples in the lab notes – both manual calculations and computer demonstrations
Note in the previous slide: The reciprocal model captures a non-linear, sharply diminishing absolute impact of X on Y as X grows,
while the lin-log model maintains a constant impact on Y for each percentage increase in X, leading to diminishing absolute changes in Y if Y grows with X