Vec Intro - Introduction To Vector Error-Correction Models: Description Remarks and Examples References Also See
Vec Intro - Introduction To Vector Error-Correction Models: Description Remarks and Examples References Also See
Vec Intro - Introduction To Vector Error-Correction Models: Description Remarks and Examples References Also See
com
vec intro — Introduction to vector error-correction models
Description
Stata has a suite of commands for fitting, forecasting, interpreting, and performing inference
on vector error-correction models (VECMs) with cointegrating variables. After fitting a VECM, the
irf commands can be used to obtain impulse–response functions (IRFs) and forecast-error variance
decompositions (FEVDs). The table below describes the available commands.
Fitting a VECM
vec [TS] vec Fit vector error-correction models
This manual entry provides an overview of the commands for VECMs; provides an introduction
to integration, cointegration, estimation, inference, and interpretation of VECM models; and gives an
example of how to use Stata’s vec commands.
1
2 vec intro — Introduction to vector error-correction models
Because Nielsen (2001) has shown that the methods implemented in varsoc can be used to choose
the order of the autoregressive process, no separate vec command is needed; you can simply use
varsoc. veclmar tests that the residuals have no serial correlation, and vecnorm tests that they are
normally distributed.
All the irf routines described in [TS] irf are available for estimating, interpreting, and managing
estimated IRFs and FEVDs for VECMs.
Remarks are presented under the following headings:
Introduction to cointegrating VECMs
What is cointegration?
The multivariate VECM specification
Trends in the Johansen VECM framework
VECM estimation in Stata
Selecting the number of lags
Testing for cointegration
Fitting a VECM
Fitting VECMs with Johansen’s normalization
Postestimation specification testing
Impulse–response functions for VECMs
Forecasting with VECMs
What is cointegration?
Standard regression techniques, such as ordinary least squares (OLS), require that the variables
be covariance stationary. A variable is covariance stationary if its mean and all its autocovariances
are finite and do not change over time. Cointegration analysis provides a framework for estimation,
inference, and interpretation when the variables are not covariance stationary.
Instead of being covariance stationary, many economic time series appear to be “first-difference
stationary”. This means that the level of a time series is not stationary but its first difference is. First-
difference stationary processes are also known as integrated processes of order 1, or I(1) processes.
Covariance-stationary processes are I(0). In general, a process whose dth difference is stationary is
an integrated process of order d, or I(d).
The canonical example of a first-difference stationary process is the random walk. This is a variable
xt that can be written as
xt = xt−1 + t (1)
where the t are independently and identically distributed (i.i.d.) with mean zero and a finite variance
σ 2 . Although E[xt ] = 0 for all t, Var[xt ] = T σ 2 is not time invariant, so xt is not covariance
stationary. Because ∆xt = xt − xt−1 = t and t is covariance stationary, xt is first-difference
stationary.
These concepts are important because, although conventional estimators are well behaved when
applied to covariance-stationary data, they have nonstandard asymptotic distributions and different
rates of convergence when applied to I(1) processes. To illustrate, consider several variants of the
model
yt = a + bxt + et (2)
Throughout the discussion, we maintain the assumption that E[et ] = 0.
vec intro — Introduction to vector error-correction models 3
where ξt and ζt are i.i.d. disturbances over time that are correlated with each other. Because t is
I(1), (3) and (4) imply that both xt and yt are I(1). The condition that |ρ| < 1 implies that νt and
yt + αxt are I(0). Thus yt and xt cointegrate, and (1, α) is the cointegrating vector.
Using a bit of algebra, we can rewrite (3) and (4) as
where δ = (1−ρ)/(α−β), zt = yt +αxt , and η1t and η2t are distinct, stationary, linear combinations
of ξt and ζt . This representation is known as the vector error-correction model (VECM). One can think
of zt = 0 as being the point at which yt and xt are in equilibrium. The coefficients on zt−1 describe
how yt and xt adjust to zt−1 being nonzero, or out of equilibrium. zt is the “error” in the system,
and (5) and (6) describe how system adjusts or corrects back to the equilibrium. As ρ goes to 1, the
system degenerates into a pair of correlated random walks. The VECM parameterization highlights
this point, because δ → 0 as ρ → 1.
4 vec intro — Introduction to vector error-correction models
If we knew α, we would know zt , and we could work with the stationary system of (5) and (6).
Although knowing α seems silly, we can conduct much of the analysis as if we knew α because
there is an estimator for the cointegrating parameter α that converges to its true value at a faster rate
than the estimator for the adjustment parameters β and δ .
The definition of a bivariate cointegrating relation requires simply that there exist a linear combination
of the I(1) variables that is I(0). If yt and xt are I(1) and there are two finite real numbers a 6= 0
and b 6= 0, such that ayt + bxt is I(0), then yt and xt are cointegrated. Although there are two
parameters, a and b, only one will be identifiable because if ayt + bxt is I(0), so is cayt + cbxt
for any finite, nonzero, real number c. Obtaining identification in the bivariate case is relatively
simple. The coefficient on yt in (4) is unity. This natural construction of the model placed the
necessary identification restriction on the cointegrating vector. As we discuss below, identification in
the multivariate case is more involved.
If yt is a K × 1 vector of I(1) variables and there exists a vector β, such that βyt is a vector
of I(0) variables, then yt is said to be cointegrating of order (1,0) with cointegrating vector β. We
say that the parameters in β are the parameters in the cointegrating equation. For a vector of length
K , there may be at most K − 1 distinct cointegrating vectors. Engle and Granger (1987) provide a
more general definition of cointegration, but this one is sufficient for our purposes.
p−1
X
∆yt = v + Πyt−1 + Γi ∆yt−i + t (8)
i=1
Pj=p Pj=p
where Π = j=1 Aj − Ik and Γi = − j=i+1 Aj . The v and t in (7) and (8) are identical.
Engle and Granger (1987) show that if the variables yt are I(1) the matrix Π in (8) has rank
0 ≤ r < K , where r is the number of linearly independent cointegrating vectors. If the variables
cointegrate, 0 < r < K and (8) shows that a VAR in first differences is misspecified because it omits
the lagged level term Πyt−1 .
Assume that Π has reduced rank 0 < r < K so that it can be expressed as Π = αβ0 , where α
and β are both r × K matrices of rank r. Without further restrictions, the cointegrating vectors are
not identified: the parameters (α, β) are indistinguishable from the parameters (αQ, βQ−10 ) for any
r × r nonsingular matrix Q. Because only the rank of Π is identified, the VECM is said to identify
the rank of the cointegrating space, or equivalently, the number of cointegrating vectors. In practice,
the estimation of the parameters of a VECM requires at least r2 identification restrictions. Stata’s vec
command can apply the conventional Johansen restrictions discussed below or use constraints that
the user supplies.
The VECM in (8) also nests two important special cases. If the variables in yt are I(1) but not
cointegrated, Π is a matrix of zeros and thus has rank 0. If all the variables are I(0), Π has full rank
K.
vec intro — Introduction to vector error-correction models 5
There are several different frameworks for estimation and inference in cointegrating systems.
Although the methods in Stata are based on the maximum likelihood (ML) methods developed by
Johansen (1988, 1991, 1995), other useful frameworks have been developed by Park and Phillips (1988,
1989); Sims, Stock, and Watson (1990); Stock (1987); and Stock and Watson (1988); among others.
The ML framework developed by Johansen was independently developed by Ahn and Reinsel (1990).
Maddala and Kim (1998) and Watson (1994) survey all these methods. The cointegration methods
in Stata are based on Johansen’s maximum likelihood framework because it has been found to be
particularly useful in several comparative studies, including Gonzalo (1994) and Hubrich, Lütkepohl,
and Saikkonen (2001).
p−1
X
0
∆yt = αβ yt−1 + Γi ∆yt−i + v + δt + t (9)
i=1
where δ is a K × 1 vector of parameters. Because (9) models the differences of the data, the constant
implies a linear time trend in the levels, and the time trend δt implies a quadratic time trend in the
levels of the data. Often we may want to include a constant or a linear time trend for the differences
without allowing for the higher-order trend that is implied for the levels of the data. VECMs exploit
the properties of the matrix α to achieve this flexibility.
Because α is a K × r rank matrix, we can rewrite the deterministic components in (9) as
v = αµ + γ (10a)
δt = αρt + τt (10b)
where µ and ρ are r × 1 vectors of parameters and γ and τ are K × 1 vectors of parameters. γ
is orthogonal to αµ, and τ is orthogonal to αρ; that is, γ0 αµ = 0 and τ0 αρ = 0, allowing us to
rewrite (9) as
p−1
X
∆yt = α(β0 yt−1 + µ + ρt) + Γi ∆yt−i + γ + τ t + t (11)
i=1
The plots on the graph indicate that all the series are trending and potential I(1) processes. In a
competitive market, the current and past prices contain all the information available, so tomorrow’s
price will be a random walk from today’s price. Some researchers may opt to use [TS] dfgls to
investigate the presence of a unit root in each series, but the test for cointegration we use includes the
case in which all the variables are stationary, so we defer formal testing until we test for cointegration.
The time trends in the data appear to be approximately linear, so we will specify trend(constant)
when modeling these series, which is the default with vec.
The next graph shows just Dallas’ and Houston’s data, so we can more carefully examine their
relationship.
vec intro — Introduction to vector error-correction models 7
12.2
12
11.8
11.6
11.4
11.2
Except for the crash at the end of 1991, housing prices in Dallas and Houston appear closely
related. Although average prices in the two cities will differ because of resource variations and other
factors, if the housing markets become too dissimilar, people and businesses will migrate, bringing
the average housing prices back toward each other. We therefore expect the series of average housing
prices in Houston to be cointegrated with the series of average housing prices in Dallas.
We will use two lags for this bivariate model because the Hannan–Quinn information criterion (HQIC)
method, Schwarz Bayesian information criterion (SBIC) method, and sequential likelihood-ratio (LR)
test all chose two lags, as indicated by the “*” in the output.
The reader can verify that when all four cities’ data are used, the LR test selects three lags, the
HQIC method selects two lags, and the SBIC method selects one lag. We will use three lags in our
four-variable model.
8 vec intro — Introduction to vector error-correction models
5%
maximum trace critical
rank parms LL eigenvalue statistic value
0 6 576.26444 . 46.8252 15.41
1 9 599.58781 0.24498 0.1785* 3.76
2 10 599.67706 0.00107
Besides presenting information about the sample size and time span, the header indicates that test
statistics are based on a model with two lags and a constant trend. The body of the table presents test
statistics and their critical values of the null hypotheses of no cointegration (line 1) and one or fewer
cointegrating equations (line 2). The eigenvalue shown on the last line is used to compute the trace
statistic in the line above it. Johansen’s testing procedure starts with the test for zero cointegrating
equations (a maximum rank of zero) and then accepts the first null hypothesis that is not rejected.
In the output above, we strongly reject the null hypothesis of no cointegration and fail to reject
the null hypothesis of at most one cointegrating equation. Thus we accept the null hypothesis that
there is one cointegrating equation in the bivariate model.
Using all four series and a model with three lags, we find that there are two cointegrating
relationships.
. vecrank austin dallas houston sa, lag(3)
Johansen tests for cointegration
Trend: constant Number of obs = 165
Sample: 1990m4 - 2003m12 Lags = 3
5%
maximum trace critical
rank parms LL eigenvalue statistic value
0 36 1107.7833 . 101.6070 47.21
1 43 1137.7484 0.30456 41.6768 29.68
2 48 1153.6435 0.17524 9.8865* 15.41
3 51 1158.4191 0.05624 0.3354 3.76
4 52 1158.5868 0.00203
Fitting a VECM
vec estimates the parameters of cointegrating VECMs. There are four types of parameters of interest:
1. The parameters in the cointegrating equations β
2. The adjustment coefficients α
3. The short-run coefficients
4. Some standard functions of β and α that have useful interpretations
vec intro — Introduction to vector error-correction models 9
Although all four types are discussed in [TS] vec, here we discuss only types 1–3 and how they
appear in the output of vec.
Having determined that there is a cointegrating equation between the Dallas and Houston series,
we now want to estimate the parameters of a bivariate cointegrating VECM for these two series by
using vec.
. vec dallas houston
Vector error-correction model
Sample: 1990m3 - 2003m12 No. of obs = 166
AIC = -7.115516
Log likelihood = 599.5878 HQIC = -7.04703
Det(Sigma_ml) = 2.50e-06 SBIC = -6.946794
Equation Parms RMSE R-sq chi2 P>chi2
D_dallas
_ce1
L1. -.3038799 .0908504 -3.34 0.001 -.4819434 -.1258165
dallas
LD. -.1647304 .0879356 -1.87 0.061 -.337081 .0076202
houston
LD. -.0998368 .0650838 -1.53 0.125 -.2273988 .0277251
D_houston
_ce1
L1. .5027143 .1068838 4.70 0.000 .2932258 .7122028
dallas
LD. -.0619653 .1034547 -0.60 0.549 -.2647327 .1408022
houston
LD. -.3328437 .07657 -4.35 0.000 -.4829181 -.1827693
Cointegrating equations
Equation Parms chi2 P>chi2
_ce1
dallas 1 . . . . .
houston -.8675936 .0214231 -40.50 0.000 -.9095821 -.825605
_cons -1.688897 . . . . .
10 vec intro — Introduction to vector error-correction models
The header contains information about the sample, the fit of each equation, and overall model
fit statistics. The first estimation table contains the estimates of the short-run parameters, along with
their standard errors, z statistics, and confidence intervals. The two coefficients on L. ce1 are the
parameters in the adjustment matrix α for this model. The second estimation table contains the
estimated parameters of the cointegrating vector for this model, along with their standard errors, z
statistics, and confidence intervals.
Using our previous notation, we have estimated
b = (−0.304, 0.503)
α b = (1, −0.868)
β v
b = (0.0056, 0.0034)
and
b= −0.165 −0.0998
Γ
−0.062 −0.333
Overall, the output indicates that the model fits well. The coefficient on houston in the cointegrating
equation is statistically significant, as are the adjustment parameters. The adjustment parameters in
this bivariate example are easy to interpret, and we can see that the estimates have the correct
signs and imply rapid adjustment toward equilibrium. When the predictions from the cointegrating
equation are positive, dallas is above its equilibrium value because the coefficient on dallas in the
cointegrating equation is positive. The estimate of the coefficient [D dallas]L. ce1 is −.3. Thus
when the average housing price in Dallas is too high, it quickly falls back toward the Houston level.
The estimated coefficient [D houston]L. ce1 of .5 implies that when the average housing price in
Dallas is too high, the average price in Houston quickly adjusts toward the Dallas level at the same
time that the Dallas prices are adjusting.
e0)
β0 = (Ir , β
_ce1
austin 1 . . . . .
dallas -1.30e-17 . . . . .
houston -.2623782 .1893625 -1.39 0.166 -.6335219 .1087655
sa -1.241805 .229643 -5.41 0.000 -1.691897 -.7917128
_cons 5.577099 . . . . .
_ce2
austin -1.41e-18 . . . . .
dallas 1 . . . . .
houston -1.095652 .0669898 -16.36 0.000 -1.22695 -.9643545
sa .2883986 .0812396 3.55 0.000 .1291718 .4476253
_cons -2.351372 . . . . .
The Johansen identification scheme has placed four constraints on the parameters in β:
[ ce1]austin=1, [ ce1]dallas=0, [ ce2]austin=0, and [ ce2]dallas=1. (The computa-
tional method used imposes zero restrictions that are numerical rather than exact. The values −3.48e–
17 and −1.26e–17 are indistinguishable from zero.) We interpret the results of the first equation as
indicating the existence of an equilibrium relationship between the average housing price in Austin
and the average prices of houses in Houston and San Antonio.
The Johansen normalization restricted the coefficient on dallas to be unity in the second
cointegrating equation, but we could instead constrain the coefficient on houston. Both sets of
restrictions define just-identified models, so fitting the model with the latter set of restrictions will
yield the same maximized log likelihood. To impose the alternative set of constraints, we use the
constraint command.
. constraint define 1 [_ce1]austin = 1
. constraint define 2 [_ce1]dallas = 0
. constraint define 3 [_ce2]austin = 0
. constraint define 4 [_ce2]houston = 1
12 vec intro — Introduction to vector error-correction models
_ce1
austin 1 . . . . .
dallas 0 (omitted)
houston -.2623784 .1876727 -1.40 0.162 -.6302102 .1054534
sa -1.241805 .2277537 -5.45 0.000 -1.688194 -.7954157
_cons 5.577099 . . . . .
_ce2
austin 0 (omitted)
dallas -.9126985 .0595804 -15.32 0.000 -1.029474 -.7959231
houston 1 . . . . .
sa -.2632209 .0628791 -4.19 0.000 -.3864617 -.1399802
_cons 2.146094 . . . . .
Only the estimates of the parameters in the second cointegrating equation have changed, and the
new estimates are simply the old estimates divided by −1.095652 because the new constraints are
just an alternative normalization of the same just-identified model. With the new normalization, we
can interpret the estimates of the parameters in the second cointegrating equation as indicating an
equilibrium relationship between the average house price in Houston and the average prices of houses
in Dallas and San Antonio.
.4
Predicted cointegrated equation
−.2 0
−.4 .2
Although the large shocks apparent in the graph of the levels have clear effects on the predictions
from the cointegrating equations, our only concern is the negative trend in the first cointegrating
equation since the end of 2000. The graph of the levels shows that something put a significant brake
on the growth of housing prices after 2000 and that the growth of housing prices in San Antonio
slowed during 2000 but then recuperated while Austin maintained slower growth. We suspect that
this indicates that the end of the high-tech boom affected Austin more severely than San Antonio.
This difference is what causes the trend in the first cointegrating equation. Although we could try to
account for this effect with a more formal analysis, we will proceed as if the cointegrating equations
are stationary.
We can use vecstable to check whether we have correctly specified the number of cointegrating
equations. As discussed in [TS] vecstable, the companion matrix of a VECM with K endogenous
variables and r cointegrating equations has K − r unit eigenvalues. If the process is stable, the moduli
of the remaining r eigenvalues are strictly less than one. Because there is no general distribution
14 vec intro — Introduction to vector error-correction models
theory for the moduli of the eigenvalues, ascertaining whether the moduli are too close to one can
be difficult.
. vecstable, graph
Eigenvalue stability condition
Eigenvalue Modulus
1 1
1 1
-.6698661 .669866
.3740191 + .4475996i .583297
.3740191 - .4475996i .583297
-.386377 + .395972i .553246
-.386377 - .395972i .553246
.540117 .540117
-.0749239 + .5274203i .532715
-.0749239 - .5274203i .532715
-.2023955 .202395
.09923966 .09924
−1 −.5 0 .5 1
Real
The VECM specification imposes 2 unit moduli
Because we specified the graph option, vecstable plotted the eigenvalues of the companion
matrix. The graph of the eigenvalues shows that none of the remaining eigenvalues appears close to
the unit circle. The stability check does not indicate that our model is misspecified.
Here we use veclmar to test for serial correlation in the residuals.
. veclmar, mlag(4)
Lagrange-multiplier test
1 56.8757 16 0.00000
2 31.1970 16 0.01270
3 30.6818 16 0.01477
4 14.6493 16 0.55046
The results clearly indicate serial correlation in the residuals. The results in Gonzalo (1994) indicate
that underspecifying the number of lags in a VECM can significantly increase the finite-sample bias
in the parameter estimates and lead to serial correlation. For this reason, we refit the model with five
lags instead of three.
. vec austin dallas houston sa, lags(5) rank(2) noetable bconstraints(1/4)
Iteration 1: log likelihood = 1200.5402
(output omitted )
Iteration 20: log likelihood = 1203.9465
Vector error-correction model
Sample: 1990m6 - 2003m12 No. of obs = 163
AIC = -13.79075
Log likelihood = 1203.946 HQIC = -13.1743
Det(Sigma_ml) = 4.51e-12 SBIC = -12.27235
Cointegrating equations
Equation Parms chi2 P>chi2
_ce1
austin 1 . . . . .
dallas 0 (omitted)
houston -.6525574 .2047061 -3.19 0.001 -1.053774 -.2513407
sa -.6960166 .2494167 -2.79 0.005 -1.184864 -.2071688
_cons 3.846275 . . . . .
_ce2
austin 0 (omitted)
dallas -.932048 .0564332 -16.52 0.000 -1.042655 -.8214409
houston 1 . . . . .
sa -.2363915 .0599348 -3.94 0.000 -.3538615 -.1189215
_cons 2.065719 . . . . .
Comparing these results with those from the previous model reveals that
1. there is now evidence that the coefficient [ ce1]houston is not equal to zero,
2. the two sets of estimated coefficients for the first cointegrating equation are different, and
3. the two sets of estimated coefficients for the second cointegrating equation are similar.
The assumption that the errors are independently, identically, and normally distributed with zero
mean and finite variance allows us to derive the likelihood function. If the errors do not come from
a normal distribution but are just independently and identically distributed with zero mean and finite
variance, the parameter estimates are still consistent, but they are not efficient.
16 vec intro — Introduction to vector error-correction models
We use vecnorm to test the null hypothesis that the errors are normally distributed.
. qui vec austin dallas houston sa, lags(5) rank(2) bconstraints(1/4)
. vecnorm
Jarque-Bera test
Skewness test
Kurtosis test
The results indicate that we can strongly reject the null hypothesis of normally distributed errors.
Most of the errors are both skewed and kurtotic.
.015
.01
.005
0 10 20 30 0 10 20 30
step
Graphs by irfname, impulse variable, and response variable
The graphs indicate that an orthogonalized shock to the average housing price in Austin has a
permanent effect on the average housing price in San Antonio but that an orthogonalized shock to
the average price of housing in Dallas has a transitory effect. According to this model, unexpected
shocks that are local to the Austin housing market will have a permanent effect on the housing market
in San Antonio, but unexpected shocks that are local to the Dallas housing market will have only a
transitory effect on the housing market in San Antonio.
. tsset
time variable: t, 1990m1 to 2003m12
delta: 1 month
. fcast compute m1_, step(24)
. fcast graph m1_austin m1_dallas m1_houston m1_sa
2004m1 2004m7 2005m1 2005m7 2006m1 2004m1 2004m7 2005m1 2005m7 2006m1
95% CI forecast
As expected, the widths of the confidence intervals grow with the forecast horizon.
References
Ahn, S. K., and G. C. Reinsel. 1990. Estimation for partially nonstationary multivariate autoregressive models. Journal
of the American Statistical Association 85: 813–823.
Becketti, S. 2013. Introduction to Time Series Using Stata. College Station, TX: Stata Press.
Engle, R. F., and C. W. J. Granger. 1987. Co-integration and error correction: Representation, estimation, and testing.
Econometrica 55: 251–276.
Gonzalo, J. 1994. Five alternative methods of estimating long-run equilibrium relationships. Journal of Econometrics
60: 203–233.
Granger, C. W. J. 1981. Some properties of time series data and their use in econometric model specification. Journal
of Econometrics 16: 121–130.
Granger, C. W. J., and P. Newbold. 1974. Spurious regressions in econometrics. Journal of Econometrics 2: 111–120.
Hamilton, J. D. 1994. Time Series Analysis. Princeton: Princeton University Press.
Hubrich, K., H. Lütkepohl, and P. Saikkonen. 2001. A review of systems cointegration tests. Econometric Reviews
20: 247–318.
Johansen, S. 1988. Statistical analysis of cointegration vectors. Journal of Economic Dynamics and Control 12:
231–254.
. 1991. Estimation and hypothesis testing of cointegration vectors in Gaussian vector autoregressive models.
Econometrica 59: 1551–1580.
. 1995. Likelihood-Based Inference in Cointegrated Vector Autoregressive Models. Oxford: Oxford University
Press.
Lütkepohl, H. 2005. New Introduction to Multiple Time Series Analysis. New York: Springer.
vec intro — Introduction to vector error-correction models 19
Maddala, G. S., and I.-M. Kim. 1998. Unit Roots, Cointegration, and Structural Change. Cambridge: Cambridge
University Press.
Nielsen, B. 2001. Order determination in general vector autoregressions. Working paper, Department of Economics,
University of Oxford and Nuffield College. http://ideas.repec.org/p/nuf/econwp/0110.html.
Park, J. Y., and P. C. B. Phillips. 1988. Statistical inference in regressions with integrated processes: Part I. Econometric
Theory 4: 468–497.
. 1989. Statistical inference in regressions with integrated processes: Part II. Econometric Theory 5: 95–131.
Paulsen, J. 1984. Order determination of multivariate autoregressive time series with unit roots. Journal of Time Series
Analysis 5: 115–127.
Phillips, P. C. B. 1986. Understanding spurious regressions in econometrics. Journal of Econometrics 33: 311–340.
Phillips, P. C. B., and S. N. Durlauf. 1986. Multiple time series regressions with integrated processes. Review of
Economic Studies 53: 473–495.
Sims, C. A., J. H. Stock, and M. W. Watson. 1990. Inference in linear time series models with some unit roots.
Econometrica 58: 113–144.
Stock, J. H. 1987. Asymptotic properties of least squares estimators of cointegrating vectors. Econometrica 55:
1035–1056.
Stock, J. H., and M. W. Watson. 1988. Testing for common trends. Journal of the American Statistical Association
83: 1097–1107.
Tsay, R. S. 1984. Order selection in nonstationary autoregressive models. Annals of Statistics 12: 1425–1433.
Watson, M. W. 1994. Vector autoregressions and cointegration. In Vol. 4 of Handbook of Econometrics, ed. R. F.
Engle and D. L. McFadden. Amsterdam: Elsevier.
Also see
[TS] vec — Vector error-correction models
[TS] irf — Create and analyze IRFs, dynamic-multiplier functions, and FEVDs