Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Dealing with structural breaks

K. Patterson and TC Mills, Palgrave Handbook of Econometrics

This chapter is concerned with methodological issues related to estimation, testing and computation in the context of structural changes in the linear models. A central theme of the review is the interplay between structural ...

Dealing with Structural Breaks∗ Pierre Perron Boston University This version: April 20, 2005 Abstract This chapter is concerned with methodological issues related to estimation, testing and computation in the context of structural changes in the linear models. A central theme of the review is the interplay between structural change and unit root and on methods to distinguish between the two. The topics covered are: methods related to estimation and inference about break dates for single equations with or without restrictions, with extensions to multi-equations systems where allowance is also made for changes in the variability of the shocks; tests for structural changes including tests for a single or multiple changes and tests valid with unit root or trending regressors, and tests for changes in the trend function of a series that can be integrated or trendstationary; testing for a unit root versus trend-stationarity in the presence of structural changes in the trend function; testing for cointegration in the presence of structural changes; and issues related to long memory and level shifts. Our focus is on the conceptual issues about the frameworks adopted and the assumptions imposed as they relate to potential applicability. We also highlight the potential problems that can occur with methods that are commonly used and recent work that has been done to overcome them. ∗ This paper was prepared for the Palgrave Handbook of Econometrics, Vol. 1: Econometric Theory. For useful comments on an earlier draft, I wish to thank Jushan Bai, Songjun Chun, Ai Deng, Mohitosh Kejriwal, Dukpa Kim, Eiji Kurozumi, Zhongjun Qu, Jonathan Treussard, Tim Vogelsang, Tatsuma Wada, Tomoyoshi Yabu, Yunpeng Zhang, Jing Zhou. 1 Introduction This chapter is concerned with methodological issues related to estimation, testing and computation for models involving structural changes. The amount of work on this subject over the last 50 years is truly voluminous in both the statistics and econometrics literature. Accordingly, any survey article is bound by the need to focus on specific aspects. Our aim is to review developments in the last fifteen years as they relate to econometric applications based on linear models, with appropriate mention of prior work to better understand the historical context and important antecedents. During this recent period, substantial advances have been made to cover models at a level of generality that allows a host of interesting practical applications. These include models with general stationary regressors and errors that can exhibit temporal dependence and heteroskedasticity, models with trending variables and possible unit roots, cointegrated models and long memory processes, among others. Advances in these contexts have been made pertaining to the following topics: computational aspects of constructing estimates, their limit distributions, tests for structural changes, and methods to determine the number of changes present. These recent developments related to structural changes have paralleled developments in the analysis of unit root models. One reason is that many of the tools used are similar. In particular, heavy use is made in both literatures of functional central limit theorems or invariance principles, which have fruitfully been used in many areas of econometrics. At the same time, a large literature has addressed the interplay between structural changes and unit roots, in particular the fact that both classes of processes contain similar qualitative features. For example, most tests that attempt to distinguish between a unit root and a (trend) stationary process will favor the unit root model when the true process is subject to structural changes but is otherwise (trend) stationary within regimes specified by the break dates. Also, most tests trying to assess whether structural change is present will reject the null hypothesis of no structural change when the process has a unit root component but with constant model parameters. As we can see, there is an intricate interplay between unit root and structural changes. This creates particular difficulties in applied work, since both are of definite practical importance in economic applications. A central theme of this review relates to this interplay and to methods to distinguish between the two. The topics addressed in this review are the following. Section 2 provides interesting historical notes on structural change, unit root and long memory tests which illustrate the intricate interplay involved when trying to distinguish between these three features. Section 1 3 reviews methods related to estimation and inference about break dates. We start with a general linear regression model that allows multiple structural changes in a subset of the coefficients (a partial change model) with the estimates obtained by minimizing the sum of squared residuals. Special attention is given to the set of assumptions used to obtain the relevant results and their relevance for practical applications (Section 3.1). We also include a discussion of results applicable when linear restrictions are imposed (3.2), methods to obtain estimates of the break dates that correspond to global minimizers of the objective function (3.3), the limit distributions of such estimates, including a discussion of benefits and potential drawbacks that arise from the adoption of a special asymptotic framework that considers shifts of shrinking magnitudes (3.4). Section 3.5 briefly discusses an alternative estimation strategy based on estimating the break dates sequentially, and Section 3.6 discusses extensions of most of these issues to a general multi-equations system, which also allows changes in the covariance matrix of the errors. Section 4 considers tests for structural changes. We start in Section 4.1 with methods based on scaled functions of partial sums of appropriate residuals. The CUSUM test is probably the best known example but the class includes basically all methods available for general models prior to the early nineties. Despite their wide appeal, these tests suffer from an important drawback, namely that power is non-monotonic, in the sense that the power can decrease and even go to zero as the magnitude of the change increases (4.2). Section 4.3 discusses tests that directly allow for a single break in the regression underlying their construction, including a class of optimal tests that have found wide appeal in practice (4.3.1), but which are also subject to non-monotonic power when two changes affect the system (4.3.2), a result which points to the usefulness of tests for multiple structural changes discussed in Section 4.4. Tests for structural changes in the linear model subject to restrictions on the parameters are discussed in Section 4.5 and extensions of the methods to multivariate systems are presented in Section 4.6. Tests valid when the regressors are unit root processes and the errors are stationary, i.e., cointegrated systems, are reviewed in Section 4.7, while Section 4.8 considers recent developments with respect to tests for changes in a trend function when the noise component of the series is either a stationary or a unit root process. Section 5 addresses the topic of testing for a unit root versus trend-stationarity in the presence of structural changes in the trend function. The motivation, issues and frameworks are presented in Section 5.1, while Section 5.2 discusses results related to the effect of changes in the trend on standard unit root tests. Methods to test for a unit root allowing for a change 2 at a known date are reviewed in Section 5.3, while Section 5.4 considers the case of breaks occurring at unknown dates including problems with commonly used methods and recent proposals to overcome them (Section 5.4.2). Section 6 tackles the problem of testing for cointegration in the presence of structural changes in the constant and/or the cointegrating vector. We review first single equation methods (Section 6.1) and then, in Section 6.2, methods based on multi-equations systems where the object of interest is to determine the number of cointegrating vectors. Finally, Section 7 presents concluding remarks outlining a few important topics for future research and briefly reviews similar issues that arise in the context of long memory processes, an area where issues of structural changes (in particular level shifts) have played an important role recently, especially in light of the characterization of the time series properties of stock return volatility. Our focus is on conceptual issues about the frameworks adopted and the assumptions imposed as they relate to potential applicability. We also highlight problems that can occur with methods that are commonly used and recent work that has been done to overcome them. Space constraints are such that a detailed elicitation of all procedures discussed is not possible and the reader should consult the original work for details needed to implement them in practice. Even with a rich agenda, this review inevitably has to leave out a wide range of important work. The choice of topic is clearly closely related to the author’s own past and current work, and it is, accordingly, not an unbiased review, though we hope that a balanced treatment has been achieved to provide a comprehensive picture of how to deal with breaks in linear models. Important parts of the literature on structural change that are not covered include, among others, the following: methods related to the so-called on-line approach where the issue is to detect whether a change sas occurred in real time; results pertaining to non-linear models, in particular to tests for structural changes in a Generalized Method of Moment framework; smooth transition changes and threshold models; non parametric methods to estimate and detect changes; Bayesian methods; issues related to forecasting in the presence of structural changes; theoretical results and methods related to specialized cases that are not of general interest in economics; structural change in seasonal models; and bootstrap methods. The reader interested in further historical developments and methods not covered in this survey can consult the books by Clements and Hendry (1999), Csörgő and Horváth (1997), Krämer and Sonnberger (1986), Hackl and Westlund (1991), Hall (2005), Hatanaka 3 and Yamada (2003), Maddala nd Kim (1998), Tong (1990) and the following review articles: Bhattacharya (1994), Deshayes and Picard (1986), Hackl and Westlund (1989), Krishnaiah and Miao (1988), Perron (1994), Pesaran et al. (1985), Shaban (1980), Stock (1994), van Dijk et al. (2002) and Zacks (1983). 2 Introductory Historical Notes It will be instructive to start with some interesting historical notes concerning early tests for structural change. Consider a univariate time series, {yt ; t = 1, ..., T }, which under the null hypothesis is independently and identically distributed with mean µ and finite variance. Under the alternative hypothesis, yt is subject to a one time change in mean at some unknown date Tb , i.e., (1) yt = µ1 + µ2 1(t > Tb ) + et where et ∼ i.i.d. (0, σ 2e ) and 1(·) denotes the indicator function. Quandt (1958, 1960) had introduced what is now known as the Sup F test (assuming normally distributed errors), i.e., the likelihood ratio test for a change in parameters evaluated at the break date that maximizes the likelihood function. However, the limit distribution was then unknown. Quandt (1960) had shown that it was far from being a chi-square distribution and resorted to tabulate finite sample critical values for selected cases. Following earlier work by Chernoff and Zacks (1964) and Kander and Zacks (1966), an alternative approach was advocated by Gardner (1969) steemming from a suggestion by Page (1955, 1957) to use partial sums of demeaned data to analyze structural changes (see more on this below). The test considered is Bayesian in nature and, under the alternative, assigns weights pt as the prior probability that a change occurs at date t (t = 1, ..., T ). Assuming Normal errors and an unknown value of σ 2e , this strategy leads to the test " T #2 T X X −1 (yj − ȳ) pt Q = σ̂−2 e T t=1 j=t+1 PT 2 −1 2 where ȳ = T t=1 yt , is the sample average, and σ̂ e = T t=1 (yt − ȳ) is the sample variance of the data. With a prior that assigns equal weight to all observations, i.e. pt = 1/T , the test reduces to " T #2 T X X −2 (yj − ȳ) Q = σ̂ −2 e T PT −1 t=1 j=t+1 Under the null hypothesis, the test can be expressed as a ratio of quadratic forms in Normal variates and standard numerical method can be used to evaluate its distribution (e.g., Imhof, 4 1961, though Gardner originally analyzed the case with σ 2e known). The limit distribution of the statistic Q was analyzed by MacNeill (1974). He showed that Z 1 B0 (r)2 dr Q⇒ 0 where B0 (r) = W (r) − rW (1) is a Brownian bridge, and noted that percentage point had already been derived by Anderson and Darling (1952) in the context of goodness of fit tests. MacNeill (1978) extended the procedure to test for a change in a polynomial trend function of the form p X β i,t ti + et yt = i=0 where β i,t = β i + δ i 1(t > Tb ) The test of no change (δ i = 0 for all i) is then −2 Qp = σ̂ −2 e T T X t=1 " T X êj j=t+1 #2 P with σ̂2e = T −1 Tt=1 ê2t and êt the residuals from a regression of yt on {1, t, ..., tp }. The limit distribution is given by Z 1 Q⇒ Bp (r)2 dr 0 where Bp (r) is a generalized Brownian bridge. MacNeill (1978) computed the critical values by exact numerical methods up to six decimals accuracy (showing, for p = 0, the critical values of Anderson and Darling (1952) to be very accurate). The test was extended to allow dependence in the errors et by Perron (1991) and Tang and MacNeill (1993) (see also Kulperger, 1987a,b, Jandhyala and MacNeill, 1989, Jandhyala and Minogue, 1993, and Antoch et al., 1997). In particular, Perron (1991) shows that, under general conditions, the same limit distribution obtains using the statistic Q∗p = ĥe (0)−1 T −2 T X t=1 " T X j=t+1 êj #2 where ĥe (0) is a consistent estimate of (2π times) the spectral density function at frequency zero of et . 5 Even though, little of this filtered through the econometrics literature, the statistic Q∗p is well known to applied economists. It is the so-called KPSS test for testing the null hypothesis of stationarity versus the alternative of a unit root, see Kwiatkowski et al. (1992). More precisely, Qp is the Lagrange Multiplier (LM) and locally best invariant (LBI) test for testing the null hypothesis that σ 2u = 0 in the model yt = p X β i,t ti + rt + et i=0 rt = rt−1 + ut with ut ∼ i.i.d. N (0, σ 2u ) and et ∼ i.i.d. N (0, σ 2e ). Q∗p is then the corresponding large sample counterpart that allows correlation. Kwiatkowski et al. (1992) provided critical values for p = 0 and 1 using simulations (which are less precise than the critical values of Anderson and Darling, 1952, and MacNeill, 1978). In the econometrics literature, several extensions of this test have been proposed; in particular for testing the null hypothesis of cointegration versus the alternative of no cointegration (Nyblom and Harvey, 2000) and testing whether any part of a sample shows a vector of series to be cointegrated (Qu, 2004). Note also that the same test can be given the interpretation of a LBI for parameter constancy versus the alternative that the parameters follow a random walk (e.g., Nyblom and Mäkeläinen, 1983, Nyblom, 1989, Nabeya and Tanaka, 1988, Jandhyala and MacNeill, 1992, Hansen, 1992b). The same statistic is also the basis for a test of the null hypothesis of no-cointegration when considering functional of its reciprocal (Breitung, 2002). So what are we to make of all of this? The important message to learn from the fact that the same statistic can be applied to tests for stationarity versus either unit root or structural change is that the two issues are linked in important ways. Evidence in favor of unit roots can be a manifestation of structural changes and vice versa. This was indeed an important message of Perron (1989, 1990); see also Rappoport and Reichlin (1989). In this survey, we shall return to this problem and see how it introduces severe complications when dealing with structural changes and unit roots. It is also of interest to go back to the work by Page (1955, 1957) who had proposed to P use partial sums of demeaned data to test for structural change. Let Sr = rj=1 (yj − ȳ), his procedure for a two-sided test for change in the mean is based on the following quantities ¸ ∙ ¸ ∙ max Sr − min Si and max min Si − Sr 0≤r≤T 0≤i<r 0≤r≤T 0≤i<r and looks whether either exceeds a threshold (which, in the symmetric case, is the same). So we reject the null hypothesis if the partial sum rises enough from its previous minimum 6 or falls enough from its previous maximum. Nadler and Robbins (1971) showed that this procedure is equivalent to looking at the statistic ¸ ∙ RS = max Sr − min Sr 0≤r≤T 0≤r≤T i.e., to assess whether the range of the sequence of partial sums is large enough. But this is also exactly the basis of the popular rescaled range procedure used to test the null hypothesis of short-memory versus the alternative of long memory (see, in particular, Hurst, 1951, Mandelbrot and Taqqu, 1979, Bhattacharya et al., 1983, and Lo, 1991). This is symptomatic of the same problem discussed above from a slightly different angle; structural change and long memory imply similar features in the data and, accordingly, are hard to distinguish. In particular, evidence for long memory can be caused by the presence of structural changes, and vice versa. The intuition is basically the same as the message in Perron (1990), i.e., level shifts induce persistent features in the data. This problem has recently received a lot of attention, especially in the finance literature concerning the characteristics of stock returns volatility (see, in particular, Diebold and Inoue, 2001, Gourieroux and Jasiak, 2001, Granger and Hyung, 2004, Lobato and Savin, 1998, and Perron and Qu, 2004). 3 Estimation and Inference about Break Dates In this section we discuss issues related to estimation and inference about the break dates in a linear regression framework. The emphasis is on describing methods that are most useful in applied econometrics, explaining the relevance of the conditions imposed and sketching some important theoretical steps that help to understand particular assumptions made. Following Bai (1997a) and Bai and Perron (1998), the main framework of analysis can be described by the following multiple linear regression with m breaks (or m + 1 regimes): yt = x0t β + zt0 δ j + ut , t = Tj−1 + 1, ..., Tj , (2) for j = 1, ..., m + 1. In this model, yt is the observed dependent variable at time t; both xt (p × 1) and zt (q × 1) are vectors of covariates and β and δj (j = 1, ..., m + 1) are the corresponding vectors of coefficients; ut is the disturbance at time t. The indices (T1 , ..., Tm ), or the break points, are explicitly treated as unknown (the convention that T0 = 0 and Tm+1 = T is used). The purpose is to estimate the unknown regression coefficients together with the break points when T observations on (yt , xt , zt ) are available. This is a partial 7 structural change model since the parameter vector β is not subject to shifts and is estimated using the entire sample. When p = 0, we obtain a pure structural change model where all the model’s coefficients are subject to change. Note that using a partial structural change models where only some coefficients are allowed to change can be beneficial both in terms of obtaining more precise estimates and also in having can be more powerful tests. The multiple linear regression system (2) may be expressed in matrix form as Y = Xβ + Z̄δ + U, where Y = (y1 , ..., yT )0 , X = (x1 , ..., xT )0 , U = (u1 , ..., uT )0 , δ = (δ 01 , δ 02 , ..., δ 0m+1 )0 , and Z̄ is the matrix which diagonally partitions Z at (T1 , ..., Tm ), i.e. Z̄ = diag(Z1 , ..., Zm+1 ) with Zi = (zTi−1 +1 , ..., zTi )0 . We denote the true value of a parameter with a 0 superscript. In 0 0 particular, δ 0 = (δ 01 , ..., δ 0m+1 )0 and (T10 , ..., Tm0 ) are used to denote, respectively, the true values of the parameters δ and the true break points. The matrix Z̄ 0 is the one which diagonally partitions Z at (T10 , ..., Tm0 ). Hence, the data-generating process is assumed to be Y = Xβ 0 + Z̄ 0 δ 0 + U. (3) The method of estimation considered is based on the least-squares principle. For each mpartition (T1 , ..., Tm ), the associated least-squares estimates of β and δ j are obtained by minimizing the sum of squared residuals 0 (Y − Xβ − Z̄δ) (Y − Xβ − Z̄δ) = m+1 X Ti X [yt − x0t β − zt0 δ i ]2 . i=1 t=Ti−1 +1 Let β̂({Tj }) and δ̂({Tj }) denote the estimates based on the given m-partition (T1 , ..., Tm ) denoted {Tj }. Substituting these in the objective function and denoting the resulting sum of squared residuals as ST (T1 , ..., Tm ), the estimated break points (T̂1 , ..., T̂m ) are such that (T̂1 , ..., T̂m ) = argmin(T1 ,...,Tm ) ST (T1 , ..., Tm ), (4) where the minimization is taken over some set of admissible partitions (see below). Thus the break-point estimators are global minimizers of the objective function. The regression parameter estimates are the estimates associated with the m-partition {T̂j }, i.e. β̂ = β̂({T̂j }), δ̂ = δ̂({T̂j }). This framework includes many contributions made in the literature as special cases depending on the assumptions imposed; e.g., single change, changes in the mean of a stationary 8 process, etc. However, the fact that the method of estimation is based on the least-squares principle implies that, even if changes in the variance of ut are allowed, provided they occur at the same dates as the breaks in the parameters of the regression, such changes are not exploited to increase the precision of the break date estimators. This is due to the fact that the least-squares method imposes equal weights on all residuals. Allowing different weights, as needed when accounting for changes in variance, requires adopting a quasi-likelihood framework, see below. 3.1 The assumptions and their relevance To obtain theoretical results about the consistency and limit distribution of the break dates, some conditions need to be imposed on the regressors, the errors, the set of admissible partitions and the break dates. To our knowledge, the most general set of assumptions, as far as applications are concerned, are those in Perron and Qu (2005). Some are simply technical (e.g., invertibility requirements), while others restrict the potential applicability of the results. Hence, it is useful to discuss the latter. • Assumption on the regressors: Let wt = (x0t , zt0 )0 , for i = 0, ..., m, (1/li ) Qi (v) a non-random positive definite matrix uniformly in v ∈ [0, 1]. PTi0 +[li v] t=Ti0 +1 wt wt0 →p This assumption allows the distribution of the regressors to vary across regimes. It, however, requires the data to be weakly stationary stochastic processes. It can, however, be relaxed substantially, though the technical proofs then depend on the nature of the relaxation. For instance the scaling used forbids trending regressors, unless they are of the form {1, (t/T ), ..., (t/T )p }, say, for a polynomial trend of order p. Casting trend functions in this form can deliver useful results in many cases. However, there are instances where specifying trends in unscaled form, i.e., {1, t, ..., tp }, can deliver much better results, especially if level and trend slope changes occur jointly. Results using unscaled trends with p = 1 are presented in Perron and Zhu (2005). A comparison of their results with other trend specifications is presented in Deng and Perron (2005). Another important restriction is implied by the requirement that the limit be a fixed matrix, as opposed to permitting it to be stochastic. This, along with the scaling, precludes integrated processes as regressors (i.e., unit roots). In the single break case, this has been relaxed by Bai, Lumsdaine and Stock (1998) who considered, among other things, structural changes in cointegrated relationships. Consistency still applies but the rate of convergence and limit distributions of the estimates are different. Another context in which integrated 9 regressors play a role is the case of changes in persistence. Chong (2001) considered an AR(1) model where the autoregressive coefficient takes a value less than one before some break date and value one after, or vice versa. He showed consistency of the estimate of the break date and derived the limit distribution. When the move is from stationarity to unit root, the rate of convergence is the same as in the stationary case (though the limit distribution is different), but interestingly, the rate of convergence is faster when the change is from a unit root to a stationary process. No results are yet available for multiple structural changes in regressions involving integrated regressors, though work is in progress on this issue. The problem here is more challenging because the presence of regressors with a unit root, whose coeffients are subject to change, implies break date estimates with limit distributions that are not independent, hence all break dates need to be evaluated jointly. The sequence {wt ut } satisfies the following set of conditions. • Assumptions on the errors: Let the Lr -norm of a random matrix X be defined by P P kXkr = ( i j E |Xij |r )1/r for r ≥ 1. (Note that kXk is the usual matrix norm or the Euclidean norm of a vector.) With {Fi : i = 1, 2, ..} a sequence of increasing σ-fields, it is assumed that {wi ui , Fi } forms a Lr -mixingale sequence with r = 2 + δ for some δ > 0. That is, there exist nonnegative constants {ci : i ≥ 1} and {ψj : j ≥ 0} such that ψj ↓ 0 as j → ∞ and for all i ≥ 1 and j ≥ 0, we have: (a) kE(wi ui |Fi−j )kr ≤ ci ψj , (b) kwi ui − E(wi ui |Fi+j )kr ≤ ci ψj+1 . Also assume (c) maxi ci ≤ K < ∞, (d) P∞ 1+k ψj < ∞, (e) kzi k2r < M < ∞ and kui k2r < N < ∞ for some K, M, N > 0. j=0 j This imposes mild restrictions on the vector wt ut and permits a wide class of potential correlation and heterogeneity (including conditional heteroskedasticity) and also allows lagged dependent variables. It rules out errors that have unit roots. In this latter case, if the regressors are stationary (or satisfy the Assumption on the regressors stated above), the estimates of the break dates are inconsistent (see Nunes et al., 1995). However, unit root errors can be of interest; for example when testing for a change in the deterministic component of the trend function for an integrated series, in which case the estimates are consistent (see Perron and Zhu, 2005). The set of conditions listed above are not the weakest possible. For example, Lavielle and Moulines (2000) allow the errors to be strongly dependent, i.e., long memory processes such as fractionally integrated ones are permitted. They, however, consider only the case of multiple changes in the mean. Technically, what is important is to be able to establish a generalized Hájek-Rényi (1955) type inequality for the zero mean variables zt ut , as well as a Functional Central Limit Theorem and a Law of Large Numbers. 10 • Assumption on the minimization problem: The minimization problem defined by (4) is taken over all possible partitions such that Ti − Ti−1 ≥ T for some > 0. This requirement was introduced in Bai and Perron (1998) only for the case where lagged dependent variables were allowed. When serial correlation in the errors was allowed they introduced the requirement that the errors be independent of the regressors at all leads and lags. This is obviously a strong assumption which is often violated in practice. The assumption on the errors listed above are much weaker, in particular concerning the relation between the errors and regressors. This weakening comes at the cost of a mild strengthening on the assumption about the regressors and the introduction of the restriction on the minimization problem. Note that the latter is also imposed in Lavielle and Moulines (2000), though they note that it can be relaxed with stronger conditions on zt ut or by constraining the estimates to lie in a compact set. • Assumption on the break dates: Ti0 = [T λ0i ] , where 0 < λ01 < ... < λ0m < 1. This assumption specifies that the break dates are asymptotically distinct. While it is standard, it is surprisingly the most controversial for some. The reason is that it dictates the asymptotic framework adopted. With this condition, when the sample size T increases, all segments increase in length in the same proportions to each other. Oftentimes, an asymptotic analysis is viewed as a thought experiment about what would happen if we were able to collect more and more data in the future. If one adheres to this view, then the last regime should increase in length (assuming no other break will occur in the future) and all other segments then become a negligible proportion of the total sample. Hence, as T increases, we would find ourselves with a single segment, in which case the framework becomes useless. The fact is that any asymptotic analysis is simply a device to enable us to get useful information about the structure, which can help us understand the finite sample distributions, and hopefully to deliver good approximations. The adoption of any asymptotic framework should only be evaluated on this basis, no matter how ad hoc it may seem at first sight. Here, with say a sample of size 100 and 3 breaks occurring at dates 25, 50 and 75, all segments are a fourth of the total sample. It therefore makes sense to use an asymptotic framework whereby this feature is preserved. The same comments apply to contexts in which some parameters are made local to some boundary as the sample size increases. No claim whatsoever is made that the parameter would actually change if more data were collected, yet such a device has been found to be of great use and to provide very useful approximations. This applies to local 11 asymptotic power function, roots local to unity or shrinking size of shifts as we will discuss later. Having said that, it does not mean that the asymptotic framework that is adopted in this literature is the only one useful or even the best. For example, it is conceivable that an asymptotic theory whereby more and more data are added keeping a fixed span of data would be useful as well. However, such a continuous time limit distribution has not yet appeared in the structural change context. Under these conditions, the main theoretical results are that the break fractions λ0i are consistently estimated, i.e., λ̂i ≡ (T̂i /T ) →p λ0i and that the rate of convergence is T . More precisely, for every ε > 0, there exists a C < ∞, such that for large T , P (|T (λ̂i − λ0i )| > C∆−2 i ) < ε (5) for every i = 1, ..., m, where ∆i = δ i+1 − δi . Note that the estimates of the break dates are not consistent themselves, but the differences between the estimates and the true values are bounded by some constant, in probability. Also, this implies that the estimates of the other parameters have the same distribution as would prevail if the break dates were known. Kurozumi and Arai (2004) obtain a similar result with I(1) regressors for a cointegrated model subject to a change in some parameters of the cointegrating vector. They show the estimate of the break fraction obtained by minimizing the sum of squared residuals from the static regression to converge at a fast enough rate for the estimates of the parameters of the model to asymptotically unaffected by the estimation of the break date. 3.2 Allowing for restrictions on the parameters Perron and Qu (2005) approach the issues of multiple structural changes in a broader framework whereby arbitrary linear restrictions on the parameters of the conditional mean can be imposed in the estimation. The class of models considered is y = Z̄δ + u where Rδ = r with R a k by (m + 1)q matrix with rank k and r, a k dimensional vector of constants. The assumptions are the same as discussed above. Note first that there is no need for a distinction between variables whose coefficients are allowed to change and those whose coefficients are not allowed to change. A partial structural change model can be obtained as a special case 12 by specifying restrictions that impose some coefficients to be identical across all regimes. This is a useful generalization since it permits a wider class of models of practical interests; for example, a model which specifies a number of states less than the number of regimes (with two states, the coefficients would be the same in odd and even regimes). Or it could be the case that the value of the parameters in a specific segment is known. Also, a subset of coefficients may be allowed to change over only a limited number of regimes. Perron and Qu (2005) show that the same consistency and rate of convergence results hold. Moreover, an interesting result is that the limit distribution (to be discussed below) of the estimates of the break dates are unaffected by the imposition of valid restrictions. They document, however, that improvements can be obtained in finite samples. But the main advantage of imposing restrictions is that much more powerful tests are possible. 3.3 Method to Compute Global Minimizers We now briefly discuss issues related to the estimation of such models, in particular when multiple breaks are allowed. What are needed are global minimizers of the objective function (4). A standard grid search procedure would require least squares operations of order O(T m ) and becomes prohibitive when the number of breaks is greater than 2, even for relatively small samples. Bai and Perron (2003a) discuss a method based on a dynamic programming algorithm that is very efficient. Indeed, the additional computing time needed to estimate more than two break dates is marginal compared to the time needed to estimate a two break model. The basis of the method, for specialized cases, is not new and was considered by Guthery (1974), Bellman and Roth (1969) and Fisher (1958). A comprehensive treatment was also presented in Hawkins (1976). Consider the case of a pure structural change model. The basic idea of the approach becomes fairly intuitive once it is realized that, with a sample of size T , the total number of possible segments is at most T (T + 1)/2 and is therefore of order O(T 2 ). One then needs a method to select which combination of segments (i.e., which partition of the sample) yields a minimal value of the objective function. This is achieved efficiently using a dynamic programming algorithm. For models with restrictions (including the partial structural change model), an iterative procedure is available, which in most cases requires very few iterations (see Bai and Perron, 2003, and Perron and Qu, 2005, who make available Gauss codes to perform these and other tasks). Hence, even with large samples, the computing cost to estimate models with multiple structural changes should be considered minimal. 13 3.4 The limit distribution of the estimates of the break dates With the assumptions on the regressors, the errors and given the asymptotic framework adopted, the limit distributions of the estimates of the break dates are independent of each other. Hence, for each break date, the analysis becomes exactly the same as if a single break has occurred. The intuition behind this feature is first that the distance between each break increases at rate T as the sample size increases. Also, the mixing conditions on the regressors and errors impose a short memory property so that events that occur a long enough time apart are independent. This independence property is unlikely to hold if the data are integrated but such an analysis is yet to be completed. We shall not reproduce the results in details but simply describe the main qualitative feature and the practical relevance of the required assumptions. The reader is referred to Bai (1997a) and Bai and Perron (1998, 2003a), in particular. Also, confidence intervals for the break dates need not be based on the limit distributions of the estimates. Other approaches are possible, for example by inverting a suitable test (e.g., Elliott and Müller, 2004, for an application in the linear model using a locally best invariant test). For a review of alternative methods, see Siegmund (1988). The limit distribution of the estimates of the break dates depends on: a) the magnitude of the change in coefficients (with larger changes leading to higher precision, as expected), b) the (limit) sample moment matrices of the regressors for the segments prior to and after the true break date (which are allowed to be different); c) the so-called ‘long-run’ variance of {wt ut }, which involves potential serial correlation in the errors (and which again is allowed to be different prior to and after the break); d) whether the regressors are trending or not. In all cases, all relevant nuisance parameters can be consistently estimated and the appropriate confidence intervals constructed. A feature of interest is that the confidence intervals need not be symmetric given that the data and errors can have different properties before and after the break. To get an idea of the importance of particular assumptions needed to derive the limit distribution, it is instructive to look at a simple case with i.i.d. errors ut and a single break (for details, see Bai, 1997a). Then the estimate of the break satisfies, £ ¤ T̂1 = arg min SSR(T1 ) = arg max SSR(T10 ) − SSR(T1 ) Using the fact that, given the rate of convergence result (5), the inequality |T̂1 − T10 | < C∆−2 is satisfied with probability one in large samples (here, ∆ = δ 2 − δ 1 ). Hence, we can restrict 14 the search over the compact set C(∆) = {T1 : |T1 − T10 | < C∆−2 }. Then for T1 < T10 , 0 SSR(T10 ) − SSR(T1 ) = −∆0 T1 X 0 T1 X zt zt0 ∆ + 2∆0 t=T1 +1 zt ut + op (1) (6) zt ut + op (1) (7) t=T1 +1 and, for T1 > T10 , SSR(T10 ) 0 − SSR(T1 ) = −∆ T1 X zt zt0 ∆ t=T10 +1 T1 X 0 − 2∆ t=T10 +1 The problem is that, with |T1 − T10 | bounded, we cannot apply a Law of Large Numbers or a Central Limit Theorem to approximate the sums above with something that does not depend on the exact distributions of zt and ut . Furthermore, the distributions of these sums depend on the exact location of the break. Now let 0 W1 (m) = −∆ 0 X zt zt0 ∆ 0 + 2∆ t=m+1 for m < 0 and W2 (m) = −∆0 m X 0 X zt ut t=m+1 zt zt0 ∆ + 2∆0 m X zt ut t=1 t=1 for m > 0. Finally, let W (m) = W1 (m) if m < 0, and W (m) = W2 (m) if m > 0 (with W (0) = 0). Now, assuming a strictly stationary distribution for the pair {zt , ut }, we have that SSR(T10 ) − SSR(T1 ) = W (T1 − T10 ) + op (1) i.e., the assumption of strict stationarity allows us to get rid of the dependence of the distribution on the exact location of the break. Assuming further that (∆0 zt )2 ± (∆0 zt )ut has a continuous distribution ensures that W (m) has a unique maximum. So that T̂1 − T10 →d arg max W (m). m An important early treatment of this result for a sequence of i.i.d. random variables is Hinkley (1970). See also Feder (1975) for segmented regressions that are continuous at the time of break, Bhattacharya (1987) for maximum likelihood estimates in a multi-parameter case and Bai (1994) for linear processes. Now the issue is that of getting rid of the dependence of this limit distribution on the exact distribution of the pair (zt , ut ). Looking at (6) and (7), what we need is for the 15 difference T1 − T10 to increase as the sample size increases, then a Law of Large Numbers and a Functional Central Limit Theorem can be applied. The trick is to realize that from the convergence rate result (5), the rate of convergence of the estimate will be slower if the change in the parameters ∆i gets smaller as the sample size increases, but does so slowly enough for the estimated break fraction to remain consistent. Early applications of this framework are Yao (1987) in the context of a change in distribution for a sequence of i.i.d. random variables, and Picard (1985) for a change in an autoregressive process. Letting ∆ = ∆T to highlight the fact the change in the parameters depends on the sample size, this leads to the specification ∆T = ∆0 vT where vT is such that vT → 0 and T (1/2)−α vT → ∞ for some α ∈ (0, 1/2). Under these specifications, we have from (5) that T̂1 − T10 = Op (T 1−2α ). Hence, we can restrict the search to those values T1 such that T1 = T10 + [svT−2 ] for some fixed s. We can write (6) as 0 0 SSR(T10 ) − SSR(T1 ) = −∆00 vT2 T1 X zt zt0 ∆ + 2∆00 vT T1 X zt ut + op (1) t=T1 +1 t=T1 +1 The next steps depend on whether the zt includes trending regressors. Without trending regressors, the following assumptions are imposed (in the case with ut is i.i.d.) 0 • Assumptions for limit distribution: Let ∆Ti0 = Ti0 − Ti−1 , then as ∆Ti0 → ∞: a) 0 0 0 0 PTi−1 +[s∆Ti ] 0 PTi−1 +[s∆Ti ] 2 (∆Ti0 )−1 t=T zt zt →p sQi , b) (∆Ti0 )−1 t=T ut →p sσ 2i 0 +1 0 +1 i−1 i−1 These imply that 0 +[s∆T 0 ] Ti−1 i (∆Ti0 )−1/2 X 0 +1 t=Ti−1 zt ut ⇒ Bi (s) where Bi (s) is a multivariate Gaussian process on [0, 1] with mean zero and covariance E[Bi (s)Bi (u)] = min{s, u}σ 2i Qi . Hence, for s < 0 SSR(T10 ) − SSR(T10 + [svT−2 ]) = −|s|∆00 Q1 ∆0 + 2(∆00 Q1 ∆0 )1/2 W1 (−s) + op (1) where W1 (−s) is a Weiner process defined on (0, ∞). A similar analysis holds for the case s > 0 and for more general assumptions on ut . But this suffices to make clear that under these assumptions, the limit distribution of the estimate of the break date no longer depends on the exact distribution of zt and ut but only on quantities that can be consistently estimated. For details, see Bai (1997) and Bai and Perron (1998, 2003a). With trending regressors, the assumption stated above is violated but a similar result is still possible (assuming trends of 16 the form (t/T )) and the reader is referred to Bai (1997a) for the case where zt is a polynomial time trend. So, what do we learn from these asymptotic results? First, for large shifts, the distributions of the estimates of the break dates depend on the exact distributions of the regressors and errors even if the sample is large. When shifts are small, we can expect the distributions of the estimates of the break dates to be insensitive to the exact nature of the distributions of the regressors and errors. The question is then, how small do the changes have to be? There is no clear cut solution to this problem and the answer is case specific. The simulations in Bai and Perron (2005) show that the shrinking shifts asymptotic framework provides useful approximations to the finite sample distribution of the estimated break dates, but their simulation design uses normally distributed errors and regressors. The coverage rates are adequate, in general, unless the shifts are quite small in which case the confidence interval is too narrow. The method of Elliott and Müller (2004), based on inverting a test, works better in that case. However, with such small breaks, tests for structural change will most likely fail to detect a change, in which case most practitioners would not pursue the analysis further and consider the construction of confidence intervals. On the other hand, Deng and Perron (2005) show that the shrinking shift asymptotic framework leads to a poor approximation in the context of changes in a linear trend function and that the limit distribution based on a fixed magnitude of shift is highly preferable. 3.5 Estimating Breaks one at a time Bai (1997b) and Bai and Perron (1998) showed that it is possible to consistently estimate all break fractions sequentially, i.e., one at a time. This is due to the following result. When estimating a single break model in the presence of multiple breaks, the estimate of the break fraction will converge to one of the true break fractions, the one that is dominant in the sense that taking it into account allows the greatest reduction in the sum of squared residuals. Then, allowing for a break at the estimated value, a second one break model can be applied which will consistently estimate the second dominating break, and so on (in the case of two breaks that are equally dominant, the estimate will converge with probability 1/2 to either break). Fu and Cornow (1990) presented an early account of this property for a sequence of Bernoulli random variables when the probability of obtaining a 0 or a 1 is subject to multiple structural changes (see also, Chong, 1995). Bai (1997b) considered the limit distribution of the estimates and shows that they are not the same as those obtained when estimating all break dates simultaneously. In particular, 17 except for the last estimated break date, the limit distributions of the estimates of the break dates depend on the parameters in all segments of the sample (when the break dates are estimated simultaneously, the limit distribution of a particular break date depends on the parameters of the adjacent regimes only). To remedy this problem, Bai (1997b) suggested a procedure called ‘repartition’. This amounts to re-estimating each break date conditional on the adjacent break dates. For example, let the initial estimates of the break dates be denoted by (T̂1a , ..., T̂ma ). The second round estimate for the ith break date is obtained by fitting a a a (with + 1 and ending at date T̂i+1 one break model to the segment starting at date T̂i−1 a = T ). The estimates obtained from this repartition the convention that T̂0a = 0 and T̂m+1 procedure have the same limit distributions as those obtained simultaneously, as discussed above. 3.6 Estimation in a system of regressions The problem of estimating structural changes in a system of regressions is relatively recent. Bai et al. (1998) considered asymptotically valid inference for the estimate of a single break date in multivariate time series allowing stationary or integrated regressors as well as trends. They show that the width of the confidence interval decreases in an important way when series having a common break are treated as a group and estimation is carried using a quasi maximum likelihood (QML) procedure. Also, Bai (2000) considers the consistency, rate of convergence and limiting distribution of estimated break dates in a segmented stationary VAR model estimated again by QML when the breaks can occur in the parameters of the conditional mean, the covariance matrix of the error term or both. Hansen (2003) considers multiple structural changes in a cointegrated system, though his analysis is restricted to the case of known break dates. To our knowledge, the most general framework is that of Qu and Perron (2005) who consider models of the form yt = (I ⊗ zt0 )Sβ j + ut for Tj−1 + 1 ≤ t ≤ Tj (j = 1, ..., m + 1), where yt is an n-vector of dependent variables and zt is a q-vector that includes the regressors from all equations. The vector of errors ut has mean 0 and covariance matrix Σj . The matrix S is of dimension nq by p with full column rank. Though, in principle it is allowed to have entries that are arbitrary constants, it is usually a selection matrix involving elements that are 0 or 1 and, hence, specifies which regressors appear in each equation. The set of basic parameters in regime j consists of the p vector β j and of Σj . They also allow for the imposition of a set of r restrictions of 18 the form g(β, vec(Σ)) = 0, where β = (β 01 , ..., β 0m+1 )0 , Σ = (Σ1 , ..., Σm+1 ) and g(·) is an r dimensional vector. Both within- and cross-equation restrictions are allowed, and in each case within or across regimes. The assumptions on the regressors zt and the errors ut are similar to those discussed in Section 3.1 (properly extended for the multivariate nature of the problem). Hence, the framework permits a wide class of models including VAR, SUR, linear panel data, change in means of a vector of stationary processes, etc. Models with integrated regressors (i.e, models with cointegration) are not permitted. Allowing for general restrictions on the parameters β j and Σj permits a very wide range of special cases that are of practical interest: a) partial structural change models where only a subset of the parameters are subject to change, b) block partial structural change models where only a subset of the equations are subject to change; c) changes in only some element of the covariance matrix Σj (e.g., only variances in a subset of equations); d) changes in only the covariance matrix Σj , while β j is the same for all segments; e) ordered break models where one can impose the breaks to occur in a particular order across subsets of equations; etc. The method of estimation is again QML (based on Normal errors) subject to the restrictions. They derive the consistency, rate of convergence and limit distribution of the estimated break dates. They obtain a general result stating that, in large samples, the restricted likelihood function can be separated in two parts: one that involves only the break dates and the true values of the coefficients, so that the estimates of the break dates are not affected by the restrictions imposed on the coefficients; the other involving the parameters of the model, the true values of the break dates and the restrictions, showing that the limiting distributions of these estimates are influenced by the restrictions but not by the estimation of the break dates. The limit distribution results for the estimates of the break dates are qualitatively similar to those discussed above, in particular they depend on the true parameters of the model. Though only root-T consistent estimates of (β, Σ) are needed to construct asymptotically valid confidence intervals, it is likely that more precise estimates of these parameters will lead to better finite sample coverage rates. Hence, it is recommended to use the estimates obtained imposing the restrictions even though imposing restrictions does not have a first-order effect on the limiting distributions of the estimates of the break dates. To make estimation possible in practice, for any number of breaks, they present an algorithm which extends the one discussed in Bai and Perron (2003a) using, in particular, an iterative GLS procedure to construct the likelihood function for all possible segments. The theoretical analysis shows how substantial efficiency gains can be obtained by casting 19 the analysis in a system of regressions. In addition, the result of Bai et al. (1998), that when a break is common across equations the precision increases in proportion to the number of equations, is extended to the multiple break case. More importantly, the precision of the estimate of a particular break date in one equation can increase when the system includes other equations even if the parameters of the latter are invariant across regimes. All that is needed is that the correlation between the errors be non-zero. While surprising, this result is ex-post fairly intuitive since a poorly estimated break in one regression affects the likelihood function through both the residual variance of that equation and the correlation with the rest of the regressions. Hence, by including ancillary equations without breaks, additional forces are in play to better pinpoint the break dates. Qu and Perron (2005) also consider a novel (to our knowledge) aspect to the problem of multiple structural changes labelled “locally ordered breaks”. Suppose one equation is a policy-reaction function and the other is some market-clearing equation whose parameters are related to the policy function. According to the Lucas critique, if a change in policy occurs, it is expected to induce a change in the market equation but the change may not be simultaneous and may occur with a lag, say because of some adjustments due to frictions or incomplete information. However, it is expected to take place soon after the break in the policy function. Here, the breaks across the two equations are “ordered” in the sense that we have the prior knowledge that the break in one equation occurs after the break in the other. The breaks are also “local” in the sense that the time span between their occurrence is expected to be short. Hence, the breaks cannot be viewed as occurring simultaneously nor can the break fractions be viewed as asymptotically distinct. An algorithm to estimate such models is presented. Also, a framework to analyze the limit distribution of the estimates is introduced. Unlike the case with asymptotically distinct breaks, here the distributions of the estimates of the break dates need to be considered jointly. 4 Testing for structural change In this section, we review testing procedures related to structural changes. The following issues are covered: tests obtained without modelling any break, tests for a single structural change obtained by explicitly modelling a break, the problem of non monotonic power functions, and tests for multiple structural changes, tests valid with I(1) regressors, and tests for a change in slope valid allowing the noise component to be I(0) or I(1). 20 4.1 Tests for a single change without modelling the break Historically, tests for structural change were first devised based on procedures that did not estimate a break point explicitly. The main reason is that the distribution theory for the estimates of the break dates (obtained using a least-squares or likelihood principle) was not available and the problem was solved only for few special cases (see, e.g., Hawkins, 1977, Kim and Siegmund, 1989). Most tests proposed were of the form of partial sums of residuals. We have already discussed in Section 2, the Q test based on the average of partial sums of residuals (e.g., demeaned data for a change in mean) and the rescaled range test based on the range of partial sums of similarly demeaned data. Another statistic which has played an important role in theory and applications is the CUSUM test proposed by Brown, Durbin and Evans (1975). This test is based on the maximum of partial sums of recursive residuals. More precisely, for a linear regression with k regressors yt = x0t β + ut it is defined by ¯ ¯ Pr ¯ t=k+1 e v t¯ ¯ /(1 + 2 r − k ) CUSU M = max ¯¯ √ k+1<r≤T σ̂ T − k ¯ T −k where σ̂ 2 is a consistent estimate of the variance of ut (usually the sum of squared OLS residuals although, to increase power, one can use the sum of squared demeaned recursive residuals, as suggested by Harvey, 1975) and e vt are the recursive residuals defined by vet = (yt − x0t β̂ t−1 )/ft 0 ft = (1 + x0t (Xt−1 Xt−1 )xt )1/2 where Xt−1 contains the observations on the regressors up to time t − 1 and β̂ t−1 is the OLS estimate of β using data up to time t − 1. For an extensive review of the use of recursive methods in the analysis of structural change, see Dufour (1982) (see also Dufour and Kiviet, 1996, for finite sample inference in a regression model with a lagged dependent variable). The limit distribution of the CUSUM test can be expressed in terms of the maximum of a weighted Wiener process, i.e., ¯ ¯ ¯ W (r) ¯ ¯ CUSUM ⇒ sup ¯¯ ¯ 0≤r≤1 1 + 2r where W (r) is a unit Wiener process defined on (0, 1), see Sen (1982). Also, it was shown by Kramer, Ploberger and Alt (1988) that the limit distribution remains valid even if lagged 21 dependent variables are present as regressors. Furthermore, Ploberger and Kramer (1992) showed that using OLS residuals instead of recursive residuals yields a valid test, though the limit distribution under the null hypothesis is different (expressed in terms of a Brownian bridge, W (r) − rW (1), instead of a Wiener process). Their simulations showed the OLS based CUSUM test to have higher power except for shifts that occur early in the sample (the standard CUSUM tests having small power for late shifts). An alternative, also suggested by Brown, Durbin and Evans (1975), is the CUSUM of squares test. It takes the form: ¯ ¯ ¯ (r) r − k ¯ ¯ CUSSQ = max ¯¯ST − k+1<r≤T T − k¯ where (r) ST = à r X t=k+1 ! à vet2 / T X t=k+1 vet2 ! Ploberger and Kramer (1990) considered the local power functions of the CUSUM and CUSUM of squares. The former has non-trivial local asymptotic power unless the mean regressor is orthogonal to all structural changes. On the other hand, the latter has only trivial local power (i.e., power equal to size) for local changes that specify a one-time change in the coefficients (see also Deshayes and Picard, 1986). This suggests that the CUSUM test should be preferred, a conclusion we shall revisit below. Another variant using partial sums is the fluctuations test of Ploberger, Kramer and Kontrus (1989) which looks at the maximum difference between the OLS estimate of β using the full sample and the OLS estimates using subsets of the sample from the first observation to some date t, ranging from t = k to T . A similar test for a change in the slope of a linear trend function is analyzed in Chu and White (1992). Also, Chu, Hornik and Kuan (1995) looked at the maximum of moving sums of recursive and least-squares residuals. 4.2 Non monotonic power functions in finite samples All tests discussed above are consistent for given fixed values in the relevant set of alternative hypotheses. All (except the CUSUM of squares) are, however, subject to the following problem. For a given sample size, the power function can be non monotonic in the sense that it can decrease and even reach a zero value as the alternative considered becomes further away from the null value. This was shown by Perron (1991) for the Q statistic and extended to a wide range of tests in a comprehensive analysis by Vogelsang (1999). 22 This was illustrated using a basic shift in mean process or a shift in the slope of a linear trend (for some statistics designed for that alternative). In the change in mean case, with a single shift occurring, it was shown that the power of the tests discussed above eventually decreases as the magnitude of the shift increases and can reach zero. This decrease in power can be especially pronounced and effective with smaller mean shifts when a lagged dependent variable is included as a regressor to account for potential serial correlation in the errors. The basic reason for this feature is the need to estimate the variance of the errors (or the spectral density function at frequency zero when correlation in the errors is allowed) to properly scale the statistics. Since no break is directly modelled, one needs to estimate this variance using least-squares or recursive residuals that are ‘contaminated’ by the shift under the alternative. As the shift gets larger, the estimate of the scale gets inflated with a resulting loss in power. With a lagged dependent variable, the problem is exacerbated because the shift induces a bias of the autoregressive coefficient towards one (Perron, 1989, 1990). See Vogelsang (1999) for a detailed treatment that explains how each test is differently affected, that also provides empirical illustrations of this problem showing its practical relevance. Crainiceanu and Vogelsang (2001) also show how the problem is exacerbated when using estimates of the scale factor that allow for correlation, e.g., weighted sums of the autocovariance function. The usual methods to select the bandwidth (e.g., Andrews, 1991) will choose a value that is severely biased upward and lead to a decrease in power. With change in slope, the bandwidth increases at rate T and the tests become inconsistent. This is a troubling feature since tests that are consistent and have good local asymptotic properties can perform rather badly globally. In simulations reported in Perron (2005), this feature does not occur for the CUSUM of squares test. This leads us to the curious conclusion that the test with the worst local asymptotic property (see above) has the better global behavior. Methods to overcome this problem have been suggested by Altissimo and Corradi (2003) and Juhl and Xiao (2005). They suggest using non-parametric or local averaging methods where the mean is estimated using data in a neighborhood of a particular data point. The resulting estimates and tests are, however, very sensitive to the bandwidth used. A large one leads to properly sized tests in finite samples but with low power, and a small bandwidth leads to better power but large size distortions. There is currently no reliable method to appropriately chose this parameter in the context of structural changes. 23 4.3 Tests that allow for a single break The discussion above suggests that to have better tests for the null hypothesis of no structural change versus the alternative hypothesis that changes are present, one should consider statistics that are based on a regression that allows for a break. As discussed in the introduction, the suggestion by Quandt (1958, 1960) was to use the likelihood ratio test evaluated at the break date that maximizes this likelihood function. This is a non-standard problem since one parameter is only identified under the alternative hypothesis, namely the break date (see Davies, 1977, 1987, King and Shively, 1993, Andrews and Ploberger, 1994, and Hansen, 1996). The problem raised by Quandt was treated under various degrees of specificity by Deshayes and Picard (1984b), Worsley (1986), James, James and Siegmund (1987), Hawkins (1987), Kim and Siegmund (1989), Horvath (1995) and generalized by Andrews (1993a). The basic method advocated by Davies (1977), for the case in which a nuisance parameter is present only under the alternative, is to use the maximum of the likelihood ratio test over all possible values of the parameter in some pre-specified set as a test statististic. In the case of a single structural change occurring at some unknown date, this translates into the following statistic sup LRT (λ1 ) λ1 ∈Λ where LR(λ1 ) denotes the value of the likelihood ratio evaluated at some break point T1 = [T λ1 ] and the maximization is restricted over break fractions that are in Λ = [ 1 , 1 − 2 ], some subset of the unit interval [0, 1] with 1 being the lower bound and 1 − 2 the upper bound. The limit distribution of the statistic is given by sup LRT (λ1 ) ⇒ sup Gq (λ1 ) λ1 ∈Λ λ1 ∈Λ where Gq (λ1 ) = [λ1 Wq (1) − Wq (λ1 )]0 [λ1 Wq (1) − Wq (λ1 )] λ1 (1 − λ1 ) (8) with Wq (λ) a vector of independent Wiener processes of dimension q, the number of coefficients that are allowed to change (this result holds with non-trending data). Not surprisingly, the limit distribution depends on q but it also depends on Λ . This is important since the restriction that the search for a maximum value be restricted is not simply a technical requirement. It influences the properties of the test in an important way. In particular, Andrews (1993a) shows that if 1 = 2 = 0 so that no restrictions are imposed, the test diverges to 24 infinity under the null hypothesis (an earlier statement of this result in a more specialized context can be found in Deshayes and Picard, 1984a). This means that critical values grow and the power of the test decreases as 1 and 2 get smaller. Hence, the range over which we search for a maximum must be small enough for the critical values not to be too large and for the test to retain descent power, yet large enough to include break dates that are potential candidates. In the single break case, a popular choice is 1 = 2 = .15. Andrews (1993a) tabulates critical values for a range of dimensions q and for intervals of the form [ , 1 − ]. This does not imply, however, that one is restricted to imposing equal trimming at both ends of the sample. This is because the limit distribution depends on 1 and 2 only through the parameter γ = 2 (1 − 1 )/( 1 (1 − 2 )). Hence, the critical values for a symmetric trimming are also valid for some asymmetric trimmings. To better understand these results, it is useful to look at the simple one-time shift in mean of some variable yt specified by (1). For a given break date T1 = [T λ1 ], the Wald test is asymptotically equivalent to the LR test and is given by WT (λ1 ) = SSR(1, T ) − SSR(1, T1 ) − SSR(T1 + 1, T ) [SSR(1, T1 ) + SSR(T1 + 1, T )]/T where SSR(i, j) is the sum of squared residuals from regressing yt on a constant using data from date i to date j, i.e. à ! ! à j j j j X X 1 X 1 X et − yt − yt = et SSR(i, j) = j − i j − i t=i t=i t=i t=i Note that the denominator converges to σ 2 and the numerator is given by T X t=1 = ∙ à T 1X et − et T t=1 !2 − T1 X t=1 à T1 1 X et − et T1 t=1 !2 − T X t=T1 +1 à T X 1 et − et T − T1 t=T 1 !2 µ ¶¸−1 à T1 T T1 T1 −1/2 X T1 T − T1 −1/2 X T T 1− et et − T T T T t=1 t=T +1 1 !2 P1 P after some algebra. If T1 /T → λ1 ∈ (0, 1), we have T −1/2 Tt=1 et ⇒ σW (λ1 ), T −1/2 Tt=T1 +1 et = P P 1 T −1/2 Tt=1 et − T −1/2 Tt=1 et ⇒ σ[W (1) − W (λ1 )] and the limit of the Wald test is WT (λ1 ) ⇒ = 1 [λ1 W (1) − λ1 W (λ1 ) − (1 − λ1 )W (λ1 )]2 λ1 (1 − λ1 ) 1 [λ1 W (1) − W (λ1 )]2 λ1 (1 − λ1 ) 25 which is equivalent to (8) for q = 1. Andrews (1993a) also considered tests based on the maximal value of the Wald and LM tests and shows that they are asymptotically equivalent, i.e., they have the same limit distribution under the null hypothesis and under a sequence of local alternatives. All tests are also consistent and have non trivial local asymptotic power against a wide range of alternatives, namely for which the parameters of interest are not constant over the interval specified by Λ . This does not mean, however, that they all have the same behavior in finite samples. Indeed, the simulations of Vogelsang (1999) for the special case of a change in mean, showed the sup LMT test to be seriously affected by the problem of non monotonic power, in the sense that, for a fixed sample size, the power of the test can rapidly decrease to zero as the change in mean increases 1 . This is again because the variance of the errors is estimated under the null hypothesis of no change. Hence, we shall not discuss it any further. In the context of Model (2) with i.i.d. errors, the LR and Wald tests have similar properties, so we shall discuss the Wald test. For a single change, it is defined by (up to a scaling by q): µ ¶ 0 T − 2q − p δ̂ H 0 (H(Z̄ 0 MX Z̄)−1 H 0 )−1 Rδ̂ (9) sup WT (λ1 ; q) = sup k SSRk λ1 ∈Λ λ1 ∈Λ where H is the conventional matrix such that (Hδ)0 = (δ 01 −δ 02 ) and MX = I −X(X 0 X)−1 X 0 . Here SSRk is the sum of squared residuals under the alternative hypothesis, which depends on the break date T1 . One thing that is very useful with the sup WT test is that the break point that maximizes the Wald test is the same as the estimate of the break point, T̂1 ≡ [T λ̂1 ], obtained by minimizing the sum of squared residuals provided the minimization problem (4) is restricted to the set Λ , i.e., sup WT (λ1 ; q) = WT (λ̂1 ; q) λ1 ∈Λ When serial correlation and/or heteroskedasticity in the errors is permitted, things are different since the Wald test must be adjusted to account for this. In this case, it is defined by µ ¶ 1 T − 2q − p 0 0 ∗ δ̂ H (H V̂ (δ̂)H 0 )−1 H δ̂, (10) WT (λ1 ; q) = T k where V̂ (δ̂) is an estimate of the variance covariance matrix of δ̂ that is robust to serial correlation and heteroskedasticity; i.e., a consistent estimate of 1 V (δ̂) = plimT →∞ T (Z̄ 0 MX Z̄)−1 Z̄ 0 MX ΩMX Z̄(Z̄ 0 MX Z̄)−1 (11) Note that what Vogelsang (1998b) actually refers to as the sup Wald test for the static case is actually the sup LM test. For the dynamic case, it does correspond to the Wald test. 26 For example, one could use the method of Andrews (1991) based on weighted sums of autocovariances. Note that it can be constructed allowing identical or different distributions for the regressors and the errors across segments. This is important because if a variance shift occurs at the same time and is not taken into account, inference can be distorted (see, e.g., Pitarakis, 2004). In some instances, the form of the statistic reduces in an interesting way. For example, consider a pure structural change model where the explanatory variables are such that plimT −1 Z̄ 0 ΩZ̄ = hu (0)plimT −1 Z̄ 0 Z̄ with hu (0) the spectral density function of the errors ut evaluated at the zero frequency. In that case, we have the asymptotically equivalent P test (σ̂ 2 /ĥu (0))WT (λ1 ; q), with σ̂ 2 = T −1 Tt=1 û2t and ĥu (0) a consistent estimate of hu (0). Hence, the robust version of the test is simply a scaled version of the original statistic. This is the case, for instance, when testing for a change in mean as in Garcia and Perron (1996). The computation of the robust version of the Wald test (10) can be involved especially if a data dependent method is used to construct the robust asymptotic covariance matrix of δ̂. Since the break fractions are T -consistent even with correlated errors, an asymptotically equivalent version is to first take the supremum of the original Wald test, as in (9), to obtain the break points, i.e. imposing Ω = σ2 I. The robust version of the test is obtained by evaluating (10) and (11) at these estimated break dates, i.e., using WT∗ (λ̂1 ; q) instead of supλ1 ∈Λ WT∗ (λ1 ; q), where λ̂1 is obtained by minimizing the sum of squared residuals over the set Λ . This will be especially helpful in the context of testing for multiple structural changes. 4.3.1 Optimal tests The sup-LR or sup-Wald tests are not optimal, except in a very restrictive sense. Andrews and Ploberger (1994) consider a class of tests that are optimal, in the sense that they maximize a weighted average power. Two types of weights are involved. The first applies to the parameter that is only identified under the alternative. It assigns a weight function J(λ1 ) that can be given the interpretation of a prior distribution over the possible break dates or break fractions. The other is related to how far the alternative value is from the null hypothesis within an asymptotic framework that treats alternative values as being local to the null hypothesis. The dependence of a given statistic on this weight function occurs only through a single scalar parameter c. The higher the value of c, the more distant is the alternative value from the null value, and vice versa. The optimal test is then a weighted function of the standard Wald, LM or LR statistics for all permissible fixed break dates. 27 Using either of the three basic statistics leads to tests that are asymptotically equivalent. Here, we shall proceed with the version based on the Wald test (and comment briefly on the version based on the LM test). The class of optimal statistics is of the following exponential form: ¾ ½ Z 1 c −q/2 WT (λ1 ) dJ(λ1 ) exp Exp-WT (c) = (1 + c) 21+c where we recall that q is the number of parameters that are subject to change, and WT (λ1 ) is the standard Wald test defined in our context as in (9). To implement this test in practice, one needs to specify J(λ1 ) and c. A natural choice for J(λ1 ) is to specify it so that equal weights are given to all break fractions in some trimmed interval [ 1 , 1− 2 ]. For the parameter c, one version sets c = 0 and puts greatest weight on alternatives close to the null value, i.e., on small shifts; the other version specifies c = ∞, in which case greatest weight is put on large changes. This leads to two statistics that have found wide appeal. When c = ∞, the test is of an exponential form, viz. ⎛ ⎞ µ µ ¶¶ T −[T 2 ] X T1 1 ⎠ WT exp Exp-WT (∞) = log ⎝T −1 2 T T1 =[T 1 ]+1 When c = 0, the test takes the form of an average of the Wald tests and is often referred to as the Mean-WT test. It is given by Mean-WT = Exp-WT (0) = T −1 T −[T X2 ] T1 =[T 1 ]+1 WT µ T1 T ¶ The limit distributions of the tests are µZ 1− 2 ¶ ¶ 1 exp Gq (λ1 ) dλ1 2 µ Exp-WT (∞) ⇒ log Z 1− 2 1 Gq (λ1 )dλ1 Mean-WT ⇒ 1 Andrews and Ploberger (1994) presented critical values for both tests for a range of values for symmetric trimmings 1 = 2 , though as stated above they can be used for some non symmetric trimmings as well. Simulations reported in Andrews, Lee and Ploberger (1996) show that the tests perform well in practice. Relative to other tests discussed above, the Mean-WT has highest power for small shifts, though the test Exp-WT (∞) performs better for moderate to large shifts. None of them uniformly dominates the Sup-WT test and they 28 recommend the use of the Exp-WT (∞) form of the test, referred to as the Exp-Wald test below. As mentioned above both tests can equally be implemented (with the same asymptotic critical values) with the LM or LR tests replacing the Wald test. As noted by Andrews and Ploberger (1994), the Mean-LM test is closely related to Gardner’s test (discussed in Section 2). This is because, in the change in mean case, the LM test takes the form of a scaled partial sums. Given the poor properties of this test, especially with respect to large shifts when the power can reach zero, we do not recommend the asymptotically optimal tests based on the LM version. In our context, tests based on the Wald or LR statistics have similar properties. Elliott and Müller (2003) consider optimal tests for a class of models involving nonconstant coefficients which, however, rule out one-time abrupt changes. The optimality criterion relates to changes that are in a local neighborhood of the null values, i.e., for small changes. Their procedure is accordingly akin to locally best invariant tests for random variations in the parameters. The suggested procedure does not explicitly model breaks and the test is then of the ‘function of partial sums type’. It has not been documented if the test suffers from non-monotonic power. They show via simulations, with small breaks, that their test also has power against a one-time change. The simulations can also be interpreted as providing support for the conclusion that the Sup, Mean and Exp tests tailored to a one-time change also have power nearly as good as the optimal test for random variation in the parameter. For optimal tests in a Generalized Method of Moments framework, see Sowell (1996). 4.3.2 Non monotonicity in power The Sup-Wald and Exp-Wald tests have monotonic power when only one break occurs under the alternative. As shown in Vogelsang (1999), the Mean-Wald test can exhibit a nonmonotonic power function, though the problem has not been shown to be severe. All of these, however, suffer from some important power problems when the alternative is one that involves two breaks. Simulations to that effect are presented in Vogelsang (1997) in the context of testing for a shift in trend. This suggests a general principle, which remains, however, just a conjecture at this point. The principle is that any (or most) tests will exhibit non monotonic power functions if the number of breaks present under the alternative hypothesis is greater than the number of breaks explicitly accounted for in the construction of the tests. This suggests that, even though a single break test is consistent against multiple 29 breaks, substantial power gains can result from using tests for multiple structural changes. These are discussed below. 4.4 Tests for multiple structural changes The literature on tests for multiple structural changes is relatively scarce. Andrews, Lee and Ploberger (1996) studied a class of optimal tests. The Avg-W and Exp-W tests remain asymptotically optimal in the sense defined above. The test Exp-WT (c) is optimal in finite samples with fixed regressors and known variance of the residuals. Their simulations, which pertain to a single change, show the test constructed with an estimate of the variance of the residuals to have power close to the known variance case. The problem, however, with these tests in the case of multiple structural changes is practical implementation. The Avg-W and Exp-W tests require the computation of the W -test over all permissible partitions of the sample, hence the number of tests that need to be evaluated is of the order O(T m ), which is already very large with m = 2 and prohibitively large when m > 2. Consider instead the Sup-W test. With i.i.d. errors, maximizing the Wald statistic with respect to admissible break points is equivalent to minimizing the sum of squared residuals when the search is restricted to the same possible partitions of the sample. As discussed in Section 3.3, this maximization problem can be solved with a very efficient algorithm. This is the approach taken by Bai and Perron (1998) (an earlier analysis with two breaks was given in Garcia and Perron, 1996). To this date, no one knows the extent of the power loss, if any, in using the sup-W type test compared with the Avg-W and Exp-W tests. To the author’s knowledge, no simulations have been presented, presumably because of the prohibitive cost of constructing the Avg-W and Exp-W tests. In the context of model (2) with i.i.d. errors, the Wald test for testing the null hypothesis of no change versus the alternative hypothesis of k changes is given by WT (λ1 , ..., λk ; q) = µ T − (k + 1)q − p k ¶ 0 δ̂ H 0 (H(Z̄ 0 MX Z̄)−1 H 0 )−1 H δ̂ SSRk where H now is the matrix such that (Hδ)0 = (δ 01 − δ 02 , ..., δ 0k − δ 0k+1 ) and MX = I − X(X 0 X)−1 X 0 . Here, SSRk is the sum of squared residuals under the alternative hypothesis, which depends on (T1 , ..., Tk ). Note that one can allow different variance across segments when construction SSRk , see Bai and Perron (2003a) for details. The sup-W test is defined by WT (λ1 , ..., λk ; q) = WT (λ̂1 , ..., λ̂k ; q) sup (λ1 ,...,λk )∈Λk, 30 where Λ = {(λ1 , ..., λk ); |λi+1 − λi | ≥ , λ1 ≥ , λk ≤ 1 − } and (λ̂1 , ..., λ̂k ) = (T̂1 /T, ..., T̂k /T ), with (T̂1 , ..., T̂k ) the estimates of the break dates obtained by minimizing the sum of squared residuals by searching over partitions defined by the set Λ . This set dictates the minimal length of a segment. In principle, this minimal length could be different across the sample but then critical values would need to be computed on a case by case basis. When serial correlation and/or heteroskedasticity in the residuals is allowed, the test is µ ¶ 1 T − (k + 1)q − p 0 0 ∗ δ̂ H (H V̂ (δ̂)H 0 )−1 H δ̂, WT (λ1 , ..., λk ; q) = T k with V̂ (δ̂) as defined by (11). Again, the asymptotically equivalent version with the Wald test evaluated at the estimates (λ̂1 , ..., λ̂k ) is used to make the problem tractable. The limit distribution of the tests under the null hypothesis is the same in both cases, namely, def W (λ1 , ..., λk ; q) sup supWT (k; q) ⇒ sup Wk,q = (λ1 ,...,λk )∈Λ with def W (λ1 , ..., λk ; q) = k 1 X [λi Wq (λi+1 ) − λi+1 Wq (λi )]0 [λi Wq (λi+1 ) − λi+1 Wq (λi )] . k i=1 λi λi+1 (λi+1 − λi ) again, assuming non-trending data. Critical values for = 0.05, k ranging from 1 to 9 and for q ranging from 1 to 10, are presented in Bai and Perron (1998). Bai and Perron (2003b) present response surfaces to get critical values, based on simulations for this and the following additional cases (all with q ranging from 1 to 10): = .10 (k = 1, ..., 8), = .15 (k = 1, ..., 5), = .20 (k = 1, 2, 3) and = .25 (k = 1, 2). The full set of tabulated critical values is available on the author’s web page (the same sources also contain critical values for other tests discussed below). The importance of the choice of for the size and power of the test is discussed in Bai and Perron (2003a, 2005). Also discussed in Bai and Perron (2003a) are variations in the exact construction of the test that allow one to impose various restrictions on the nature of the errors and regressors, which can help improve power. 4.4.1 Double maximum tests Often, one may not wish to pre-specify a particular number of breaks to make inference. For such instances, a test of the null hypothesis of no structural break against an unknown 31 number of breaks given some upper bound M can be used. These are called the ‘double maximum tests’. The first is an equal-weight version defined by UD max WT (M, q) = max1≤m≤M WT (λ̂1 , ..., λ̂m ; q), where λ̂j = T̂j /T (j = 1, .., m) are the estimates of the break points obtained using the global minimization of the sum of squared residuals. This UD max test can be given a Bayesian interpretation in which the prior assigns equal weights to the possible number of changes (see, e.g., Andrews, Lee and Ploberger, 1996). The second test applies weights to the individual tests such that the marginal p-values are equal across values of m and is denoted W D max FT (M, q) (see Bai and Perron, 1998, for details). The choice M = 5 should be sufficient for most applications. In any event, the critical values vary little as M is increased beyond 5. Double Maximum tests can play a significant role in testing for structural changes and it are arguably the most useful tests to apply when trying to determine if structural changes are present. While the test for one break is consistent against alternatives involving multiple changes, its power in finite samples can be rather poor. First, there are types of multiple structural changes that are difficult to detect with a test for a single change (for example, two breaks with the first and third regimes the same). Second, as discussed above, tests for a particular number of changes may have non monotonic power when the number of changes is greater than specified. Third, the simulations of Bai and Perron (2005) show that the power of the double maximum tests is almost as high as the best power that can be achieved using the test that accounts for the correct number of breaks. All these elements strongly point to their usefulness. 4.4.2 Sequential tests Bai and Perron (1998) also discuss a test of versus + 1 breaks, which can be used as the basis of a sequential testing procedure. For the model with breaks, the estimated break points denoted by (T̂1 , ..., T̂ ) are obtained by a global minimization of the sum of squared residuals. The strategy proceeds by testing for the presence of an additional break in each of the ( + 1) segments (obtained using the estimated partition T̂1 , ..., T̂ ). The test amounts to the application of ( + 1) tests of the null hypothesis of no structural change versus the alternative hypothesis of a single change. It is applied to each segment containing the observations T̂i−1 + 1 to T̂i (i = 1, ..., + 1). We conclude for a rejection in favor of a model with ( + 1) breaks if the overall minimal value of the sum of squared residuals (over all segments where an additional break is included) is sufficiently smaller than the sum of squared residuals from the breaks model. The break date thus selected is the one associated 32 with this overall minimum. More precisely, the test is defined by: WT ( + 1| ) = {ST (T̂1 , ..., T̂ ) − min inf ST (T̂1 , ..., T̂i−1 , τ , T̂i , ..., T̂ )}/σ̂ 2 , 1≤i≤ +1 τ ∈Λi,η (12) where ST (·) denotes the sum of squared residuals, and Λi, = {τ ; T̂i−1 + (T̂i − T̂i−1 ) ≤ τ ≤ T̂i − (T̂i − T̂i−1 ) }, (13) and σ̂ 2 is a consistent estimate of σ 2 under the null hypothesis and also, preferably, under the alternative. Note that for i = 1, ST (T̂1 , ..., T̂i−1 , τ , T̂i , ..., T̂ ) is understood as ST (τ , T̂1 , ..., T̂ ) and for i = + 1 as ST (T̂1 , ..., T̂ , τ ). It is important to note that one can allow different distributions across segments for the regressors and the errors. The limit distribution of the test is related to the limit distribution of a test for a single change. Bai (1999) considers the same problem of testing for versus + 1 breaks while allowing the breaks to be global minimizers of the sum of squared residuals under both the null and alternative hypotheses. This leads to the likelihood ratio test defined by: sup LRT ( + 1| ) = ST (T̂1 , ..., T̂ ) − ST (T̂1∗ , ..., T̂ ∗+1 ) ST (T̂1∗ , ..., T̂ ∗+1 )/T where {T̂1 , ..., T̂ } and {T̂1∗ , ..., T̂ ∗+1 } are the sets of and +1 breaks obtained by minimizing the sum of squared residuals using and + 1 breaks models, respectively. The limit distribution of the test is different and is given by: sup LRT ( + 1| ) ⇒ max{ξ 1 , ..., ξ where ξ 1 , ..., ξ +1 +1 } are independent random variables with the following distribution q X Bi,j (s) ξ i = sup s(1 − s) ηi ≤s≤1−η i j=1 with Bi,j (s) independent standard Brownian bridges on [0, 1] and η i = /(λ0i − λ0i−1 ). Bai (1999) discusses a method to compute the asymptotic critical values and also extends the results to the case of trending regressors. These tests can form the basis of a sequential testing procedure. One simply needs to apply the tests successively starting from = 0, until a non-rejection occurs. The estimate of the number of breaks thus selected will be consistent provided the significance level used decreases at an appropriate rate. The simulation results of Bai and Perron (2005) show 33 that such an estimate of the number of breaks is much better than those obtained using information criteria as suggested by, among others, Liu et al. (1997) and Yao (1998) (see also, Perron, 1997b). But for the reasons discussed above (concerning the problems with tests that allow a number of breaks smaller than the true value), such a sequential procedure should not be applied mechanically. It is easy to have cases where the procedure stops too early. The recommendation is to first use a double maximum test to ascertain if any break is at all present. The sequential tests can then be used starting at some value greater than 0 to determine the number of breaks. An alternative sequential method is provided by Altissimo and Corradi (2003) for the case of multiple changes in mean. It consists in testing for a single break using the maximum of the absolute value of the partial sums of demeaned data. One then estimate the break date by minimizing the sum of squared residuals and continue the procedure conditional on the break date previously found, until a non-rejection occurs. They derive an appropriate bound to use a critical values for the procedure to yield a strongly consistent estimate of the number of breaks. It is unclear, however, how the procedure can be extended to the more general case with general regressors. 4.5 Tests for restricted structural changes As discussed in Section 3.2, Perron and Qu (2005) consider estimation of structural change models subject to restrictions. Consider testing the null hypothesis of 0 break versus an alternative with k breaks. Recall that the restrictions are Rδ = r. Define 0 δ H 0 (H Ve (e δ)H 0 )− He δ, WT (λ1 , ..., λk ; q) = e (14) δ) is where e δ is the restricted estimate of δ obtained using the partition {λ1 , ..., λk }, and Ve (e an estimate of the variance covariance matrix of e δ that may be constructed to be robust to heteroskedasticity and serial correlation in the errors. As usual, for a matrix A, A− denotes the generalized inverse of A. Such a generalized inverse is needed since, in general, the covariance matrix of e δ will be singular given that restrictions are imposed. Again, instead of using the sup WT (λ1 , ..., λk ; q) statistic where the supremum is taken over all possible partitions in the set Λ , we consider the asymptotically equivalent test that evaluates the e1 , ..., λ ek ; q). Wald test at the restricted estimate, i.e., WT (λ The restrictions can alternatively be parameterized by the relation δ = Sθ + s where S is a q(k + 1) by d matrix, with d the number of basic parameters in the column 34 vector θ, and s is a q(k + 1) vector of constants. Then WT (λ̂1 , ..., λ̂k ; q, S) ⇒ sup W (λ1 , ..., λk ; q, S) |λi −λi−1 |>ε with W (λ1 , ..., λk ; q, S) = W ∗0 S[S 0 (Λ ⊗ Iq )S]−1 S 0 H 0 [HS(S 0 (Λ ⊗ Iq )S 0 )−1 H 0 S 0 ]− HS[S 0 (Λ ⊗ Iq )S]−1 S 0 W ∗ where Λ = diag(λ1 , λ2 − λ1 , ..., 1 − λk ), Iq is the standard identity matrix of dimension q and the q(k + 1) vector W ∗ is defined by W ∗ = [Wq (λ1 ), Wq (λ2 ) − Wq (λ1 ), ..., Wq (1) − Wq (λk )] with Wq (r) a q vector of independent unit Wiener processes. The limit distribution depends on the exact nature of the restrictions so that it is not possible to tabulate critical values that are valid in general. Perron and Qu (2005) discuss a simulation algorithm to compute the relevant critical values given some restrictions. Imposing valid restrictions results in tests with much improved power. 4.6 Tests for structural changes in multivariate systems Bai et al. (1998) considered a sup Wald test for a single change in a multivariate system. Bai (2000) and Qu and Perron (2005) extend the analysis to the context of multiple structural changes. They consider the case where only a subset of the coefficients is allowed to change, whether it be the parameters of the conditional mean, the covariance matrix of the errors, or both. The tests are based on the maximized value of the likelihood ratio over permissible partitions assuming uncorrelated and homoskedastic errors. As above, the tests can be corrected to allow for serial correlation and heteroskedasticity when testing for changes in the parameters of the conditional mean assuming no change in the covariance matrix of the errors. The results are similar to those obtained in Bai and Perron (1998). The limit distributions are identical and depend only on the number of coefficients allowed to change, and the number of times that they are allowed to do so. However, when the tests involve potential changes in the covariance matrix of the errors, the limit distributions are only valid assuming a Normal distribution for these errors. This is because, in this case, the limit distributions of the tests depend on the higher-order moments of the errors’ distribution. Without the 35 assumption of Normality, additional parameters are present which take different forms for different distributions. Hence, testing becomes case specific even in large samples. It is not yet known how assuming Normality affects the size of the tests when it is not valid. An important advantage of the general framework analyzed by Qu and Perron (2005) is that it allows studying changes in the variance of the errors in the presence of simultaneous changes in the parameters of the conditional mean, thereby avoiding inference problem when changes in variance are studied in isolation. Also, it allows for the two types of changes to occur at different dates, thereby avoiding problems related to tests for changes in the paremeters when, for example, a change in variance occurs at some other date (see, e.g., Pitarakis, 2004). Tests using the quasi-likelihood based method of Qu and Perron (2005) are especially important in light of Hansen’s (2000) analysis. First note that, the limit distribution of the Sup, Mean and Exp type tests in a single equation system have the stated limit distribution under the assumption that the regressors and the variance of the errors have distributions that are stable across the sample. For example, the mean of the regressors or the variance of the errors cannot undergo a change at some date. Hansen (2000) shows that when this condition is not satisfied the limit distribution changes and the test can be distorted. His asymptotic results pertaining to the local asymptotic analysis show, however, the sup-Wald test to be little affected in terms of size and power. The finite sample simulations show that if the errors are homoskedastic, the size distortions are quite mild (over and above that applying with i.i.d. regressors, given that he uses a very small sample of T = 50). The distortions are, however, quite severe when a change in variance occurs. But both problems of changes in the distribution of the regressors and the variance of the errors can easily be handled using the framework of Qu and Perron (2005). If a change in the variance of the residuals in a concern, one can perform a test for no change in some parameters of the conditional model allowing for a change in variance since the tests are based on a likelihood ratio approach. If changes in the marginal distribution of some regressors is a concern, one can use a multi-equations system with equations for these regressors. Whether this is preferable to Hansen’s (2000) bootstrap method remains an open question. Note, however, that in the context of multiple changes it is not clear if that method is computationaly feasible, especially for the heteroskedastic case. 36 4.7 Tests valid with I(1) regressors With I(1) regressors, the case of interest is that of a system of cointegrated variables. The goal is then to test whether the cointegrating relationship has changed and to estimate the break dates and form confidence intervals for them. Consider, for simplicity, the following case with an intercept and m I(1) regressors y2t : y1t = a + βy2t + ut (15) where ut is I(0) so that y1t and y2t are cointegrated with cointegrating vector (1, −β). To our knowledge, the only contribution concerning the consistency and limit distribution of the estimates of the break dates is that of Bai et al. (1998). They consider a single break in a multi-equations system and show the estimates obtained by maximizing the likelihood function to be consistent. They also obtain a limit distribution under a shrinking shifts scenario with the shift in the constant a decreasing at rate T b1 for some b1 ∈ (0, 1/2) and the shift in β decreasing at rate T b2 for some b2 ∈ (1/2, 1). Under this scenario the rate of convergence is the same as in the stationary case (since the coefficients on the I(1) variables are assumed to shrink at a faster rate). For testing, an early contribution in this area is Hansen (1992a). He considers tests of the null hypothesis of no change in both coefficients (for an extension to partial changes, see Kuo, 1998, who considers tests for changes in intercept only and tests for changes in all coefficients of the cointegrating vector). The tests considered are the sup and Mean LM tests directed against an alternative of a one time change in the coefficients. He also considers a version of the LM test directed against the alternative that the coefficients are random walk processes denoted Lc . The latter is an extension of Gardner’s (1969) Q-test to the multivariate cointegration context, which is based on the average of the partial sums of the scores and the use of a full sample estimate of the conditional variance of these scores. For related results with respect to LM tests for parameter constancy in cointegrated regressions, see Quintos and Phillips (1993). Gregory et al. (1996) study the finite sample properties of Hansen’s (1992a) tests in the context of a linear quadratic model with costs of adjustments. They show that power can be low when the cost of adjustment is high and suggest a simple transformation of the dependent variable that can increase power. They also consider the behavior of standard residuals based tests of the null hypothesis of no cointegration and show that their power reduces considerably when structural breaks are present in the cointegrating relation. Again, this is simply a manifestation of the fact that unit root tests have little power when the process 37 is stationary around a trend function that changes. Moreover, since Hansen’s (1992a) tests can also be viewed as a test for the null hypothesis of stationarity, in this context it can also be viewed as a test for the null hypothesis of cointegration versus the alternative of no cointegration. Note, however, that the sup and Mean Wald test will also reject when no structural change is present and the system is not cointegrated. Hence, the application of such tests should be interpreted with caution. No test are available for the null hypothesis of no change in the coefficients a and β allowing the errors to be I(0) or I(1). This is because when the errors are I(1), we have a spurious regression and the parameters are not identified. To be able to properly interpret the tests, they should be used in conjunction with tests for the presence or absence of cointegration allowing shifts in the coefficients (see, Section 6). The same comments apply to other tests discussed below. Consider now a cointegrated VAR system written in the following error correction format 0 0 ) of dimension n = m + 1, with yt0 = (y1t , y2t 0 ∆yt = µ + αB yt−1 + p X Γi ∆yt−i + ut (16) i=1 where B (n×r) is the cointegrating matrix and α (n×r) the adjustment matrix (hence, there are r cointegrating vectors). Under the null hypothesis, both are assumed constant, while under the alternative either one or both are assumed to exhibit a one time change at some unknown date T1 . For the case of a triangular system with the restriction that B 0 − [Ir , B ∗ ], Seo (1998) considers the Sup, Mean and Exp versions of the LM test for the following three cases: 1) the constant vector µ is excluded (and the data are assumed non-trending), 2) the constant µ is included but the data are not trending, 3) the constant µ is included and the data are trending. The Sup and Mean LM tests in this cointegrated VAR setup are shown to have a similar asymptotic distribution as the Sup and Mean LM tests of Hansen (1992a) for the case of a change in all coefficients. See also Hao (1996) who also considers the Lc tests for no cointegration allowing for a one time change in intercept at some unknown date using the maximal value overall possible break dates. Hansen and Johansen (1999) also consider a VAR process 2 . Then, the MLE (based on Normal errors) of the cointegrating matrix B are the eigenvectors corresponding to the r largest eigenvalues of the system −1 S01 | = 0 |λS11 − S10 S00 2 A contribution related to multiple structural changes occurring at known dates in the context of cointegrated VAR processes is Hansen (2003), in which case all tests have the usual chi-square distribution. 38 where Sij = T −1 T X Rit Rjt (17) j=1 with R0t (resp., R1t ) the residuals from a regression of ∆yt (resp., yt−1 ) on a constant and lags of ∆yt . Hansen and Johansen (1999) show that instability in α and/or B will manifest themselves in the form of instability eigenvalues’ estimates when evaluated using different samples. They therefore suggest the use of the recursive estimates of λ. Their test takes the form of the fluctuations test of Ploberger et al. (1989) and will have power when either α or B change (see also Quintos, 1997). They also suggest a test that allows the detection of changes in β, an extension of the Lc test of Hansen (1992a) that can be constructed using recursive estimates of λ. Interestingly, Quintos (1997) documents that such tests over-reject the null hypothesis of no structural change when the cointegrating rank is over specified, i.e., when the number of stochastic trends, or unit root components, is under specified. This is the multivariate equivalent of the problem discussed in Section 2, namely that structural change and unit roots can easily be confounded. She proposes a test for the stability of the cointegrating rank. However, when the alternative hypothesis is of a greater rank (less unit roots), the tests will not have power if structural change is present. This is the dilemma faced when trying to assess jointly the correct rank of a cointegrating system and whether structural change is present in the cointegrating vectors. Another contribution, again based on functions of partial sums, is Hao and Inder (1996) who consider the CUSUM test based on OLS residuals from a cointegrating regression. >From this brief review, most tests available are seen to be of the LM type. Given our earlier discussion, these can be expected to have non-monotonic power since they do not explicitly allow for any break. However, no simulation study is available to substantiate this claim and show its relevance in practice. More work is needed in that direction and in considering Wald or LR type tests in a multiple structural changes context. Also, these tests are valid if the cointegrating rank is well specified. As discussed above, a rejection can be due to an over specification of this rank. The problem of jointly determining whether the cointegrating rank is appropriate and whether the system is structurally stable is an important avenue of further research. A potential, yet speculative, approach to determining if the data suggest structural changes in a cointegrating relationship or a spurious regression is the following. Suppose that one is willing to put an upper bound M (say 5) on the possible number of breaks. One can then use a multiple structural change test as discussed in Section 4.4. The reason is that 39 if the system is cointegrated with less than M breaks, the tests can be used to consistently estimate the number of breaks. However, if the regression is spurious, the number of breaks selected will always (in large enough samples) be the maximum number of breaks allowed. The same occurs when an information criterion is used to select the number of breaks (see, Nunes et al., 1996, and Perron, 1997b). Hence, selecting the maximum permissible number of breaks can be symptomatic of the presence of I(1) errors. Of course, more work is needed to turn this argument into a rigorous procedure. 4.8 Tests valid whether the errors are I(1) or I(0) We now consider the issue of testing for structural change when the errors in (2) may have a unit root. In the general case with arbitrary regressors, this question is of little interest. If the regressors are I(0) and the errors I(1), the estimates of the break dates will be inconsistent and so will the tests. This is simply due to the fact that the variability in the errors masks any potential shifts. With I(1) regressors, we have a cointegrated system when the errors are I(0), and a spurious regression with I(1) errors. In general, only the former is of interest. The problem of testing for structural changes in a linear model with errors that are either I(0) or I(1) is, however, of substantial interest when the regression is on a polynomial time trend. The leading case is testing for changes in the mean or slope of a linear trend, a question of substantial interest with economic data. We shall use this example to illustrate the main issues involved. Consider the following structure for some variable yt (t = 1, ..., T ) yt = β 0 + β 1 t + ut (18) where the errors follow a (possibly nonstationary) AR(k) process A(L)ut = et with A(L) = (1 − αL)A∗ (L) and the roots of A∗ (L) all outside the unit circle. If α = 1, the series contains an autoregressive unit root, while if |α| < 1, it is trend-stationary. The Q statistic is defined by #2 " T T X X ûj Q∗1 = ĥe (0)−1 T −2 t=1 j=t+1 P PT −1 where ĥu (0) = m t=τ +1 ût ût−τ , w(m, τ ) is some weight τ −m w(m, τ )R̂u (τ ) with R̂u (τ ) = T function with m/T → 0 (e.g., w(m, τ ) = 1 − |τ |/m if |τ | < m and 0 otherwise) and ût are 40 the least-squares residuals from estimating (18) by OLS. Then if |α| < 1, Z 1 ∗ B1 (r)2 dr Q1 ⇒ (19) 0 where ∙ ¸ ∙ ¸ Z 1 Z 1 B1 (r) = W (r) + 2 W (1) − 3 W (s) ds r − 3 W (1) − 2 W (s) ds r2 0 0 On the other hand, if α = 1, (m/T ) Q∗1 ⇒ Z 1 ∙Z 0 r 0 W1∗ ¸2 Z 1 W1∗ (r)2 dr (s) ds drÁκ Z ¸ ∙Z 1 ¸ Z 1 sW (s) ds +6r W (s) ds − 2 sW (s) ds 0 with W1∗ (r) = W (r)−4 ∙Z 0 R1 1 W (s) ds − (3/2) 0 1 0 0 and κ = −1 K (s) ds where K (τ /m) = ω (m, τ ) (see, Perron, 1991). Hence, the limit distribution is not only different under both the I(1) and I(0) cases, but the scaling needed is different. If one does not have prior knowledge about whether the series is integrated or not, one would need to use the statistic (m/T ) Q∗1 and reject using the critical values in the I(1) case in order to have a test that has asymptotic size no greater than some prespecified level in all cases. But this would entail a test with zero asymptotic size whenever the series is stationary. As suggested by Perron (1991), a solution is to base the test on a regression that parametrically account for the serial correlation in ut , namely yt = β 0 + β 1 t + k X αj yt−j + et (20) j=1 Since the errors are uncorrelated, one uses the statistic #2 " T T X X −2 êj QD1 = σ̂ −2 e T t=k+1 j=t+1 P where σ̂ 2e = T −1 Tt=k+1 ê2t with êt the residuals from estimating (20) by OLS (since the dynamics is taken into account parametrically, there is no need to scale with an estimate of the “long-run” variance). When |α| < 1, result (19) still holds, while when α = 1, we have ¸2 Z 1∙ Z r ∗ B1 (r) + H (1) QD1 ⇒ W1 (s) ds dr 0 0 41 R1 R1 where H (1) = 0 W1∗ (s) dW (s) Á 0 W1∗ (s)2 ds. The limit distributions are different but now the scaling for the convergence is the same. Hence, a conservative procedure is to use the largest of the two sets of critical values, which correspond to those from the limit distribution that applies to the I(1) case. The test is then somewhat asymptotically conservative in the I(0) case but power is still non-trivial. Perron (1991) discusses the power function in details and shows that it is non-monotonic, in that the test has zero power for large shifts (of course, when testing for a shift in level, the test has little power, if any, when the errors are I(1)). A natural extension is to explicitly model breaks and consider a regression of the form (21) yt = β 0 + β 1 t + γ 1 DUt + γ 2 DTt + ut where DUt = 1(t > T1 ) and DTt = 1(t > T1 )(t − T1 ). One can then use any of the tests advocated by Andrews and Ploberger (1996), though they may not be optimal with I(1) errors. These tests, however, also have different rates of convergence under the null hypothesis for I(0) and I(1) errors when based on regression (21). To remedy this problem, one can use a dynamic regression of the form yt = β 0 + β 1 t + γ 1 DUt + γ 2 DTt + k X αj yt−j + et j=1 This is the approach taken by Vogelsang (1997). He considers the Sup, Mean and Exp Wald tests and shows that they have well defined limit distributions under both I(0) and I(1) errors, which are, however, different. Again, at any significance level, the critical values are larger in the I(1) case and these are to be used to ensure tests with an asymptotic size no greater than pre-specified in both cases. Interestingly, Vogelsang’s results show that the Sup and Exp Wald tests have monotonic power functions but that the Mean-Wald test does not, the decrease in power being especially severe in the case of a level shift (this is due to the fact that the Sup and Exp tests assign most weight to the correct date, unlike the Mean test). Banerjee et al. (1992) also consider a Sup Wald test for a change in any one or more coefficients in a regression of yt on {1, t, yt−1 , ∆yt−1 , ..., ∆yt−k } assuming yt to be I(1). Vogelsang (2001) takes a different approach to obtain a statistic that has the same rate of convergence under both the I(0) and I(1) cases (see also Vogelsang 1998a,b). Let WT (λ1 ) be the Wald statistic for testing that γ 1 = γ 2 = 0 in (21). The statistic considered is P SWT (λ1 ) = WT (λ1 )[s2u /(100T −1 s2z )] exp(−bJT (m)) where s2u = T −1 PT t=1 û2t with ût the OLS residuals from regression (21), s2z = T −1 42 PT t=1 v̂t2 where v̂t are the OLS residuals from the following partial sum regression version of (21) ytp = β 0 t + β 1 ((t2 + t)/2) + γ 1 DTt + γ 2 [DTt2 + DTt ]/2 + vt (22) P where ytp = tj=1 yj and JT (m) is a unit root test that has a non-degenerate limit distribution in the I(1) case and converges to 0 in the I(0) case. Consider first the case with I(0) errors. We have WT (λ1 ), s2u and T −1 s2z all Op (1), hence P SWT (λ1 ) = Op (1), which does not depend on b . If the errors are I(1), WT (λ1 ) = Op (T ), T −1 s2u = Op (T ) and T −1 s2z = Op (1), hence P SWT (λ1 ) = Op (1) again. The trick is then to set b at the value which makes the critical values the same in both cases for any prescribed significance level. One can then use the Sup, Mean or Exponential version of the Wald test, though neither of the three have any optimal property in this context (another version based directly on the partial sums regression (22) is also discussed). Perron and Yabu (2005) consider an alternative approach which leads to more powerful tests. Consider the following special case of (21) for illustration yt = β 0 + β 1 t + γ 2 DTt + ut (23) so that the goal is to test for a shift in the slope of the trend function with both segments joined at the time of break. Assume that the errors are generated by an AR(1) process of the form (24) ut = αut−1 + et (an extension to the more general case is also discussed). If α = 1, the errors are I(1) and if |α| < 1, the errors are I(0). Consider the infeasible GLS regression yt∗ = β ∗0 + β 1 t∗ + γ 2 DTt∗ + et where for any variable, a ∗ indicates the quasi-differenced data, e.g., yt∗ = (1 − αL)yt . For a fixed break point T1 , the Wald test would be the best test to use and the limit distribution would be chi-square in both the I(1) and I(0) cases. However, if one used a standard estimate of α to construct a test based on the feasible GLS regression (e.g., α̂ obtained by estimating (24) with ut replaced by ût , the OLS residuals from (23)), the limit distribution would be different in both cases. Perron and Yabu (2005) show, however, that the same chi-square distribution prevails if one replace α̂ by a truncated version given by ⎧ ⎨ α̂ if T δ |α̂ − 1| > d α̂S = ⎩ 1 if T δ |α̂ − 1| ≤ d 43 for some δ ∈ (0, 1) and some d > 0. Theoretical arguments presented in Perron and Yabu (2004) show that δ = 1/2 is the preferred choice. Also, finite sample improvements are possible if one replaces α̂ by a median unbiased estimate (e.g., Andrews, 1993b) or the estimate proposed by Roy and Fuller (2001; see also Roy et al., 2004). When the break date is unknown, the limit distributions of the Sup, Mean or Exp Wald tests are no longer the same for the I(0) and I(1) cases. However, for the Mean version, the asymptotic critical values are very close (for all common significance levels). Hence, with this version, there is no need for an adjustment. Simulations show that for this case, a value d = 2 leads to tests with good finite sample properties and a power function that is close to that which could be obtained using the infeasible GLS regression, unless the value of α is close to but not equal to one. The issue of testing for structural changes in the trend function of a time series without having to take a stand on whether the series is I(1) or I(0) is of substantial practical importance. As discussed above, some useful recent developments have been made. Much remains to be done, however. First, none of the procedures proposed have been shown to have some optimality property. Second, there is still a need to extend the analysis to the multiple structural changes case with unknown break dates. 4.9 Testing for change in persistence A problem involving structural change and the presence of I(0) and I(1) processes relates to the quite recent literature on change in persistence. What is meant, in most cases, by this is that a process can switch at some date from being I(0) to being I(1), or vice versa. This has been an issue of substantial empirical interest, especially concerning inflation rate series (e.g., Barsky, 1987, Burdekin and Siklos, 1999), short-term interest rates (e.g., Mankiw et., 1987), government budget deficits (e.g., Hakkio and Rush, 1991) and real output (e.g., Delong and Summers, 1988). As discussed in Section 3.1, Chong (2001) derived the limit distribution of the estimate of the break date obtained by minimizing the sum of squared residuals from a regression that allows the coefficient on the lagged dependent variable to change at some unknown date. However, he provided no procedure to test whether a change has occurred and in which direction. A discussed in this review, tests for structural change started with statistics based on partial sums of the data (or some appropriate residuals, in general) as in the Q test of Gardner (1969), and tests of the null hypothesis of stationarity versus a unit root process started with the same statistic. Interestingly, we are again back to Gardner (1969) when 44 devising procedures to test for a change in persistence. Kim (2000) and Busetti and Taylor (2001) consider testing the null hypothesis that the series is I(0) throughout the sample versus the alternative that it switches from I(0) to I(1) or vice versa. The statistic used is the ratio of the unscaled Gardner’s (1969) Q test over P the post and pre-break samples. With the partial sums Si,t = tj=i+1 ûj , where ût are the residuals from a regression of the data yt on a constant (non-trending series) or on a constant and time trend (for trending series), it is defined by P (T − T1 )−2 Tt=T1 +1 ST21 ,t (25) ΞT (T1 ) = P 1 2 T1−2 Tt=1 S1,t Under the null hypothesis of I(0) throughout the sample, both the numerator and denominator are Op (1). Consider an alternative with the process being I(0) in the first sample and I(1) is the second, the numerator is then Op (T 2 ) and one rejects for large values. If the alternative is reversed, the denominator is Op (T 2 ) and one rejects for small values. Hence, with a known break date, a two sided test provides a consistent test against both alternatives. For the case with an unknown break date, Kim (2001) considers the sup, Mean or Exp functionals of the sequence ΞT (T1 ) with, as usual, a set specifying a range for permissible values of T1 /T (he suggests [0.2, 0.8]). The test is then consistent for a change from I(0) to I(1) but inconsistent for a change from I(1) to I(0). Busetti and Taylor (2005) note that maximizing the reciprocal of the test ΞT (T1 ) provides a consistent test against the alternative of a change from I(1) to I(0). Hence, their suggestion is to use the maximum of the test based ΞT (T1 ) and its reciprocal (whether the sup, Mean of Exp functional is used). Interestingly, Leybourne and Taylor (2004) suggest scaling both the numerator and denominator of (25) by an estimate of the long-run variance constructed from the respective sub-samples, in which case the test is then exactly the ratio of the Q tests applied to each sub-samples. No version of the test will deliver a consistent estimate of the break date and they suggest using the ratio of the post-break to pre-break sample variances of the residuals ût . They show consistency of the estimate but no limit distribution is obtained, thereby preventing making inference about the break date. Another issue related to this class of tests is the fact that they reject the null hypothesis often when the process is actually I(1) throughout the sample. This is due to the fact that, though the statistic ΞT (T1 ) is Op (1) in this case, the limit distribution is quite different from that prevailing in the constant I(0) case, with quantiles that are greater. Harvey et al. (2004) use the same device suggested by Vogelsang (1998, 2001) to solve the problem by multiplying the test by exp(−bJT ) with JT a unit root test that has a non-degenerate limit 45 distribution in the constant I(1) case and that converges to zero in the constant I(0) case. For a given size of the test, one can then select b so that the critical values are the same in both cases (see Section 4.8). Busetti and Taylor (2005) also consider locally best invariant (LBI) tests. As discussed in this review, this class of tests has important problems (e.g., non monotonic power) and here is no exception. Consider the LBI test for a change from I(0) to I(1), the form of the statistic is then given by: T X −2 2 −2 St,T (26) σ̂ (T − T1 ) t=T1 +1 PT −1 2 where σ̂ 2 = T t=1 ût is the estimate of the variance of the residuals using the full sample. Under the alternative, σ̂ 2 is Op (T ) and, hence, the test is Op (T ) under the alternative. Busetti and Harvey (2005) also consider using the sup, Mean or Exp functionals of the original Q test applied to the post break data only. It is similar to the test (26) but with P a scaling based on σ̂ 21 (T1 ) = (T − T1 )−1 Tt=T1 +1 û2t , an estimate of the variance based on the post-break data. This test has similar properties. In fact using the Q test itself applied to the whole sample would be consistent against a change from I(0) to I(1), showing that this class of tests will reject the null hypothesis with probability one in large samples if the process is I(1) throughout the sample. Both the LBI and the post break Q tests have a scaling that is Op (T ) when the alternative is true whatever break date is used. Consider now P1 2 ût . The test would then have the same instead the statistic (26) scaled by σ̂ 20 (T1 ) = T1−1 Tt=1 limit distribution under the constant I(0) null hypothesis but would be Op (T 2 ) under the alternative and, hence, more powerful. This illustrates once again, a central problem with LBI or LM type tests in the context of structural changes. The scaling factor is evaluated under the null hypothesis, which implies an inflated estimate when the alternative is true and a consequent loss of power. Leybourne et al. (2003) consider instead the null hypothesis that the process is I(1) throughout the sample, with the same alternatives that it can switch from I(1) to I(0), or vice versa. Their test for a change from I(0) to I(1) is based on the minimal value of the unit root test ADF GLS (T1 ), the ADF test proposed by Elliott et al. (1996) constructed using observations up to time T1 (labelled recursive test). Since this test does not use all information in the data for any given particular break date, they also consider using a similar unit root test from a full sample regression in which the coefficient on the lagged level is constrained to be zero in the post-break sample (labelled sequential). To test against the alternative hypothesis of a change from I(1) to I(0), the same procedures are applied to the 46 data arranged in reverse order. When the direction of the change is unknown, they consider the minimal value of the pair of statistics for each case. These tests will, however, reject when the process has no change and is I(0) throughout the sample. To remedy this problem, Leybourne et al. (2003) consider an alternative procedure when the null hypothesis is I(1) throughout the sample. It is the ratio of the minimal value of the pre-break sample variance of the residuals constructed from the original series relative to the minimal value of the same statistic constructed using time reversed data. The test has a well defined limit distribution under the null hypothesis of constant I(1), rejects when there is a shift and has a limit value of 1 when the process is I(0) throughout, which implies a non-rejection asymptotically in the latter case. Kurozumi (2004) considers a test constructed upon the LM principle. He shows that the test is asymptotically equivalent to the sum of the t-statistics on α1 and α2 in the regression yt−1 + α2 1(t ≤ T1 )e yt−1 + ∆e yt = α1 1(t ≤ T1 )e k X ci ∆e yt−i + et i=1 where yet are OLS detrended data (he also considers a version with GLS detrended data but his simulations show no power improvement). This test performs rather poorly and he recommends using a regression with a fitted mean that is allowed to change at the break date, even though this results in a test with lower local asymptotic power. With an unknown break date, one takes the minimal value of the tests over the range of permissible break dates. Deng and Perron (2005b) take the null hypothesis to be I(1) throughout and they follow the approach suggested by Elliott et al. (1996) in specifying a the null hypothesis as involving an autoregressive parameter (in the I(0) subsample) that is local to unity. They derive the Gaussian local power envelop and a feasible test that achieves this power envelop. It is shown that the test has higher power than those of Leybourne et al. (2003) and Kurozumi (2004), according to both the local power function and to the finite sample power (via simulations). But they also find a curious feature. The test is consistent when only a constant is included but inconsistent when a constant and a time trend are included. This is really a theoretical artifact that has little impact on finite sample power but is interesting nevertheless. Under the null hypothesis, the support of the test is the positive real line and one reject for small values. When a fitted trend is included, the limit distribution of the test is exactly zero. Most of the mass it at zero but there is a very small tail to the right, so that the probability of rejecting does not go to one for all possible significance levels. 47 5 Unit Root Versus Trend Stationarity in the Presence of Structural Change in the Trend Function As discussed throughout this review, structural changes and unit root non-stationarity share similar features in the sense that most tests for structural changes will reject in the presence of a unit root in the errors, and vice versa, tests of stationarity versus unit root will reject in the presence of structural changes. We now discuss methods to test the null hypothesis of a unit root in the presence of structural changes in the trend function. 5.1 The motivation, issues and framework To motivate the problem addressed, it is useful to step back and look at some basic properties of unit root and trend-stationary processes. Consider a trending series generated by (27) yt = µ + βt + ut where ∆ut = C(L)et (28) P P ∞ j with et ∼ i.i.d. (0, σ 2e ) and C(L) = ∞ j=0 cj L such that j=1 j|cj | < ∞ and c0 = 1. A popular trend-cycle decomposition is that suggested by Beveridge and Nelson (1981). The trend is specified as the long run forecast of the series conditional on current information, which results in the following τ t = µ + βt + C(1) t X ej j=1 P∞ e e while the cycle is given by ct = C(L)e cj Lj where e cj = Σ∞ t with C(L) = i=j+1 ci . Here the j=0 e trend has two components, a deterministic one (a linear trend) and a stochastic one specified by a random walk weighted by C(1). Hence, the trend exhibits changes every period in the form of level shifts. Note that if one considered a process which is potentially integrated of order 2, the trend would exhibit changes in both level and slope every period. When the process has no unit root, C(1) = 0 and the trend is a linear deterministic function of time. Within this framework, one can view the unit root versus trend-stationary problem as addressing the following question: do the data support the view that the trend is changing every period or never? The empirical analysis of Nelson and Plosser (1982) provided strong evidence that, if the comparison is restricted to these polar cases, the data support the view that a trend which ‘always’ changes is a better description than a trend that ‘never’ 48 changes (using a variety of US macroeconomic variables, furthermore many other studies have reached similar conclusions for other series and other countries). The question is then why restrict the comparison to ‘never’ or ‘always’ ? Would it not be preferable to make a comparison between ‘always’ and ‘sometimes’ ? Ideally, then, the proper question to ask would be ‘what is the frequency of permanent shocks?’. This is a question for which no satisfactory framework has been provided and, as such, it still remains a very important item for further research. The basic motivation for the work initiated by Perron (1989, 1990) is to take a stand on what is ‘sometimes’ (see also Rappoport and Reichlin, 1989). The specific number chosen then becomes case-specific. His argument was that in many cases of interest, especially with historical macroeconomic time series (including those analyzed by Nelson and Plosser, 1982), the relevant number of changes is relatively small, in many cases only one. These changes are then associated with important historical events: the Great Crash (US and Canada, 1929, change in level), the oil price shock (G7 countries, 1973, change in slope); World War II (European countries, change in level and slope), World War I (United Kingdom, 1917, change in level), and so on. As far as statistical modelling is concerned, the main conceptual issue is to view such changes as possibly stochastic but of a different nature than shocks that occur every period, i.e., drawn from a different distribution. However, the argument that such large changes are infrequent makes it difficult to specify and estimate a probability distribution for them. The approach is then to model these infrequent large changes in the trend as structural changes. The question asked by unit root tests is then: ‘do the data favor a view that the trend is ‘always’ changing or is changing at most occasionally?’ or ‘if allowance is made for the possibility of some few large permanent changes in the trend function, is a unit root present in the structure of the stochastic component?’. Note that two important qualifications need to be made. First, the setup allows but does not impose such large changes. Second, by “permanent” what should be understood is not that it will last forever but that, given a sample of data, the change is still in effect. For instance, the decrease in the slope of the trend function after 1973 for US real GDP is still in effect (see, Perron and Wada, 2005). When allowance is made for a one-time change in the trend function, Perron (1989, 1990) specified two versions of four different structures: 1) a change in level for a non-trending series; and for trending series, 2) a change in level, 3) a change in slope, and 4) a change in both level and slope. For each of the four cases, two different versions allow for different transition effects. Following the terminology in Box and Tiao (1975), the first is labelled 49 the “additive outlier model” and specifies that the change to the new trend function occurs instantaneously. The second is labelled the “innovational outlier model” and specifies that the change to the new trend function is gradual. Of course, in principle, there is an infinity of ways to model gradual changes following the occurrence of a “big shock”. One way out of this difficulty is to suppose that the variables respond to the “big shocks” the same way as they respond to so-called “regular shocks” (shocks associated with the stationary noise component of the series). This is the approach taken in the modelization of the “innovational outlier model”, following the treatment of intervention analyses in Box and Tiao (1975). The distinction between the additive and innovational outlier models is important not only because the assumed transition paths are different but also because the statistical procedures to test for unit roots are different. The additive outlier models for each of the four specifications for the types of changes occurring at a break date T1 are specified as follows: Model (AO-0) Model (AO-A) Model (AO-B) Model (AO-C) yt = µ1 + (µ2 − µ1 ) DUt + ut yt = µ1 + βt + (µ2 − µ1 ) DUt + ut yt = µ1 + β 1 t + (β 2 − β 1 ) DTt∗ + ut yt = µ1 + β 1 t + (µ2 − µ1 ) DUt + (β 2 − β 1 ) DTt∗ + ut where DUt = 1, DTt∗ = t−T1 if t > T1 and 0 otherwise, and ut is specified by (28). Under the null hypothesis C(1) 6= 0, while under the alternative hypothesis, C(1) = 0. Alternatively, one can define the autoregressive polynomial A(L) = (1 − L)C(L)−1 . The null hypothesis then specifies that a root of the autoregressive polynomial is one, i.e., that we can write A(L) = (1 − L)A∗ (L) where all the roots of A∗ (L) are outside the unit circle. Under the alternative hypothesis of stationary fluctuations around the trend function, all the roots of A(L) are strictly outside the unit circle. Model (AO-B) was found to be useful for the analysis of postwar quarterly real GNP for the G7 countries and Model (AO-0) for some exchange rate series as well as the US real interest rate, among others. It is important to note that changes in the trend function are allowed to occur under both the null and alternative hypotheses. The innovational outlier models are easier to characterize by describing them separately under the null and alternative hypotheses. Note also that the innovational outlier versions have been considered only for Models (A) and (C) in the case of trending series. The basic reason is that the innovational outlier version of Model (B) does not lend itself easily to 50 empirical applications using linear estimation methods. Under the null hypothesis, we have: Model (IO-0-UR) yt = yt−1 + C (L) (et + δD (T1 )t ) Model (IO-A-UR) yt = yt−1 + b + C (L) (et + δD (T1 )t ) Model (IO-C-UR) yt = yt−1 + b + C (L) (et + δD (T1 )t + ηDUt ) where D (T1 )t = 1 if t = T1 + 1 and 0 otherwise. Under this specification, the immediate impact of the change in the intercept is δ while the long run impact is C (1) δ. Similarly, under Model (IO-C), the immediate impact of the change in slope is η while the long run impact is C (1) η. Under the alternative hypothesis of stationary fluctuations, the specifications are: Model (IO-0-TS) yt = µ + C (L)∗ (et + θDUt ) Model (IO-A-TS) yt = µ + βt + C (L)∗ (et + θDUt ) Model (IO-C-TS) yt = µ + βt + C (L)∗ (et + θDUt + γDTt∗ ) where C(L)∗ = (1 − L)−1 C(L). The immediate impact of the change in the intercept of the trend function is θ while the long run impact is C (1)∗ θ, and the immediate impact of the change in slope is γ while the long run impact is C (1)∗ γ. 5.2 The effect of structural change in trend on standard unit root tests A standard unit root test used in applied research is the so-called augmented Dickey-Fuller (1979) test, which is based on the t-statistic for testing that α = 1 in the following regression yt = µ + βt + αyt−1 + k X ci ∆yt−i + et i=1 with the trend regressor excluded when dealing with non-trending series. A central message of the work by Perron (1989, 1990) is that, when the true process involves structural changes in the trend function, the power of such unit root tests can dramatically be reduced. In particular, it was shown that if a level shift is present, the estimate of the autoregressive coefficient (α when k = 0) is asymptotically biased towards 1. If a change in slope is present, its limit value is 1. It was shown that this translates into substantial power losses. Simulations presented in Perron (1994) show the power reduction to increase as k is increased (see also the theoretical analysis of Montañés and Reyes, 2000, who also show that the power problem remains with the Phillips-Perron (1988) type unit root test). For a more precise and complete theoretical analysis, see Montañés and Reyes (1998, 1999). Under 51 the null hypothesis, the large sample distribution is unaffected by the presence of a level shift (Montañés and Reyes, 1999) and the test is asymptotically conservative in the presence of a change in slope. It can, however, have a liberal size if the break occurs very early in the sample (λ1 < .15) as documented by Leybourne et al. (1998) and Leybourne and Newbold (2000). Intuitively, the latter result can be understood by thinking about the early observations as outliers such that the series reverts back to the mean in effect for the rest of the sample. The latter problem is, however, specific to the Dickey-Fuller (1979) type unit root test, which is based on the conditional likelihood function, discarding the first observations (see, Lee, 2000) 3 . It has also been documented that the presence of stuctural breaks in trend affects tests of the null hypothesis of stationarity (e.g., the Q or KPSS test) by inducing size distortions towards rejecting the null hypothesis too often (e.g., Lee et al., 1997). This is consistent with the effect on unit root tests in the sense that when trying to distinguish the two hypotheses, the presence of structural changes induces a bias in favor of the unit root representation. It is important to discuss these results in relation to the proper way to specify alternative unit root tests. The main result is that large enough changes in level and/or slope will induce a reduction in the power of standard unit root tests. Small shifts, especially in level, are likely to reduce power only slightly. Hence, what is important is to account for the large shifts, not all of them if the others are small. Consider analyzing the US real GDP over, say, the period 1900-1980. Within this sample, one can identify two shifts related to the 1929 crash (change in level) and the post 1973 productivity slowdown (change in slope). However the post 73 sample would, here, consists of only a small proportion of the total sample and the shift in slope in this period is unlikely to induce a bias and need not be accounted for. Hence, the testing strategy discussed below need not make a statement about the precise number of changes. It should rather be viewed as a device to remove biases induced by shifts large enough to cause an important reduction in power. 5.3 Testing for a unit root allowing for changes at known dates. The IO models under the null and alternative hypotheses can be nested in the a way which specifies the regression from which the statistics will be constructed as follows: yt = µ + θDUt + βt + γDTt∗ + δD (T1 )t + αyt−1 + k X ci ∆yt−i + et (29) i=1 3 Kim et al, 2004, study what happens when the trend regressor is absent and the series has a broken trend with the coefficients on the trend and shift in slope shrinking to zero as the sample size increases. 52 for a value of the truncation lag parameter k chosen to be large enough as to provide a good approximation (for methods on how to choose k, see Ng and Perron, 1995, 2001). For Model (IO-0), the regressors (t, DTt∗ ) are not present, while for Model (IO-A), the regressor (DTt∗ ) is not present. The null hypothesis imposes the following restrictions on the coefficients. For Model (IO—0), these are α = 1, θ = µ = 0 and, in general, δ 6= 0 (if there is a change in the intercept). For Model (IO-A), the restrictions are α = 1, β = θ = 0 and again, in general, δ 6= 0, while for Model (IO-C), α = 1, β = γ = 0. Under the alternative hypothesis, we have the following specifications: |α| < 1 and, in general, δ = 0. These restrictions are, however, not imposed by most testing procedures. The test statistic used is the t-statistic for testing the null hypothesis that α = 1 versus the alternative hypothesis that |α| < 1, denoted tα (λ1 ) with λ1 = T1 /T . It is important to note that, provided the specified break date corresponds to the true break date, the statistic is invariant to the parameters of the trend-function, including those related to the changes in level and slope (for an analysis of the case when the break date is mis-specified, see Hecq and Urbain, 1993, Montañés, 1997, Montañés and Olloqui, 1999, and Montañés et al., 2005, who also consider the effect of choosing the wrong specification for the type of break). The limit distribution of the test under the null hypothesis is R1 ∗ W (r, λ1 )dW (r) (30) tα (λ1 ) ⇒ hR0 i1/2 1 ∗ (r, λ )2 dr W 1 0 where W ∗ (r, λ1 ) is the residual function from a projection of a Wiener process W (r) on the relevant continuous time versions of the deterministic components ({1, 1(r > λ1 )} for Model (IO-0), {1, 1(r > λ1 ), r} for Model (IO-A) and {1, 1(r > λ1 ), r, 1(r > λ1 )(r − λ1 )} for Model (IO-C)). Tabulated critical values can be found in Perron (1989, 1990). See also Carrion i Silvestre et al. (1999). For the additive outlier models, the procedures are different and consist of a two-step approach. In the first step, the trend function of the series is estimated and removed from the original series via the following regressions for Model (AO-0) to (AO-C), respectively: yt = µ + γDUt + yet yt = µ + βt + γDUt + yet yt = µ + βt + γDTt∗ + yet yt = µ + βt + θDUt + γDTt∗ + yet 53 where yet is accordingly defined as the detrended series. The next step differs according to whether or not the first step involves DUt , the dummy associated with a change in intercept. For Models (AO-0), (AO-A) and (AO-C), the test is based on the value of the t-statistic for testing that α = 1 in the following autoregression: yt−1 + yet = αe k X dj D (T1 )t−j + j=0 k X ai ∆e yt−i + et i=1 Details about the need to introduce the current value and lags of the dummies D (Tb )t can be found in Perron and Vogelsang (1992b). The limit distributions of the tests are then the same as for the IO case. There is no need to introduce the dummies in the second step regression for Model (AO-B) where no change in level is involved and the two segments of the trend are joined at the time of break. The limit distribution is, however, different; see Perron and Vogelsang (1993a, 1993b). Again, in all cases, the tests are invariant to the change in level or slope provided the break date is correctly specified. These unit root tests with known break dates have been extended in the following directions. Kunitomo and Sato (1995) derive the limit distribution of the likelihood ratio tests for multiple structural changes in the AO case. Amsler and Lee (1995) consider a LM type test in the context of a shift in level of the AO type. Saikkonen and Lütkepohl (2001) also consider cases with a level shift of the AO type, though they allow for general forms of shifts which can be indexed by some unknown parameter to be estimated. Following Elliott et al. (1996), they propose a GLS-type detrending procedure, which is however based on an AR(p) process for the noise. On the basis of simulation results, they recommend using GLS detrending under the null hypothesis instead of a local alternative as done in Elliott et al. (1996). Lanne et al. (2002) propose a finite sample modification which is akin to a pre-whitening device. Let the detrended series be e e−γ eDUt − βt ytGLS = yt − µ and the estimate of the autoregressive polynomial of the first difference ∆ut be eb(L) (all estimates being obtained from the GLS procedure). With the filtered series defined as w et = eb(L)ytGLS , the test is then the t-statistic for testing that α = 1 in the regression w et = µ + αw et−1 + πeb(L)D (T1 )t + k X GLS ai ∆yt−i + et . i=1 Note that the limit distribution does not depend on the break date. This is because the data are detrended using a GLS approach under the null hypothesis of a unit root (or more 54 generally under a sequence of alternatives that are local to a unit root) and the level shift regressor is, in the terminology of Elliott et al. (1996), a slowly evolving trend, in which case, the limit distribution is the same as it would be if it was excluded (loosely speaking, the level shift becomes a one-time dummy). Hence, the limit distribution of the test is the same as that of Elliott et al. (1996) for their unit root test when only a constant is included as deterministic regressor. Lanne and Lütkepohl (2002) show that this test has better size and power than the test proposed in Perron (1990) and the LM test of Amsler and Lee (1995). A similar procedure for level shifts of the IO type is presented in Lütkepohl, Müller and Saikkonen (2001). 5.4 Testing for a unit root allowing for changes at unknown dates The methodology adopted by Perron (1989, 1990) was criticized by, among others, Christiano (1992), on the ground that using a framework whereby the break is treated as fixed is inappropriate. The argument is that the choice of the break date is inevitably linked to the historical record and, hence, involves an element of data-mining. He showed that if one did a systematic search for a break when the series is actually a unit root process without break, using fixed break critical value would entail a test with substantial size distortions. While the argument is correct, it is difficult to quantify the extent of the ‘data-mining’ problem in Perron’s (1989) study. Indeed, no systematic search was done, the break dates were selected as obvious candidates (the Crash of 1929 and the productivity slowdown after 1973) and the same break date was used for all series. Given the intractability of correctly assessing the right p-values for the tests reported, the ensuing literature addressed the problem by adopting a completely agnostic approach where a complete and systematic search was done. While this leads to tests with the correct asymptotic size (under some conditions to be discussed), it obviously implies a reduction in power. We shall return to the practical importance of this point. An avenue taken by Banerjee et al. (1992) was to consider rolling and recursive tests. Both perform standard unit root tests without breaks, the former using a sample of fixed length (much smaller than the full sample) that moves sequentially from some starting date to the end of the sample. The latter considers a fixed starting date for all tests and increases the sample used (from some minimal value to the full sample). In each case, one then considers the minimal value of the unit root test and rejects the null hypothesis of a unit root if this minimal value is small enough. Asymptotically, such procedures will correctly reject the null hypothesis if the alternative is true but the fact that all tests are based on 55 sub-samples means that not all information in the data is used and consequently one can expect a loss of power. An alternative strategy, more closely related to the methodology of Perron (1989) was adopted by Zivot and Andrews (1992) as well as Banerjee et al. (1992). They consider the IO type specification and a slightly different regression that does not involve the one-time dummy when a shift in level is allowed under the alternative hypothesis. For example, for Model C, the regression is yt = µ + θDUt + βt + γDTt∗ + αyt−1 + k X ci ∆yt−i + et (31) i=1 and the test considered is the minimal value of the t-statistic for testing that α = 1 over all possible break dates in some pre-specified range for the break fraction [ , 1 − ] where a popular choice for is 0.15. Denote the resulting test by t∗α = inf λ1 ∈[ ,1− ] tα (λ1 ) where tα (λ1 ) is the t-statistic for testing α = 1 in (31) when the break date T1 = [T λ1 ] is used. The limit distribution of the test is R1 ∗ W (r, λ1 )dW (r) ∗ (32) tα ⇒ inf hR0 i1/2 λ1 ∈[ ,1− ] 1 ∗ 2 W (r, λ1 ) dr 0 with W ∗ (r, λ1 ) as defined in (30). Perron (1997a) extended their theoretical results by showing, using projection arguments, that trimming for the possible values of λ1 was unnecessary and that one could minimize over all possible break dates 4 . For the Nelson-Plosser (1982) data set, Zivot and Andrews (1992) reported fewer rejections compared to what was reported in Perron (1989) using a known break date assumption. These rejections should be viewed as providing stronger evidence against the unit root but a failure to reject does not imply a reversal of Perron’s (1989) conclusions. This is a commonly found mis-conception in the literature, which overlooks the fact that a failure to reject may simply be due to tests with low power. Zivot and Andrews’ (1992) extension involves, however, a substantial methodological difference. The null hypothesis considered is that of a unit root process with no break while the alternative hypothesis is a stationary process with a break. Hence, there is an asymmetric treatment of the specification of the trend under the null and alternative hypotheses. In particular, limit result (32) is not valid if a break is present under the null hypothesis. 4 Perron (1997a) also showed how the weak convergence result could be obtained using the usual sup metric instead of the hybrid metric adopted in Zivot and Andrews (1992). 56 Vogelsang and Perron (1998) show that, in this case, t∗α diverges to −∞ when a shift in slope is present. This implies that a rejection can be due to the presence of a unit root process with a breaking trend. The reason for this is the following. With a fixed break date, the statistic tα (λ1 ) from regression (29) is invariant to the values of the parameters of the trend function under both the null and alternative hypotheses. When searching over a range of values for the break date (only one of which corresponding to the true value), this invariance no longer holds. In the case of Model A with only a level shift and for non-trending series with a change in mean considered by Perron and Vogelsang (1992a), the statistic t∗α is asymptotically invariant to the value of the level shift but not in finite samples. Simulations reported by Perron and Vogelsang (1992b) show size distortions that increase with the magnitude of the level shift. They argue, however, that substantial size distortions are in effect only when implausibly large shifts occur and that the problem is not important in practice. Vogelsang and Perron (1998) make the same arguments for the case of a shift in slope. Even though in practice the distortions may be small, it nevertheless remains a problematic feature of this approach and we consider recent attempts below which do not have this problem. Perron and Vogelsang (1992a), for the non-trending case, and Perron (1997a), for the trending case, extend the analysis of Zivot and Andrews (1992). They consider tests for both the IO and AO cases based on the minimal value of the t-statistic for testing that α = 1, and also tests based on tα (λ1 ) with T1 selected by maximizing the absolute value of the t-statistic on the coefficient of the appropriate shift dummy, DUt if only a level shift is present and DTt∗ if a slope change is present (see also Christiano, 1992, and Banerjee et al., 1992). For the IO case, they also suggest using regression (29) instead of (31) which includes the one-time dummy D (Tb )t since that would be the right regression to use with a known break date. They derive the limit distribution under the null hypothesis of a unit root and no break (in which case it does not matter if the one-time dummy D (Tb )t is incorporated). Perron (1997a) also considers tests where the break date is selected by minimizing or maximizing the value of the t-statistic on the slope dummy, which allows one to impose a priori the restriction of a direction for the change in slope and provides a more powerful test. Carrion-i-Silvestre et al. (2004) consider statistics which jointly test the null hypothesis and the zero value of appropriate deterministic regressors, extending the likelihood ratio test of Dickey and Fuller (1981). 57 5.4.1 Extensions and other approaches We now briefly review some extensions and alternative approaches and return below to an assessment of the various methods discussed above. Unless stated otherwise, all work described below specifies the null hypothesis as a unit root process with no break in trend. Perron and Rodríguez (2003) consider tests for trending series with a shift in slope in the AO framework. Following Elliott et al. (1996), they derive the asymptotic local power envelop and show that using GLS detrended series (based on a local alternative) yields tests with power close to the envelop. For the non-trending case, Clemente, Montañés and Reyes (1998) extend the results of Perron and Vogelsang (1992a) to the case with two breaks. A similar extension is provided by Lumsdaine and Papell (1997) for the case of trending series. Generalizations to multiple breaks include the following. Ohara (1999) extends the Zivot and Andrews (1992) approach to the general case with m breaks, though only critical values for the two break case are presented. Ohara (1999) also proves an interesting generalization of a result in Perron (1989) to the effect that, if a unit root test allowing for m1 changes in slope is performed on a series having m0 changes with m0 > m1 , then the least-squares estimate of α converges to one. This provides theoretical support for Rule 6 stated in Campbell and Perron (1991), which states that ‘a non-rejection of the unit root hypothesis may be due to misspecification of the deterministic components included as regressors’. Kapetanios (2005) also deals with the multiple break case but considers the following strategy, based on the sequential method of Bai (1997b) and Bai and Perron (1998) (see section 3.5). First, denote the set of t-statistics for a unit root over all possible one break partitions by τ 1 . Choose the break date that minimizes the sum of squared residuals. Then impose that break and insert an additional break over all permissible values (given some imposed trimming) and store the associated unit root tests in the set τ 2 , then choose the additional break that minimizes the sum of squared residuals. Continue in this fashion until an m break model is fitted and m sets of unit root tests are obtained. The unit root test selected is then the one that is minimal over all m sets. The limit distribution is, however, not derived, and the critical values are obtained through simulation with T = 250. Saikonnen and Lütkepohl (2002) extend their tests for a level shift with a known break date (of a general form possibly indexed by some unknown parameter) to the case of a shift occurring at an unknown date. It can be performed in both the AO and IO frameworks and the resulting procedure is basically the same as discussed in Section 5.3 for the known break date case. This is because, with a GLS detrending procedure based on a specification that is local to a unit root, the limit distribution of the test is the same whatever the break point 58 is selected to be. Hence, one can substitute any estimate of the break date without affecting the limit null distribution of the test. They recommend using a unit root specification for the detrending (as opposed to using a local alternative as in Elliott et al., 1996) since it leads to tests with finite sample sizes that are robust to departures of the estimate of the break date from its true value. Of course, power is highly sensitive to an incorrectly estimated break date. Lanne et al. (2003) assess the properties of the tests when different estimates of the break date are used. A substantial drawback of their approach is that they found the test to have non-monotonic power, in the sense that the larger the shift in level the lower the power in rejecting the unit root. Also, the power is sensitive to departures from the exact specification for the type of change, and power can be reduced substantially if allowance is made for a general shift indexed by some parameter when the shift is actually an abrupt one. Consider now testing the null hypothesis of stationarity. Tests of the type proposed by Kwiatkowski et al. (1992) will reject the null hypothesis with probability one in large enough samples if the process is affected by structural changes in mean and/or slope but is otherwise stationary within regimes. This follows in an obvious way once one realizes that the KPSS test is also a consistent test for structural change (see, nevertheless, simulations in Lee et al., 1997). In order not to incorrectly reject the null hypothesis of stationarity, modifications are therefore necessary. Kurozumi (2002), Lee and Strazicich (2001b) and Busetti and Harvey (2001, 2003) consider testing the null hypothesis of stationarity versus the alternative of a unit root in the presence of a single break for the specifications described above (see also, Harvey and Mills, 2003). Their test is an extension of the Q-statistic of Gardner (1969), or equivalently the KPSS test as discussed in Section 2. The test is constructed using leastsquares residuals from a regression incorporating the appropriate dummy variables. They provide critical values for the known break date case. When the break is unknown, things are less satisfactory. To ensure the consistency of the test, Lee and Strazicich (2001b) and Busetti and Harvey (2001, 2003) consider the minimal value (as opposed to the maximal value) of the statistics over all permissible break dates. Since the test rejects for large values, this implies the need to resort to the value of the statistic at the break point that permits the least-favorable outcome against the alternative. Hence, it results in a procedure with low power. Kurozumi (2002) as well as Busetti and Harvey (2001, 2003) also consider using the estimate of the break date that minimizes the sum of squared residuals from the relevant regression under the null hypothesis. Since, the estimate of the break fraction is then consistent, one can use critical values corresponding to the known break date case. They 59 show, however, that the need to estimate the break date induces substantial power losses. Busetti (2002) extended this approach to a multivariate setting, where the null hypothesis is that a set of series all share a common trend subject to a change and a stationary noise function, the alternative being that one or more series have unit root noise components. Also of related interest is the study by Kim et al. (2002) who study unit root tests with a break in innovation variance following the work by Hamori and Tokihisa (1997). The issue of unit roots and trend breaks has also been addressed using a Bayesian framework with results that are generally in agreement with those of Perron (1989), see Zivot and Phillips (1994), Wang and Zivot (2000) and Marriott and Newbold (2000). 5.4.2 Problems and recent proposals Theoretical results by Vogelsang and Perron (1998) and simulation results reported in Perron and Vogelsang (1992a), Lee and Strazicich (2001a), Harvey et al. (2001) and Nunes et al. (1997) yield the following conclusions about the tests when a break is present under the null hypothesis. For the IO case when a slope shift is present, both versions using the break date by minimizing the unit root test or maximizing the absolute value of the t-statistic on the coefficient of the slope dummy, yield tests with similar features, namely an asymptotic size 100%. In the presence of a level shift, the asymptotic size is correct but liberal distortions occur when the level shift is large. When the one time dummy D (T1 )t is included in the regression, the source of the problem is that the break point selected with highest probability (which increases as the magnitude of the break increases) is T10 −1, i.e., one period before the true break; and it is for this choice of the break date that the tests have most size distortions. Lee and Strazicich (2001a) show that the problem is the same as if the one time dummy D (T1 )t was excluded when considering the known break date case. Their result also implies that, when unit root tests are performed using a regression of the form (31) without the one time dummy D (T1 )t , the correct break date is selected but the tests are still affected by size distortions (which was also documented by simulations). In cases with only a level shift, Harvey et al. (2001) suggest evaluating the unit root t-statistic at the break date selected by maximizing the absolute value of the t-statistic on the coefficient of the level shift plus one, and show that the tests then have correct size even for large breaks. For the AO type models, the following features apply. When the break date is selected by minimizing the unit root test, similar size distortions apply. However, when the break date is selected by maximizing the absolute value of the t-statistic on the relevant shift dummy, the tests have the correct size even for large breaks, and the correct break date is selected in 60 large samples. Vogelsang and Perron (1998) argue that the limit distribution of the unit root tests is then that corresponding to the known break date case. They suggest, nevertheless, to use the asymptotic critical values corresponding to the no break case since this leads to a test having asymptotic size no greater than that specified for all magnitudes of the break, even though this implies a conservative procedure when a break is present. An alternative testing procedure, which naturally follows from the structural change literature reviewed in Section 3, is to evaluate the unit root test at the break date selected by minimizing the sum of squared residuals from the appropriate regression. Interesting simulations pertaining to the IO case are presented in Lee and Strazicich (2001). They show that if one uses the usual asymptotic critical values that apply for the no break case under the null hypothesis, the tests are conservative when a break is present (provided the one time dummy D (T1 )t is included in the regression). They correctly note, however, that the limit null distribution when no break is present depends on the limit distribution of the estimated break date which may depend on nuisance parameters. Hatanaka and Yamada (1999) present useful theoretical results for the IO regression (though they specify the data generating process to be of the AO type). They show that, when a change in slope is present, the estimate of the break fraction λ1 , obtained by minimizing the sum of squared residuals, is consistent and that the rate of convergence is T in both the I(1) and I(0) cases. They also show that this rate of convergence is sufficient to ensure that the null limit distribution of the unit root test is the same as when the break date is known. Hence, one need only use the critical values for the known break date case that pertains to the estimated break date. The test has accordingly more power since the critical values are smaller in absolute value (they also consider a two break model and show the estimates of the break dates to be asymptotically independent). The problem, however, is that the results apply provided there is a break in the slope under the null hypothesis. Indeed, if no break is present, the known break date limit distribution no longer applies; and if the break is small, it is likely to provide a poor approximation to the finite sample distribution. Hatanaka and Yamada (1999) present simulations results calibrated to slope changes in Japanese real GDP that show the estimates of the break dates to have a distribution with fat tails and the unit root test accordingly shows size distortions. For the AO case, the work of Kim and Perron (2005) leads to the following results based on prior work by Perron and Zhu (2005). Under the null hypothesis of a unit root, if a slope change is present, the rate of convergence of the estimate of the break date obtained by minimizing the sum of squared residuals is not fast enough to lead to a limit distribution 61 for the unit root tests (evaluated at this estimate of the break date) that is the same as in the known break date case. They, however, show that a simple modification yields a similar result as in the IO case. It involves performing the unit root test by trimming or eliminating data points in a neighborhood of the estimated break date. This again leads to unit root tests with higher power. Let us summarize the above discussion. First, in the unknown break date case, the invariance properties with respect to the parameters of the trend no longer apply as in the known break date case. Popular methods based on evaluating the unit root test at the value of the break date that minimizes it or maximizes the absolute value of the t-statistic on the coefficient of the relevant dummy variable suffer from problems of liberal size distortions when a large break is present (except with the latter method to select the break date in the IO case) and little if any when the break is small. When the break is large, evaluating the unit root test at the break date that minimizes the sum of squared residuals leads to a procedure with correct size and better power. So this suggests a two step procedure that requires in the first step a test for a change in the trend function that is valid whether a unit root is present or not, i.e., under both the null and alternative hypotheses. In this context, the work of Perron and Yabu (2005) becomes especially relevant. This is the approach taken by Kim and Perron (2005). They use a pre-test for a change in trend valid whether the series is I(1) or I(0). Upon a rejection, the unit root test is evaluated at the estimate of the break date that minimizes the sum of squared residuals from the relevant regression. If the test does not reject, a standard Dickey-Fuller test is applied. This is shown to yield unit root tests with good size properties overall and better power. In cases where only level shifts are present, similar improvements are possible even though, with a fixed magnitude of shift, the estimate of the break date is not consistent under the null hypothesis of a unit root. 6 Testing for Cointegration Allowing for Structural Changes We now discuss issues related to testing for cointegration when allowing for structural changes. We first consider, in Section 6.1, single equation methods involving systems with one cointegrating vector. Here tests have been considered with the null hypothesis as nocointegration and the alternative as cointegration, and vice versa. In Section 6.2, we consider the multivariate case, where the issue is mainly determining the correct number of cointegrating vectors. Since many of the issues are similar to the case of testing for unit roots allowing structural breaks, our discussion will be brief and outline the main results and procedures suggested. 62 6.1 Single equation methods Consider an n dimensional vector of variables yt = (y1t , y2t ) with y1t a scalar, and y2t an n − 1 vector. We suppose that the sub-system y2t is not cointegrated. Then the issue is to determine whether or not there exists a cointegrating vector for the full system yt . Consider the following static regression (33) y1t = α + βy2t + ut The system is cointegrated if there exists a β such that the errors ut are I(0). Hence, a popular method is to estimate this static regression by OLS and perform a unit root test on the estimated residuals (see, Phillips and Ouliaris, 1990). Here the null hypothesis is no-cointegration and the alternative is cointegration. Another approach is to use the Error Correction Model (ECM) representation given by: ∆y1t = bzt−1 + k X di ∆y2t + et i=1 where zt = y1t − βy2t is the equilibrium error. In practice, one needs to replace β by an estimate that is consistent when there is cointegration. The test can then be carried using the t-statistic for testing that b = 0 (see, e.g., Banerjee et al., 1986). When adopting the reverse null and alternative hypotheses, a statistic that has been suggested is, again, Gardner’s (1969) Q test (see Shin, 1994). It can be constructed using the OLS residuals from the static regression when the regressors are strictly exogenous, or, more generally, the residuals from a regression augmented with leads and lags of the firstdifferences of the regressors, as suggested by Saikkonen (1991) and Stock and Watson (1993). Of course, many other procedures are possible. Here, structural changes can manifest themselves in several ways. First, there can be structural changes in the trend functions of the series without a change in the cointegrating relationship (i.e., a change in the marginal distributions of the series). Campos et al. (1996) have documented that shifts in levels do not affect the size of tests of the null hypothesis of no cointegration, for both the ECM based test and the test based on the residuals from the static regression. However, they affect the power of the latter, though not of the former. If all regressors have a common break in the slope of their trend function, the tests can be liberal and reject the null hypothesis of no-cointegration too often, though different tests are affected differently (Leybourne and Newbold, 2003). This is related to what has been labelled as co-breaking processes. Changes in the variance of the errors ut can also induce size distortions if it occurs early enough in the sample (e.g., Noh and Kim, 2003). 63 Second, structural changes can manifest themselves through changes in the long-run relationship (33), either in the form of a change in the intercept, or a change in the cointegrating vector. Here, the power of standard tests for the null hypothesis of no-cointegration can have substantially reduced power as documented by Gregory et al. (1996) and Gregory and Hansen (1996a). An early contribution that proposed tests for the null hypothesis of no-cointegration allowing for the possibility of a change in the long-run relation is that of Gregory and Hansen (1996a). They extend the residual-based tests by incorporating appropriate dummies in regression (33) and taking as the resulting test-statistic the minimal value over all possible break dates. Cases covered are: 1) allowing a change in the level α; 2) allowing for a similar change in level when regression (33) includes a time trend; 3) allowing for changes in both the level α and the cointegrating vector β (with no trend); 4) the case allowing for a change in the level and slope of an included trend and of the cointegrating vector is analyzed in Gregory and Hansen (1986b). The limit distributions of the various tests are derived under the null hypothesis that the series are not cointegrated and are individually I(1) processes with a stable deterministic trend component. As in the case of tests for unit roots, the value of the break date associated with the minimal value of a given statistic is not, in general, a consistent estimate of the break date if a change is present. Cook (2004) shows the size of the tests to be affected (towards excessive rejections) when the series are not cointegrated and are individually I(1) processes with a change in trend. The issue of allowing the possible change in trend under both the null and alternative hypotheses does arise in the context of testing the null hypothesis of no-cointegration. Indeed, under the null hypothesis, the model is a spurious one and the parameters of the cointegrating vector are not identified. It might be possible to identify a change in the slope of a trend under the null hypothesis, but this case is seldom of empirical interest. This means that no further gains in power are possible by trying to exploit the fact that a change in specification occurs under both the null and alternative hypotheses, as was done for unit root tests. Such gains are, however, possible, when adopting cointegration as the null hypothesis. Concerning tests that takes the null hypothesis to be cointegration, the contributions include Bartley et al. (2001), Carrion-i-Silvestre and Sanso (2004) and Arai and Kurozumi (2005). All are based on various modifications of Gardner’s (1969) Q statistic as used by Shin (1994) without structural breaks. The general framework used is to specify the cointegrating relationship by y1t = α1 + α2 (1 > T1 ) + γ 1 t + γ 2 (t − T1 )1(t > T1 ) + β 1 y2t + β 1 y2t 1(t > T1 ) + ut 64 (34) The required residuals to construct the Q test are based on transformed regressions that allow the construction of asymptotically optimal estimates of the cointegrating vector. Bartley et al. (2001) consider only a change in the level and slope of the trend and use the canonical cointegrating regression approach suggested by Park (1992) to estimate the cointegrating vector β. The break date is selected by minimizing the sum of squared residuals from the canonical cointegrating regression. They argue that the resulting estimate of the break fraction is consistent and that the limit distribution of the test corresponds to that applying in the known break date case. The simulations supports this assertion. Carrion-i-Silvestre and Sanso (2004) and Arai and Kurozumi (2005) extend the analysis to cover more cases, in particular allowing for a change in the cointegrating vector. In the case of strictly exogenous regressors, they construct the Q test using residuals from the static regression (34) (scaled appropriately with an estimate of the long-run variance of the errors, which allows for serial correlation). In the general case without strictly exogenous regressors, both recommend using the residuals from regression (34) augmented with leads and lags of the first-differences of y2t (Carrion-i-Silvestre and Sanso (2004) show that the use of the Fully Modified estimates of Phillips and Hansen (1990) leads to tests with very poor finite sample properties). Both select the break date by minimizing the sum of squared residuals from the appropriate regression, following the work of Kurozumi and Arai (2004) who show that the estimate of the break fraction in this model converges at least at rate T 1/2 . This permits limit critical values corresponding to the known break date case. They also consider selecting the break date as the value which minimizes the Q statistic but do not recommend its use given that the resulting tests then suffers from large size distortions in finite samples. A caveat about the approach discussed above is the fact that for the suggested methods to be valid, there must be a change in the cointegrating relationship, if cointegration actually holds. This is because the search for the potential break date is restricted to break fractions that are bounded, in large samples, from the boundaries 0 and 1. Hence, when there is no change the limit value cannot be 0 or 1, the estimate is inconsistent and has a non-degenerate limit distribution, which in turn affects the limit distribution of the test (i.e., it does not correspond to the one that would prevail if no break was present). But to ascertain whether a break is present, one needs to know if there is cointegration, which is actually the object of the test. Tests of whether a change in structure has occurred (as reviewed in Section 4.7) will reject the null hypothesis of no change when a change actually occurs in a cointegrating relationship, and will also reject if the system is simply not cointegrated. Hence, we are led to a circular argument. The test procedure needs to allow for the possibility of a change and 65 not impose it. It may be possible to relax the restriction on the search for the break date by allowing all possible values. In the context of cointegrated I(1) regressors, it is, however, unknown at this point, if the estimate of the break fraction would converge to 0 or 1 when no change is present. 6.2 Methods based on a mutivariate framework We now consider tests that have been proposed when the variables are analyzed jointly as a system. Here, the results available in the literature are quite fragmentary and much of it pertains to a single break at a known date. Also, different treatments are possible by allowing for a change in the trend function of the original series (i.e., the marginal processes), or in allowing for a change in the cointegrating relation. One of the early contribution is that of Inoue (1999). It allows for a one time shift in the trend function of the series at some unknown date, either in level for non-trending series and for both level and slope in trending series. He considers an AO type framework and also an IO type regression when only a shift in intercept is allowed in the VAR. The specification of the null and alternative hypotheses follow Zivot and Andrews (1992) and Gregory and Hansen (1996), in that the shifts are allowed only under the alternative hypothesis. Hence, the null hypothesis is that the system contains no break and no more than r cointegrating vectors, and the alternative hypothesis is that the data can exhibit a change in trend and that the cointegrating rank is r + 1, or greater than r. The breaks are assumed to occur at the same date for all series. Under the alternative hypothesis, the series are not assumed to be co-breaking, in the sense that the cointegrating vector that reduces the non-stationarity in the stochastic component also eliminates the non-stationarity in the deterministic trend. He considers the trace and maximal eigenvalue tests of Johansen (1988, 1991) with data appropriately detrended allowing for a shift in trend, and the resulting statistic is based on the maximal values over all permissible break dates. It is unclear what are the properties of the tests when the null hypothesis is true with data that have broken trends. Also, although the parameter r can be selected arbitrarily, the procedures cannot be applied sequentially to determine the cointegrating rank of the system. This is because the breaks are not allowed under the null hypothesis, only under the alternative. So if one starts with, say, r = 0, breaks are allowed for alternatives such that the cointegrating rank is greater than 0. But, upon a rejection, if one then wants to test the null of rank 1 versus an alternative with rank greater than 1, one needs to impose no break under the null hypothesis of rank 1, a contradiction from what was specified in the earlier step. 66 Saikkonen and Lütkepohl (2000a) also considers a test of the null hypothesis of r cointegrating vectors versus the alternative that this number is greater than r, allowing for a break in the trend function of the individual series under both the null and alternative hypotheses. They, however, only consider a level shift (in trending or non trending series) occurring at some known date. To estimate the coefficients of the trend component of the series, they use a similar GLS procedure, as discussed in Section 5.3, appropriately extended for the multivariate nature of the problem. This detrending method imposes the null hypothesis. Hence, the effect of level shifts is negligible in large samples and the limit distribution of the test is the same as the standard (no-break) cointegration test of Lütkepohl and Saikkonen (2000) and Saikkonen and Lütkepohl (2000b). Once the detrended data is obtained the test is based on the eigenvalues of a reduced rank problem where restrictions implied by the process and the breaks are not imposed. Johansen et al. (2000) consider a more general problem but still with known break dates. They consider multiple structural changes in the following VAR of order k, ⎞ ⎛ k−1 X yt−1 ⎠ + µj + Γi ∆yt−i + et ∆yt = (Π, Πj ) ⎝ t i=1 for Tj−1 + k < t ≤ Tj for j = 1, ..., m. Hence, there are m breaks which can affect the constant and the coefficients of the trend. Various tests for the rank of the cointegrating matrix are proposed (imposing or not various restrictions on the deterministic components). Since the estimates of the coefficients of the trend are estimated from a maximum-likelihood type approach (following Johansen, 1988, 1991), the limit distribution depends on the exact specification of the deterministic components and on the true break dates. Asymptotic critical values are presented via a response surface analysis. For the special case of a single shift in level, Lütkepohl et al. (2003) compare the two approaches of Saikkonen and Lütkepohl (2000a) and Johansen et al. (2000). They show that the former has higher local asymptotic power. However, the finite sample size-adjusted power is very similar. They recommend using the method of Saikkonen and Lütkepohl (2000a) on the basis of better size properties in finite samples and also on the fact that they view having a limit distribution free of the break dates to be advantageous. A problem with this argument is that the non-dependence of the limit distribution on the break date with the procedure of Saikkonen and Lütkepohl (2000a) no longer holds in more general models, especially when slope shifts are involved. Indeed, no result is yet available for this approach with a GLS type detrending procedure when slope shifts are present. 67 Lütkepohl et al. (2004) extend the analysis of Saikkonen and Lütkepohl (2002), which pertained to testing for a unit root allowing for a change in the level of a series occurring at an unknown date (see Section 5.4.1). The GLS type procedure discussed above is used to estimate the coefficients of the deterministic components. Once the series are detrended, the cointegration tests of Johansen (1988) can be used. In the unit root case, with a GLS detrending procedure that imposes the null hypothesis, the change in mean reduces to an outlier in the first-differenced series. Here, thing are more complex and a consistent estimate of the break date is preferable. Estimating the break date has, however, no effect on the limit null distribution of the test statistic since, here again, it does not depend on the true value of the break date. In it useful to consider in more detail the issue of estimating the break date. The n vector of data yt is assumed to be generated by yt = µ + θDUt + δt + xt where DUt = 1(t > T1 ) and xt is a noise component generated by a VAR, with the following ECM representation, k X ∆xt = Πxt−1 + Γi ∆xt−i + et i=1 Here, the presence of cointegration implies the decomposition Π = αβ 0 with β the n × r matrix of cointegrating vectors. Hence, we also have the following ECM representation for yt ∆yt = ν + αβ 0 (yt−1 − δ(t − 1) − θDUt−1 ) + k X i=1 Γi ∆yt−i + k X γ i ∆DUt−i + et (35) i=1 This ECM representation will be affected by a level shift if β 0 θ 6= 0, otherwise only the impulse dummies ∆DUt−i are present. In most cases of interest, we have β 0 δ = 0, which specifies that the same linear combinations that eliminate the stochastic non-stationarity also eliminate the non-stationarity induced by the trend. The condition β 0 θ = 0 can be interpreted in the same way, i.e., if some variables are affected by changes in trend, the linear combination of the data specified by the cointegrating vectors will be free of structural breaks. This is often referred to as ‘co-breaking’. Hence, the condition β 0 θ 6= 0 requires that the series be non co-breaking, which may be unappealing in many cases. Lütkepohl et al. (2004) estimate the break date by minimizing the determinant of the sample covariance matrix of the estimates of the errors et . They show the estimate of the break fraction to 68 converge at rate T , though no limit distribution is given since this rate is enough to guarantee that the limit distribution of the test be independent of the break date. Note that the search for the break date is restricted to an interval that excludes a break fraction occurring near the beginning or the end of the sample. This is important, since it makes the procedure valid conditional on shifts in level occurring. Without shifts, the true break fraction is 0 or 1, which are excluded from the search. Hence, in this case the estimated break fractions will converge to some random variable. But given that a GLS type detrending is done, this has no impact of the limit distribution of the rank test. A similar result holds when co-breaking shifts are present, though Lütkepohl et al. (2004) argue that if the shifts are large enough, they can be captured by the impulse dummies ∆DUt−i (for more details on estimation of break dates in this framework, see Saikkonen et al., 2004). All contributions discussed above do not address the problem of a potential shift in the cointegrating vector. A recent analysis by Andrade et al. (2005) deals with this in the context of a one-time change. The object is to test the null hypothesis of r cointegrating vectors versus the alternative that this value is greater than r. They allow the change in the cointegrating relationship to occur under both the null and alternative hypotheses and the number of cointegrating vectors is the same in both regimes. This allows a sequential procedure to determine the rank. The issues are addressed using the following generalized ECM ∆yt = 1(t ≤ T1 )[α0 β 00 yt−1 − δ 0 dt ] + 1(t > T1 )[α1 β 01 (yt−1 − yT1 ) − δ 1 dt ] + k X Γi ∆yt−i + et i=1 where dt is a vector of deterministic components (usually the null set or a constant). Note that the data is re-normalized after the break to start again at 0. This is done since otherwise the variance of β 01 yt−1 would increase after the break given that it depends on the value of β 01 yT1 . They note that the estimation of this model by maximum likelihood is quite involved and suggest a simpler principle components analysis. Let β i⊥ be a matrix such that β 0i β i⊥ = 0, and suppose that the loading factors (or adjustment matrices) are constant, i.e., α0 = α1 , the test for the null hypothesis that the cointegrating rank is r is based on testing that γ 0 = γ 1 = 0 in the following system h i h i 0 0 0 0 ∆yt = 1(t ≤ T1 ) γ 0 β̂ 0⊥ yt−1 + αβ̂ 0 yt−1 + 1(t > T1 ) γ 1 β̂ 1⊥ yt−1 + αβ̂ 1 (yt−1 − yT1 ) + k X Γi ∆yt−i + et i=1 69 0 0 where β̂ 1 and β̂ i⊥ are estimates obtained from the principle components analysis. The statistic is based on a multivariate Fisher-type statistic modified to eliminate the effect of nuisance parameters on the limit distribution under the null hypothesis. They also consider a version that is valid when the break date is unknown, based on the maximal values over a specified range for the break date, and present a test to evaluate how many cointegrating vectors are subject to change across regimes. When both the cointegrating matrix β and the loading factors α are allowed to change, a more involved testing procedure is offered, which applies, however, only to the known break date case. An interesting recent contribution is that of Qu (2004). It proposes a procedure to detect whether cointegration (or stationarity in the scalar case) is present in any part of the sample, more precisely whether there is evidence in any part of the sample that a system is cointegrated with a higher cointegrating rank than the rest of the sample. The test procedure is based on a multivariate generalization of Gardner’s (1969) Q test as used in Breitung (2002). The main device used is that if one or more sub-samples have a different cointegrating rank, one can find them by searching, in an iterative fashion, over all possible partitions of the sample with three segments or two breaks. The relevant limit distributions are derived allowing the possibility of imposing some structure if desired (e.g., that the change occurs at the beginning or end of the sample). He also discusses how to consistently estimate the break dates or the boundaries of the regimes when a change has been detected. A modification is also suggested to improve the finite sample performance of the test. This approach also permit testing for changes in persistence with the null hypothesis specified as an I(1) process throughout the sample. It also permits detecting whether cointegration is present when the cointegrating vector changes at some unknown possibly multiple dates. 7 Conclusions This review has discussed a large amount of research that has been done in the last fifteen years or so pertaining to issues related to structural changes and to try to distinguish between structural changes and unit roots. But still, some important questions remain to be addressed: limit distributions of estimates of break dates in a cointegrated system with multiple structural changes, issues of non-monotonic power functions for tests of structural change and how to alleviate the problems, evaluating the frequency of permanent shocks; just to name a few. Research currently under progress is trying to address these and other issues. One recent area of research where similar tools have been applied is related to distinguish70 ing between long-memory processes and short-memory processes with structural changes, in particular level shifts. This is especially important in financial economics, where it is widely documented that various measures of stock return volatility exhibit properties similar to those of a long-memory process (e.g., Ding et al., 1993, Granger and Ding, 1995 and Lobato and Savin, 1998). For reviews of the literature on purely long-memory processes, see Robinson (1994a), Beran (1994) and Baillie (1996). A mentioned in Section 2, a popular test for long-memory is the rescaled-range test. Yet, interestingly, Gardner’s (1969) Q test makes yet another appearance. Indeed, it was, along with a slight modification, also proposed to test this problem by Giraitis et al. (2003). So we have the same test acting with the null hypothesis of a stable short-run memory process versus an alternative that is either structural change, a unit root or long-memory. This goes a long way showing how the three problems are inter-related. One of the most convincing evidence that stock market volatility may be better characterized by a short-memory process affected by occasional level shifts is that of Perron and Qu (2004). They show that the behavior of the log-periodogram estimate of the long-memory parameter (the fractional differencing coefficient), as a function of the number of frequencies used in the regression, is very different for the two types of processes. The pattern found with data on daily SP500 return series (absolute or square root returns) is very close to what is expected with a short-memory process with level shifts. They also present a test which rejects the null hypothesis of long memory. Given that unit root and long memory processes share similar features, it is not surprising that many of the same problems are being addressed with similar findings. Along the lines of Perron (1989) for unit roots, it has been documented that short-memory processes with level shifts will exhibit properties that make standard tools conclude that long memory is present (e.g., Diebold and Inoue, 2001, Engle and Smith, 1999, Gourieroux and Jasiak, 2001, Granger and Ding, 1996, Granger and Hyung, 2004, Lobato and Savin, 1998, and Teverosovky and Taqqu, 1997). Some papers have also documented the fact that long-memory processes will induce, similar to unit root processes, a rejection of the null hypothesis of no-structural change when using standard structural change tests; for the CUSUM and the Sup-Wald test applied to a change in a polynomial trend, see Wright (1998) and Krämer and Sibbertsen (2002). Results about the rate of convergence of the estimated break fraction in a single mean shift model can be found in Kuan and Hsu (1998). When there is structural change, the estimate is consistent but the rate of convergence depends on d. When d ∈ (0, 1/2) and 71 there is no change, the limit value is not 0 or 1 but rather the estimate of the break fraction converges to a random variable, suggesting a spurious change, exactly as in the unit root case (see Nunes et al., 1995, and Bai, 1998). For results related to multiple structural changes in mean, see Lavielle and Moulines (2000). A test for a single structural change occurring at some known date in the linear regression model is discussed in Hidalgo and Robinson (1996). It is essentially a Wald test for testing that the coefficients are the same in both regimes, which accounts for the long-memory correlation pattern in the residuals. Lazarová (2005) presents a test for the case of a single change in the parameters of a linear regression model occurring at an unknown date. The test follows the “fluctuations tests” approach of Ploberger et al. (1989) with different metrics used to weight the differences in the estimates for each permissible break dates (giving special attention to the Sup and Mean functionals). The limit distribution depends on nuisance parameters and a bootstrap procedure is suggested to obtain the relevant critical values. Related to the problem of change in persistence (see Section 4.9), Beran and Terrin (1996) present a test for a change in the long-memory parameter, based on the maximal difference, across potential break dates, of appropriately weighted sums of autocovariances. Related to unit root tests allowing for a change in the trend function, Gil-Alana (2004) extends Robinson’s (1994b) test to allow for a one-time change occurring at a known date. For a review of some related results, see Sibbertsen (2004). The literature on structural changes in the context of long memory processes is quite new and few results are available. Still, there is a large demand for empirical applications. Given the nature of the problems and series analyzed, it is important to have procedures that are valid for multiple structural changes. For example, with many financial time series, it is the case that allowing for structural breaks reduces considerably the estimates of the long-memory parameters within regimes (e.g., Granger and Hyung, 2004, for stock return volatility). Are the reductions statistically significant? Are the reductions big enough that one can consider the process as being of a short-memory nature within regimes? Is there significant evidence of structural changes? Is the long-memory parameter stable across regimes? The econometrics and statistics literatures have a long way to go to provide reliable tools to answer these questions. Given that the issues are similar to the structural change versus unit root problem, our hope is that this survey will provide a valuable benchmark to direct research in specific directions and to alert researchers of the potential merits and drawbacks of various approaches. 72 References [1] Altissimo, F., Corradi, V., 2003. Strong rules for detecting the number of breaks in a time series. Journal of Econometrics 117, 207-244. [2] Amsler, C., Lee, J., 1995. An LM test for a unit root in the presence of a structural change. Econometric Theory 11, 359-368. [3] Anderson, T.W., Darling, D.A., 1952. Asymptotic theory of certain ‘goodness of fit’ criteria based on stochastic processes. The Annals of Mathematical Statistics 23, 193212. [4] Andrade, P., Bruneau, C., Gregoir, S., 2005. Testing for the cointegration rank when some cointegrating directions are changing. Journal of Econometrics 124, 269-310. [5] Andrews, D.W.K., 1991. Heteroskedasticity and autocorrelation consistent covariance matrix estimation. Econometrica 59, 817-858. [6] Andrews, D.W.K., 1993a. Tests for parameter instability and structural change with unknown change point. Econometrica 61, 821-856 (Corrigendum, 71, 395-397). [7] Andrews, D.W.K., 1993b. Exactly median-unbiasedestimation of first-order autoregressive/unit root models. Econometrica 61, 139-165. [8] Andrews, D.W.K., Lee, I., Ploberger, W., 1996. Optimal change point tests for normal linear regression. Journal of Econometrics 70, 9-38. [9] Andrews, D.W.K., Ploberger, W., 1994. Optimal tests when a nuisance parameter is present only under the alternative. Econometrica 62, 1383-1414. [10] Antoch, J., Hušková, M., Prášková, Z., 1997. Effect of dependence on statistics for determination of change. Journal of Statistical Planning and Inference 60, 291-310. [11] Arai, Y., Kurozumi, E., 2005. Testing the null hypothesis of cointegration with structural breaks. Unpublished manuscript, Hitotsubashi University. [12] Bai, J., 1994. Least squares estimation of a shift in linear processes. Journal of Time Series Analysis 15, 453-472. [13] Bai, J., 1997a. Estimation of a change point in multiple regression models. Review of Economic and Statistics 79, 551-563. [14] Bai, J., 1997b. Estimating multiple breaks one at a time. Econometric Theory 13, 315-352. [15] Bai, J., 1998. A note on spurious break. Econometric Theory 14, 663-669. [16] Bai., J., 1999. Likelihood ratio tests for multiple structural changes. Journal of Econometrics 91, 299-323. 73 [17] Bai, J., 2000. Vector autoregressive models with structural changes in regression coefficients and in variance-covariance matrices. Annals of Economics and Finance 1, 303-339. [18] Bai, J., Lumsdaine, R.L., Stock, J.H., 1998. Testing for and dating breaks in multivariate time series. Review of Economic Studies 65, 395-432. [19] Bai, J., Perron, P., 1998. Estimating and testing linear models with multiple structural changes. Econometrica 66, 47-78. [20] Bai, J., Perron, P., 2003a. Computation and analysis of multiple structural change models. Journal of Applied Econometrics 18, 1-22. [21] Bai, J., Perron, P., 2003b. Critical values for multiple structural change tests. Econometrics Journal 6, 72-78. [22] Bai, J., Perron, P., 2005. Multiple structural change models: a simulation analysis. Forthcoming in Econometric Essays, D. Corbea, S. Durlauf and B. E. Hansen (eds.), Cambridge University Press. [23] Baillie, R.T., 1996. Long memory processes and fractional integration in econometrics. Journal of Econometrics 73, 5-59. [24] Banerjee, A., Dolado, J.J., Hendry, D.F., Smith, G.W., 1986. Exploring equlibrium relationships in econometrics through static models: some Monte Carlo evidence. Oxford Bulletin of Economics and Statistics 48, 253-277. [25] Banerjee, A., Lumsdaine, R.L., Stock, J.H., 1992. Recursive and sequential tests of the unit-root and trend-break hypotheses: theory and international evidence. Journal of Business and Economic Statistics 10, 271-287. [26] Barsky, R.B., 1987. The Fisher hypothesis and the forecastibility and persistence of inflation. Journal of Monetary Economics 19, 3-24. [27] Bartley, W.A., Lee, J., Strazicich, M.C., 2001, Testing the null of cointegration in the presence of a structural break. Economics Letters 73, 315-323. [28] Bellman, R., Roth, R., 1969. Curve fitting by segmented straight lines. Journal of the American Statistical Association 64, 1079-1084. [29] Beran, J., 1994. Statistics for Long Memory Processes. New York: Chapman & Hall. [30] Beran, J., Terrin, N., 1996. Testing for a change of the long-memory parameter. Biometrika 83, 627-638. [31] Beveridge, S., Nelson, C.R., 1981. A New Approach to Decomposition of Economic Time Series into Permanent and Transitory Components with Particular Attention to Measurement of the ‘Business Cycle’. Journal of Monetary Economics 7, 151-74. 74 [32] Bhattacharya, P.K., 1987. Maximum likelihood estimation of a change-point in the distribution of independent random variables, general multiparameter case. Journal of Multivariate Analysis 23, 183-208. [33] Bhattacharya, P.K., 1994. Some aspects of change-point analysis. In Carlstein, E., Müller, H.-G., Siegmund, D. (eds.), Change Point Problems, IMS Lecture Notes Monograph Series, vol. 23, 28-56. [34] Bhattacharya, R.N., Gupta, V.K., Waymire, E., 1983. The Hurst effect under trends. Journal of Applied Probability 20, 649-662. [35] Box, G.E.P., Tiao, G.C., 1975. Intervention analysis with applications to economic and environmental problems. Journal of the American Statistical Association 70, 70-79. [36] Breitung, J., 2002. Nonparametric tests for unit roots and cointegration. Journal of Econometrics 108, 343-363. [37] Brown, R.L., Durbin, J., Evans, J.M., 1975. Techniques for testing the constancy of regression relationships over time. Journal of the Royal Statistical Society B 37, 149163. [38] Burdekin, R.C.K., Siklos, P.L., 1999. Exchange rate regimes and shifts in inflation persistence: does nothing else matter?. Journal of Money, Credit and Banking 31, 235-247. [39] Busetti, F., 2002. Testing for (common) stochastic trends in the presence of structural breaks. Journal of Forecasting 21, 81-105. [40] Busetti, F., Harvey, A.C., 2001. Testing for the presence of a random walk in series with structural breaks. Journal of Time Series Analysis 22, 127-150. [41] Busetti, F., Harvey, A.C., 2003. Further comments on stationarity tests in series with structural breaks at unknown points. Journal of Time Series Analysis 24, 137-140. [42] Busetti, F., Taylor, A.M.R., 2001. Tests stationarity against a change in persistence. Discussion Paper 01-13, Department of Economics, University of Birmingham. [43] Busetti, F., Taylor, A.M.R., 2005. Tests stationarity against a change in persistence. Journal of Econometrics. Forthcoming. [44] Campbell, J.Y., Perron, P., 1991. Pitfalls and opportunities: what macroeconomists should know about unit roots. NBER Macroeconomics Annual, Vol. 6, Blanchard, O.J., Fisher, S. (eds.), 141-201. [45] Campos, J., Ericsson, N.R., Hendry, D.F., 1996. Cointegration tests in the presence of structural breaks. Journal of Econometrics 70, 187-220. 75 [46] Carion-i-Silvestre, J.L., Sansó-i-Rosselló, A.S., 2004. Testing the null hypothesis of cointegration with structural breaks. Unpublished manuscript, Departament d’Econometria, Estadística i Economia Espanyola, Universitat de Barcelona. [47] Carion-i-Silvestre, J.L., Sansó-i-Rosselló, A.S., Artis, M., 2004. Joint hypothesis specification for unit root tests with a structural break. Unpublished manuscript, Departament d’Econometria, Estadística i Economia Espanyola, Universitat de Barcelona. [48] Carion-i-Silvestre, J.L., Sansó-i-Rosselló, A.S., Ortuño, M.A., 1999. Response surface estimates for the Dickey-Fuller test with structural breaks. Economics Letters 63, 279283. [49] Chernoff, H., Zacks, S., 1964. Estimating the current mean of a normal distribution which is subject to changes in time. The Annals of Mathematical Statistics 35, 9991018. [50] Chong, T.T.L., 1995. Partial parameter consistency in a misspecified structural change model. Economics Letters 49, 351-357. [51] Chong, T.T.L., 2001. Structural change in AR(1) models. Econometric Theory 17, 87-155. [52] Christiano, L.J., 1992. Searching for breaks in GNP. Journal of Business and Economic Statistics 10, 237-250. [53] Chu, C.-S.J., Hornik, K., Kuan, C.-M., 1995. MOSUM tests for parameter constancy. Biometrika 82, 603-617. [54] Chu, C.-S. J., White, H., 1992, A direct test for changing trend. Journal of Business and Economic Statistics 10, 289-299. [55] Clemente, J., Montañés, A., Reyes, M., 1998. Testing for a unit root in variables with a double change in the mean. Economics Letters 59, 175-182. [56] Clements, M.P., Hendry, D.F., 1999. Forecasting Non-stationary Economic Time Series. MIT Press, Cambridge, MA. [57] Cook, S., 2004. Spurious rejection by cointegration tests incorporating structural change in the cointegrating relationship. Unpublished manuscript, Department of Economics, University of Wales Swansea. [58] Crainiceanu, C.M., Vogelsang, T.J., 2001. Spectral density bandwidth choice: source of nonmonotonic power for tests of a mean shift in a time series. Unpublished manuscript, Department of Economics, Cornell University. [59] Csörgő, M., Horváth, L., 1997. Limit theorems in change-point analysis. Wiley Series in Probability and Statistics, New York: John Wiley. 76 [60] Davies, R.B., 1977. Hypothesis testing when a nuisance parameter is present only under the alternative. Biometrika 64, 247-254. [61] Davies, R.B., 1987. Hypothesis testing when a nuisance parameter is present only under the alternative. Biometrika 74, 33-43. [62] DeLong, J.B., Summers, L.H., 1988. How does macroeconomic policy affect output?. Brookings Papers on Economic Activity 2, 433-494. [63] Deng, A., Perron, P., 2005a. A comparison of alternative asymptotic frameworks to analyze structural change in a linear time trend. Manuscript in preparation, Department of Economics, Boston University. [64] Deng, A., Perron, P., 2005b. A locally asymptotic point optimal test that is inconsistent: the case of a change in persistence. Unpublished manuscript, Department of Economics, Boston University. [65] Deshayes, J., Picard, D., 1984a. Principe d’invariance sur le processus de vraisemblance. Annales de l’Institut Henri Poincaré, Probabilités et Statistiques 20, 1-20. [66] Dehayes, J., Picard, D., 1984b. Lois asymptotiques des tests et estimateurs de rupture dans un modèle statistique classique. Annales de l’Institut Henri Poincaré, Probabilités et Statistiques 20, 309-327. [67] Deshayes, J., Picard, D., 1986. Off-line statistical analysis of change point models using non-parametric and likelihood methods. In Basseville, M., Beneviste, A., (eds), Detection of Abrupt Changes in Signals and Dynamical Systems (Lecture Notes in Control and Information Sciences 77), 103-168, Berlin: Springer. [68] Dickey, D.A., Fuller, W.A., 1979. Distribution of the estimators for autoregressive time Series with a unit Root. Journal of the American Statistical Association 74, 427-431. [69] Dickey, D.A., Fuller, W.A., 1981. Likelihood ratio statistics for autoregressive time series with a unit root. Econometrica 49, 1057-1072. [70] Diebold, F., Inoue, A., 2001. Long memory and regime switching. Journal of Econometrics 105, 131-159. [71] Ding, Z., Engle, R.F., Granger, C.W.J., 1993. A long memory property of stock market returns and a new model. Journal of Empirical Finance 1, 83-106. [72] Dufour, J.M., 1982. Recusrive stability analysis of linear regression relationships: an exploratory methodology. Journal of Econometrics 19, 31-76. [73] Dufour, J.-M., Kiviet, J.F., 1996. Exact tests for structural change in first-order dynamic models. Journal of Econometrics 70, 39-68. 77 [74] Elliott, G., Müller, U.K., 2003. Optimally testing general breaking processes in linear time series models. Unpublished manuscript. Department of Economics, University of California at San Diego. [75] Elliott, G., Müller, U.K., 2004. Confidence sets for the date of a single break in linear time series regressions. Unpublished manuscript, Department of Economics, University of California at San Diego. [76] Elliott, G., Rothenberg, T.J., Stock, J.H., 1996. Efficient tests for an autoregressive unit root. Econometrica 64, 813-836. [77] Engle, R.F., Smith, A.D., 1999. Stochastic permanent breaks. Review of Economics and Statistics 81, 533-574. [78] Feder, P.I., 1975. On asymptotic distribution theory in segmented regression problems: identified case. Annals of Statistics 3, 49-83. [79] Fisher, W.D., 1958. On grouping for maximum homogeneity. Journal of the American Statistical Association 53, 789-798. [80] Fu, Y-X., Curnow, R.N., 1990. Maximum likelihood estimation of multiple change points. Biometrika 77, 563-573. [81] Garcia, R., Perron, P., 1996. An analysis of the real interest rate under regime shifts. Review of Economics and Statistics 78, 111-125. [82] Gardner, L.A., 1969. On detecting changes in the mean of normal variates. The Annals of Mathematical Statistics 40, 116-126. [83] Giraitis, L., Kokoszka, P., Leipus, R., Teyssière, G., 2003. Rescaled variance and related tests for long memory in volatility and level. Journal of Econometrics 112, 265-294 (Corrigendum, 126, 571-572). [84] Gil-Alana, L.A., 2004. A joint test of fractional integration and structural breaks at a known period of time. Journal of Time Series Analysis 25, 691-700. [85] Gourieroux, C., Jasiak, J., 2001. Memory and infrequent breaks. Economics Letters 70, 29-41. [86] Granger, C.W.J., Ding, Z., 1996. Varieties of long memory models. Journal of Econometrics 73, 61-77. [87] Granger, C.W.J., Hyung N., 2004. Occasional structural breaks and long memory with an application to the S&P 500 absolute stock returns. Journal of Empirical Finance 11, 399-421. [88] Gregory, A.W., Hansen, B.E., 1996a. Residual-based tests for cointegration in models with regime shifts. Journal of Econometrics 70, 99-126. 78 [89] Gregory, A.W., Hansen, B.E., 1996b. Tests for cointegration in models with regime and trend shifts. Oxford Bulletin of Economics and Statistics 58, 555-560. [90] Gregory, A.W., Nason, J.M., Watt, D.G., 1996. Testing for structural breaks in cointegrated relationships. Journal of Econometrics 71, 321-341. [91] Guthery, S.B., 1974. Partition Regression. Journal of the American Statistical Association 69, 945-947. [92] Hackl, P., Westlund, A.H., 1989. Statistical analysis of ‘structural change’: an annotated bibliography. Empirical Economics 14, 167-192. [93] Hackl, P., Westlund, A.H., 1991 eds. Economic Structural Change: Analysis and Forecasting. Springer-Verlag: Berlin. [94] Hakkio, C.S.., Rush, M., 1991. Is the budget deficit too large?. Economic Inquiry 29, 429-445. [95] Hall, A.R., 2005. Generalized Methods of Moments. Oxford University Press. [96] Hamori, S., Tokihisa, A., 1997. Testing for a unit root in the presence of a variance shift. Economics Letter 57, 245-253. [97] Hansen, B.E., 1992a. Tests for parameter instability in regressions with I(1) processes. Journal of Business and Economic Statistics 10, 321-335. [98] Hansen, B.E., 1992b. Testing for parameter instability in linear models. Journal of Policy Modeling 14, 517-533. [99] Hansen, B.E., 1996. Inference when a nuisance parameter is not identified under the null hypothesis. Econometrica 64, 413-430. [100] Hansen, B.E., 2000. Testing for structural change in conditional models. Journal of Econometrics 97, 93-115. [101] Hansen, H., Johansen, S., 1999. Some tests for parameter constancy in cointegrated VAR-models. Econometrics Journal 306-333. [102] Hansen, P.R., 2003. Structural changes in the cointegrated vector autoregressive model. Journal of Econometrics 114, 261-295. [103] Harvey, D.I., Leybourne, S.J., Taylor, A.M.R., 2004. Modified tests for a change in persistence. Unpublished manuscript, Department of Economics, University of Birmingham. [104] Hájek, J., Rényi, A., 1955. Generalization of an inequality of Kolmogorov. Acta Math. Acad. Sci. Hungar. 6, 281-283. 79 [105] Hao, K., 1996. Testing for structural change in cointegrated regression models: some comparisons and generalizations. Econometric Reviews 15, 401-429. [106] Hao, K., Inder, B., 1996. Diagnostic test for structural change in cointegrated regression models. Economics Letters 50, 179-187. [107] Harvey, A.C., 1975. Comment on the paper by Brown, Durbin and Evans. Journal of the Royal Statistical Society B 37, 179-180. [108] Harvey, D.I., Leybourne, S.J., Newbold, P., 2001. Innovational outlier unit root tests with an endogeneously determined break in level. Oxford Bulletin of Economics and Statistics 63, 559-575. [109] Harvey, D.I., Mills, T.C., 2003. A note on Busetti-Harvey tests for stationarity in series with structural breaks. Journal of Time Series Analysis 24, 159-164. [110] Hatanaka, M., Yamada, K., 1999. A unit root test in the presence of structural changes in I(1) and I(0) Models. In Cointegration, Causality, and Forecasting: A Festschrift in Honour of Clive W.J. Granger, Engle, R.F., White, H., eds., Oxford University Press. [111] Hatanaka, M., Yamada, K., 2003. Co-trending: A Statistical System Analysis of Economic Trends. Springer-Verlag: Tokyo. [112] Hawkins, D.L., 1987. A test for a change point in a parametric model based on a maximal Wald-type statistic. Sankhya 49, 368-376. [113] Hawkins, D.M., 1976. Point estimation of the parameters of piecewise regression models. Applied Statistics 25, 51-57. [114] Hawkins, D.M., 1977. Testing a sequence of observations for a shift in location. Journal of the American Stastistical Association 72, 180-186. [115] Hecq, A., Urbain, J.P., 1993. Misspecification tests, unit roots and level shifts. Economics Letters 43, 129-135. [116] Hidalgo, J., Robinson, P.M., 1996. Testing for structural change in a long-memory environment. Journal of Econometrics 70, 159-174. [117] Hinkley, D.V., 1970. Inference about the change-point in a sequence of random variables. Biometrika 57, 1-17. [118] Horváth, L., 1995. Detecting changes in linear regression. Statistics 26, 189-208. [119] Hurst, H., 1951. Long term storage capacity of reservoirs. Transactions of the American Society of Civil Engineers 116, 770-799. [120] Imhof, J.P., 1961. Computing the distribution of quadratic forms in normal variables. Biometrika 48, 419-426. 80 [121] Inoue, A., 1999. Tests of cointegrating rank with a trend-break. Journal of Econometrics 90, 215-237. [122] James, B., James, K.L., Siegmund, D., 1987. Test for a change-point. Biometrika 74, 71-83. [123] Jandhyala, V.K., MacNeill, I.B., 1989. Residual partial sum limit process for regression models with applications to detecting parameter changes at unknown times. Stochastic Processes and their Applications 33, 309-323. [124] Jandhyala, V.K., MacNeill, I.B., 1992. On testing for the constancy of regression coefficients under random walk and change-point alternatives. Econometric Theory 8, 501-517. [125] Jandhyala, V.K., Minogue, C.D., 1993. Distributions of Bayes-type change-point statistics under polynomial regression. Journal of Statistical Planning and Inference 37, 271-290. [126] Johansen, S., 1988. Statistical analysis of cointegrating vectors. Journal of Economic Dynamics and Control 12, 231-254. [127] Johansen, S., 1991. Estimation and hypothesis testing of cointegration vectors in Gaussian vector autoregressive models. Econometrica 59, 1551-1580. [128] Johansen, S., Mosconi, R., Nielsen, B., 2000. Cointegration analysis in the presence of structural breaks in the deterministic trend. Econometrics Journal 3, 216-249. [129] Juhl, T., Xiao, Z., 2005. Tests for changing mean with monotonic power. Unpublished manuscript, Department of Economics, Boston College. [130] Kander, Z., Zacks, S., 1966. Test procedures for possible changes in parameters of statistical distributions occurring at unknown time points. The Annals of Mathematical Statistics 37, 1196-1210. [131] Kapetanios, G., 2005. Unit-root testing against the alternative hypothesis of up to m structural breaks. Journal of Time Series Analysis 26, 123-133. [132] Kim, D., Perron, P., 2005. Unit Root Tests with a Consistent Break Fraction Estimator. Manuscript, Department of Economics, Boston University. [133] Kim, H.-J., Siegmund, D., 1989. The likelihood ratio test for a change-point in simple linear regression. Biometrika 76, 409-423. [134] Kim, J.Y., 2000. Detection of change in persistence of a llinear time series. Journal of Econometrics 95, 97-116 (corrigendum, 2002, 109, 389-392) [135] Kim, T.-H., Leybourne, S.J., Newbold, P., 2002. Unit root tests with a break in innovation variance. Journal of Econometrics 109, 365-387. 81 [136] Kim, T.-H., Leybourne, S.J., Newbold, P., 2004. Behavior of Dickey-Fuller unit root tests under trend misspecification. Journal of Time Series Analysis 25, 755-764. [137] King, M.L., Shiveley, T.S., 1993. Locally optimal testing when a nuisance parameter is present only under the alternative. Review of Economics and Statistics 75, 1-7. [138] Krämer, W., Ploberger, W., Alt, R., 1988. Testing for structural change in dynamic models. Econometrica 56, 1355-1369. [139] Krämer, W., Sonnberger, H., 1986. The Linear Regression Model Under Test. PhysicaVerlag: Heidelberg. [140] Krämer, W., Sibbertsen, P., 2002. Testing for structural changes in the presence of long memory. International Journal of Business and Economics 1, 235-242. [141] Krishnaiah, P.R., Miao, B.Q., 1988. Review about estimation of change points. In Handbook of Statistics, vol. 7, Krishnaiah, P.R., Rao, C.R. (eds.). New York: Elsevier. [142] Kuan, C.-M., Hsu, C.-C., 1998. Change-point estimation of fractionally integrated processes. Journal of Time Series Analysis 19, 693-708. [143] Kulperger, R.J., 1987a. On the residuals of autoregressive processes and polynomial regression. Stochastic Processes and their Applications 21, 107-118. [144] Kulperger, R.J., 1987b. Some remarks on regression residuals with autoregressive errors and their residual processes. Journal of Applied Probability 24, 668-678. [145] Kunitomo, N., Sato, S., 1995. Tables of limiting distributions useful for testing unit roots and co-integration with multiple structural changes. Manuscript, Department of Economics, University of Tokyo. [146] Kuo, B.-S., 1998. Test for partial parameter stability in regressions with I(1) processes. Journal of Econometrics 86, 337-368. [147] Kurozumi, E., 2002. Testing for stationarity with a break. Journal of Econometrics 108, 63-99. [148] Kurozumi, E., 2004. Detection of structural change in the long-run persistence in a univariate time series. Unpublished manuscript, Department of Economics, Hitotsubashi University. [149] Kurozumi, E., Arai, Y., 2004. Efficient estimation and inference in cointegrating regressions with structural breaks. Unpublished manuscript, Department of Economics, Hitotsubashi University. [150] Kwiatkowski, D., Phillips, P.C.B., Schmidt, P., Shin, Y., 1992. Testing the null hypothesis of stationarity against the alternative of a unit root: how sure are we that economic time series have a unit root. Journal of Econometrics 54, 159-178. 82 [151] Lanne, M., Lütkepohl, H., 2002. Unit root tests for time series with level shifts: a comparison of different proposals. Economics Letters 75, 109-114. [152] Lanne, M., Lütkepohl, H., Saikkonen, P., 2002. Comparison of unit root tests for time series with level shifts. Journal of Time Series Analysis 23, 667-685. [153] Lanne, M., Lütkepohl, H., Saikkonen, P., 2003. Test procedures for unit roots in time series with level shifts at unknown time. Oxford Bulletin of Economics and Statistics 65, 91-115. [154] Lavielle, M., Moulines, E., 2000. Least-squares estimation of an unknown number of shifts in a time series. Journal of Time Series Analysis 21, 33-59. [155] Lazarová, Š., 2005. Testing for structural change in regression with long memory errors. Forthcoming in the Journal of Econometrics. [156] Lee, J., 2000. On the end-point issue in unit root tests in the presence of a structural break. Economics Letters 68, 7-11. [157] Lee, J., Huang, C.J., Shin, Y., 1997. On stationary tests in the presence of structural breaks. Economics Letters 55, 165-172. [158] Lee, J., Strazicich, M.C., 2001a. Break point estimation and spurious rejections with endogenous unit root tests. Oxford Bulletin of Economics and Statistics 63, 535-558. [159] Lee, J., Strazicich, M.C., 2001b. Testing the null of stationarity in the presence of a structural break. Applied Economics Letters 8, 377-382. [160] Leybourne, S., Kim, T.-H., Smith, V., Newbold, P., 2003. Tests for a change in persistence against the null of difference-stationarity. Econometrics Journal 6, 291-311. [161] Leybourne, S.J., Mills, T.C., Newbold, P., 1998. Spurious rejections by Dickey-Fuller tests in the presence of a break under the null. Journal of Econometrics 87, 191-203. [162] Leybourne, S.J., Newbold, P., 2000. Behavior of the standard and symmetric DickeyFuller type tests when there is a break under the null hypothesis. Econometrics Journal 3, 1-15. [163] Leybourne, S.J., Newbold, P., 2003. Spurious rejections by cointegration tests induced by structural breaks. Applied Economics 35, 1117-1121. [164] Leybourne, S.J., Taylor, A.M.R., 2004. On tests for changes in persistence. Economics Letters 84, 107-115. [165] Leybourne, S.J., Taylor, A.M.R., Kim, T.-H., 2003. An unbiased test for a change in persistence. Unpublished manuscript, Department of Economics, University of Birmingham. 83 [166] Liu, J., Wu, S., Zidek, J.V., 1997. On segmented multivariate regressions. Statistica Sinica 7, 497-525. [167] Lo, A., 1991. Long-term memory in stock market prices. Econometrica 59, 1279-1313. [168] Lobato, I.N., Savin, N.E., 1998. Real and spurious long-memory properties of stockmarket data. Journal of Business and Economics Statistics 16, 261-268. [169] Lumsdaine, R.L., Papell, D.H., 1997. Multiple trend breaks and the unit root hypothesis. Review of Economics and Statistics 79, 212-218. [170] Lütkepohl, H., Saikkonen, P., 2000. Testing for the cointegrating rank of a VAR process with a time trend. Journal of Econometrics 95, 177-198. [171] Lütkepohl, H., Saikkonen, P., Trenkler, C., 2003. Comparison of tests for cointegrating rank of a VAR process with a structural shift. Journal of Econometrics 113, 201-229. [172] Lütkepohl, H., Muller, C., Saikkonen, P., 2001. Unit root tests for time series with a structural break when the break point is known. In Nonlinear Statistical Modelling: Essays in Honor of Takeshi Amemiya (eds. Hsiao, C., Morimune, K., Powell, J.). Cambridge: Cambridge University Press, 327-348. [173] Lütkepohl, H., Saikkonen, P., Trenkler, C., 2004. Testing for the cointegrating rank of a VAR process with level shift at unknown time. Econometrica 72, 647-662. [174] MacNeill, I.B., 1974. Tests for change of parameter at unknown time and distributions on some related functionals of Brownian motion. Annals of Statistics 2, 950-962. [175] MacNeill, I.B., 1978. Properties of sequences of partial sums of polynomial regression residuals with applications to tests for change of regression at unknown times. The Annals of Statistics 6, 422-433. [176] Maddala, G.S., Kim, I.M., 1998. Unit Roots, Cointegration and Structural Change. Cambridge University Press, Cambridge. [177] Mandelbrot, B.B., Taqqu, M.S., 1979. Robust R/S analysis of long run serial correlation. In Proceedings of the 42nd Session of the International Statistical Institute, Vol. 2, 69-99. [178] Mankiw, N.G., Miron, J.A., Weil, D.N., 1987. The adjustment of expectations to change in regime: a study of the founding of the federal reserve. American Economic Review 77, 358-374. [179] Marriott, J., Newbold, P., 2000. The strength of evidence for unit autoregressive roots and structural breaks: a Bayesian perspective. Journal of Econometrics 98, 1-25. [180] Montañés, A., 1997. Level shifts, unit roots and misspecification of the breaking date. Economics Letters 54, 7-13. 84 [181] Montañés, A., Olloqui, I., 1999. Misspecification of the breaking date in segmented trend variables: effect on the unit root tests. Economics Letters 65, 301-307. [182] Montañés, A., Olloqui, I., Calvo, E., 2005. Selection of the break in the Perron-type tests. Forthcoming in the Journal of Econometrics. [183] Montañés, A., Reyes, M., 1998. Effect of a shift in the trend function on Dickey-Fuller unit root tests. Econometric Theory 14, 355-363. [184] Montañés, A., Reyes, M., 1999. The asymptotic behavior of the Dickey-Fuller tests under the crash hypothesis. Statistics and Probability Letters 42, 81-89. [185] Montañés, A., Reyes, M., 2000. Structural breaks, unit roots and methods for removing the autocorrelation pattern. Statistics and Probablity Letters 48, 401-409. [186] Nabeya, S., Tanaka, K., 1988. Asymptotic theory of a test for the constancy of regression coefficients against the random walk alternative. Annals of Statistics 16, 218-235. [187] Nadler, J., Robbins, N.B., 1971. Some characteristics of Page’s two-sided procedure for detecting a change in a location parameter. The Annals of Mathematical Statistics 42, 538-551. [188] Nelson, C.R., Plosser, C.I., 1982. Trends and Random Walks in Macroeconomics Time Series: Some Evidence and Implications. Journal of Monetary Economics 10, 139-162. [189] Ng, S., Perron, P., 1995. Unit root tests in ARMA models with data dependent methods for selection of the truncation lag. Journal of the American Statistical Association 90, 268-281. [190] Ng, S., Perron, P., 2001. Lag length selection and the construction of unit root tests with good size and power. Econometrica 69, 1519-1554. [191] Noh, J., Kim, T.-H., 2003. Behavior of cointegration tests in the presence of structural breaks in variance. Applied Economics Letters 10, 999-1002. [192] Nunes, L. C., Kuan, C.-M., Newbold, P., 1995. Spurious Break, Econometric Theory 11, 736-749. [193] Nunes, L.C., Newbold, P., Kuan, C.-M., 1996. Spurious number of breaks. Economics Letters 50, 175-178. [194] Nunes, L.C., Newbold, P., Kuan, C.-M., 1997. Testing for unit roots with breaks: evidence on the great crash and the unit root hypothesis reconsidered. Oxford Bulletin of Economics and Statistics 59, 435-448. [195] Nyblom, J., 1989. Testing the constancy of parameters over time. Journal of the American Statistical Association 84, 223-230. 85 [196] Nyblom, J., Harvey, A.C., 2000. Tests of common stochstic trends. Econometric Theory 16, 176-199. [197] Nyblom, J., Mäkeläinen, T., 1983. Comparisons of tests for the presence of random walk coefficients in a simple linear model. Journal of the American Statistical Association 78, 856-864. [198] Ohara, H.I., 1999. A unit root test with multiple trend breaks: a theory and application to US and Japanese macroeconomic time-series. The Japanese Economic Review 50, 266-290. [199] Page, E. S., 1955. A test for a change in a parameter occurring at an unknown point. Biometrika 42, 523-527. [200] Page, E. S., 1957. On problems in which a change in a parameter occurs at an unknown point. Biometrika 44, 248-252. [201] Park, J.Y., Canonical cointegrating regressions. Econometrica 60, 119-143. [202] Perron, P., 1989. The great crash, the oil price shock and the unit root hypothesis. Econometrica 57, 1361-1401. [203] Perron, P., 1990. Testing for a Unit Root in a Time Series with a Changing Mean. Journal of Business and Economic Statistics 8, 153-162. [204] Perron, P., 1991, A test for changes in a polynomial trend function for a dynamic time series. Research Memorandum No. 363, Econometric Research Program, Princeton University. [205] Perron, P., 1994. Trend, unit root and structural change in macroeconomic time series. In Cointegration for the Applied Economist, Rao, B.B. (ed.), 1994, Basingstoke: Macmillan Press, 113-146. [206] Perron, P., 1997a. Further evidence from breaking trend functions in macroeconomic variables. Journal of Econometrics 80, 355-385. [207] Perron, P., 1997b. L’estimation de modèles avec changements structurels multiples. Actualité Économique 73, 457-505. [208] Perron, P., 2005. A note on the finite sample power function of the dynamic cusum and cusum of squares tests. Manuscript in preparation, Department of Economics, Boston University. [209] Perron, P., Qu, Z., 2004. An analytical evaluation of the log-periodogram estimate in the presence of level shifts and its implications for stock return volatility. Manuscript, Department of Economics, Boston University. 86 [210] Perron, P., Qu, Z., 2005. Estimating restricted structural change models. Forthcoming in Journal of Econometrics. [211] Perron, P., Rodríguez, G.H., 2003. GLS detrending, efficient unit root tests and structural change. Journal of Econometrics 115, 1-27. [212] Perron, P., Vogelsang, T,J., 1992a. Nonstationarity and level shifts with an application to purchasing power parity. Journal of Business and Economic Statistics 10, 301-320. [213] Perron, P., Vogelsang, T,J., 1992b. Testing for a unit root in a time series with a changing mean: corrections and extensions. Journal of Business and Economic Statistics 10, 467-470. [214] Perron, P., Vogelsang, T,J., 1993a. The great crash, the oil price shock and the unit root hypothesis: erratum. Econometrica 61, 248-249. [215] Perron, P., Vogelsang, T,J., 1993b. A note on the additive outlier model with breaks. Revista de Econometria 13, 181-201. [216] Perron, P., Wada, T., 2005. Trends and cycles: a new approach and explanations of some old puzzles. Manuscript, Department of Economics, Boston University. [217] Perron, P., Yabu, T., 2004. Estimating deterministic trends with an integrated or stationary noise component. Manuscript, Department of Economics, Boston University. [218] Perron, P., Yabu, T., 2005. Testing for shifts in trend with an integrated or stationary noise component. Manuscript in preparation, Department of Economics, Boston University. [219] Perron, P., Zhu, X., 2005. Structural breaks with stochastic and deterministic trends. Forthcoming in the Journal of Econometrics. [220] Pesaran, H.M., Smith, R.P., Yeo, J.S., 1985. Testing for structural stability and predictive failure: a review. The Manchester School of Economic & Social Studies 53, 281-295. [221] Phillips, P.C.B., Ouliaris, S., 1990. Asymptotic properties of residual based tests for cointegration. Econometrica 58, 165-193. [222] Phillips, P.C.B., Perron, P., 1988. Testing for a unit root in time series regression. Biometrika 75, 335-346. [223] Picard, D., 1985. Testing and estimating change-points in time series. Journal of Applied Probability 17, 841-867. [224] Pitarakis, J.-Y., 2004. Least squares estimation and tests of breaks in mean and variance under misspecification. Econometrics Journal 7, 32-54. 87 [225] Ploberger, W., Krämer, W., 1990. The local power of the cusum and cusum of squares tests. Econometric Theory 6, 335-347. [226] Ploberger, W., Krämer, W., 1992. The CUSUM test with OLS residuals. Econometrica 60, 271-285. [227] Ploberger, W., Krämer, W., Kontrus, K., 1989. A new test for structural stability in the linear regression model. Journal of Econometrics 40, 307-318. [228] Qu, Z., 2004. Searching for cointegration in a dynamic system. Manuscript, Department of Economics, Boston University. [229] Qu, Z., Perron, P., 2005. Estimating and testing multiple structural changes in multivariate regressions. Manuscript, Department of Economics, Boston University. [230] Quandt, R.E., 1958. The estimation of the parameters of a linear regression system obeying two separate regimes. Journal of the American Statistical Association 53, 873880. [231] Quandt, R.E., 1960. Tests of the hypothesis that a linear regression system obeys two separate regimes. Journal of the American Statistical Association 55, 324-330. [232] Quintos, C.E., 1997. Stability tests in error correction models. Journal of Econometrics 82, 289-315. [233] Qunitos, C.E., Phillips, P.C.B., 1993. Parameter constancy in cointegrated regressions. Empirical Economics 18, 675-706. [234] Rappoport, P., Reichlin, L, 1989. Segmented trends and non-stationary time series. Economic Journal 99, 168-177. [235] Robinson, P.M., 1994a. Time series with strong dependence, in: C. Sims, ed., Advances in econometrics, 6th world congress (Cambridge University Press, Cambridge), 47-95. [236] Robinson, P.M., 1994b. Efficient tests of nonstationary hypotheses. Journal of the American Statistical Association 89, 1420-1437. [237] Roy, A., Falk, B., Fuller, W.A., 2004. Testing for trend in the presence of autoregressive errors. Journal of the American Statistical Association 99, 1082-1091. [238] Roy, A., Fuller, W.A., 2001. Estimation for autoregressive time series with a root near 1. Journal of Business and Economic Statistics 19, 482-493. [239] Saikkonen, P., 1991. Asymptotically efficient estimation of cointegrated regressions. Econometric Theory 7, 1-21. [240] Saikkonen, P., Lütkepohl, H., 2000a. Testing for the cointegrating rank of a VAR process with structural shifts. Journal of Business and Economic Statistics 18, 451464. 88 [241] Saikkonen, P., Lütkepohl, H., 2000b. Trend adjustment prior to testing for the cointegrating rank of a vector autoregressive process. Journal of Time Series Analysis 21, 435-456. [242] Saikkonen, P., Lütkepohl, H., 2001. Testing for unit roots in time series with level shifts. Allgemeines Statistisches Archiv 85, 1-25. [243] Saikkonen, P., Lütkepohl, H., 2002. Testing for a unit root in a time series with a level shift at unknown time. Econometric Theory 18, 313-348. [244] Saikkonen, P., Lütkepohl, H., Trenkler, C., 2004. Break date estimation and cointegration testing in VAR processes with level shift. Manuscript, Humboldt University Berlin and University of Helsinki. [245] Sen, P.K., 1980.Asymptotic theory of some tests for a possible change in the regression slope occurring at an unknown time point. Zeitschrift für Wahrscheinlichkeitstheorie und verwandte Gebiete 52, 203-218. [246] Sen, P.K., 1982. Invariance principles for recursive residuals. The Annals of Statistics 10, 307-312. [247] Seo, B., 1998. Tests for structural change in cointegrated systems. Econometric Theory 14, 222-259. [248] Shaban, S.A., 1980. Change point problem and two-phase regression: an annotated bibliography. International Statistical Review 48, 83-93. [249] Shin, Y., 1994. A residual-based test of the null of cointegration against the alternative of no cointegration. Econometric Theory 10, 91-115. [250] Sibbertsen, P., 2004. Long memory versus structural breaks: an overview. Statistical Papers 45, 465-515. [251] Siegmund, D., 1988. Confidence sets in change-point problems. International Statistical Review 56, 31-48. [252] Sowell, F., 1996. Optimal tests for parameter instability in the generalized method of moments framework. Econometrica 64, 1085-1107. [253] Stock, J.H., 1994. Unit roots, structural breaks and trends. In Handbook of Econometrics, vol. 4 (Engle, R.F., MacFaden, D., eds), Elsevier, 2740-2841. [254] Stock, J.H., Watson. M.W., 1993. A simple estimator of cointegrating vectors in highre order integrated systems. Econometrica 64, 783-820. [255] Tang, S. M., MacNeill, I.B., 1993. The effect of serial correlation on tests for parameter change at unknown time. Annals of Statistics 21, 552-575. 89 [256] Teverovsky, V., Taqqu, M., 1997. Testing for long-range dependence in the presence of shifting means or a slowly declining trend, using a variance-type estimator. Journal of Time Series Analysis 18, 279-304. [257] Tong, H., 1990. Non-linear Time Series, A Dynamical System Approach. Oxford University Press. [258] van Dijk, D., Terasvirta, T., Franses, P.H., 2002. Smooth transition autoregressive models - a survey of recent developments. Econometric Reviews 21, 1-47. [259] Vogelsang, T.J., 1997. Wald-type tests for detecting breaks in the trend function of a dynamic time series. Econometric Theory 13, 818-849. [260] Vogelsang, T.J., 1998a. Trend function hypothesis testing in the presence of serial correlation. Econometrica 66, 123-148. [261] Vogelsang, T.J., 1998b. Testing for a shift in mean without having to estimate serialcorrelation parameters. Journal of Business and Economic Statistics 16, 73-80. [262] Vogelsang, T.J., 1999. Sources of nonmonotonic power when testing for a shift in mean of a dynamic time series. Journal of Econometrics 88, 283-299. [263] Vogelsang, T.J., 2001. Testing for a shift in trend when serial correlation is of unknown form. Manuscript, Department of Economics, Cornell University. [264] Vogelsang, T.J., Perron, P., 1998. Additional tests for a unit root allowing the possibility of breaks in the trend function. International Economic Review 39, 1073-1100. [265] Wang, J., Zivot, E., 2000. A Bayesian time series model of multiple structural change in level, trend and variance. Journal of Business and Economic Statistics 18, 374-386. [266] Worsley, K.J., 1986. Confidence regions and test for a change-point in a sequence of exponential family random variables. Biometrika 73, 91-104. [267] Wright, J.H., 1998. Testing for a structural break at unknown date with long-memory disturbances. Journal of Time Series Analysis 19, 369-376. [268] Yao, Y.-C., 1987. Approximating the distribution of the maximum likelihood estimate of the change-point in a sequence of independent random variables. Annals of Statistics 15, 1321-1328. [269] Yao, Y-C., 1988. Estimating the number of change-points via Schwarz’ criterion. Statistics and Probability Letters 6, 181-189. [270] Zacks, S., 1983. Survey of classical and Bayesian approaches to the change-point problem: fixed and sequential procedures of testing and estimation. In Recent Advances in Statistics, Rivzi, M.H., Rustagi, J.S., Siegmund, D. (eds.). New York: Academic Press. 90 [271] Zivot, E., Andrews, D.W.K., 1992. Further evidence on the great crash, the oil price shock and the unit root hypothesis. Journal of Business and Economic Statistics 10, 251-270. [272] Zivot, E., Phillips, P.B.C., 1994. A Bayesian analysis of trend determination in economic time series. Econometric Reviews 13, 291-336. 91