Unit root

Last updated

In probability theory and statistics, a unit root is a feature of some stochastic processes (such as random walks) that can cause problems in statistical inference involving time series models. A linear stochastic process has a unit root if 1 is a root of the process's characteristic equation. Such a process is non-stationary but does not always have a trend.

Contents

If the other roots of the characteristic equation lie inside the unit circle—that is, have a modulus (absolute value) less than one—then the first difference of the process will be stationary; otherwise, the process will need to be differenced multiple times to become stationary. [1] If there are d unit roots, the process will have to be differenced d times in order to make it stationary. [2] Due to this characteristic, unit root processes are also called difference stationary. [3] [4]

Unit root processes may sometimes be confused with trend-stationary processes; while they share many properties, they are different in many aspects. It is possible for a time series to be non-stationary, yet have no unit root and be trend-stationary. In both unit root and trend-stationary processes, the mean can be growing or decreasing over time; however, in the presence of a shock, trend-stationary processes are mean-reverting (i.e. transitory, the time series will converge again towards the growing mean, which was not affected by the shock) while unit-root processes have a permanent impact on the mean (i.e. no convergence over time). [5]

If a root of the process's characteristic equation is larger than 1, then it is called an explosive process, even though such processes are sometimes inaccurately called unit roots processes.

The presence of a unit root can be tested using a unit root test.

Definition

Consider a discrete-time stochastic process , and suppose that it can be written as an autoregressive process of order p:

Here, is a serially uncorrelated, zero-mean stochastic process with constant variance . For convenience, assume . If is a root of the characteristic equation, of multiplicity 1:

then the stochastic process has a unit root or, alternatively, is integrated of order one, denoted . If m = 1 is a root of multiplicity r, then the stochastic process is integrated of order r, denoted I(r).

Example

The first order autoregressive model, , has a unit root when . In this example, the characteristic equation is . The root of the equation is .

If the process has a unit root, then it is a non-stationary time series. That is, the moments of the stochastic process depend on . To illustrate the effect of a unit root, we can consider the first order case, starting from y0 = 0:

By repeated substitution, we can write . Then the variance of is given by:

The variance depends on t since , while . Note that the variance of the series is diverging to infinity with t.

There are various tests to check for the existence of a unit root, some of them are given by:

  1. The Dickey–Fuller test (DF) or augmented Dickey–Fuller (ADF) tests
  2. Testing the significance of more than one coefficients (f-test)
  3. The Phillips–Perron test (PP)
  4. Dickey Pantula test

In addition to autoregressive (AR) and autoregressive–moving-average (ARMA) models, other important models arise in regression analysis where the model errors may themselves have a time series structure and thus may need to be modelled by an AR or ARMA process that may have a unit root, as discussed above. The finite sample properties of regression models with first order ARMA errors, including unit roots, have been analyzed. [6] [7]

Estimation when a unit root may be present

Often, ordinary least squares (OLS) is used to estimate the slope coefficients of the autoregressive model. Use of OLS relies on the stochastic process being stationary. When the stochastic process is non-stationary, the use of OLS can produce invalid estimates. Granger and Newbold called such estimates 'spurious regression' results: [8] high R2 values and high t-ratios yielding results with no real (in their context, economic) meaning.

To estimate the slope coefficients, one should first conduct a unit root test, whose null hypothesis is that a unit root is present. If that hypothesis is rejected, one can use OLS. However, if the presence of a unit root is not rejected, then one should apply the difference operator to the series. If another unit root test shows the differenced time series to be stationary, OLS can then be applied to this series to estimate the slope coefficients.

For example, in the AR(1) case, is stationary.

In the AR(2) case, can be written as where L is a lag operator that decreases the time index of a variable by one period: . If , the model has a unit root and we can define ; then

is stationary if . OLS can be used to estimate the slope coefficient, .

If the process has multiple unit roots, the difference operator can be applied multiple times.

Properties and characteristics of unit-root processes

Unit root hypothesis

The diagram above depicts an example of a potential unit root. The red line represents an observed drop in output. Green shows the path of recovery if the series has a unit root. Blue shows the recovery if there is no unit root and the series is trend-stationary. The blue line returns to meet and follow the dashed trend line while the green line remains permanently below the trend. The unit root hypothesis also holds that a spike in output will lead to levels of output higher than the past trend. Unit root hypothesis diagram.svg
The diagram above depicts an example of a potential unit root. The red line represents an observed drop in output. Green shows the path of recovery if the series has a unit root. Blue shows the recovery if there is no unit root and the series is trend-stationary. The blue line returns to meet and follow the dashed trend line while the green line remains permanently below the trend. The unit root hypothesis also holds that a spike in output will lead to levels of output higher than the past trend.

Economists debate whether various economic statistics, especially output, have a unit root or are trend-stationary. [9] A unit root process with drift is given in the first-order case by

where c is a constant term referred to as the "drift" term, and is white noise. Any non-zero value of the noise term, occurring for only one period, will permanently affect the value of as shown in the graph, so deviations from the line are non-stationary; there is no reversion to any trend line. In contrast, a trend-stationary process is given by

where k is the slope of the trend and is noise (white noise in the simplest case; more generally, noise following its own stationary autoregressive process). Here any transient noise will not alter the long-run tendency for to be on the trend line, as also shown in the graph. This process is said to be trend-stationary because deviations from the trend line are stationary.

The issue is particularly popular in the literature on business cycles. [10] [11] Research on the subject began with Nelson and Plosser whose paper on GNP and other output aggregates failed to reject the unit root hypothesis for these series. [12] Since then, a debate—entwined with technical disputes on statistical methods—has ensued. Some economists [13] argue that GDP has a unit root or structural break, implying that economic downturns result in permanently lower GDP levels in the long run. Other economists argue that GDP is trend-stationary: That is, when GDP dips below trend during a downturn it later returns to the level implied by the trend so that there is no permanent decrease in output. While the literature on the unit root hypothesis may consist of arcane debate on statistical methods, the hypothesis carries significant practical implications for economic forecasts and policies.

See also

Notes

  1. "Trend-Stationary vs. Difference-Stationary Processes - MATLAB & Simulink". uk.mathworks.com. Archived from the original on 2016-06-08. Retrieved 2016-06-05.
  2. "EViews Help". Archived from the original on 2020-05-27. Retrieved 2020-05-28.
  3. "Differencing and unit root tests" (PDF). Archived (PDF) from the original on 2016-10-18.
  4. "Non-Stationary Series" (PDF). Archived (PDF) from the original on 2014-06-11.
  5. Heino Bohn Nielsen. "Non-Stationary Time Series and Unit Root Tests" (PDF). Archived (PDF) from the original on 2016-11-30.
  6. Sargan, J. D.; Bhargava, Alok (1983). "Testing residuals from least squares regressions for being generated by the Gaussian random walk". Econometrica . 51 (1): 153–174. doi:10.2307/1912252. JSTOR   1912252.
  7. Sargan, J. D.; Bhargava, Alok (1983). "Maximum Likelihood Estimation of Regression Models with First Order Moving Average Errors when the Root Lies on the Unit Circle". Econometrica. 51 (3): 799–820. doi:10.2307/1912159. JSTOR   1912159.
  8. Granger, C. W. J.; Newbold, P. (1974). "Spurious regressions in econometrics". Journal of Econometrics . 2 (2): 111–120. CiteSeerX   10.1.1.353.2946 . doi:10.1016/0304-4076(74)90034-7.
  9. Krugman, Paul (March 3, 2009). "Roots of evil (wonkish)". The New York Times. Archived from the original on September 5, 2015. Retrieved February 7, 2017.
  10. Hegwood, Natalie; Papell, David H. (2007). "Are Real GDP Levels Trend, Difference, or Regime-Wise Trend Stationary? Evidence from Panel Data Tests Incorporating Structural Change" (PDF). Southern Economic Journal. 74 (1): 104–113. doi:10.1002/j.2325-8012.2007.tb00829.x. JSTOR   20111955. Archived (PDF) from the original on 2022-06-14. Retrieved 2021-08-14.
  11. Lucke, Bernd (2005). "Is Germany's GDP trend-stationary? A measurement-with-theory approach" (PDF). Jahrbücher für Nationalökonomie und Statistik. 225 (1): 60–76. doi:10.1515/jbnst-2005-0105. S2CID   209856533. Archived (PDF) from the original on 2013-12-24. Retrieved 2013-07-29.
  12. Nelson, Charles R.; Plosser, Charles I. (1982). "Trends and Random Walks in Macroeconomic Time Series: Some Evidence and Implications". Journal of Monetary Economics . 10 (2): 139–162. doi:10.1016/0304-3932(82)90012-5.
  13. Olivier Blanchard Archived 2009-08-26 at the Wayback Machine with the International Monetary Fund makes the claim that after a banking crisis "on average, output does not go back to its old trend path, but remains permanently below it."

Related Research Articles

In statistics, the Gauss–Markov theorem states that the ordinary least squares (OLS) estimator has the lowest sampling variance within the class of linear unbiased estimators, if the errors in the linear regression model are uncorrelated, have equal variances and expectation value of zero. The errors do not need to be normal, nor do they need to be independent and identically distributed. The requirement that the estimator be unbiased cannot be dropped, since biased estimators exist with lower variance. See, for example, the James–Stein estimator, ridge regression, or simply any degenerate estimator.

In mathematics and statistics, a stationary process is a stochastic process whose unconditional joint probability distribution does not change when shifted in time. Consequently, parameters such as mean and variance also do not change over time. If you draw a line through the middle of a stationary process then it should be flat; it may have 'seasonal' cycles around the trend line, but overall it does not trend up nor down.

In the statistical analysis of time series, autoregressive–moving-average (ARMA) models provide a parsimonious description of a (weakly) stationary stochastic process in terms of two polynomials, one for the autoregression (AR) and the second for the moving average (MA). The general ARMA model was described in the 1951 thesis of Peter Whittle, Hypothesis testing in time series analysis, and it was popularized in the 1970 book by George E. P. Box and Gwilym Jenkins.

In statistics, econometrics, and signal processing, an autoregressive (AR) model is a representation of a type of random process; as such, it is used to describe certain time-varying processes in nature, economics, behavior, etc. The autoregressive model specifies that the output variable depends linearly on its own previous values and on a stochastic term ; thus the model is in the form of a stochastic difference equation which should not be confused with a differential equation. Together with the moving-average (MA) model, it is a special case and key component of the more general autoregressive–moving-average (ARMA) and autoregressive integrated moving average (ARIMA) models of time series, which have a more complicated stochastic structure; it is also a special case of the vector autoregressive model (VAR), which consists of a system of more than one interlocking stochastic difference equation in more than one evolving random variable.

In statistics and econometrics, and in particular in time series analysis, an autoregressive integrated moving average (ARIMA) model is a generalization of an autoregressive moving average (ARMA) model. To better comprehend the data or to forecast upcoming series points, both of these models are fitted to time series data. ARIMA models are applied in some cases where data show evidence of non-stationarity in the sense of mean, where an initial differencing step can be applied one or more times to eliminate the non-stationarity of the mean function. When the seasonality shows in a time series, the seasonal-differencing could be applied to eliminate the seasonal component. Since the ARMA model, according to the Wold's decomposition theorem, is theoretically sufficient to describe a regular wide-sense stationary time series, we are motivated to make stationary a non-stationary time series, e.g., by using differencing, before we can use the ARMA model. Note that if the time series contains a predictable sub-process, the predictable component is treated as a non-zero-mean but periodic component in the ARIMA framework so that it is eliminated by the seasonal differencing.

In the statistical analysis of time series, a trend-stationary process is a stochastic process from which an underlying trend can be removed, leaving a stationary process. The trend does not have to be linear.

In statistics, ordinary least squares (OLS) is a type of linear least squares method for choosing the unknown parameters in a linear regression model by the principle of least squares: minimizing the sum of the squares of the differences between the observed dependent variable in the input dataset and the output of the (linear) function of the independent variable.

Cointegration is a statistical property of a collection (X1X2, ..., Xk) of time series variables. First, all of the series must be integrated of order d (see Order of integration). Next, if a linear combination of this collection is integrated of order less than d, then the collection is said to be co-integrated. Formally, if (X,Y,Z) are each integrated of order d, and there exist coefficients a,b,c such that aX + bY + cZ is integrated of order less than d, then X, Y, and Z are cointegrated. Cointegration has become an important property in contemporary time series analysis. Time series often have trends—either deterministic or stochastic. In an influential paper, Charles Nelson and Charles Plosser (1982) provided statistical evidence that many US macroeconomic time series (like GNP, wages, employment, etc.) have stochastic trends.

Vector autoregression (VAR) is a statistical model used to capture the relationship between multiple quantities as they change over time. VAR is a type of stochastic process model. VAR models generalize the single-variable (univariate) autoregressive model by allowing for multivariate time series. VAR models are often used in economics and the natural sciences.

In statistics, a unit root test tests whether a time series variable is non-stationary and possesses a unit root. The null hypothesis is generally defined as the presence of a unit root and the alternative hypothesis is either stationarity, trend stationarity or explosive root depending on the test used.

In statistics, the Dickey–Fuller test tests the null hypothesis that a unit root is present in an autoregressive (AR) time series model. The alternative hypothesis is different depending on which version of the test is used, but is usually stationarity or trend-stationarity. The test is named after the statisticians David Dickey and Wayne Fuller, who developed it in 1979.

Difference in differences is a statistical technique used in econometrics and quantitative research in the social sciences that attempts to mimic an experimental research design using observational study data, by studying the differential effect of a treatment on a 'treatment group' versus a 'control group' in a natural experiment. It calculates the effect of a treatment on an outcome by comparing the average change over time in the outcome variable for the treatment group to the average change over time for the control group. Although it is intended to mitigate the effects of extraneous factors and selection bias, depending on how the treatment group is chosen, this method may still be subject to certain biases.

In statistics, an augmented Dickey–Fuller test (ADF) tests the null hypothesis that a unit root is present in a time series sample. The alternative hypothesis is different depending on which version of the test is used, but is usually stationarity or trend-stationarity. It is an augmented version of the Dickey–Fuller test for a larger and more complicated set of time series models.

In statistics, Wold's decomposition or the Wold representation theorem, named after Herman Wold, says that every covariance-stationary time series can be written as the sum of two time series, one deterministic and one stochastic.

In statistics, autoregressive fractionally integrated moving average models are time series models that generalize ARIMA (autoregressive integrated moving average) models by allowing non-integer values of the differencing parameter. These models are useful in modeling time series with long memory—that is, in which deviations from the long-run mean decay more slowly than an exponential decay. The acronyms "ARFIMA" or "FARIMA" are often used, although it is also conventional to simply extend the "ARIMA(p, d, q)" notation for models, by simply allowing the order of differencing, d, to take fractional values. Fractional differencing and the ARFIMA model were introduced in the early 1980s by Clive Granger, Roselyne Joyeux, and Jonathan Hosking.

In statistics, principal component regression (PCR) is a regression analysis technique that is based on principal component analysis (PCA). More specifically, PCR is used for estimating the unknown regression coefficients in a standard linear regression model.

In time series analysis, the moving-average model, also known as moving-average process, is a common approach for modeling univariate time series. The moving-average model specifies that the output variable is cross-correlated with a non-identical to itself random-variable.

An error correction model (ECM) belongs to a category of multiple time series models most commonly used for data where the underlying variables have a long-run common stochastic trend, also known as cointegration. ECMs are a theoretically-driven approach useful for estimating both short-term and long-term effects of one time series on another. The term error-correction relates to the fact that last-period's deviation from a long-run equilibrium, the error, influences its short-run dynamics. Thus ECMs directly estimate the speed at which a dependent variable returns to equilibrium after a change in other variables.

In statistics, the Johansen test, named after Søren Johansen, is a procedure for testing cointegration of several, say k, I(1) time series. This test permits more than one cointegrating relationship so is more generally applicable than the Engle–Granger test which is based on the Dickey–Fuller test for unit roots in the residuals from a single (estimated) cointegrating relationship.

In statistics and econometrics, the ADF-GLS test is a test for a unit root in an economic time series sample. It was developed by Elliott, Rothenberg and Stock (ERS) in 1992 as a modification of the augmented Dickey–Fuller test (ADF).