Autoregressive conditional heteroskedasticity

Last updated

In econometrics, the autoregressive conditional heteroskedasticity (ARCH) model is a statistical model for time series data that describes the variance of the current error term or innovation as a function of the actual sizes of the previous time periods' error terms; [1] often the variance is related to the squares of the previous innovations. The ARCH model is appropriate when the error variance in a time series follows an autoregressive (AR) model; if an autoregressive moving average (ARMA) model is assumed for the error variance, the model is a generalized autoregressive conditional heteroskedasticity (GARCH) model. [2]

Contents

ARCH models are commonly employed in modeling financial time series that exhibit time-varying volatility and volatility clustering, i.e. periods of swings interspersed with periods of relative calm. ARCH-type models are sometimes considered to be in the family of stochastic volatility models, although this is strictly incorrect since at time t the volatility is completely predetermined (deterministic) given previous values. [3]

Model specification

To model a time series using an ARCH process, let denote the error terms (return residuals, with respect to a mean process), i.e. the series terms. These are split into a stochastic piece and a time-dependent standard deviation characterizing the typical size of the terms so that

The random variable is a strong white noise process. The series is modeled by

,
where and .

An ARCH(q) model can be estimated using ordinary least squares. A method for testing whether the residuals exhibit time-varying heteroskedasticity using the Lagrange multiplier test was proposed by Engle (1982). This procedure is as follows:

  1. Estimate the best fitting autoregressive model AR(q) .
  2. Obtain the squares of the error and regress them on a constant and q lagged values:
    where q is the length of ARCH lags.
  3. The null hypothesis is that, in the absence of ARCH components, we have for all . The alternative hypothesis is that, in the presence of ARCH components, at least one of the estimated coefficients must be significant. In a sample of T residuals under the null hypothesis of no ARCH errors, the test statistic T'R² follows distribution with q degrees of freedom, where is the number of equations in the model which fits the residuals vs the lags (i.e. ). If T'R² is greater than the Chi-square table value, we reject the null hypothesis and conclude there is an ARCH effect in the ARMA model. If T'R² is smaller than the Chi-square table value, we do not reject the null hypothesis.

GARCH

If an autoregressive moving average (ARMA) model is assumed for the error variance, the model is a generalized autoregressive conditional heteroskedasticity (GARCH) model. [2]

In that case, the GARCH (p, q) model (where p is the order of the GARCH terms and q is the order of the ARCH terms ), following the notation of the original paper, is given by

Generally, when testing for heteroskedasticity in econometric models, the best test is the White test. However, when dealing with time series data, this means to test for ARCH and GARCH errors.

Exponentially weighted moving average (EWMA) is an alternative model in a separate class of exponential smoothing models. As an alternative to GARCH modelling it has some attractive properties such as a greater weight upon more recent observations, but also drawbacks such as an arbitrary decay factor that introduces subjectivity into the estimation.

GARCH(p, q) model specification

The lag length p of a GARCH(p, q) process is established in three steps:

  1. Estimate the best fitting AR(q) model
    .
  2. Compute and plot the autocorrelations of by
  3. The asymptotic, that is for large samples, standard deviation of is . Individual values that are larger than this indicate GARCH errors. To estimate the total number of lags, use the Ljung–Box test until the value of these are less than, say, 10% significant. The Ljung–Box Q-statistic follows distribution with n degrees of freedom if the squared residuals are uncorrelated. It is recommended to consider up to T/4 values of n. The null hypothesis states that there are no ARCH or GARCH errors. Rejecting the null thus means that such errors exist in the conditional variance.

NGARCH

NAGARCH

Nonlinear Asymmetric GARCH(1,1) (NAGARCH) is a model with the specification: [6] [7]

,
where and , which ensures the non-negativity and stationarity of the variance process.

For stock returns, parameter is usually estimated to be positive; in this case, it reflects a phenomenon commonly referred to as the "leverage effect", signifying that negative returns increase future volatility by a larger amount than positive returns of the same magnitude. [6] [7]

This model should not be confused with the NARCH model, together with the NGARCH extension, introduced by Higgins and Bera in 1992. [8]

IGARCH

Integrated Generalized Autoregressive Conditional heteroskedasticity (IGARCH) is a restricted version of the GARCH model, where the persistent parameters sum up to one, and imports a unit root in the GARCH process. [9] The condition for this is

.

EGARCH

The exponential generalized autoregressive conditional heteroskedastic (EGARCH) model by Nelson & Cao (1991) is another form of the GARCH model. Formally, an EGARCH(p,q):

where , is the conditional variance, , , , and are coefficients. may be a standard normal variable or come from a generalized error distribution. The formulation for allows the sign and the magnitude of to have separate effects on the volatility. This is particularly useful in an asset pricing context. [10] [11]

Since may be negative, there are no sign restrictions for the parameters.

GARCH-M

The GARCH-in-mean (GARCH-M) model adds a heteroskedasticity term into the mean equation. It has the specification:

The residual is defined as:

QGARCH

The Quadratic GARCH (QGARCH) model by Sentana (1995) is used to model asymmetric effects of positive and negative shocks.

In the example of a GARCH(1,1) model, the residual process is

where is i.i.d. and

GJR-GARCH

Similar to QGARCH, the Glosten-Jagannathan-Runkle GARCH (GJR-GARCH) model by Glosten, Jagannathan and Runkle (1993) also models asymmetry in the ARCH process. The suggestion is to model where is i.i.d., and

where if , and if .

TGARCH model

The Threshold GARCH (TGARCH) model by Zakoian (1994) is similar to GJR GARCH. The specification is one on conditional standard deviation instead of conditional variance:

where if , and if . Likewise, if , and if .

fGARCH

Hentschel's fGARCH model, [12] also known as Family GARCH, is an omnibus model that nests a variety of other popular symmetric and asymmetric GARCH models including APARCH, GJR, AVGARCH, NGARCH, etc.

COGARCH

In 2004, Claudia Klüppelberg, Alexander Lindner and Ross Maller proposed a continuous-time generalization of the discrete-time GARCH(1,1) process. The idea is to start with the GARCH(1,1) model equations

and then to replace the strong white noise process by the infinitesimal increments of a Lévy process , and the squared noise process by the increments , where

is the purely discontinuous part of the quadratic variation process of . The result is the following system of stochastic differential equations:

where the positive parameters , and are determined by , and . Now given some initial condition , the system above has a pathwise unique solution which is then called the continuous-time GARCH (COGARCH) model. [13]

ZD-GARCH

Unlike GARCH model, the Zero-Drift GARCH (ZD-GARCH) model by Li, Zhang, Zhu and Ling (2018) [14] lets the drift term in the first order GARCH model. The ZD-GARCH model is to model , where is i.i.d., and

The ZD-GARCH model does not require , and hence it nests the Exponentially weighted moving average (EWMA) model in "RiskMetrics". Since the drift term , the ZD-GARCH model is always non-stationary, and its statistical inference methods are quite different from those for the classical GARCH model. Based on the historical data, the parameters and can be estimated by the generalized QMLE method.

Spatial GARCH

Spatial GARCH processes by Otto, Schmid and Garthoff (2018) [15] are considered as the spatial equivalent to the temporal generalized autoregressive conditional heteroscedasticity (GARCH) models. In contrast to the temporal ARCH model, in which the distribution is known given the full information set for the prior periods, the distribution is not straightforward in the spatial and spatiotemporal setting due to the interdependence between neighboring spatial locations. The spatial model is given by and

where denotes the -th spatial location and refers to the -th entry of a spatial weight matrix and for . The spatial weight matrix defines which locations are considered to be adjacent.

Gaussian process-driven GARCH

In a different vein, the machine learning community has proposed the use of Gaussian process regression models to obtain a GARCH scheme. [16] This results in a nonparametric modelling scheme, which allows for: (i) advanced robustness to overfitting, since the model marginalises over its parameters to perform inference, under a Bayesian inference rationale; and (ii) capturing highly-nonlinear dependencies without increasing model complexity.[ citation needed ]

Related Research Articles

In statistics, the Gauss–Markov theorem states that the ordinary least squares (OLS) estimator has the lowest sampling variance within the class of linear unbiased estimators, if the errors in the linear regression model are uncorrelated, have equal variances and expectation value of zero. The errors do not need to be normal for the theorem to apply, nor do they need to be independent and identically distributed.

<span class="mw-page-title-main">Ordinary least squares</span> Method for estimating the unknown parameters in a linear regression model

In statistics, ordinary least squares (OLS) is a type of linear least squares method for choosing the unknown parameters in a linear regression model by the principle of least squares: minimizing the sum of the squares of the differences between the observed dependent variable in the input dataset and the output of the (linear) function of the independent variable. Some sources consider OLS to be linear regression.

In general relativity, the Gibbons–Hawking–York boundary term is a term that needs to be added to the Einstein–Hilbert action when the underlying spacetime manifold has a boundary.

<span class="mw-page-title-main">Simple linear regression</span> Linear regression model with a single explanatory variable

In statistics, simple linear regression (SLR) is a linear regression model with a single explanatory variable. That is, it concerns two-dimensional sample points with one independent variable and one dependent variable and finds a linear function that, as accurately as possible, predicts the dependent variable values as a function of the independent variable. The adjective simple refers to the fact that the outcome variable is related to a single predictor.

In statistics, the Breusch–Pagan test, developed in 1979 by Trevor Breusch and Adrian Pagan, is used to test for heteroskedasticity in a linear regression model. It was independently suggested with some extension by R. Dennis Cook and Sanford Weisberg in 1983. Derived from the Lagrange multiplier test principle, it tests whether the variance of the errors from a regression is dependent on the values of the independent variables. In that case, heteroskedasticity is present.

In statistics, generalized least squares (GLS) is a method used to estimate the unknown parameters in a linear regression model. It is used when there is a non-zero amount of correlation between the residuals in the regression model. GLS is employed to improve statistical efficiency and reduce the risk of drawing erroneous inferences, as compared to conventional least squares and weighted least squares methods. It was first described by Alexander Aitken in 1935.

<span class="mw-page-title-main">Electromagnetic stress–energy tensor</span> Type of stress-energy tensor

In relativistic physics, the electromagnetic stress–energy tensor is the contribution to the stress–energy tensor due to the electromagnetic field. The stress–energy tensor describes the flow of energy and momentum in spacetime. The electromagnetic stress–energy tensor contains the negative of the classical Maxwell stress tensor that governs the electromagnetic interactions.

In statistics, stochastic volatility models are those in which the variance of a stochastic process is itself randomly distributed. They are used in the field of mathematical finance to evaluate derivative securities, such as options. The name derives from the models' treatment of the underlying security's volatility as a random process, governed by state variables such as the price level of the underlying security, the tendency of volatility to revert to some long-run mean value, and the variance of the volatility process itself, among others.

A ratio distribution is a probability distribution constructed as the distribution of the ratio of random variables having two other known distributions. Given two random variables X and Y, the distribution of the random variable Z that is formed as the ratio Z = X/Y is a ratio distribution.

<span class="mw-page-title-main">Truncated normal distribution</span> Type of probability distribution

In probability and statistics, the truncated normal distribution is the probability distribution derived from that of a normally distributed random variable by bounding the random variable from either below or above. The truncated normal distribution has wide applications in statistics and econometrics.

The topic of heteroskedasticity-consistent (HC) standard errors arises in statistics and econometrics in the context of linear regression and time series analysis. These are also known as heteroskedasticity-robust standard errors, Eicker–Huber–White standard errors, to recognize the contributions of Friedhelm Eicker, Peter J. Huber, and Halbert White.

In financial econometrics, an autoregressive conditional duration model considers irregularly spaced and autocorrelated intertrade durations. ACD is analogous to GARCH. In a continuous double auction waiting times between two consecutive trades vary at random.

In regression, mean response and predicted response, also known as mean outcome and predicted outcome, are values of the dependent variable calculated from the regression parameters and a given value of the independent variable. The values of these two responses are the same, but their calculated variances are different. The concept is a generalization of the distinction between the standard error of the mean and the sample standard deviation.

Financial models with long-tailed distributions and volatility clustering have been introduced to overcome problems with the realism of classical financial models. These classical models of financial time series typically assume homoskedasticity and normality cannot explain stylized phenomena such as skewness, heavy tails, and volatility clustering of the empirical asset returns in finance. In 1963, Benoit Mandelbrot first used the stable distribution to model the empirical distributions which have the skewness and heavy-tail property. Since -stable distributions have infinite -th moments for all , the tempered stable processes have been proposed for overcoming this limitation of the stable distribution.

<span class="mw-page-title-main">Errors-in-variables models</span> Regression models accounting for possible errors in independent variables

In statistics, errors-in-variables models or measurement error models are regression models that account for measurement errors in the independent variables. In contrast, standard regression models assume that those regressors have been measured exactly, or observed without error; as such, those models account only for errors in the dependent variables, or responses.

In theoretical physics, massive representations of an extended supersymmetry algebra called BPS states have mass equal to the supersymmetry central charge Z. Quantum mechanically, if the supersymmetry remains unbroken, exact equality to the modulus of Z exists. Their importance arises as the supermultiplets shorten for generic massive representations, with stability and mass formula exact.

In econometrics, the Park test is a test for heteroscedasticity. The test is based on the method proposed by Rolla Edward Park for estimating linear regression parameters in the presence of heteroscedastic error terms.

In mathematics, Ricci calculus constitutes the rules of index notation and manipulation for tensors and tensor fields on a differentiable manifold, with or without a metric tensor or connection. It is also the modern name for what used to be called the absolute differential calculus, developed by Gregorio Ricci-Curbastro in 1887–1896, and subsequently popularized in a paper written with his pupil Tullio Levi-Civita in 1900. Jan Arnoldus Schouten developed the modern notation and formalism for this mathematical framework, and made contributions to the theory, during its applications to general relativity and differential geometry in the early twentieth century.

<span class="mw-page-title-main">Homoscedasticity and heteroscedasticity</span> Statistical property

In statistics, a sequence of random variables is homoscedastic if all its random variables have the same finite variance; this is also known as homogeneity of variance. The complementary notion is called heteroscedasticity, also known as heterogeneity of variance. The spellings homoskedasticity and heteroskedasticity are also frequently used. Skedasticity comes from the Ancient Greek word skedánnymi, meaning “to scatter”. Assuming a variable is homoscedastic when in reality it is heteroscedastic results in unbiased but inefficient point estimates and in biased estimates of standard errors, and may result in overestimating the goodness of fit as measured by the Pearson coefficient.

In machine learning, diffusion models, also known as diffusion probabilistic models or score-based generative models, are a class of latent variable generative models. A diffusion model consists of three major components: the forward process, the reverse process, and the sampling procedure. The goal of diffusion models is to learn a diffusion process that generates a probability distribution for a given dataset from which we can then sample new images. They learn the latent structure of a dataset by modeling the way in which data points diffuse through their latent space.

References

  1. Engle, Robert F. (1982). "Autoregressive Conditional Heteroskedasticity with Estimates of the Variance of United Kingdom Inflation". Econometrica . 50 (4): 987–1007. doi:10.2307/1912773. JSTOR   1912773.
  2. 1 2 Bollerslev, Tim (1986). "Generalized Autoregressive Conditional Heteroskedasticity". Journal of Econometrics . 31 (3): 307–327. CiteSeerX   10.1.1.468.2892 . doi:10.1016/0304-4076(86)90063-1. S2CID   8797625.
  3. Brooks, Chris (2014). Introductory Econometrics for Finance (3rd ed.). Cambridge: Cambridge University Press. p. 461. ISBN   9781107661455.
  4. Lanne, Markku; Saikkonen, Pentti (July 2005). "Non-linear GARCH models for highly persistent volatility" (PDF). The Econometrics Journal. 8 (2): 251–276. doi:10.1111/j.1368-423X.2005.00163.x. JSTOR   23113641. S2CID   15252964.
  5. Bollerslev, Tim; Russell, Jeffrey; Watson, Mark (May 2010). "Chapter 8: Glossary to ARCH (GARCH)" (PDF). Volatility and Time Series Econometrics: Essays in Honor of Robert Engle (1st ed.). Oxford: Oxford University Press. pp. 137–163. ISBN   9780199549498 . Retrieved 27 October 2017.
  6. 1 2 Engle, Robert F.; Ng, Victor K. (1993). "Measuring and testing the impact of news on volatility" (PDF). Journal of Finance. 48 (5): 1749–1778. doi: 10.1111/j.1540-6261.1993.tb05127.x . SSRN   262096. It is not yet clear in the finance literature that the asymmetric properties of variances are due to changing leverage. The name "leverage effect" is used simply because it is popular among researchers when referring to such a phenomenon.
  7. 1 2 Posedel, Petra (2006). "Analysis Of The Exchange Rate And Pricing Foreign Currency Options On The Croatian Market: The Ngarch Model As An Alternative To The Black Scholes Model" (PDF). Financial Theory and Practice. 30 (4): 347–368. Special attention to the model is given by the parameter of asymmetry [theta (θ)] which describes the correlation between returns and variance.6 ...
    6 In the case of analyzing stock returns, the positive value of [theta] reflects the empirically well known leverage effect indicating that a downward movement in the price of a stock causes more of an increase in variance more than a same value downward movement in the price of a stock, meaning that returns and variance are negatively correlated
  8. Higgins, M.L; Bera, A.K (1992). "A Class of Nonlinear Arch Models". International Economic Review. 33 (1): 137–158. doi:10.2307/2526988. JSTOR   2526988.
  9. Caporale, Guglielmo Maria; Pittis, Nikitas; Spagnolo, Nicola (October 2003). "IGARCH models and structural breaks". Applied Economics Letters. 10 (12): 765–768. doi:10.1080/1350485032000138403. ISSN   1350-4851.
  10. St. Pierre, Eilleen F. (1998). "Estimating EGARCH-M Models: Science or Art". The Quarterly Review of Economics and Finance. 38 (2): 167–180. doi:10.1016/S1062-9769(99)80110-0.
  11. Chatterjee, Swarn; Hubble, Amy (2016). "Day-Of-The-Whieek Effect In Us Biotechnology Stocks—Do Policy Changes And Economic Cycles Matter?". Annals of Financial Economics. 11 (2): 1–17. doi:10.1142/S2010495216500081.
  12. Hentschel, Ludger (1995). "All in the family Nesting symmetric and asymmetric GARCH models". Journal of Financial Economics. 39 (1): 71–104. CiteSeerX   10.1.1.557.8941 . doi:10.1016/0304-405X(94)00821-H.
  13. Klüppelberg, C.; Lindner, A.; Maller, R. (2004). "A continuous-time GARCH process driven by a Lévy process: stationarity and second-order behaviour". Journal of Applied Probability. 41 (3): 601–622. doi:10.1239/jap/1091543413. hdl: 10419/31047 . S2CID   17943198.
  14. Li, D.; Zhang, X.; Zhu, K.; Ling, S. (2018). "The ZD-GARCH model: A new way to study heteroscedasticity" (PDF). Journal of Econometrics . 202 (1): 1–17. doi:10.1016/j.jeconom.2017.09.003.
  15. Otto, P.; Schmid, W.; Garthoff, R. (2018). "Generalised spatial and spatiotemporal autoregressive conditional heteroscedasticity". Spatial Statistics . 26 (1): 125–145. arXiv: 1609.00711 . doi:10.1016/j.spasta.2018.07.005. S2CID   88521485.
  16. Platanios, E.; Chatzis, S. (2014). "Gaussian process-mixture conditional heteroscedasticity". IEEE Transactions on Pattern Analysis and Machine Intelligence . 36 (5): 889–900. arXiv: 1211.4410 . doi:10.1109/TPAMI.2013.183. PMID   26353224. S2CID   10424638.

Further reading