Distributed lag

Last updated

In statistics and econometrics, a distributed lag model is a model for time series data in which a regression equation is used to predict current values of a dependent variable based on both the current values of an explanatory variable and the lagged (past period) values of this explanatory variable. [1] [2]

Contents

The starting point for a distributed lag model is an assumed structure of the form

or the form

where yt is the value at time period t of the dependent variable y, a is the intercept term to be estimated, and wi is called the lag weight (also to be estimated) placed on the value i periods previously of the explanatory variable x. In the first equation, the dependent variable is assumed to be affected by values of the independent variable arbitrarily far in the past, so the number of lag weights is infinite and the model is called an infinite distributed lag model. In the alternative, second, equation, there are only a finite number of lag weights, indicating an assumption that there is a maximum lag beyond which values of the independent variable do not affect the dependent variable; a model based on this assumption is called a finite distributed lag model.

In an infinite distributed lag model, an infinite number of lag weights need to be estimated; clearly this can be done only if some structure is assumed for the relation between the various lag weights, with the entire infinitude of them expressible in terms of a finite number of assumed underlying parameters. In a finite distributed lag model, the parameters could be directly estimated by ordinary least squares (assuming the number of data points sufficiently exceeds the number of lag weights); nevertheless, such estimation may give very imprecise results due to extreme multicollinearity among the various lagged values of the independent variable, so again it may be necessary to assume some structure for the relation between the various lag weights.

The concept of distributed lag models easily generalizes to the context of more than one right-side explanatory variable.

Unstructured estimation

The simplest way to estimate parameters associated with distributed lags is by ordinary least squares, assuming a fixed maximum lag , assuming independently and identically distributed errors, and imposing no structure on the relationship of the coefficients of the lagged explanators with each other. However, multicollinearity among the lagged explanators often arises, leading to high variance of the coefficient estimates.

Structured estimation

Structured distributed lag models come in two types: finite and infinite. Infinite distributed lags allow the value of the independent variable at a particular time to influence the dependent variable infinitely far into the future, or to put it another way, they allow the current value of the dependent variable to be influenced by values of the independent variable that occurred infinitely long ago; but beyond some lag length the effects taper off toward zero. Finite distributed lags allow for the independent variable at a particular time to influence the dependent variable for only a finite number of periods.

Finite distributed lags

The most important structured finite distributed lag model is the Almon lag model. [3] This model allows the data to determine the shape of the lag structure, but the researcher must specify the maximum lag length; an incorrectly specified maximum lag length can distort the shape of the estimated lag structure as well as the cumulative effect of the independent variable. The Almon lag assumes that k + 1 lag weights are related to n + 1 linearly estimable underlying parameters (n < k) aj according to

for

Infinite distributed lags

The most common type of structured infinite distributed lag model is the geometric lag, also known as the Koyck lag. In this lag structure, the weights (magnitudes of influence) of the lagged independent variable values decline exponentially with the length of the lag; while the shape of the lag structure is thus fully imposed by the choice of this technique, the rate of decline as well as the overall magnitude of effect are determined by the data. Specification of the regression equation is very straightforward: one includes as explanators (right-hand side variables in the regression) the one-period-lagged value of the dependent variable and the current value of the independent variable:

where . In this model, the short-run (same-period) effect of a unit change in the independent variable is the value of b, while the long-run (cumulative) effect of a sustained unit change in the independent variable can be shown to be

Other infinite distributed lag models have been proposed to allow the data to determine the shape of the lag structure. The polynomial inverse lag [4] [5] assumes that the lag weights are related to underlying, linearly estimable parameters aj according to

for

The geometric combination lag [6] assumes that the lags weights are related to underlying, linearly estimable parameters aj according to either

for or

for

The gamma lag [7] and the rational lag [8] are other infinite distributed lag structures.

Distributed lag model in health studies

Distributed lag models were introduced into health-related studies in 2002 by Zanobetti and Schwartz. [9] The Bayesian version of the model was suggested by Welty in 2007. [10] Gasparrini introduced more flexible statistical models in 2010 [11] that are capable of describing additional time dimensions of the exposure-response relationship, and developed a family of distributed lag non-linear models (DLNM), a modeling framework that can simultaneously represent non-linear exposure-response dependencies and delayed effects. [12]

The distributed lag model concept was first to applied to longitudinal cohort research by Hsu in 2015, [13] studying the relationship between PM2.5 and child asthma, and more complicated distributed lag method aimed to accommodate longitudinal cohort research analysis such as Bayesian Distributed Lag Interaction Model [14] by Wilson have been subsequently developed to answer similar research questions.

See also

Related Research Articles

In mathematics, and more specifically in linear algebra, a linear map is a mapping between two vector spaces that preserves the operations of vector addition and scalar multiplication. The same names and the same definition are also used for the more general case of modules over a ring; see Module homomorphism.

<span class="mw-page-title-main">Linear span</span> In linear algebra, generated subspace

In mathematics, the linear span (also called the linear hull or just span) of a set S of vectors (from a vector space), denoted span(S), is defined as the set of all linear combinations of the vectors in S. For example, two linearly independent vectors span a plane. The linear span can be characterized either as the intersection of all linear subspaces that contain S, or as the smallest subspace containing S. The linear span of a set of vectors is therefore a vector space itself. Spans can be generalized to matroids and modules.

A weight function is a mathematical device used when performing a sum, integral, or average to give some elements more "weight" or influence on the result than other elements in the same set. The result of this application of a weight function is a weighted sum or weighted average. Weight functions occur frequently in statistics and analysis, and are closely related to the concept of a measure. Weight functions can be employed in both discrete and continuous settings. They can be used to construct systems of calculus called "weighted calculus" and "meta-calculus".

<span class="mw-page-title-main">Lattice model (physics)</span>

In mathematical physics, a lattice model is a mathematical model of a physical system that is defined on a lattice, as opposed to a continuum, such as the continuum of space or spacetime. Lattice models originally occurred in the context of condensed matter physics, where the atoms of a crystal automatically form a lattice. Currently, lattice models are quite popular in theoretical physics, for many reasons. Some models are exactly solvable, and thus offer insight into physics beyond what can be learned from perturbation theory. Lattice models are also ideal for study by the methods of computational physics, as the discretization of any continuum model automatically turns it into a lattice model. The exact solution to many of these models includes the presence of solitons. Techniques for solving these include the inverse scattering transform and the method of Lax pairs, the Yang–Baxter equation and quantum groups. The solution of these models has given insights into the nature of phase transitions, magnetization and scaling behaviour, as well as insights into the nature of quantum field theory. Physical lattice models frequently occur as an approximation to a continuum theory, either to give an ultraviolet cutoff to the theory to prevent divergences or to perform numerical computations. An example of a continuum theory that is widely studied by lattice models is the QCD lattice model, a discretization of quantum chromodynamics. However, digital physics considers nature fundamentally discrete at the Planck scale, which imposes upper limit to the density of information, aka Holographic principle. More generally, lattice gauge theory and lattice field theory are areas of study. Lattice models are also used to simulate the structure and dynamics of polymers.

<span class="mw-page-title-main">Affine space</span> Euclidean space without distance and angles

In mathematics, an affine space is a geometric structure that generalizes some of the properties of Euclidean spaces in such a way that these are independent of the concepts of distance and measure of angles, keeping only the properties related to parallelism and ratio of lengths for parallel line segments.

<span class="mw-page-title-main">Quantum group</span> Algebraic construct of interest in theoretical physics

In mathematics and theoretical physics, the term quantum group denotes one of a few different kinds of noncommutative algebras with additional structure. These include Drinfeld–Jimbo type quantum groups, compact matrix quantum groups, and bicrossproduct quantum groups. Despite their name, they do not themselves have a natural group structure, though they are in some sense 'close' to a group.

In probability theory, a compound Poisson distribution is the probability distribution of the sum of a number of independent identically-distributed random variables, where the number of terms to be added is itself a Poisson-distributed variable. The result can be either a continuous or a discrete distribution.

In statistics, multicollinearity is a phenomenon in which one predictor variable in a multiple regression model can be linearly predicted from the others with a substantial degree of accuracy. In this situation, the coefficient estimates of the multiple regression may change erratically in response to small changes in the model or the data. Multicollinearity does not reduce the predictive power or reliability of the model as a whole, at least within the sample data set; it only affects calculations regarding individual predictors. That is, a multivariable regression model with collinear predictors can indicate how well the entire bundle of predictors predicts the outcome variable, but it may not give valid results about any individual predictor, or about which predictors are redundant with respect to others.

In linear algebra, an eigenvector or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted by , is the factor by which the eigenvector is scaled.

In statistics, Poisson regression is a generalized linear model form of regression analysis used to model count data and contingency tables. Poisson regression assumes the response variable Y has a Poisson distribution, and assumes the logarithm of its expected value can be modeled by a linear combination of unknown parameters. A Poisson regression model is sometimes known as a log-linear model, especially when used to model contingency tables.

Verma modules, named after Daya-Nand Verma, are objects in the representation theory of Lie algebras, a branch of mathematics.

Linear Programming Boosting (LPBoost) is a supervised classifier from the boosting family of classifiers. LPBoost maximizes a margin between training samples of different classes and hence also belongs to the class of margin-maximizing supervised classification algorithms. Consider a classification function

In statistics, principal component regression (PCR) is a regression analysis technique that is based on principal component analysis (PCA). More specifically, PCR is used for estimating the unknown regression coefficients in a standard linear regression model.

In applied mathematics, the Atkinson–Mingarelli theorem, named after Frederick Valentine Atkinson and A. B. Mingarelli, concerns eigenvalues of certain Sturm–Liouville differential operators.

Log-linear analysis is a technique used in statistics to examine the relationship between more than two categorical variables. The technique is used for both hypothesis testing and model building. In both these uses, models are tested to find the most parsimonious model that best accounts for the variance in the observed frequencies.

<span class="mw-page-title-main">Poisson point process</span> Type of random mathematical object

In probability, statistics and related fields, a Poisson point process is a type of random mathematical object that consists of points randomly located on a mathematical space with the essential feature that the points occur independently of one another. The Poisson point process is often called simply the Poisson process, but it is also called a Poisson random measure, Poisson random point field or Poisson point field. This point process has convenient mathematical properties, which has led to its being frequently defined in Euclidean space and used as a mathematical model for seemingly random processes in numerous disciplines such as astronomy, biology, ecology, geology, seismology, physics, economics, image processing, and telecommunications.

Structured sparsity regularization is a class of methods, and an area of research in statistical learning theory, that extend and generalize sparsity regularization learning methods. Both sparsity and structured sparsity regularization methods seek to exploit the assumption that the output variable to be learned can be described by a reduced number of variables in the input space . Sparsity regularization methods focus on selecting the input variables that best describe the output. Structured sparsity regularization methods generalize and extend sparsity regularization methods, by allowing for optimal selection over structures like groups or networks of input variables in .

<span class="mw-page-title-main">Representation theory of semisimple Lie algebras</span>

In mathematics, the representation theory of semisimple Lie algebras is one of the crowning achievements of the theory of Lie groups and Lie algebras. The theory was worked out mainly by E. Cartan and H. Weyl and because of that, the theory is also known as the Cartan–Weyl theory. The theory gives the structural description and classification of a finite-dimensional representation of a semisimple Lie algebra ; in particular, it gives a way to parametrize irreducible finite-dimensional representations of a semisimple Lie algebra, the result known as the theorem of the highest weight.

<span class="mw-page-title-main">Birth process</span> Type of continuous process in probability theory

In probability theory, a birth process or a pure birth process is a special case of a continuous-time Markov process and a generalisation of a Poisson process. It defines a continuous process which takes values in the natural numbers and can only increase by one or remain unchanged. This is a type of birth–death process with no deaths. The rate at which births occur is given by an exponential random variable whose parameter depends only on the current value of the process

Tau functions are an important ingredient in the modern theory of integrable systems, and have numerous applications in a variety of other domains. They were originally introduced by Ryogo Hirota in his direct method approach to soliton equations, based on expressing them in an equivalent bilinear form. The term Tau function, or -function, was first used systematically by Mikio Sato and his students in the specific context of the Kadomtsev–Petviashvili equation and related integrable hierarchies. It is a central ingredient in the theory of solitons. Tau functions also appear as matrix model partition functions in the spectral theory of Random Matrices, and may also serve as generating functions, in the sense of combinatorics and enumerative geometry, especially in relation to moduli spaces of Riemann surfaces, and enumeration of branched coverings, or so-called Hurwitz numbers.

References

  1. Cromwell, Jeff B.; et al. (1994). Multivariate Tests For Time Series Models. SAGE Publications. ISBN   0-8039-5440-9.
  2. Judge, George G.; Griffiths, William E.; Hill, R. Carter; Lee, Tsoung-Chao (1980). The Theory and Practice of Econometrics. New York: Wiley. pp. 637–660. ISBN   0-471-05938-2.
  3. Almon, Shirley, "The distributed lag between capital appropriations and net expenditures," Econometrica 33, 1965, 178-196.
  4. Mitchell, Douglas W., and Speaker, Paul J., "A simple, flexible distributed lag technique: the polynomial inverse lag," Journal of Econometrics 31, 1986, 329-340.
  5. Gelles, Gregory M., and Mitchell, Douglas W., "An approximation theorem for the polynomial inverse lag," Economics Letters 30, 1989, 129-132.
  6. Speaker, Paul J., Mitchell, Douglas W., and Gelles, Gregory M., "Geometric combination lags as flexible infinite distributed lag estimators," Journal of Economic Dynamics and Control 13, 1989, 171-185.
  7. Schmidt, Peter (1974). "A modification of the Almon distributed lag". Journal of the American Statistical Association. 69 (347): 679–681. doi:10.1080/01621459.1974.10480188.
  8. Jorgenson, Dale W. (1966). "Rational distributed lag functions". Econometrica . 34 (1): 135–149. doi:10.2307/1909858. JSTOR   1909858.
  9. Zanobetti, Antonella; Schwartz, Joel; Samoli, Evi; Gryparis, Alexandros; Touloumi, Giota; Atkinson, Richard; Le Tertre, Alain; Bobros, Janos; Celko, Martin; Goren, Ayana; Forsberg, Bertil (January 2002). "The temporal pattern of mortality responses to air pollution: a multicity assessment of mortality displacement". Epidemiology. 13 (1): 87–93. doi: 10.1097/00001648-200201000-00014 . ISSN   1044-3983. PMID   11805591. S2CID   25181383.
  10. Welty, L. J.; Peng, R. D.; Zeger, S. L.; Dominici, F. (March 2009). "Bayesian distributed lag models: estimating effects of particulate matter air pollution on daily mortality". Biometrics. 65 (1): 282–291. doi:10.1111/j.1541-0420.2007.01039.x. ISSN   1541-0420. PMID   18422792.
  11. Gasparrini, A; Armstrong, B; Kenward, M G (2010-09-20). "Distributed lag non-linear models". Statistics in Medicine. 29 (21): 2224–2234. doi:10.1002/sim.3940. ISSN   0277-6715. PMC   2998707 . PMID   20812303.
  12. "Distributed Lag Non-Linear Models [R package dlnm version 2.4.6]". cran.r-project.org. 2021-06-15. Retrieved 2021-09-17.
  13. Leon Hsu, Hsiao-Hsien; Mathilda Chiu, Yueh-Hsiu; Coull, Brent A.; Kloog, Itai; Schwartz, Joel; Lee, Alison; Wright, Robert O.; Wright, Rosalind J. (2015-11-01). "Prenatal Particulate Air Pollution and Asthma Onset in Urban Children. Identifying Sensitive Windows and Sex Differences". American Journal of Respiratory and Critical Care Medicine. 192 (9): 1052–1059. doi:10.1164/rccm.201504-0658OC. ISSN   1073-449X. PMC   4642201 . PMID   26176842.
  14. Wilson, Ander; Chiu, Yueh-Hsiu Mathilda; Hsu, Hsiao-Hsien Leon; Wright, Robert O.; Wright, Rosalind J.; Coull, Brent A. (July 2017). "Bayesian distributed lag interaction models to identify perinatal windows of vulnerability in children's health". Biostatistics. 18 (3): 537–552. doi:10.1093/biostatistics/kxx002. ISSN   1465-4644. PMC   5862289 . PMID   28334179.