In probability theory, a distribution is said to be stable if a linear combination of two independentrandom variables with this distribution has the same distribution, up tolocation and scale parameters. A random variable is said to be stable if its distribution is stable. The stable distribution family is also sometimes referred to as the Lévy alpha-stable distribution, after Paul Lévy, the first mathematician to have studied it.[1][2]
Of the four parameters defining the family, most attention has been focused on the stability parameter, (see panel). Stable distributions have , with the upper bound corresponding to the normal distribution, and to the Cauchy distribution. The distributions have undefined variance for , and undefined mean for . The importance of stable probability distributions is that they are "attractors" for properly normed sums of independent and identically distributed (iid) random variables. The normal distribution defines a family of stable distributions. By the classical central limit theorem the properly normed sum of a set of random variables, each with finite variance, will tend toward a normal distribution as the number of variables increases. Without the finite variance assumption, the limit may be a stable distribution that is not normal. Mandelbrot referred to such distributions as "stable Paretian distributions",[3][4][5] after Vilfredo Pareto. In particular, he referred to those maximally skewed in the positive direction with as "Pareto–Lévy distributions",[1] which he regarded as better descriptions of stock and commodity prices than normal distributions.[6]
Definition
A non-degenerate distribution is a stable distribution if it satisfies the following property:
Let X1 and X2 be independent realizations of a random variableX. Then X is said to be stable if for any constants a > 0 and b > 0 the random variable aX1 + bX2 has the same distribution as cX + d for some constants c > 0 and d. The distribution is said to be strictly stable if this holds with d = 0.[7]
Such distributions form a four-parameter family of continuous probability distributions parametrized by location and scale parameters μ and c, respectively, and two shape parameters and , roughly corresponding to measures of asymmetry and concentration, respectively (see the figures).
The characteristic function of any probability distribution is the Fourier transform of its probability density function . The density function is therefore the inverse Fourier transform of the characteristic function:[8]
Although the probability density function for a general stable distribution cannot be written analytically, the general characteristic function can be expressed analytically. A random variable X is called stable if its characteristic function can be written as[7][9] where sgn(t) is just the sign of t and μ ∈ R is a shift parameter, , called the skewness parameter, is a measure of asymmetry. Notice that in this context the usual skewness is not well defined, as for the distribution does not admit 2nd or higher moments, and the usual skewness definition is the 3rd central moment.
The reason this gives a stable distribution is that the characteristic function for the sum of two independent random variables equals the product of the two corresponding characteristic functions. Adding two random variables from a stable distribution gives something with the same values of and , but possibly different values of μ and c.
Not every function is the characteristic function of a legitimate probability distribution (that is, one whose cumulative distribution function is real and goes from 0 to 1 without decreasing), but the characteristic functions given above will be legitimate so long as the parameters are in their ranges. The value of the characteristic function at some value t is the complex conjugate of its value at −t as it should be so that the probability distribution function will be real.
In the simplest case , the characteristic function is just a stretched exponential function; the distribution is symmetric about μ and is referred to as a (Lévy) symmetric alpha-stable distribution, often abbreviated SαS.
When and , the distribution is supported on [μ, ∞).
The parameter c > 0 is a scale factor which is a measure of the width of the distribution while is the exponent or index of the distribution and specifies the asymptotic behavior of the distribution.
Parametrizations
The parametrization of stable distributions is not unique. Nolan [10] tabulates 11 parametrizations seen in the literature and gives conversion formulas. The two most commonly used parametrizations are the one above (Nolan's "1") and the one immediately below (Nolan's "0").
The parametrization above is easiest to use for theoretical work, but its probability density is not continuous in the parameters at .[11] A continuous parametrization, better for numerical work, is[7] where:
The ranges of and are the same as before, γ (like c) should be positive, and δ (like μ) should be real.
In either parametrization one can make a linear transformation of the random variable to get a random variable whose density is . In the first parametrization, this is done by defining the new variable:
For the second parametrization, simply use independent of . In the first parametrization, if the mean exists (that is, ) then it is equal to μ, whereas in the second parametrization when the mean exists it is equal to
The distribution
A stable distribution is therefore specified by the above four parameters. It can be shown that any non-degenerate stable distribution has a smooth (infinitely differentiable) density function.[7] If denotes the density of X and Y is the sum of independent copies of X: then Y has the density with
The asymptotic behavior is described, for , by:[7] where Γ is the Gamma function (except that when and , the tail does not vanish to the left or right, resp., of μ, although the above expression is 0). This "heavy tail" behavior causes the variance of stable distributions to be infinite for all . This property is illustrated in the log–log plots below.
When , the distribution is Gaussian (see below), with tails asymptotic to exp(−x2/4c2)/(2c√π).
One-sided stable distribution and stable count distribution
When and , the distribution is supported on [μ, ∞). This family is called one-sided stable distribution.[12] Its standard distribution (μ=0) is defined as
, where
Let , its characteristic function is . Thus the integral form of its PDF is (note: )
The double-sine integral is more effective for very small .
Consider the Lévy sum where , then Y has the density where . Set to arrive at the stable count distribution.[13] Its standard distribution is defined as
The stable count distribution is the conjugate prior of the one-sided stable distribution. Its location-scale family is defined as
,
It is also a one-sided distribution supported on . The location parameter is the cut-off location, while defines its scale.
Its mean is and its standard deviation is . It is hypothesized that VIX is distributed like with and (See Section 7 of [13]). Thus the stable count distribution is the first-order marginal distribution of a volatility process. In this context, is called the "floor volatility".
Another approach to derive the stable count distribution is to use the Laplace transform of the one-sided stable distribution, (Section 2.4 of [13])
Let , and one can decompose the integral on the left hand side as a product distribution of a standard Laplace distribution and a standard stable count distribution,
This is called the "lambda decomposition" (See Section 4 of [13]) since the right hand side was named as "symmetric lambda distribution" in Lihn's former works. However, it has several more popular names such as "exponential power distribution", or the "generalized error/normal distribution", often referred to when .
The n-th moment of is the -th moment of , and all positive moments are finite.
Stable distributions are closed under convolution.
Stable distributions are closed under convolution for a fixed value of . Since convolution is equivalent to multiplication of the Fourier-transformed function, it follows that the product of two stable characteristic functions with the same will yield another such characteristic function. The product of two stable characteristic functions is given by:
Since Φ is not a function of the μ, c or variables it follows that these parameters for the convolved function are given by:
In each case, it can be shown that the resulting parameters lie within the required intervals for a stable distribution.
The Generalized Central Limit Theorem
The Generalized Central Limit Theorem (GCLT) was an effort of multiple mathematicians (Berstein, Lindeberg, Lévy, Feller, Kolmogorov, and others) over the period from 1920 to 1937. [14] The first published complete proof (in French) of the GCLT was in 1937 by Paul Lévy.[15] An English language version of the complete proof of the GCLT is available in the translation of Gnedenko and Kolmogorov's 1954 book.[16]
A non-degenerate random variableZis α-stable for some 0 < α ≤ 2 if and only if there is an independent, identically distributed sequence of random variablesX1, X2, X3, ... and constantsan > 0, bn ∈ ℝ with
an (X1 + ... + Xn) − bn → Z.
Here → means the sequence of random variable sums converges in distribution; i.e., the corresponding distributions satisfyFn(y) → F(y) at all continuity points ofF.
In other words, if sums of independent, identically distributed random variables converge in distribution to some Z, then Z must be a stable distribution.
Special cases
There is no general analytic solution for the form of f(x). There are, however three special cases which can be expressed in terms of elementary functions as can be seen by inspection of the characteristic function:[7][9][17]
For the distribution reduces to a Gaussian distribution with variance σ2 = 2c2 and mean μ; the skewness parameter has no effect.
For and the distribution reduces to a Cauchy distribution with scale parameter c and shift parameter μ.
For and the distribution reduces to a Lévy distribution with scale parameter c and shift parameter μ.
Note that the above three distributions are also connected, in the following way: A standard Cauchy random variable can be viewed as a mixture of Gaussian random variables (all with mean zero), with the variance being drawn from a standard Lévy distribution. And in fact this is a special case of a more general theorem (See p.59 of [18]) which allows any symmetric alpha-stable distribution to be viewed in this way (with the alpha parameter of the mixture distribution equal to twice the alpha parameter of the mixing distribution—and the beta parameter of the mixing distribution always equal to one).
A general closed form expression for stable PDFs with rational values of is available in terms of Meijer G-functions.[19] Fox H-Functions can also be used to express the stable probability density functions. For simple rational numbers, the closed form expression is often in terms of less complicated special functions. Several closed form expressions having rather simple expressions in terms of special functions are available. In the table below, PDFs expressible by elementary functions are indicated by an E and those that are expressible by special functions are indicated by an s.[18]
Some of the special cases are known by particular names:
For and , the distribution is a Landau distribution (L) which has a specific usage in physics under this name.
For and the distribution reduces to a Holtsmark distribution with scale parameter c and shift parameter μ.
Also, in the limit as c approaches zero or as α approaches zero the distribution will approach a Dirac delta functionδ(x−μ).
Series representation
The stable distribution can be restated as the real part of a simpler integral:[20]
Expressing the second exponential as a Taylor series, this leads to: where . Reversing the order of integration and summation, and carrying out the integration yields: which will be valid for x≠μ and will converge for appropriate values of the parameters. (Note that the n=0 term which yields a delta function in x−μ has therefore been dropped.) Expressing the first exponential as a series will yield another series in positive powers of x−μ which is generally less useful.
For one-sided stable distribution, the above series expansion needs to be modified, since and . There is no real part to sum. Instead, the integral of the characteristic function should be carried out on the negative axis, which yields:[21][12]
Parameter estimation
In addition to the existing tests for normality and subsequent parameter estimation, a general method which relies on the quantiles was developed by McCulloch and works for both symmetric and skew stable distributions and stability parameter .[22]
Simulation of stable variates
There are no analytic expressions for the inverse nor the CDF itself, so the inversion method cannot be used to generate stable-distributed variates.[11][13] Other standard approaches like the rejection method would require tedious computations. An elegant and efficient solution was proposed by Chambers, Mallows and Stuck (CMS),[23] who noticed that a certain integral formula[24] yielded the following algorithm:[25]
generate a random variable uniformly distributed on and an independent exponential random variable with mean 1;
for compute:
for compute: where
This algorithm yields a random variable . For a detailed proof see.[26]
To simulate a stable random variable for all admissible values of the parameters , , and use the following property: If then is . For (and ) the CMS method reduces to the well known Box-Muller transform for generating Gaussian random variables.[27] While other approaches have been proposed in the literature, including application of Bergström[28] and LePage[29] series expansions, the CMS method is regarded as the fastest and the most accurate.
Applications
Stable distributions owe their importance in both theory and practice to the generalization of the central limit theorem to random variables without second (and possibly first) order moments and the accompanying self-similarity of the stable family. It was the seeming departure from normality along with the demand for a self-similar model for financial data (i.e. the shape of the distribution for yearly asset price changes should resemble that of the constituent daily or monthly price changes) that led Benoît Mandelbrot to propose that cotton prices follow an alpha-stable distribution with equal to 1.7.[6]Lévy distributions are frequently found in analysis of critical behavior and financial data.[9][30]
The Lévy distribution of solar flare waiting time events (time between flare events) was demonstrated for CGRO BATSE hard x-ray solar flares in December 2001. Analysis of the Lévy statistical signature revealed that two different memory signatures were evident; one related to the solar cycle and the second whose origin appears to be associated with a localized or combination of localized solar active region effects.[31]
Other analytic cases
A number of cases of analytically expressible stable distributions are known. Let the stable distribution be expressed by , then:
The STABLE program for Windows is available from John Nolan's stable webpage: http://www.robustanalysis.com/public/stable.html. It calculates the density (pdf), cumulative distribution function (cdf) and quantiles for a general stable distribution, and performs maximum likelihood estimation of stable parameters and some exploratory data analysis techniques for assessing the fit of a data set.
The GNU Scientific Library which is written in C has a package randist, which includes among the Gaussian and Cauchy distributions also an implementation of the Levy alpha-stable distribution, both with and without a skew parameter.
libstable is a C implementation for the Stable distribution pdf, cdf, random number, quantile and fitting functions (along with a benchmark replication package and an R package).
R Package 'stabledist' by Diethelm Wuertz, Martin Maechler and Rmetrics core team members. Computes stable density, probability, quantiles, and random numbers.
Julia provides package StableDistributions.jl which has methods of generation, fitting, probability density, cumulative distribution function, characteristic and moment generating functions, quantile and related functions, convolution and affine transformations of stable distributions. It uses modernised algorithms improved by John P. Nolan.[10]
Related Research Articles
In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval [0, 1] or in terms of two positive parameters, denoted by alpha (α) and beta (β), that appear as exponents of the variable and its complement to 1, respectively, and control the shape of the distribution.
In the general theory of relativity, the Einstein field equations relate the geometry of spacetime to the distribution of matter within it.
In probability theory and statistics, the inverse gamma distribution is a two-parameter family of continuous probability distributions on the positive real line, which is the distribution of the reciprocal of a variable distributed according to the gamma distribution.
In electromagnetism, the electromagnetic tensor or electromagnetic field tensor is a mathematical object that describes the electromagnetic field in spacetime. The field tensor was first used after the four-dimensional tensor formulation of special relativity was introduced by Hermann Minkowski. The tensor allows related physical laws to be written concisely, and allows for the quantization of the electromagnetic field by the Lagrangian formulation described below.
In mathematics, the Gaussian or ordinary hypergeometric function2F1(a,b;c;z) is a special function represented by the hypergeometric series, that includes many other special functions as specific or limiting cases. It is a solution of a second-order linear ordinary differential equation (ODE). Every second-order linear ODE with three regular singular points can be transformed into this equation.
The noncentral t-distribution generalizes Student's t-distribution using a noncentrality parameter. Whereas the central probability distribution describes how a test statistic t is distributed when the difference tested is null, the noncentral distribution describes how t is distributed when the null is false. This leads to its use in statistics, especially calculating statistical power. The noncentral t-distribution is also known as the singly noncentral t-distribution, and in addition to its primary use in statistical inference, is also used in robust modeling for data.
A theoretical motivation for general relativity, including the motivation for the geodesic equation and the Einstein field equation, can be obtained from special relativity by examining the dynamics of particles in circular orbits about the Earth. A key advantage in examining circular orbits is that it is possible to know the solution of the Einstein Field Equation a priori. This provides a means to inform and verify the formalism.
The covariant formulation of classical electromagnetism refers to ways of writing the laws of classical electromagnetism in a form that is manifestly invariant under Lorentz transformations, in the formalism of special relativity using rectilinear inertial coordinate systems. These expressions both make it simple to prove that the laws of classical electromagnetism take the same form in any inertial coordinate system, and also provide a way to translate the fields and forces from one frame to another. However, this is not as general as Maxwell's equations in curved spacetime or non-rectilinear coordinate systems.
In physics, Maxwell's equations in curved spacetime govern the dynamics of the electromagnetic field in curved spacetime or where one uses an arbitrary coordinate system. These equations can be viewed as a generalization of the vacuum Maxwell's equations which are normally formulated in the local coordinates of flat spacetime. But because general relativity dictates that the presence of electromagnetic fields induce curvature in spacetime, Maxwell's equations in flat spacetime should be viewed as a convenient approximation.
In the theory of general relativity, a stress–energy–momentum pseudotensor, such as the Landau–Lifshitz pseudotensor, is an extension of the non-gravitational stress–energy tensor that incorporates the energy–momentum of gravity. It allows the energy–momentum of a system of gravitating matter to be defined. In particular it allows the total of matter plus the gravitating energy–momentum to form a conserved current within the framework of general relativity, so that the total energy–momentum crossing the hypersurface of any compact space–time hypervolume vanishes.
The Newman–Penrose (NP) formalism is a set of notation developed by Ezra T. Newman and Roger Penrose for general relativity (GR). Their notation is an effort to treat general relativity in terms of spinor notation, which introduces complex forms of the usual variables used in GR. The NP formalism is itself a special case of the tetrad formalism, where the tensors of the theory are projected onto a complete vector basis at each point in spacetime. Usually this vector basis is chosen to reflect some symmetry of the spacetime, leading to simplified expressions for physical observables. In the case of the NP formalism, the vector basis chosen is a null tetrad: a set of four null vectors—two real, and a complex-conjugate pair. The two real members often asymptotically point radially inward and radially outward, and the formalism is well adapted to treatment of the propagation of radiation in curved spacetime. The Weyl scalars, derived from the Weyl tensor, are often used. In particular, it can be shown that one of these scalars— in the appropriate frame—encodes the outgoing gravitational radiation of an asymptotically flat system.
In statistics, the multivariate t-distribution is a multivariate probability distribution. It is a generalization to random vectors of the Student's t-distribution, which is a distribution applicable to univariate random variables. While the case of a random matrix could be treated within this structure, the matrix t-distribution is distinct and makes particular use of the matrix structure.
Expected shortfall (ES) is a risk measure—a concept used in the field of financial risk measurement to evaluate the market risk or credit risk of a portfolio. The "expected shortfall at q% level" is the expected return on the portfolio in the worst of cases. ES is an alternative to value at risk that is more sensitive to the shape of the tail of the loss distribution.
A ratio distribution is a probability distribution constructed as the distribution of the ratio of random variables having two other known distributions. Given two random variables X and Y, the distribution of the random variable Z that is formed as the ratio Z = X/Y is a ratio distribution.
In financial mathematics, tail value at risk (TVaR), also known as tail conditional expectation (TCE) or conditional tail expectation (CTE), is a risk measure associated with the more general value at risk. It quantifies the expected value of the loss given that an event outside a given probability level has occurred.
In probability theory and statistics, the normal-inverse-gamma distribution is a four-parameter family of multivariate continuous probability distributions. It is the conjugate prior of a normal distribution with unknown mean and variance.
A product distribution is a probability distribution constructed as the distribution of the product of random variables having two other known distributions. Given two statistically independent random variables X and Y, the distribution of the random variable Z that is formed as the product is a product distribution.
The table of chords, created by the Greek astronomer, geometer, and geographer Ptolemy in Egypt during the 2nd century AD, is a trigonometric table in Book I, chapter 11 of Ptolemy's Almagest, a treatise on mathematical astronomy. It is essentially equivalent to a table of values of the sine function. It was the earliest trigonometric table extensive enough for many practical purposes, including those of astronomy. Since the 8th and 9th centuries, the sine and other trigonometric functions have been used in Islamic mathematics and astronomy, reforming the production of sine tables. Khwarizmi and Habash al-Hasib later produced a set of trigonometric tables.
The q-Gaussian is a probability distribution arising from the maximization of the Tsallis entropy under appropriate constraints. It is one example of a Tsallis distribution. The q-Gaussian is a generalization of the Gaussian in the same way that Tsallis entropy is a generalization of standard Boltzmann–Gibbs entropy or Shannon entropy. The normal distribution is recovered as q → 1.
In probability theory, the stable count distribution is the conjugate prior of a one-sided stable distribution. This distribution was discovered by Stephen Lihn in his 2017 study of daily distributions of the S&P 500 and the VIX. The stable distribution family is also sometimes referred to as the Lévy alpha-stable distribution, after Paul Lévy, the first mathematician to have studied it.
References
1 2 Mandelbrot, B. (1960). "The Pareto–Lévy Law and the Distribution of Income". International Economic Review. 1 (2): 79–106. doi:10.2307/2525289. JSTOR2525289.
↑ Lévy, Paul (1925). Calcul des probabilités. Paris: Gauthier-Villars. OCLC1417531.
↑ Mandelbrot, B. (1961). "Stable Paretian Random Functions and the Multiplicative Variation of Income". Econometrica. 29 (4): 517–543. doi:10.2307/1911802. JSTOR1911802.
↑ Mandelbrot, B. (1963). "The Variation of Certain Speculative Prices". The Journal of Business. 36 (4): 394–419. doi:10.1086/294632. JSTOR2350970.
↑ Fama, Eugene F. (1963). "Mandelbrot and the Stable Paretian Hypothesis". The Journal of Business. 36 (4): 420–429. doi:10.1086/294633. JSTOR2350971.
1 2 Nolan, John P. (1997). "Numerical calculation of stable densities and distribution functions". Communications in Statistics. Stochastic Models. 13 (4): 759–774. doi:10.1080/15326349708807450. ISSN0882-0287.
↑ Lévy, Paul (1937). Theorie de l'addition des variables aleatoires [Combination theory of unpredictable variables]. Paris: Gauthier-Villars.
↑ Gnedenko, Boris Vladimirovich; Kologorov, Andreĭ Nikolaevich; Doob, Joseph L.; Hsu, Pao-Lu (1968). Limit distributions for sums of independent random variables. Reading, MA: Addison-wesley.
↑ Zolotarev, V. (1995). "On Representation of Densities of Stable Laws by Special Functions". Theory of Probability and Its Applications. 39 (2): 354–362. doi:10.1137/1139025. ISSN0040-585X.
↑ Chambers, J. M.; Mallows, C. L.; Stuck, B. W. (1976). "A Method for Simulating Stable Random Variables". Journal of the American Statistical Association. 71 (354): 340–344. doi:10.1080/01621459.1976.10480344. ISSN0162-1459.
↑ Janicki, Aleksander; Kokoszka, Piotr (1992). "Computer investigation of the Rate of Convergence of Lepage Type Series to α-Stable Random Variables". Statistics. 23 (4): 365–373. doi:10.1080/02331889208802383. ISSN0233-1888.
1 2 Garoni, T. M.; Frankel, N. E. (2002). "Lévy flights: Exact results and asymptotics beyond all orders". Journal of Mathematical Physics. 43 (5): 2670–2689. Bibcode:2002JMP....43.2670G. doi:10.1063/1.1467095.
↑ Uchaikin, V. V.; Zolotarev, V. M. (1999). "Chance And Stability – Stable Distributions And Their Applications". VSP.
↑ Zlotarev, V. M. (1961). "Expression of the density of a stable distribution with exponent alpha greater than one by means of a frequency with exponent 1/alpha". Selected Translations in Mathematical Statistics and Probability (Translated from the Russian Article: Dokl. Akad. Nauk SSSR. 98, 735–738 (1954)). 1: 163–167.
This page is based on this Wikipedia article Text is available under the CC BY-SA 4.0 license; additional terms may apply. Images, videos and audio are available under their respective licenses.