Probability density function | |||
Cumulative distribution function | |||
Notation | |||
---|---|---|---|
Parameters | location (real) scale (real) | ||
Support | |||
where | |||
CDF | |||
Quantile | |||
Mean | where is the Euler–Mascheroni constant | ||
Median | |||
Mode | |||
Variance | |||
Skewness | |||
Excess kurtosis | |||
Entropy | |||
MGF | |||
CF |
In probability theory and statistics, the Gumbel distribution (also known as the type-I generalized extreme value distribution ) is used to model the distribution of the maximum (or the minimum) of a number of samples of various distributions.
This distribution might be used to represent the distribution of the maximum level of a river in a particular year if there was a list of maximum values for the past ten years. It is useful in predicting the chance that an extreme earthquake, flood or other natural disaster will occur. The potential applicability of the Gumbel distribution to represent the distribution of maxima relates to extreme value theory, which indicates that it is likely to be useful if the distribution of the underlying sample data is of the normal or exponential type. [lower-alpha 1]
The Gumbel distribution is a particular case of the generalized extreme value distribution (also known as the Fisher–Tippett distribution). It is also known as the log-Weibull distribution and the double exponential distribution (a term that is alternatively sometimes used to refer to the Laplace distribution). It is related to the Gompertz distribution: when its density is first reflected about the origin and then restricted to the positive half line, a Gompertz function is obtained.
In the latent variable formulation of the multinomial logit model — common in discrete choice theory — the errors of the latent variables follow a Gumbel distribution. This is useful because the difference of two Gumbel-distributed random variables has a logistic distribution.
The Gumbel distribution is named after Emil Julius Gumbel (1891–1966), based on his original papers describing the distribution. [1] [2]
The cumulative distribution function of the Gumbel distribution is
The standard Gumbel distribution is the case where and with cumulative distribution function
and probability density function
In this case the mode is 0, the median is , the mean is (the Euler–Mascheroni constant), and the standard deviation is
The cumulants, for n > 1, are given by
The mode is μ, while the median is and the mean is given by
where is the Euler–Mascheroni constant.
The standard deviation is hence [3]
At the mode, where , the value of becomes , irrespective of the value of
If are iid Gumbel random variables with parameters then is also a Gumbel random variable with parameters .
If are iid random variables such that has the same distribution as for all natural numbers , then is necessarily Gumbel distributed with scale parameter (actually it suffices to consider just two distinct values of k>1 which are coprime).
Theory related to the generalized multivariate log-gamma distribution provides a multivariate version of the Gumbel distribution.
Gumbel has shown that the maximum value (or last order statistic) in a sample of random variables following an exponential distribution minus the natural logarithm of the sample size [7] approaches the Gumbel distribution as the sample size increases. [8]
Concretely, let be the probability distribution of and its cumulative distribution. Then the maximum value out of realizations of is smaller than if and only if all realizations are smaller than . So the cumulative distribution of the maximum value satisfies
and, for large , the right-hand-side converges to
In hydrology, therefore, the Gumbel distribution is used to analyze such variables as monthly and annual maximum values of daily rainfall and river discharge volumes, [3] and also to describe droughts. [9]
Gumbel has also shown that the estimator r⁄(n+1) for the probability of an event — where r is the rank number of the observed value in the data series and n is the total number of observations — is an unbiased estimator of the cumulative probability around the mode of the distribution. Therefore, this estimator is often used as a plotting position.
In number theory, the Gumbel distribution approximates the number of terms in a random partition of an integer [10] as well as the trend-adjusted sizes of maximal prime gaps and maximal gaps between prime constellations. [11]
It appears in the coupon collector's problem.
In machine learning, the Gumbel distribution is sometimes employed to generate samples from the categorical distribution. This technique is called "Gumbel-max trick" and is a special example of "reparameterization tricks". [12]
In detail, let be nonnegative, and not all zero, and let be independent samples of Gumbel(0, 1), then by routine integration,That is,
Equivalently, given any , we can sample from its Boltzmann distribution by
Related equations include: [13]
Since the quantile function (inverse cumulative distribution function), , of a Gumbel distribution is given by
the variate has a Gumbel distribution with parameters and when the random variate is drawn from the uniform distribution on the interval .
In pre-software times probability paper was used to picture the Gumbel distribution (see illustration). The paper is based on linearization of the cumulative distribution function :
In the paper the horizontal axis is constructed at a double log scale. The vertical axis is linear. By plotting on the horizontal axis of the paper and the -variable on the vertical axis, the distribution is represented by a straight line with a slope 1. When distribution fitting software like CumFreq became available, the task of plotting the distribution was made easier.
In probability theory and statistics, the Weibull distribution is a continuous probability distribution. It models a broad range of random variables, largely in the nature of a time to failure or time between events. Examples are maximum one-day rainfalls and the time a user spends on a web page.
In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval [0, 1] or in terms of two positive parameters, denoted by alpha (α) and beta (β), that appear as exponents of the variable and its complement to 1, respectively, and control the shape of the distribution.
In probability theory and statistics, the gamma distribution is a versatile two-parameter family of continuous probability distributions. The exponential distribution, Erlang distribution, and chi-squared distribution are special cases of the gamma distribution. There are two equivalent parameterizations in common use:
In probability theory and statistics, the logistic distribution is a continuous probability distribution. Its cumulative distribution function is the logistic function, which appears in logistic regression and feedforward neural networks. It resembles the normal distribution in shape but has heavier tails. The logistic distribution is a special case of the Tukey lambda distribution.
In probability theory, a distribution is said to be stable if a linear combination of two independent random variables with this distribution has the same distribution, up to location and scale parameters. A random variable is said to be stable if its distribution is stable. The stable distribution family is also sometimes referred to as the Lévy alpha-stable distribution, after Paul Lévy, the first mathematician to have studied it.
Variational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning. They are typically used in complex statistical models consisting of observed variables as well as unknown parameters and latent variables, with various sorts of relationships among the three types of random variables, as might be described by a graphical model. As typical in Bayesian inference, the parameters and latent variables are grouped together as "unobserved variables". Variational Bayesian methods are primarily used for two purposes:
In probability theory and statistics, the generalized extreme value (GEV) distribution is a family of continuous probability distributions developed within extreme value theory to combine the Gumbel, Fréchet and Weibull families also known as type I, II and III extreme value distributions. By the extreme value theorem the GEV distribution is the only possible limit distribution of properly normalized maxima of a sequence of independent and identically distributed random variables. that a limit distribution needs to exist, which requires regularity conditions on the tail of the distribution. Despite this, the GEV distribution is often used as an approximation to model the maxima of long (finite) sequences of random variables.
In probability theory and statistics, the inverse gamma distribution is a two-parameter family of continuous probability distributions on the positive real line, which is the distribution of the reciprocal of a variable distributed according to the gamma distribution.
The folded normal distribution is a probability distribution related to the normal distribution. Given a normally distributed random variable X with mean μ and variance σ2, the random variable Y = |X| has a folded normal distribution. Such a case may be encountered if only the magnitude of some variable is recorded, but not its sign. The distribution is called "folded" because probability mass to the left of x = 0 is folded over by taking the absolute value. In the physics of heat conduction, the folded normal distribution is a fundamental solution of the heat equation on the half space; it corresponds to having a perfect insulator on a hyperplane through the origin.
A number of different Markov models of DNA sequence evolution have been proposed. These substitution models differ in terms of the parameters used to describe the rates at which one nucleotide replaces another during evolution. These models are frequently used in molecular phylogenetic analyses. In particular, they are used during the calculation of likelihood of a tree and they are used to estimate the evolutionary distance between sequences from the observed differences between the sequences.
Expected shortfall (ES) is a risk measure—a concept used in the field of financial risk measurement to evaluate the market risk or credit risk of a portfolio. The "expected shortfall at q% level" is the expected return on the portfolio in the worst of cases. ES is an alternative to value at risk that is more sensitive to the shape of the tail of the loss distribution.
A ratio distribution is a probability distribution constructed as the distribution of the ratio of random variables having two other known distributions. Given two random variables X and Y, the distribution of the random variable Z that is formed as the ratio Z = X/Y is a ratio distribution.
In probability theory and statistics, the normal-gamma distribution is a bivariate four-parameter family of continuous probability distributions. It is the conjugate prior of a normal distribution with unknown mean and precision.
In financial mathematics, tail value at risk (TVaR), also known as tail conditional expectation (TCE) or conditional tail expectation (CTE), is a risk measure associated with the more general value at risk. It quantifies the expected value of the loss given that an event outside a given probability level has occurred.
The term generalized logistic distribution is used as the name for several different families of probability distributions. For example, Johnson et al. list four forms, which are listed below.
In probability theory and statistics, the normal-inverse-gamma distribution is a four-parameter family of multivariate continuous probability distributions. It is the conjugate prior of a normal distribution with unknown mean and variance.
A product distribution is a probability distribution constructed as the distribution of the product of random variables having two other known distributions. Given two statistically independent random variables X and Y, the distribution of the random variable Z that is formed as the product is a product distribution.
In probability theory, to obtain a nondegenerate limiting distribution of the extreme value distribution, it is necessary to "reduce" the actual greatest value by applying a linear transformation with coefficients that depend on the sample size.
In statistics and probability theory, the nonparametric skew is a statistic occasionally used with random variables that take real values. It is a measure of the skewness of a random variable's distribution—that is, the distribution's tendency to "lean" to one side or the other of the mean. Its calculation does not require any knowledge of the form of the underlying distribution—hence the name nonparametric. It has some desirable properties: it is zero for any symmetric distribution; it is unaffected by a scale shift; and it reveals either left- or right-skewness equally well. In some statistical samples it has been shown to be less powerful than the usual measures of skewness in detecting departures of the population from normality.
In statistics, the variance function is a smooth function that depicts the variance of a random quantity as a function of its mean. The variance function is a measure of heteroscedasticity and plays a large role in many settings of statistical modelling. It is a main ingredient in the generalized linear model framework and a tool used in non-parametric regression, semiparametric regression and functional data analysis. In parametric modeling, variance functions take on a parametric form and explicitly describe the relationship between the variance and the mean of a random quantity. In a non-parametric setting, the variance function is assumed to be a smooth function.