Generalized logistic distribution

Last updated

The term generalized logistic distribution is used as the name for several different families of probability distributions. For example, Johnson et al. [1] list four forms, which are listed below.

Contents

Type I has also been called the skew-logistic distribution. Type IV subsumes the other types and is obtained when applying the logit transform to beta random variates. Following the same convention as for the log-normal distribution, type IV may be referred to as the logistic-beta distribution, with reference to the standard logistic function, which is the inverse of the logit transform.

For other families of distributions that have also been called generalized logistic distributions, see the shifted log-logistic distribution, which is a generalization of the log-logistic distribution; and the metalog ("meta-logistic") distribution, which is highly shape-and-bounds flexible and can be fit to data with linear least squares.

Definitions

The following definitions are for standardized versions of the families, which can be expanded to the full form as a location-scale family. Each is defined using either the cumulative distribution function (F) or the probability density function (ƒ), and is defined on (-∞,∞).

Type I

The corresponding probability density function is:

This type has also been called the "skew-logistic" distribution.

Type II

The corresponding probability density function is:

Type III

Here B is the beta function. The moment generating function for this type is

The corresponding cumulative distribution function is:

Type IV

Where, B is the beta function and is the standard logistic function. The moment generating function for this type is

This type is also called the "exponential generalized beta of the second type". [1]

The corresponding cumulative distribution function is:

Relationship between types

Type IV is the most general form of the distribution. The Type III distribution can be obtained from Type IV by fixing . The Type II distribution can be obtained from Type IV by fixing (and renaming to ). The Type I distribution can be obtained from Type IV by fixing . Fixing gives the standard logistic distribution.

Type IV (logistic-beta) properties

Type IV probability density functions (means=0, variances=1) Pdf of sigmoid-beta distribution.svg
Type IV probability density functions (means=0, variances=1)

The Type IV generalized logistic, or logistic-beta distribution, with support and shape parameters , has (as shown above) the probability density function (pdf):

where is the standard logistic function. The probability density functions for three different sets of shape parameters are shown in the plot, where the distributions have been scaled and shifted to give zero means and unity variances, in order to facilitate comparison of the shapes.

In what follows, the notation is used to denote the Type IV distribution.

Relationship with Gamma Distribution

This distribution can be obtained in terms of the gamma distribution as follows. Let and independently, and let . Then . [2]

Symmetry

If , then .

Mean and variance

By using the logarithmic expectations of the gamma distribution, the mean and variance can be derived as:

where is the digamma function, while is its first derivative, also known as the trigamma function, or the first polygamma function. Since is strictly increasing, the sign of the mean is the same as the sign of . Since is strictly decreasing, the shape parameters can also be interpreted as concentration parameters. Indeed, as shown below, the left and right tails respectively become thinner as or are increased. The two terms of the variance represent the contributions to the variance of the left and right parts of the distribution.

Cumulants and skewness

The cumulant generating function is , where the moment generating function is given above. The cumulants, , are the -th derivatives of , evaluated at :

where and are the digamma and polygamma functions. In agreement with the derivation above, the first cumulant, , is the mean and the second, , is the variance.

The third cumulant, , is the third central moment , which when scaled by the third power of the standard deviation gives the skewness:

The sign (and therefore the handedness) of the skewness is the same as the sign of .

Mode

The mode (pdf maximum) can be derived by finding where the log pdf derivative is zero:

This simplifies to , so that: [2]

Tail behaviour

Tail comparison: Type IV (means=0, variances=1) vs standard normal, vs standard Cauchy Comparison of log-pdf's- sigmoid-beta vs normal vs cauchy.svg
Tail comparison: Type IV (means=0, variances=1) vs standard normal, vs standard Cauchy

In each of the left and right tails, one of the sigmoids in the pdf saturates to one, so that the tail is formed by the other sigmoid. For large negative , the left tail of the pdf is proportional to , while the right tail (large positive ) is proportional to . This means the tails are independently controlled by and . Although type IV tails are heavier than those of the normal distribution (, for variance ), the type IV means and variances remain finite for all . This is in contrast with the Cauchy distribution for which the mean and variance do not exist. In the log pdf plots shown here, the type IV tails are linear, the normal distribution tails are quadratic and the Cauchy tails are logarithmic.

Exponential family properties

forms an exponential family with natural parameters and and sufficient statistics and . The expected values of the sufficient statistics can be found by differentiation of the log-normalizer: [3]

Given a data set assumed to have been generated IID from , the maximum-likelihood parameter estimate is:

where the overlines denote the averages of the sufficient statistics. The maximum-likelihood estimate depends on the data only via these average statistics. Indeed, at the maximum-likelihood estimate the expected values and averages agree:

which is also where the partial derivatives of the above maximand vanish.

Relationships with other distributions

Relationships with other distributions include:

Large shape parameters

Type IV vs normal distribution with matched mean and variance. For large values of
a
,
b
{\displaystyle \alpha ,\beta }
, the pdf's are very similar, except for very rare values of
x
{\displaystyle x}
. Sigmoid beta distribution with large shape parameters vs matched normal distribution.svg
Type IV vs normal distribution with matched mean and variance. For large values of , the pdf's are very similar, except for very rare values of .

For large values of the shape parameters, , the distribution becomes more Gaussian, with:

This is demonstrated in the pdf and log pdf plots here.

Random variate generation

Since random sampling from the gamma and beta distributions are readily available on many software platforms, the above relationships with those distributions can be used to generate variates from the type IV distribution.

Generalization with location and scale parameters

A flexible, four-parameter family can be obtained by adding location and scale parameters. One way to do this is if , then let , where is the scale parameter and is the location parameter. The four-parameter family obtained thus has the desired additional flexibility, but the new parameters may be hard to interpret because and . Moreover maximum-likelihood estimation with this parametrization is hard. These problems can be addressed as follows.

Recall that the mean and variance of are:

Now expand the family with location parameter and scale parameter , via the transformation:

so that and are now interpretable. It may be noted that allowing to be either positive or negative does not generalize this family, because of the above-noted symmetry property. We adopt the notation for this family.

If the pdf for is , then the pdf for is:

where it is understood that is computed as detailed above, as a function of . The pdf and log-pdf plots above, where the captions contain (means=0, variances=1), are for .

Maximum likelihood parameter estimation

In this section, maximum-likelihood estimation of the distribution parameters, given a dataset is discussed in turn for the families and .

Maximum likelihood for standard Type IV

As noted above, is an exponential family with natural parameters , the maximum-likelihood estimates of which depend only on averaged sufficient statistics:

Once these statistics have been accumulated, the maximum-likelihood estimate is given by:

By using the parametrization and an unconstrained numerical optimization algorithm like BFGS can be used. Optimization iterations are fast, because they are independent of the size of the data-set.

An alternative is to use an EM-algorithm based on the composition: if and . Because of the self-conjugacy of the gamma distribution, the posterior expectations, and that are required for the E-step can be computed in closed form. The M-step parameter update can be solved analogously to maximum-likelihood for the gamma distribution.

Maximum likelihood for the four-parameter family

The maximum-likelihood problem for , having pdf is:

This is no longer an exponential family, so that each optimization iteration has to traverse the whole data-set. Moreover the computation of the partial derivatives (as required for example by BFGS) is considerably more complex than for the above two-parameter case. However, all the component functions are readily available in software packages with automatic differentiation. Again, the positive parameters can be parametrized in terms of their logarithms to obtain an unconstrained numerical optimization problem.

For this problem, numerical optimization may fail unless the initial location and scale parameters are chosen appropriately. However the above-mentioned interpretability of these parameters in the parametrization of can be used to do this. Specifically, the initial values for and can be set to the empirical mean and variance of the data.

See also

Related Research Articles

In particle physics, the Dirac equation is a relativistic wave equation derived by British physicist Paul Dirac in 1928. In its free form, or including electromagnetic interactions, it describes all spin-12 massive particles, called "Dirac particles", such as electrons and quarks for which parity is a symmetry. It is consistent with both the principles of quantum mechanics and the theory of special relativity, and was the first theory to account fully for special relativity in the context of quantum mechanics. It was validated by accounting for the fine structure of the hydrogen spectrum in a completely rigorous way.

<span class="mw-page-title-main">Exponential distribution</span> Probability distribution

In probability theory and statistics, the exponential distribution or negative exponential distribution is the probability distribution of the distance between events in a Poisson point process, i.e., a process in which events occur continuously and independently at a constant average rate; the distance parameter could be any meaningful mono-dimensional measure of the process, such as time between production errors, or length along a roll of fabric in the weaving manufacturing process. It is a particular case of the gamma distribution. It is the continuous analogue of the geometric distribution, and it has the key property of being memoryless. In addition to being used for the analysis of Poisson point processes it is found in various other contexts.

<span class="mw-page-title-main">Beta distribution</span> Probability distribution

In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval [0, 1] or in terms of two positive parameters, denoted by alpha (α) and beta (β), that appear as exponents of the variable and its complement to 1, respectively, and control the shape of the distribution.

<span class="mw-page-title-main">Gamma distribution</span> Probability distribution

In probability theory and statistics, the gamma distribution is a versatile two-parameter family of continuous probability distributions. The exponential distribution, Erlang distribution, and chi-squared distribution are special cases of the gamma distribution. There are two equivalent parameterizations in common use:

  1. With a shape parameter k and a scale parameter θ
  2. With a shape parameter and an inverse scale parameter , called a rate parameter.
<span class="mw-page-title-main">Gumbel distribution</span> Particular case of the generalized extreme value distribution

In probability theory and statistics, the Gumbel distribution is used to model the distribution of the maximum of a number of samples of various distributions.

<i>F</i>-distribution Continuous probability distribution

In probability theory and statistics, the F-distribution or F-ratio, also known as Snedecor's F distribution or the Fisher–Snedecor distribution, is a continuous probability distribution that arises frequently as the null distribution of a test statistic, most notably in the analysis of variance (ANOVA) and other F-tests.

<span class="mw-page-title-main">Logistic distribution</span> Continuous probability distribution

In probability theory and statistics, the logistic distribution is a continuous probability distribution. Its cumulative distribution function is the logistic function, which appears in logistic regression and feedforward neural networks. It resembles the normal distribution in shape but has heavier tails. The logistic distribution is a special case of the Tukey lambda distribution.

In probability theory and statistics, the generalized extreme value (GEV) distribution is a family of continuous probability distributions developed within extreme value theory to combine the Gumbel, Fréchet and Weibull families also known as type I, II and III extreme value distributions. By the extreme value theorem the GEV distribution is the only possible limit distribution of properly normalized maxima of a sequence of independent and identically distributed random variables. Note that a limit distribution needs to exist, which requires regularity conditions on the tail of the distribution. Despite this, the GEV distribution is often used as an approximation to model the maxima of long (finite) sequences of random variables.

<span class="mw-page-title-main">Inverse-gamma distribution</span> Two-parameter family of continuous probability distributions

In probability theory and statistics, the inverse gamma distribution is a two-parameter family of continuous probability distributions on the positive real line, which is the distribution of the reciprocal of a variable distributed according to the gamma distribution.

<span class="mw-page-title-main">Beta prime distribution</span> Probability distribution

In probability theory and statistics, the beta prime distribution is an absolutely continuous probability distribution. If has a beta distribution, then the odds has a beta prime distribution.

<span class="mw-page-title-main">Folded normal distribution</span> Probability distribution

The folded normal distribution is a probability distribution related to the normal distribution. Given a normally distributed random variable X with mean μ and variance σ2, the random variable Y = |X| has a folded normal distribution. Such a case may be encountered if only the magnitude of some variable is recorded, but not its sign. The distribution is called "folded" because probability mass to the left of x = 0 is folded over by taking the absolute value. In the physics of heat conduction, the folded normal distribution is a fundamental solution of the heat equation on the half space; it corresponds to having a perfect insulator on a hyperplane through the origin.

Expected shortfall (ES) is a risk measure—a concept used in the field of financial risk measurement to evaluate the market risk or credit risk of a portfolio. The "expected shortfall at q% level" is the expected return on the portfolio in the worst of cases. ES is an alternative to value at risk that is more sensitive to the shape of the tail of the loss distribution.

<span class="mw-page-title-main">Generalized Pareto distribution</span> Family of probability distributions often used to model tails or extreme values

In statistics, the generalized Pareto distribution (GPD) is a family of continuous probability distributions. It is often used to model the tails of another distribution. It is specified by three parameters: location , scale , and shape . Sometimes it is specified by only scale and shape and sometimes only by its shape parameter. Some references give the shape parameter as .

A ratio distribution is a probability distribution constructed as the distribution of the ratio of random variables having two other known distributions. Given two random variables X and Y, the distribution of the random variable Z that is formed as the ratio Z = X/Y is a ratio distribution.

<span class="mw-page-title-main">Truncated normal distribution</span> Type of probability distribution

In probability and statistics, the truncated normal distribution is the probability distribution derived from a multivariate normally distributed random variable conditioned to taking values in a box, i.e.: the values of each component of the random variable are conditioned to being bounded from either below or above. The truncated normal distribution has wide applications in statistics and econometrics.

<span class="mw-page-title-main">Log-logistic distribution</span>

In probability and statistics, the log-logistic distribution is a continuous probability distribution for a non-negative random variable. It is used in survival analysis as a parametric model for events whose rate increases initially and decreases later, as, for example, mortality rate from cancer following diagnosis or treatment. It has also been used in hydrology to model stream flow and precipitation, in economics as a simple model of the distribution of wealth or income, and in networking to model the transmission times of data considering both the network and the software.

<span class="mw-page-title-main">Shifted log-logistic distribution</span>

The shifted log-logistic distribution is a probability distribution also known as the generalized log-logistic or the three-parameter log-logistic distribution. It has also been called the generalized logistic distribution, but this conflicts with other uses of the term: see generalized logistic distribution.

<span class="mw-page-title-main">Half-normal distribution</span> Probability distribution

In probability theory and statistics, the half-normal distribution is a special case of the folded normal distribution.

<span class="mw-page-title-main">Normal-inverse-gamma distribution</span>

In probability theory and statistics, the normal-inverse-gamma distribution is a four-parameter family of multivariate continuous probability distributions. It is the conjugate prior of a normal distribution with unknown mean and variance.

<span class="mw-page-title-main">Logit-normal distribution</span>

In probability theory, a logit-normal distribution is a probability distribution of a random variable whose logit has a normal distribution. If Y is a random variable with a normal distribution, and t is the standard logistic function, then X = t(Y) has a logit-normal distribution; likewise, if X is logit-normally distributed, then Y = logit(X)= log (X/(1-X)) is normally distributed. It is also known as the logistic normal distribution, which often refers to a multinomial logit version (e.g.).

References

  1. 1 2 Johnson, N.L., Kotz, S., Balakrishnan, N. (1995) Continuous Univariate Distributions, Volume 2, Wiley. ISBN   0-471-58494-0 (pages 140–142)
  2. 1 2 Leigh J. Halliwell (2018). "The Log-Gamma Distribution and Non-Normal Error". S2CID   173176687.{{cite journal}}: Cite journal requires |journal= (help)
  3. C.M.Bishop, Pattern Recognition and Machine Learning, Springer 2006.