Log-Cauchy distribution

Last updated
Log-Cauchy
Probability density function
Logcauchypdf.svg
Cumulative distribution function
Logcauchycdf.svg
Parameters (real)
(real)
Support
PDF
CDF
Mean infinite
Median
Variance infinite
Skewness does not exist
Ex. kurtosis does not exist
MGF does not exist

In probability theory, a log-Cauchy distribution is a probability distribution of a random variable whose logarithm is distributed in accordance with a Cauchy distribution. If X is a random variable with a Cauchy distribution, then Y = exp(X) has a log-Cauchy distribution; likewise, if Y has a log-Cauchy distribution, then X = log(Y) has a Cauchy distribution. [1]

Contents

Characterization

The log-Cauchy distribution is a special case of the log-t distribution where the degrees of freedom parameter is equal to 1. [2]

Probability density function

The log-Cauchy distribution has the probability density function:

where is a real number and . [1] [3] If is known, the scale parameter is . [1] and correspond to the location parameter and scale parameter of the associated Cauchy distribution. [1] [4] Some authors define and as the location and scale parameters, respectively, of the log-Cauchy distribution. [4]

For and , corresponding to a standard Cauchy distribution, the probability density function reduces to: [5]

Cumulative distribution function

The cumulative distribution function (cdf) when and is: [5]

Survival function

The survival function when and is: [5]

Hazard rate

The hazard rate when and is: [5]

The hazard rate decreases at the beginning and at the end of the distribution, but there may be an interval over which the hazard rate increases. [5]

Properties

The log-Cauchy distribution is an example of a heavy-tailed distribution. [6] Some authors regard it as a "super-heavy tailed" distribution, because it has a heavier tail than a Pareto distribution-type heavy tail, i.e., it has a logarithmically decaying tail. [6] [7] As with the Cauchy distribution, none of the non-trivial moments of the log-Cauchy distribution are finite. [5] The mean is a moment so the log-Cauchy distribution does not have a defined mean or standard deviation. [8] [9]

The log-Cauchy distribution is infinitely divisible for some parameters but not for others. [10] Like the lognormal distribution, log-t or log-Student distribution and Weibull distribution, the log-Cauchy distribution is a special case of the generalized beta distribution of the second kind. [11] [12] The log-Cauchy is actually a special case of the log-t distribution, similar to the Cauchy distribution being a special case of the Student's t distribution with 1 degree of freedom. [13] [14]

Since the Cauchy distribution is a stable distribution, the log-Cauchy distribution is a logstable distribution. [15] Logstable distributions have poles at x=0. [14]

Estimating parameters

The median of the natural logarithms of a sample is a robust estimator of . [1] The median absolute deviation of the natural logarithms of a sample is a robust estimator of . [1]

Uses

In Bayesian statistics, the log-Cauchy distribution can be used to approximate the improper Jeffreys-Haldane density, 1/k, which is sometimes suggested as the prior distribution for k where k is a positive parameter being estimated. [16] [17] The log-Cauchy distribution can be used to model certain survival processes where significant outliers or extreme results may occur. [3] [4] [18] An example of a process where a log-Cauchy distribution may be an appropriate model is the time between someone becoming infected with HIV and showing symptoms of the disease, which may be very long for some people. [4] It has also been proposed as a model for species abundance patterns. [19]

Related Research Articles

<span class="mw-page-title-main">Cauchy distribution</span> Probability distribution

The Cauchy distribution, named after Augustin Cauchy, is a continuous probability distribution. It is also known, especially among physicists, as the Lorentz distribution, Cauchy–Lorentz distribution, Lorentz(ian) function, or Breit–Wigner distribution. The Cauchy distribution is the distribution of the x-intercept of a ray issuing from with a uniformly distributed angle. It is also the distribution of the ratio of two independent normally distributed random variables with mean zero.

<span class="mw-page-title-main">Normal distribution</span> Probability distribution

In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is

<span class="mw-page-title-main">Multivariate normal distribution</span> Generalization of the one-dimensional normal distribution to higher dimensions

In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One definition is that a random vector is said to be k-variate normally distributed if every linear combination of its k components has a univariate normal distribution. Its importance derives mainly from the multivariate central limit theorem. The multivariate normal distribution is often used to describe, at least approximately, any set of (possibly) correlated real-valued random variables each of which clusters around a mean value.

<span class="mw-page-title-main">Log-normal distribution</span> Probability distribution

In probability theory, a log-normal (or lognormal) distribution is a continuous probability distribution of a random variable whose logarithm is normally distributed. Thus, if the random variable X is log-normally distributed, then Y = ln(X) has a normal distribution. Equivalently, if Y has a normal distribution, then the exponential function of Y, X = exp(Y), has a log-normal distribution. A random variable which is log-normally distributed takes only positive real values. It is a convenient and useful model for measurements in exact and engineering sciences, as well as medicine, economics and other topics (e.g., energies, concentrations, lengths, prices of financial instruments, and other metrics).

In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data. This is achieved by maximizing a likelihood function so that, under the assumed statistical model, the observed data is most probable. The point in the parameter space that maximizes the likelihood function is called the maximum likelihood estimate. The logic of maximum likelihood is both intuitive and flexible, and as such the method has become a dominant means of statistical inference.

<span class="mw-page-title-main">Gumbel distribution</span> Particular case of the generalized extreme value distribution

In probability theory and statistics, the Gumbel distribution is used to model the distribution of the maximum of a number of samples of various distributions.

<span class="mw-page-title-main">Lévy distribution</span>

In probability theory and statistics, the Lévy distribution, named after Paul Lévy, is a continuous probability distribution for a non-negative random variable. In spectroscopy, this distribution, with frequency as the dependent variable, is known as a van der Waals profile. It is a special case of the inverse-gamma distribution. It is a stable distribution.

<span class="mw-page-title-main">Directional statistics</span>

Directional statistics is the subdiscipline of statistics that deals with directions, axes or rotations in Rn. More generally, directional statistics deals with observations on compact Riemannian manifolds including the Stiefel manifold.

In probability theory and statistics, the generalized extreme value (GEV) distribution is a family of continuous probability distributions developed within extreme value theory to combine the Gumbel, Fréchet and Weibull families also known as type I, II and III extreme value distributions. By the extreme value theorem the GEV distribution is the only possible limit distribution of properly normalized maxima of a sequence of independent and identically distributed random variables. Note that a limit distribution needs to exist, which requires regularity conditions on the tail of the distribution. Despite this, the GEV distribution is often used as an approximation to model the maxima of long (finite) sequences of random variables.

von Mises distribution Probability distribution on the circle

In probability theory and directional statistics, the von Mises distribution is a continuous probability distribution on the circle. It is a close approximation to the wrapped normal distribution, which is the circular analogue of the normal distribution. A freely diffusing angle on a circle is a wrapped normally distributed random variable with an unwrapped variance that grows linearly in time. On the other hand, the von Mises distribution is the stationary distribution of a drift and diffusion process on the circle in a harmonic potential, i.e. with a preferred orientation. The von Mises distribution is the maximum entropy distribution for circular data when the real and imaginary parts of the first circular moment are specified. The von Mises distribution is a special case of the von Mises–Fisher distribution on the N-dimensional sphere.

Differential entropy is a concept in information theory that began as an attempt by Claude Shannon to extend the idea of (Shannon) entropy, a measure of average surprisal of a random variable, to continuous probability distributions. Unfortunately, Shannon did not derive this formula, and rather just assumed it was the correct continuous analogue of discrete entropy, but it is not. The actual continuous version of discrete entropy is the limiting density of discrete points (LDDP). Differential entropy is commonly encountered in the literature, but it is a limiting case of the LDDP, and one that loses its fundamental association with discrete entropy.

Expected shortfall (ES) is a risk measure—a concept used in the field of financial risk measurement to evaluate the market risk or credit risk of a portfolio. The "expected shortfall at q% level" is the expected return on the portfolio in the worst of cases. ES is an alternative to value at risk that is more sensitive to the shape of the tail of the loss distribution.

<span class="mw-page-title-main">Generalized Pareto distribution</span> Family of probability distributions often used to model tails or extreme values

In statistics, the generalized Pareto distribution (GPD) is a family of continuous probability distributions. It is often used to model the tails of another distribution. It is specified by three parameters: location , scale , and shape . Sometimes it is specified by only scale and shape and sometimes only by its shape parameter. Some references give the shape parameter as .

A ratio distribution is a probability distribution constructed as the distribution of the ratio of random variables having two other known distributions. Given two random variables X and Y, the distribution of the random variable Z that is formed as the ratio Z = X/Y is a ratio distribution.

Tail value at risk (TVaR), also known as tail conditional expectation (TCE) or conditional tail expectation (CTE), is a risk measure associated with the more general value at risk. It quantifies the expected value of the loss given that an event outside a given probability level has occurred.

<span class="mw-page-title-main">Wrapped normal distribution</span>

In probability theory and directional statistics, a wrapped normal distribution is a wrapped probability distribution that results from the "wrapping" of the normal distribution around the unit circle. It finds application in the theory of Brownian motion and is a solution to the heat equation for periodic boundary conditions. It is closely approximated by the von Mises distribution, which, due to its mathematical simplicity and tractability, is the most commonly used distribution in directional statistics.

<span class="mw-page-title-main">Wrapped Cauchy distribution</span>

In probability theory and directional statistics, a wrapped Cauchy distribution is a wrapped probability distribution that results from the "wrapping" of the Cauchy distribution around the unit circle. The Cauchy distribution is sometimes known as a Lorentzian distribution, and the wrapped Cauchy distribution may sometimes be referred to as a wrapped Lorentzian distribution.

In probability theory and statistics, an inverse distribution is the distribution of the reciprocal of a random variable. Inverse distributions arise in particular in the Bayesian context of prior distributions and posterior distributions for scale parameters. In the algebra of random variables, inverse distributions are special cases of the class of ratio distributions, in which the numerator random variable has a degenerate distribution.

<span class="mw-page-title-main">ProbOnto</span>

ProbOnto is a knowledge base and ontology of probability distributions. ProbOnto 2.5 contains over 150 uni- and multivariate distributions and alternative parameterizations, more than 220 relationships and re-parameterization formulas, supporting also the encoding of empirical and univariate mixture distributions.

In probability theory, a log-t distribution or log-Student t distribution is a probability distribution of a random variable whose logarithm is distributed in accordance with a Student's t-distribution. If X is a random variable with a Student's t-distribution, then Y = exp(X) has a log-t distribution; likewise, if Y has a log-t distribution, then X = log(Y) has a Student's t-distribution.

References

  1. 1 2 3 4 5 6 Olive, D.J. (June 23, 2008). "Applied Robust Statistics" (PDF). Southern Illinois University. p. 86. Archived from the original (PDF) on September 28, 2011. Retrieved 2011-10-18.
  2. Olosunde, Akinlolu & Olofintuade, Sylvester (January 2022). "Some Inferential Problems from Log Student's T-distribution and its Multivariate Extension". Revista Colombiana de Estadística - Applied Statistics. 45 (1): 209–229. doi: 10.15446/rce.v45n1.90672 . Retrieved 2022-04-01.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  3. 1 2 Lindsey, J.K. (2004). Statistical analysis of stochastic processes in time . Cambridge University Press. pp.  33, 50, 56, 62, 145. ISBN   978-0-521-83741-5.
  4. 1 2 3 4 Mode, C.J. & Sleeman, C.K. (2000). Stochastic processes in epidemiology: HIV/AIDS, other infectious diseases . World Scientific. pp.  29–37. ISBN   978-981-02-4097-4.
  5. 1 2 3 4 5 6 Marshall, A.W. & Olkin, I. (2007). Life distributions: structure of nonparametric, semiparametric, and parametric families . Springer. pp.  443–444. ISBN   978-0-387-20333-1.
  6. 1 2 Falk, M.; Hüsler, J. & Reiss, R. (2010). Laws of Small Numbers: Extremes and Rare Events . Springer. p.  80. ISBN   978-3-0348-0008-2.
  7. Alves, M.I.F.; de Haan, L. & Neves, C. (March 10, 2006). "Statistical inference for heavy and super-heavy tailed distributions" (PDF). Archived from the original (PDF) on June 23, 2007.
  8. "Moment". Mathworld . Retrieved 2011-10-19.
  9. Wang, Y. "Trade, Human Capital and Technology Spillovers: An Industry Level Analysis". Carleton University: 14.{{cite journal}}: Cite journal requires |journal= (help)
  10. Bondesson, L. (2003). "On the Lévy Measure of the Lognormal and LogCauchy Distributions". Methodology and Computing in Applied Probability: 243–256. Archived from the original on 2012-04-25. Retrieved 2011-10-18.
  11. Knight, J. & Satchell, S. (2001). Return distributions in finance . Butterworth-Heinemann. p.  153. ISBN   978-0-7506-4751-9.
  12. Kemp, M. (2009). Market consistency: model calibration in imperfect markets. Wiley. ISBN   978-0-470-77088-7.
  13. MacDonald, J.B. (1981). "Measuring Income Inequality". In Taillie, C.; Patil, G.P.; Baldessari, B. (eds.). Statistical distributions in scientific work: proceedings of the NATO Advanced Study Institute. Springer. p. 169. ISBN   978-90-277-1334-6.
  14. 1 2 Kleiber, C. & Kotz, S. (2003). Statistical Size Distributions in Economics and Actuarial Science . Wiley. pp.  101–102, 110. ISBN   978-0-471-15064-0.
  15. Panton, D.B. (May 1993). "Distribution function values for logstable distributions". Computers & Mathematics with Applications. 25 (9): 17–24. doi: 10.1016/0898-1221(93)90128-I .
  16. Good, I.J. (1983). Good thinking: the foundations of probability and its applications. University of Minnesota Press. p. 102. ISBN   978-0-8166-1142-3.
  17. Chen, M. (2010). Frontiers of Statistical Decision Making and Bayesian Analysis. Springer. p. 12. ISBN   978-1-4419-6943-9.
  18. Lindsey, J.K.; Jones, B. & Jarvis, P. (September 2001). "Some statistical issues in modelling pharmacokinetic data". Statistics in Medicine. 20 (17–18): 2775–278. doi:10.1002/sim.742. PMID   11523082. S2CID   41887351.
  19. Zuo-Yun, Y.; et al. (June 2005). "LogCauchy, log-sech and lognormal distributions of species abundances in forest communities". Ecological Modelling. 184 (2–4): 329–340. doi:10.1016/j.ecolmodel.2004.10.011.