Coskewness

Last updated

In probability theory and statistics, coskewness is a measure of how much three random variables change together. Coskewness is the third standardized cross central moment, related to skewness as covariance is related to variance. In 1976, Krauss and Litzenberger used it to examine risk in stock market investments. [1] The application to risk was extended by Harvey and Siddique in 2000. [2]

Contents

If three random variables exhibit positive coskewness they will tend to undergo extreme deviations at the same time, an odd number of which are in the positive direction (so all three random variables undergoing extreme positive deviations, or one undergoing an extreme positive deviation while the other two undergo extreme negative deviations). Similarly, if three random variables exhibit negative coskewness they will tend to undergo extreme deviations at the same time, an even number of which are in the positive direction (so all three random variables undergoing extreme negative deviations, or one undergoing an extreme negative deviation while the other two undergo extreme positive deviations).

Types

There are two different measures for the degree of coskewness in data.

Coskewness

For three random variables X, Y and Z, the non-trivial coskewness statistic is defined as: [3]

where E[X] is the expected value of X, also known as the mean of X, and is the standard deviation of X.

Standardized rank coskewness

Bernard, Chen, Rüschendorf and Vanduffel defined the standardized rank coskewness of three random variables X, Y and Z as: [4]

where FX (X), FY (Y) and FZ (Z) are the cumulative distribution functions of X, Y and Z, respectively.

Properties

Skewness is a special case of the coskewness when the three random variables are identical:

For two random variables, X and Y, the skewness of the sum, X + Y, is

where SX is the skewness of X and is the standard deviation of X. It follows that the sum of two random variables can be skewed (SX+Y  0) even if both random variables have zero skew in isolation (SX = 0 and SY = 0).

The coskewness between variables X and Y does not depend on the scale on which the variables are expressed. If we are analyzing the relationship between X and Y, the coskewness between X and Y will be the same as the coskewness between a + bX and c + dY, where a, b, c, and d are constants.

The standardized rank coskewness RS(X, Y, Z) satisfies the following properties: [4]

(1) −1 ≤ RS(X, Y, Z) ≤ 1.

(2) The upper bound of 1 is obtained by the copula given in (3.3) in Bernard, Chen, Rüschendorf and Vanduffel (2023). The lower bound of −1 is obtained by the copula (3.5) in the same paper.

(3) It is invariant under strictly increasing transformations, i.e., when fi, i = 1, 2, 3, are arbitrary strictly increasing functions, RS(X, Y, Z) = RS(f1 (X), f2 (Y), f3 (Z)).

(4) RS(X, Y, Z) = 0 if X, Y and Z are independent.

Example

Let X be standard normally distributed and Y be the distribution obtained by setting X=Y whenever X<0 and drawing Y independently from a standard half-normal distribution whenever X>0. In other words, X and Y are both standard normally distributed with the property that they are completely correlated for negative values and uncorrelated apart from sign for positive values. The joint probability density function is

where H(x) is the Heaviside step function and δ(x) is the Dirac delta function. The third moments are easily calculated by integrating with respect to this density:

Note that although X and Y are individually standard normally distributed, the distribution of the sum X+Y is significantly skewed. From integration with respect to density, we find that the covariance of X and Y is

from which it follows that the standard deviation of their sum is

Using the skewness sum formula above, we have

This can also be computed directly from the probability density function of the sum:

Bernard, Chen, Rüschendorf and Vanduffel (2023) found risk bounds on coskewness for some popular marginal distributions as shown in the following table. [4]

Marginal distributionsMinimum coskewnessMaximum coskewness
N(, )
Student(),
Laplace(, )
U(, )

where is the gamma function.

See also

Related Research Articles

In probability theory and statistics, kurtosis is a measure of the "tailedness" of the probability distribution of a real-valued random variable. Like skewness, kurtosis describes a particular aspect of a probability distribution. There are different ways to quantify kurtosis for a theoretical distribution, and there are corresponding ways of estimating it using a sample from a population. Different measures of kurtosis may have different interpretations.

<span class="mw-page-title-main">Normal distribution</span> Probability distribution

In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is

<span class="mw-page-title-main">Full width at half maximum</span> Concept in statistics and wave theory

In a distribution, full width at half maximum (FWHM) is the difference between the two values of the independent variable at which the dependent variable is equal to half of its maximum value. In other words, it is the width of a spectrum curve measured between those points on the y-axis which are half the maximum amplitude. Half width at half maximum (HWHM) is half of the FWHM if the function is symmetric. The term full duration at half maximum (FDHM) is preferred when the independent variable is time.

<span class="mw-page-title-main">Multivariate normal distribution</span> Generalization of the one-dimensional normal distribution to higher dimensions

In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One definition is that a random vector is said to be k-variate normally distributed if every linear combination of its k components has a univariate normal distribution. Its importance derives mainly from the multivariate central limit theorem. The multivariate normal distribution is often used to describe, at least approximately, any set of (possibly) correlated real-valued random variables each of which clusters around a mean value.

<span class="mw-page-title-main">Log-normal distribution</span> Probability distribution

In probability theory, a log-normal (or lognormal) distribution is a continuous probability distribution of a random variable whose logarithm is normally distributed. Thus, if the random variable X is log-normally distributed, then Y = ln(X) has a normal distribution. Equivalently, if Y has a normal distribution, then the exponential function of Y, X = exp(Y), has a log-normal distribution. A random variable which is log-normally distributed takes only positive real values. It is a convenient and useful model for measurements in exact and engineering sciences, as well as medicine, economics and other topics (e.g., energies, concentrations, lengths, prices of financial instruments, and other metrics).

<span class="mw-page-title-main">Error function</span> Sigmoid shape special function

In mathematics, the error function, often denoted by erf, is a complex function of a complex variable defined as:

In mathematics, a Gaussian function, often simply referred to as a Gaussian, is a function of the base form

<span class="mw-page-title-main">Rayleigh distribution</span> Probability distribution

In probability theory and statistics, the Rayleigh distribution is a continuous probability distribution for nonnegative-valued random variables. Up to rescaling, it coincides with the chi distribution with two degrees of freedom. The distribution is named after Lord Rayleigh.

<span class="mw-page-title-main">Fisher transformation</span> Statistical transformation

In statistics, the Fisher transformation of a Pearson correlation coefficient is its inverse hyperbolic tangent (artanh). When the sample correlation coefficient r is near 1 or -1, its distribution is highly skewed, which makes it difficult to estimate confidence intervals and apply tests of significance for the population correlation coefficient ρ. The Fisher transformation solves this problem by yielding a variable whose distribution is approximately normally distributed, with a variance that is stable over different values of r.

<span class="mw-page-title-main">Voigt profile</span>

The Voigt profile is a probability distribution given by a convolution of a Cauchy-Lorentz distribution and a Gaussian distribution. It is often used in analyzing data from spectroscopy or diffraction.

<span class="mw-page-title-main">Rice distribution</span> Probability distribution

In probability theory, the Rice distribution or Rician distribution is the probability distribution of the magnitude of a circularly-symmetric bivariate normal random variable, possibly with non-zero mean (noncentral). It was named after Stephen O. Rice (1907–1986).

<span class="mw-page-title-main">Chi distribution</span>

In probability theory and statistics, the chi distribution is a continuous probability distribution. It is the distribution of the positive square root of the sum of squares of a set of independent random variables each following a standard normal distribution, or equivalently, the distribution of the Euclidean distance of the random variables from the origin. It is thus related to the chi-squared distribution by describing the distribution of the positive square roots of a variable obeying a chi-squared distribution.

In probability theory, calculation of the sum of normally distributed random variables is an instance of the arithmetic of random variables, which can be quite complex based on the probability distributions of the random variables involved and their relationships.

<span class="mw-page-title-main">Inverse Gaussian distribution</span> Family of continuous probability distributions

In probability theory, the inverse Gaussian distribution is a two-parameter family of continuous probability distributions with support on (0,∞).

In statistics, the multivariate t-distribution is a multivariate probability distribution. It is a generalization to random vectors of the Student's t-distribution, which is a distribution applicable to univariate random variables. While the case of a random matrix could be treated within this structure, the matrix t-distribution is distinct and makes particular use of the matrix structure.

A ratio distribution is a probability distribution constructed as the distribution of the ratio of random variables having two other known distributions. Given two random variables X and Y, the distribution of the random variable Z that is formed as the ratio Z = X/Y is a ratio distribution.

<span class="mw-page-title-main">Half-normal distribution</span> Probability distribution

In probability theory and statistics, the half-normal distribution is a special case of the folded normal distribution.

<span class="mw-page-title-main">Exponentially modified Gaussian distribution</span> Describes the sum of independent normal and exponential random variables

In probability theory, an exponentially modified Gaussian distribution describes the sum of independent normal and exponential random variables. An exGaussian random variable Z may be expressed as Z = X + Y, where X and Y are independent, X is Gaussian with mean μ and variance σ2, and Y is exponential of rate λ. It has a characteristic positive skew from the exponential component.

In statistics and probability theory, the nonparametric skew is a statistic occasionally used with random variables that take real values. It is a measure of the skewness of a random variable's distribution—that is, the distribution's tendency to "lean" to one side or the other of the mean. Its calculation does not require any knowledge of the form of the underlying distribution—hence the name nonparametric. It has some desirable properties: it is zero for any symmetric distribution; it is unaffected by a scale shift; and it reveals either left- or right-skewness equally well. In some statistical samples it has been shown to be less powerful than the usual measures of skewness in detecting departures of the population from normality.

In probability theory and statistics, cokurtosis is a measure of how much two random variables change together. Cokurtosis is the fourth standardized cross central moment. If two random variables exhibit a high level of cokurtosis they will tend to undergo extreme positive and negative deviations at the same time.

References

  1. Friend, Irwin; Randolf Westerfield (1980). "Co-Skewness and Capital Asset Pricing". The Journal of Finance. 35 (4): 897–913. doi:10.1111/j.1540-6261.1980.tb03508.x.
  2. Jondeau, Eric; Ser-Huang Poon; Michael Rockinger (2007). Financial Modeling Under Non-Gaussian Distributions. Springer. pp. 31–32. ISBN   978-1-84628-696-4.
  3. Miller, Michael B. (2014). "Chapter 3. Basic Statistics". Mathematics and Statistics for Financial Risk Management (2nd ed.). Hoboken, New Jersey: John Wiley & Sons, Inc. pp. 53–56. ISBN   978-1-118-75029-2.
  4. 1 2 3 Bernard, Carole; Jinghui, Chen; Rüschendorf, Ludger; Vanduffel, Steven (5 May 2023). "Coskewness under dependence uncertainty". Statistics and Probability Letters. 199 (8).

Further reading