Location testing for Gaussian scale mixture distributions

Last updated

In statistics, the topic of location testing for Gaussian scale mixture distributions arises in some particular types of situations where the more standard Student's t-test is inapplicable. Specifically, these cases allow tests of location to be made where the assumption that sample observations arise from populations having a normal distribution can be replaced by the assumption that they arise from a Gaussian scale mixture distribution. The class of Gaussian scale mixture distributions contains all symmetric stable distributions, Laplace distributions, logistic distributions, and exponential power distributions, etc. [1] [2]

Introduce

tGn(x),

the counterpart of Student's t-distribution for Gaussian scale mixtures. This means that if we test the null hypothesis that the center of a Gaussian scale mixture distribution is 0, say, then tnG(x) (x  0) is the infimum of all monotone nondecreasing functions u(x)  1/2, x  0 such that if the critical values of the test are u1(1  α), then the significance level is at most α  1/2 for all Gaussian scale mixture distributions [tGn(x) = 1  tGn(x),for x < 0]. An explicit formula for tGn(x), is given in the papers in the references in terms of Student’s t-distributions, tk, k = 1, 2, ..., n. Introduce

ΦG(x):= limn  tGn(x),

the Gaussian scale mixture counterpart of the standard normal cumulative distribution function, Φ(x).

Theorem. ΦG(x) = 1/2 for 0  x < 1, ΦG(1) = 3/4, ΦG(x) = C(x/(2  x2)1/2) for quantiles between 1/2 and 0.875, where C(x) is the standard Cauchy cumulative distribution function. This is the convex part of the curve ΦG(x), x  0 which is followed by a linear section ΦG(x) = x/(23) + 1/2 for 1.3136... < x < 1.4282... Thus the 90% quantile is exactly 43/5. Most importantly,

ΦG(x) = Φ(x) for x3.

Note that Φ(3) = 0.958..., thus the classical 95% confidence interval for the unknown expected value of Gaussian distributions covers the center of symmetry with at least 95% probability for Gaussian scale mixture distributions. On the other hand, the 90% quantile of ΦG(x) is 43/5 = 1.385... > Φ1(0.9) = 1.282... The following critical values are important in applications: 0.95 = Φ(1.645) = ΦG(1.651), and 0.9 = Φ(1.282) = ΦG(1.386). [3]

For the extension of the Theorem to all symmetric unimodal distributions one can start with a classical result of Aleksandr Khinchin: namely that all symmetric unimodal distributions are scale mixtures of symmetric uniform distributions.

Open problem

The counterpart of the Theorem above for the class of all symmetric distributions, or equivalently, for the class of scale mixtures of coin flipping random variables, leads to the following problem: [4]

How many vertices of an n-dimensional unit cube can be covered by a sphere with given radius r (and varying center)? Answer this question for all positive integers n and all positive real numbers r. (Certain special cases can be easy to compute.)

Related Research Articles

<span class="mw-page-title-main">Normal distribution</span> Probability distribution

In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is

<span class="mw-page-title-main">Skewness</span> Measure of the asymmetry of random variables

In probability theory and statistics, skewness is a measure of the asymmetry of the probability distribution of a real-valued random variable about its mean. The skewness value can be positive, zero, negative, or undefined.

In probability theory, the central limit theorem (CLT) states that, under appropriate conditions, the distribution of a normalized version of the sample mean converges to a standard normal distribution. This holds even if the original variables themselves are not normally distributed. There are several versions of the CLT, each applying in the context of different conditions.

<span class="mw-page-title-main">Student's t-distribution</span> Probability distribution

In probability and statistics, Student's t distribution is a continuous probability distribution that generalizes the standard normal distribution. Like the latter, it is symmetric around zero and bell-shaped.

<span class="mw-page-title-main">Normal probability plot</span> Graphical technique in statistics

The normal probability plot is a graphical technique to identify substantive departures from normality. This includes identifying outliers, skewness, kurtosis, a need for transformations, and mixtures. Normal probability plots are made of raw data, residuals from model fits, and estimated parameters.

In probability theory and statistics, a scale parameter is a special kind of numerical parameter of a parametric family of probability distributions. The larger the scale parameter, the more spread out the distribution.

In statistics, a mixture model is a probabilistic model for representing the presence of subpopulations within an overall population, without requiring that an observed data set should identify the sub-population to which an individual observation belongs. Formally a mixture model corresponds to the mixture distribution that represents the probability distribution of observations in the overall population. However, while problems associated with "mixture distributions" relate to deriving the properties of the overall population from those of the sub-populations, "mixture models" are used to make statistical inferences about the properties of the sub-populations given only observations on the pooled population, without sub-population identity information. Mixture models are used for clustering, under the name model-based clustering, and also for density estimation.

<span class="mw-page-title-main">Multimodal distribution</span> Probability distribution with more than one mode

In statistics, a multimodaldistribution is a probability distribution with more than one mode. These appear as distinct peaks in the probability density function, as shown in Figures 1 and 2. Categorical, continuous, and discrete data can all form multimodal distributions. Among univariate analyses, multimodal distributions are commonly bimodal.

In mathematics, unimodality means possessing a unique mode. More generally, unimodality means there is only a single highest value, somehow defined, of some mathematical object.

<span class="mw-page-title-main">Hyperbolic secant distribution</span> Continuous probability distribution

In probability theory and statistics, the hyperbolic secant distribution is a continuous probability distribution whose probability density function and characteristic function are proportional to the hyperbolic secant function. The hyperbolic secant function is equivalent to the reciprocal hyperbolic cosine, and thus this distribution is also called the inverse-cosh distribution.

<span class="mw-page-title-main">Empirical distribution function</span> Distribution function associated with the empirical measure of a sample

In statistics, an empirical distribution function is the distribution function associated with the empirical measure of a sample. This cumulative distribution function is a step function that jumps up by 1/n at each of the n data points. Its value at any specified value of the measured variable is the fraction of observations of the measured variable that are less than or equal to the specified value.

<span class="mw-page-title-main">Quantile function</span> Statistical function that defines the quantiles of a probability distribution

In probability and statistics, the quantile function outputs the value of a random variable such that its probability is less than or equal to an input probability value. Intuitively, the quantile function associates with a range at and below a probability input the likelihood that a random variable is realized in that range for some probability distribution. It is also called the percentile function, percent-point function, inverse cumulative distribution function or inverse distribution function.

<span class="mw-page-title-main">Studentized range distribution</span>

In probability and statistics, studentized range distribution is the continuous probability distribution of the studentized range of an i.i.d. sample from a normally distributed population.

The Birnbaum–Saunders distribution, also known as the fatigue life distribution, is a probability distribution used extensively in reliability applications to model failure times. There are several alternative formulations of this distribution in the literature. It is named after Z. W. Birnbaum and S. C. Saunders.

<span class="mw-page-title-main">Skew normal distribution</span> Probability distribution

In probability theory and statistics, the skew normal distribution is a continuous probability distribution that generalises the normal distribution to allow for non-zero skewness.

Algorithmic inference gathers new developments in the statistical inference methods made feasible by the powerful computing devices widely available to any data analyst. Cornerstones in this field are computational learning theory, granular computing, bioinformatics, and, long ago, structural probability . The main focus is on the algorithms which compute statistics rooting the study of a random phenomenon, along with the amount of data they must feed on to produce reliable results. This shifts the interest of mathematicians from the study of the distribution laws to the functional properties of the statistics, and the interest of computer scientists from the algorithms for processing data to the information they process.

In statistics and probability theory, the nonparametric skew is a statistic occasionally used with random variables that take real values. It is a measure of the skewness of a random variable's distribution—that is, the distribution's tendency to "lean" to one side or the other of the mean. Its calculation does not require any knowledge of the form of the underlying distribution—hence the name nonparametric. It has some desirable properties: it is zero for any symmetric distribution; it is unaffected by a scale shift; and it reveals either left- or right-skewness equally well. In some statistical samples it has been shown to be less powerful than the usual measures of skewness in detecting departures of the population from normality.

<span class="mw-page-title-main">Symmetric probability distribution</span>

In statistics, a symmetric probability distribution is a probability distribution—an assignment of probabilities to possible occurrences—which is unchanged when its probability density function or probability mass function is reflected around a vertical line at some value of the random variable represented by the distribution. This vertical line is the line of symmetry of the distribution. Thus the probability of being any given distance on one side of the value about which symmetry occurs is the same as the probability of being the same distance on the other side of that value.

References

  1. Andrews, D. and C Mallows, C. (1974) “Scale mixtures of normal distributions,” Journal of the Royal Statistical Society, 36, 99–102 JSTOR   2984774
  2. West, M. (1987) "On scale mixtures of normal distributions", Biometrika , 74(3), 646648 doi : 10.1093/biomet/74.3.646
  3. Bakirov, N.K. and Székely, G. J (2005). "Students’ t-test for Gaussian scale mixtures" (alternative link) Zapiski Nauchnyh Seminarov POMI, 328, Probability and Statistics. Part 9 (editor V.N.Sudakov) 519. Reprinted (2006): Journal of Mathematical Sciences, 139 (3) 64976505 doi : 10.1007/s10958-006-0366-5 .
  4. Székely, G. J. (2004/2006). "Student’s t-test for scale mixture errors", Optimality: The Second Erich L. Lehmann Symposium, May 1922, 2004, Rice University, Ed. Rojo, J. Lecture NotesMonograph Series, Number 49, Beachwood, Ohio, Institute of Mathematical Statistics, 1018. doi : 10.1214/074921706000000365.