Studentized range distribution

Last updated
Studentized range distribution
Probability density function
StudentizedRangePDF.svg
Cumulative distribution function
StudentizedRangeCDF.svg
Parameters k > 1, the number of groups
> 0, the degrees of freedom
Support
PDF
CDF

In probability and statistics, studentized range distribution is the continuous probability distribution of the studentized range of an i.i.d. sample from a normally distributed population.

Contents

Suppose that we take a sample of size n from each of k populations with the same normal distribution N(μ, σ2) and suppose that is the smallest of these sample means and is the largest of these sample means, and suppose s² is the pooled sample variance from these samples. Then the following statistic has a Studentized range distribution.

Definition

Probability density function

Differentiating the cumulative distribution function with respect to q gives the probability density function.

Note that in the outer part of the integral, the equation

was used to replace an exponential factor.

Cumulative distribution function

The cumulative distribution function is given by [1]

Special cases

If k is 2 or 3, [2] the studentized range probability distribution function can be directly evaluated, where is the standard normal probability density function and is the standard normal cumulative distribution function.

When the degrees of freedom approaches infinity the studentized range cumulative distribution can be calculated for any k using the standard normal distribution.

Applications

Critical values of the studentized range distribution are used in Tukey's range test. [3]

The studentized range is used to calculate significance levels for results obtained by data mining, where one selectively seeks extreme differences in sample data, rather than only sampling randomly.

The Studentized range distribution has applications to hypothesis testing and multiple comparisons procedures. For example, Tukey's range test and Duncan's new multiple range test (MRT), in which the sample x1, ..., xn is a sample of means and q is the basic test-statistic, can be used as post-hoc analysis to test between which two groups means there is a significant difference (pairwise comparisons) after rejecting the null hypothesis that all groups are from the same population (i.e. all means are equal) by the standard analysis of variance. [4]

When only the equality of the two groups means is in question (i.e. whether μ1 = μ2), the studentized range distribution is similar to the Student's t distribution, differing only in that the first takes into account the number of means under consideration, and the critical value is adjusted accordingly. The more means under consideration, the larger the critical value is. This makes sense since the more means there are, the greater the probability that at least some differences between pairs of means will be significantly large due to chance alone.

Derivation

The studentized range distribution function arises from re-scaling the sample range R by the sample standard deviation s, since the studentized range is customarily tabulated in units of standard deviations, with the variable q = Rs. The derivation begins with a perfectly general form of the distribution function of the sample range, which applies to any sample data distribution.

In order to obtain the distribution in terms of the "studentized" range q, we will change variable from R to s and q. Assuming the sample data is normally distributed, the standard deviation s will be χ distributed. By further integrating over s we can remove s as a parameter and obtain the re-scaled distribution in terms of q alone.

General form

For any probability density function fX, the range probability density fR is: [2]

What this means is that we are adding up the probabilities that, given k draws from a distribution, two of them differ by r, and the remaining k  2 draws all fall between the two extreme values. If we change variables to u where is the low-end of the range, and define FX as the cumulative distribution function of fX, then the equation can be simplified:

We introduce a similar integral, and notice that differentiating under the integral-sign gives

which recovers the integral above, [lower-alpha 1] so that last relation confirms

because for any continuous cdf

Special form for normal data

The range distribution is most often used for confidence intervals around sample averages, which are asymptotically normally distributed by the central limit theorem.

In order to create the studentized range distribution for normal data, we first switch from the generic fX and FX to the distribution functions φ and Φ for the standard normal distribution, and change the variable r to s·q, where q is a fixed factor that re-scales r by scaling factor s:

Choose the scaling factor s to be the sample standard deviation, so that q becomes the number of standard deviations wide that the range is. For normal data s is chi distributed [lower-alpha 2] and the distribution function fS of the chi distribution is given by:

Multiplying the distributions fR and fS and integrating to remove the dependence on the standard deviation s gives the studentized range distribution function for normal data:

where

q is the width of the data range measured in standard deviations,
ν is the number of degrees of freedom for determining the sample standard deviation, [lower-alpha 3] and
k is the number of separate averages that form the points within the range.

The equation for the pdf shown in the sections above comes from using

to replace the exponential expression in the outer integral.

Notes

  1. Technically, the relation is only true for points where , which holds everywhere for normal data as discussed in the next section, but not for distributions whose support has an upper bound, like uniformly distributed data.
  2. Note well the absence of "squared": The text refers to the χ distribution, not the χ2 distribution.
  3. Usually , where n is the total number of all datapoints used to find the averages that are the values in the range.

Related Research Articles

Normal distribution Probability distribution

In statistics, a normal distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is

Dirac delta function Pseudo-function δ such that an integral of δ(x-c)f(x) always takes the value of f(c)

In mathematics, the Dirac delta distribution, also known as the unit impulse symbol, is a generalized function or distribution over the real numbers, whose value is zero everywhere except at zero, and whose integral over the entire real line is equal to one.

Stellar dynamics

Stellar dynamics is the branch of astrophysics which describes in a statistical way the collective motions of stars subject to their mutual gravity. The essential difference from celestial mechanics is that the number of body

In mathematics, the total variation identifies several slightly different concepts, related to the structure of the codomain of a function or a measure. For a real-valued continuous function f, defined on an interval [a, b] ⊂ R, its total variation on the interval of definition is a measure of the one-dimensional arclength of the curve with parametric equation xf(x), for x ∈ [a, b]. Functions whose total variation is finite are called functions of bounded variation.

In the theory of stochastic processes, the Karhunen–Loève theorem, also known as the Kosambi–Karhunen–Loève theorem is a representation of a stochastic process as an infinite linear combination of orthogonal functions, analogous to a Fourier series representation of a function on a bounded interval. The transformation is also known as Hotelling transform and eigenvector transform, and is closely related to principal component analysis (PCA) technique widely used in image processing and in data analysis in many fields.

Stable distribution Distribution of variables which satisfies a stability property under linear combinations

In probability theory, a distribution is said to be stable if a linear combination of two independent random variables with this distribution has the same distribution, up to location and scale parameters. A random variable is said to be stable if its distribution is stable. The stable distribution family is also sometimes referred to as the Lévy alpha-stable distribution, after Paul Lévy, the first mathematician to have studied it.

Lévy distribution

In probability theory and statistics, the Lévy distribution, named after Paul Lévy, is a continuous probability distribution for a non-negative random variable. In spectroscopy, this distribution, with frequency as the dependent variable, is known as a van der Waals profile. It is a special case of the inverse-gamma distribution. It is a stable distribution.

LSZ reduction formula

In quantum field theory, the LSZ reduction formula is a method to calculate S-matrix elements from the time-ordered correlation functions of a quantum field theory. It is a step of the path that starts from the Lagrangian of some quantum field theory and leads to prediction of measurable quantities. It is named after the three German physicists Harry Lehmann, Kurt Symanzik and Wolfhart Zimmermann.

In mathematics, the Hankel transform expresses any given function f(r) as the weighted sum of an infinite number of Bessel functions of the first kind Jν(kr). The Bessel functions in the sum are all of the same order ν, but differ in a scaling factor k along the r axis. The necessary coefficient Fν of each Bessel function in the sum, as a function of the scaling factor k constitutes the transformed function. The Hankel transform is an integral transform and was first developed by the mathematician Hermann Hankel. It is also known as the Fourier–Bessel transform. Just as the Fourier transform for an infinite interval is related to the Fourier series over a finite interval, so the Hankel transform over an infinite interval is related to the Fourier–Bessel series over a finite interval.

Noncentral <i>t</i>-distribution Probability distribution

The noncentral t-distribution generalizes Student's t-distribution using a noncentrality parameter. Whereas the central probability distribution describes how a test statistic t is distributed when the difference tested is null, the noncentral distribution describes how t is distributed when the null is false. This leads to its use in statistics, especially calculating statistical power. The noncentral t-distribution is also known as the singly noncentral t-distribution, and in addition to its primary use in statistical inference, is also used in robust modeling for data.

In mathematics and economics, transportation theory or transport theory is a name given to the study of optimal transportation and allocation of resources. The problem was formalized by the French mathematician Gaspard Monge in 1781.

In mathematical analysis, and especially in real, harmonic analysis and functional analysis, an Orlicz space is a type of function space which generalizes the Lp spaces. Like the Lp spaces, they are Banach spaces. The spaces are named for Władysław Orlicz, who was the first to define them in 1932.

Bending of plates

Bending of plates, or plate bending, refers to the deflection of a plate perpendicular to the plane of the plate under the action of external forces and moments. The amount of deflection can be determined by solving the differential equations of an appropriate plate theory. The stresses in the plate can be calculated from these deflections. Once the stresses are known, failure theories can be used to determine whether a plate will fail under a given load.

A product distribution is a probability distribution constructed as the distribution of the product of random variables having two other known distributions. Given two statistically independent random variables X and Y, the distribution of the random variable Z that is formed as the product

In statistics, the generalized Marcum Q-function of order is defined as

Coherent states have been introduced in a physical context, first as quasi-classical states in quantum mechanics, then as the backbone of quantum optics and they are described in that spirit in the article Coherent states. However, they have generated a huge variety of generalizations, which have led to a tremendous amount of literature in mathematical physics. In this article, we sketch the main directions of research on this line. For further details, we refer to several existing surveys.

Ramanujans master theorem

In mathematics, Ramanujan's master theorem is a technique that provides an analytic expression for the Mellin transform of an analytic function.

Lagrangian field theory is a formalism in classical field theory. It is the field-theoretic analogue of Lagrangian mechanics. Lagrangian mechanics is used to analyze the motion of a system of discrete particles each with a finite number of degrees of freedom. Lagrangian field theory applies to continua and fields, which have an infinite number of degrees of freedom.

In representation theory of mathematics, the Waldspurger formula relates the special values of two L-functions of two related admissible irreducible representations. Let k be the base field, f be an automorphic form over k, π be the representation associated via the Jacquet–Langlands correspondence with f. Goro Shimura (1976) proved this formula, when and f is a cusp form; Günter Harder made the same discovery at the same time in an unpublished paper. Marie-France Vignéras (1980) proved this formula, when and f is a newform. Jean-Loup Waldspurger, for whom the formula is named, reproved and generalized the result of Vignéras in 1985 via a totally different method which was widely used thereafter by mathematicians to prove similar formulas.

Stable count distribution Probability distribution

In probability theory, the stable count distribution is the conjugate prior of a one-sided stable distribution. This distribution was discovered by Stephen Lihn in his 2017 study of daily distributions of the S&P 500 and the VIX. The stable distribution family is also sometimes referred to as the Lévy alpha-stable distribution, after Paul Lévy, the first mathematician to have studied it.

References

  1. Lund, R.E.; Lund, J.R. (1983). "Algorithm AS 190: Probabilities and upper quantiles for the studentized range". Journal of the Royal Statistical Society. 32 (2): 204–210. JSTOR   2347300.
  2. 1 2 McKay, A.T. (1933). "A note on the distribution of range in samples of n". Biometrika . 25 (3): 415–420. doi:10.2307/2332292. JSTOR   2332292.
  3. "StatsExamples | table of Q distribution critical values for alpha=0.05".
  4. Pearson & Hartley (1970, Section 14.2)

Further reading