Rademacher distribution

Last updated
Rademacher
Support
PMF
CDF
Mean
Median
Mode N/A
Variance
Skewness
Ex. kurtosis
Entropy
MGF
CF

In probability theory and statistics, the Rademacher distribution (which is named after Hans Rademacher) is a discrete probability distribution where a random variate X has a 50% chance of being +1 and a 50% chance of being -1. [1]

Contents

A series (that is, a sum) of Rademacher distributed variables can be regarded as a simple symmetrical random walk where the step size is 1.

Mathematical formulation

The probability mass function of this distribution is

In terms of the Dirac delta function, as

Bounds on sums of independent Rademacher variables

There are various results in probability theory around analyzing the sum of i.i.d. Rademacher variables, including concentration inequalities such as Bernstein inequalities as well as anti-concentration inequalities like Tomaszewski's conjecture.

Concentration inequalities

Let {xi} be a set of random variables with a Rademacher distribution. Let {ai} be a sequence of real numbers. Then

where ||a||2 is the Euclidean norm of the sequence {ai}, t > 0 is a real number and Pr(Z) is the probability of event Z. [2]

Let Y = Σ xiai and let Y be an almost surely convergent series in a Banach space. The for t > 0 and s ≥ 1 we have [3]

for some constant c.

Let p be a positive real number. Then the Khintchine inequality says that [4]

where c1 and c2 are constants dependent only on p.

For p ≥ 1,

Tomaszewski’s conjecture

In 1986, Bogusław Tomaszewski proposed a question about the distribution of the sum of independent Rademacher variables. A series of works on this question [5] [6] culminated into a proof in 2020 by Nathan Keller and Ohad Klein of the following conjecture. [7]

Conjecture. Let , where and the 's are independent Rademacher variables. Then

For example, when , one gets the following bound, first shown by Van Zuijlen. [8]

The bound is sharp and better than that which can be derived from the normal distribution (approximately Pr > 0.31).

Applications

The Rademacher distribution has been used in bootstrapping.

The Rademacher distribution can be used to show that normally distributed and uncorrelated does not imply independent.

Random vectors with components sampled independently from the Rademacher distribution are useful for various stochastic approximations, for example:

Rademacher random variables are used in the Symmetrization Inequality.

Related Research Articles

<span class="mw-page-title-main">Cumulative distribution function</span> Probability that random variable X is less than or equal to x

In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable , or just distribution function of , evaluated at , is the probability that will take a value less than or equal to .

<span class="mw-page-title-main">Expected value</span> Average value of a random variable

In probability theory, the expected value is a generalization of the weighted average. Informally, the expected value is the arithmetic mean of a large number of independently selected outcomes of a random variable.

<span class="mw-page-title-main">Variance</span> Statistical measure of how far values spread from their average

In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of numbers is spread out from their average value. Variance has a central role in statistics, where some ideas that use it include descriptive statistics, statistical inference, hypothesis testing, goodness of fit, and Monte Carlo sampling. Variance is an important tool in the sciences, where statistical analysis of data is common. The variance is the square of the standard deviation, the second central moment of a distribution, and the covariance of the random variable with itself, and it is often represented by , , , , or .

In probability theory, the central limit theorem (CLT) establishes that, in many situations, for identically distributed independent samples, the standardized sample mean tends towards the standard normal distribution even if the original variables themselves are not normally distributed.

<span class="mw-page-title-main">Probability density function</span> Function whose integral over a region describes the probability of an event occurring in that region

In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample in the sample space can be interpreted as providing a relative likelihood that the value of the random variable would be equal to that sample. Probability density is the probability per unit length, in other words, while the absolute likelihood for a continuous random variable to take on any particular value is 0, the value of the PDF at two different samples can be used to infer, in any particular draw of the random variable, how much more likely it is that the random variable would be close to one sample compared to the other sample.

In probability theory, Chebyshev's inequality guarantees that, for a wide class of probability distributions, no more than a certain fraction of values can be more than a certain distance from the mean. Specifically, no more than 1/k2 of the distribution's values can be k or more standard deviations away from the mean. The rule is often called Chebyshev's theorem, about the range of standard deviations around the mean, in statistics. The inequality has great utility because it can be applied to any probability distribution in which the mean and variance are defined. For example, it can be used to prove the weak law of large numbers.

In probability theory and statistics, the moment-generating function of a real-valued random variable is an alternative specification of its probability distribution. Thus, it provides the basis of an alternative route to analytical results compared with working directly with probability density functions or cumulative distribution functions. There are particularly simple results for the moment-generating functions of distributions defined by the weighted sums of random variables. However, not all random variables have moment-generating functions.

In probability theory, the central limit theorem states that, under certain circumstances, the probability distribution of the scaled mean of a random sample converges to a normal distribution as the sample size increases to infinity. Under stronger assumptions, the Berry–Esseen theorem, or Berry–Esseen inequality, gives a more quantitative result, because it also specifies the rate at which this convergence takes place by giving a bound on the maximal error of approximation between the normal distribution and the true distribution of the scaled sample mean. The approximation is measured by the Kolmogorov–Smirnov distance. In the case of independent samples, the convergence rate is n−1/2, where n is the sample size, and the constant is estimated in terms of the third absolute normalized moment.

In probability theory, a Chernoff bound is an exponentially decreasing upper bound on the tail of a random variable based on its moment generating function or exponential moments. The minimum of all such exponential bounds forms the Chernoff or Chernoff-Cramér bound, which may decay faster than exponential. The Chernoff bound is especially useful for sums of independent random variables, such as sums of Bernoulli random variables.

In probability theory, Hoeffding's inequality provides an upper bound on the probability that the sum of bounded independent random variables deviates from its expected value by more than a certain amount. Hoeffding's inequality was proven by Wassily Hoeffding in 1963.

In probability theory, although simple examples illustrate that linear uncorrelatedness of two random variables does not in general imply their independence, it is sometimes mistakenly thought that it does imply that when the two random variables are normally distributed. This article demonstrates that assumption of normal distributions does not have that consequence, although the multivariate normal distribution, including the bivariate normal distribution, does.

In probability theory, Kolmogorov's inequality is a so-called "maximal inequality" that gives a bound on the probability that the partial sums of a finite collection of independent random variables exceed some specified bound.

<span class="mw-page-title-main">Dvoretzky–Kiefer–Wolfowitz inequality</span> Statistical inequality

In the theory of probability and statistics, the Dvoretzky–Kiefer–Wolfowitz–Massart inequality bounds how close an empirically determined distribution function will be to the distribution function from which the empirical samples are drawn. It is named after Aryeh Dvoretzky, Jack Kiefer, and Jacob Wolfowitz, who in 1956 proved the inequality

In probability theory, Bernstein inequalities give bounds on the probability that the sum of random variables deviates from its mean. In the simplest case, let X1, ..., Xn be independent Bernoulli random variables taking values +1 and −1 with probability 1/2, then for every positive ,

In mathematics, the Khintchine inequality, named after Aleksandr Khinchin and spelled in multiple ways in the Latin alphabet, is a theorem from probability, and is also frequently used in analysis. Heuristically, it says that if we pick complex numbers , and add them together each multiplied by a random sign , then the expected value of the sum's modulus, or the modulus it will be closest to on average, will be not too far off from .

In probability theory, the multidimensional Chebyshev's inequality is a generalization of Chebyshev's inequality, which puts a bound on the probability of the event that a random variable differs from its expected value by more than a specified amount.

In probability theory and statistics, the Poisson binomial distribution is the discrete probability distribution of a sum of independent Bernoulli trials that are not necessarily identically distributed. The concept is named after Siméon Denis Poisson.

In probability theory, concentration inequalities provide bounds on how a random variable deviates from some value. The law of large numbers of classical probability theory states that sums of independent random variables are, under very mild conditions, close to their expectation with a large probability. Such sums are the most basic examples of random variables concentrated around their mean. Recent results show that such behavior is shared by other functions of independent random variables.

For certain applications in linear algebra, it is useful to know properties of the probability distribution of the largest eigenvalue of a finite sum of random matrices. Suppose is a finite sequence of random matrices. Analogous to the well-known Chernoff bound for sums of scalars, a bound on the following is sought for a given parameter t:

In probability theory, Eaton's inequality is a bound on the largest values of a linear combination of bounded random variables. This inequality was described in 1974 by Morris L. Eaton.

References

  1. Hitczenko, P.; Kwapień, S. (1994). "On the Rademacher series". Probability in Banach Spaces. Progress in probability. Vol. 35. pp. 31–36. doi:10.1007/978-1-4612-0253-0_2. ISBN   978-1-4612-6682-2.
  2. Montgomery-Smith, S. J. (1990). "The distribution of Rademacher sums". Proc Amer Math Soc. 109 (2): 517–522. doi: 10.1090/S0002-9939-1990-1013975-0 .
  3. Dilworth, S. J.; Montgomery-Smith, S. J. (1993). "The distribution of vector-valued Radmacher series". Ann Probab. 21 (4): 2046–2052. arXiv: math/9206201 . doi:10.1214/aop/1176989010. JSTOR   2244710. S2CID   15159626.
  4. Khintchine, A. (1923). "Über dyadische Brüche". Math. Z. 18 (1): 109–116. doi:10.1007/BF01192399. S2CID   119840766.
  5. Holzman, Ron; Kleitman, Daniel J. (1992-09-01). "On the product of sign vectors and unit vectors". Combinatorica. 12 (3): 303–316. doi:10.1007/BF01285819. ISSN   1439-6912. S2CID   20281665.
  6. Boppana, Ravi B.; Holzman, Ron (2017-08-31). "Tomaszewski's Problem on Randomly Signed Sums: Breaking the 3/8 Barrier". arXiv: 1704.00350 [math.CO].
  7. Keller, Nathan; Klein, Ohad (2021-08-03). "Proof of Tomaszewski's Conjecture on Randomly Signed Sums". arXiv: 2006.16834 [math.CO].
  8. van Zuijlen, Martien C. A. (2011). "On a conjecture concerning the sum of independent Rademacher random variables". arXiv: 1112.4988 . Bibcode:2011arXiv1112.4988V.{{cite journal}}: Cite journal requires |journal= (help)
  9. Avron, H.; Toledo, S. (2011). "Randomized algorithms for estimating the trace of an implicit symmetric positive semidefinite matrix". Journal of the ACM. 58 (2): 8. CiteSeerX   10.1.1.380.9436 . doi:10.1145/1944345.1944349. S2CID   5827717.