McDiarmid's inequality

Last updated

In probability theory and theoretical computer science, McDiarmid's inequality (named after Colin McDiarmid [1] ) is a concentration inequality which bounds the deviation between the sampled value and the expected value of certain functions when they are evaluated on independent random variables. McDiarmid's inequality applies to functions that satisfy a bounded differences property, meaning that replacing a single argument to the function while leaving all other arguments unchanged cannot cause too large of a change in the value of the function.

Contents

Statement

A function satisfies the bounded differences property if substituting the value of the th coordinate changes the value of by at most . More formally, if there are constants such that for all , and all ,

McDiarmid's Inequality [2]   Let satisfy the bounded differences property with bounds .

Consider independent random variables where for all . Then, for any ,

and as an immediate consequence,

Extensions

Unbalanced distributions

A stronger bound may be given when the arguments to the function are sampled from unbalanced distributions, such that resampling a single argument rarely causes a large change to the function value.

McDiarmid's Inequality (unbalanced) [3] [4]   Let satisfy the bounded differences property with bounds .

Consider independent random variables drawn from a distribution where there is a particular value which occurs with probability . Then, for any ,

This may be used to characterize, for example, the value of a function on graphs when evaluated on sparse random graphs and hypergraphs, since in a sparse random graph, it is much more likely for any particular edge to be missing than to be present.

Differences bounded with high probability

McDiarmid's inequality may be extended to the case where the function being analyzed does not strictly satisfy the bounded differences property, but large differences remain very rare.

McDiarmid's Inequality (Differences bounded with high probability) [5]   Let be a function and be a subset of its domain and let be constants such that for all pairs and ,

Consider independent random variables where for all . Let and let . Then, for any ,

and as an immediate consequence,

There exist stronger refinements to this analysis in some distribution-dependent scenarios, [6] such as those that arise in learning theory.

Sub-Gaussian and sub-exponential norms

Let the th centered conditional version of a function be

so that is a random variable depending on random values of .

McDiarmid's Inequality (Sub-Gaussian norm) [7] [8]   Let be a function. Consider independent random variables where for all .

Let refer to the th centered conditional version of . Let denote the sub-Gaussian norm of a random variable.

Then, for any ,

McDiarmid's Inequality (Sub-exponential norm) [8]   Let be a function. Consider independent random variables where for all .

Let refer to the th centered conditional version of . Let denote the sub-exponential norm of a random variable.

Then, for any ,

Bennett and Bernstein forms

Refinements to McDiarmid's inequality in the style of Bennett's inequality and Bernstein inequalities are made possible by defining a variance term for each function argument. Let

McDiarmid's Inequality (Bennett form) [4]   Let satisfy the bounded differences property with bounds . Consider independent random variables where for all . Let and be defined as at the beginning of this section.

Then, for any ,

McDiarmid's Inequality (Bernstein form) [4]   Let satisfy the bounded differences property with bounds . Let and be defined as at the beginning of this section.

Then, for any ,

Proof

The following proof of McDiarmid's inequality [2] constructs the Doob martingale tracking the conditional expected value of the function as more and more of its arguments are sampled and conditioned on, and then applies a martingale concentration inequality (Azuma's inequality). An alternate argument avoiding the use of martingales also exists, taking advantage of the independence of the function arguments to provide a Chernoff-bound-like argument. [4]

For better readability, we will introduce a notational shorthand: will denote for any and integers , so that, for example,

Pick any . Then, for any , by triangle inequality,

and thus is bounded.

Since is bounded, define the Doob martingale (each being a random variable depending on the random values of ) as

for all and , so that .

Now define the random variables for each

Since are independent of each other, conditioning on does not affect the probabilities of the other variables, so these are equal to the expressions

Note that . In addition,

Then, applying the general form of Azuma's inequality to , we have

The one-sided bound in the other direction is obtained by applying Azuma's inequality to and the two-sided bound follows from a union bound.

See also

Related Research Articles

In mathematics, the Lp spaces are function spaces defined using a natural generalization of the p-norm for finite-dimensional vector spaces. They are sometimes called Lebesgue spaces, named after Henri Lebesgue, although according to the Bourbaki group they were first introduced by Frigyes Riesz.

In mathematics, an infinite series of numbers is said to converge absolutely if the sum of the absolute values of the summands is finite. More precisely, a real or complex series is said to converge absolutely if for some real number Similarly, an improper integral of a function, is said to converge absolutely if the integral of the absolute value of the integrand is finite—that is, if

In mathematical analysis, Hölder's inequality, named after Otto Hölder, is a fundamental inequality between integrals and an indispensable tool for the study of Lp spaces.

In mathematics, the uniform boundedness principle or Banach–Steinhaus theorem is one of the fundamental results in functional analysis. Together with the Hahn–Banach theorem and the open mapping theorem, it is considered one of the cornerstones of the field. In its basic form, it asserts that for a family of continuous linear operators whose domain is a Banach space, pointwise boundedness is equivalent to uniform boundedness in operator norm.

In mathematics, Fatou's lemma establishes an inequality relating the Lebesgue integral of the limit inferior of a sequence of functions to the limit inferior of integrals of these functions. The lemma is named after Pierre Fatou.

Vapnik–Chervonenkis theory was developed during 1960–1990 by Vladimir Vapnik and Alexey Chervonenkis. The theory is a form of computational learning theory, which attempts to explain the learning process from a statistical point of view.

In probability theory, the Azuma–Hoeffding inequality gives a concentration result for the values of martingales that have bounded differences.

In probability theory, a Chernoff bound is an exponentially decreasing upper bound on the tail of a random variable based on its moment generating function. The minimum of all such exponential bounds forms the Chernoff or Chernoff-Cramér bound, which may decay faster than exponential. It is especially useful for sums of independent random variables, such as sums of Bernoulli random variables.

In probability theory, Hoeffding's inequality provides an upper bound on the probability that the sum of bounded independent random variables deviates from its expected value by more than a certain amount. Hoeffding's inequality was proven by Wassily Hoeffding in 1963.

<span class="mw-page-title-main">Szemerédi regularity lemma</span>

In extremal graph theory, Szemerédi’s regularity lemma states that a graph can be partitioned into a bounded number of parts so that the edges between parts are regular. The lemma shows that properties of random graphs can be applied to general graphs like counting the copies of a given subgraph within graphs. Endre Szemerédi proved the lemma over bipartite graphs for his theorem on arithmetic progressions in 1975 and for general graphs in 1978. Variants of the lemma use different notions of regularity and apply to other mathematical objects like hypergraphs.

In the theory of probability, the Glivenko–Cantelli theorem, named after Valery Ivanovich Glivenko and Francesco Paolo Cantelli, describes the asymptotic behaviour of the empirical distribution function as the number of independent and identically distributed observations grows. Specifically, the empirical distribution function converges uniformly to the true distribution function almost surely.

In the mathematical theory of probability, a Doob martingale is a stochastic process that approximates a given random variable and has the martingale property with respect to the given filtration. It may be thought of as the evolving sequence of best approximations to the random variable based on information accumulated up to a certain time.

In mathematics, the theory of optimal stopping or early stopping is concerned with the problem of choosing a time to take a particular action, in order to maximise an expected reward or minimise an expected cost. Optimal stopping problems can be found in areas of statistics, economics, and mathematical finance. A key example of an optimal stopping problem is the secretary problem. Optimal stopping problems can often be written in the form of a Bellman equation, and are therefore often solved using dynamic programming.

<span class="mw-page-title-main">Dvoretzky–Kiefer–Wolfowitz inequality</span> Statistical inequality

In the theory of probability and statistics, the Dvoretzky–Kiefer–Wolfowitz–Massart inequality provides a bound on the worst case distance of an empirically determined distribution function from its associated population distribution function. It is named after Aryeh Dvoretzky, Jack Kiefer, and Jacob Wolfowitz, who in 1956 proved the inequality

In mathematics, Doob's martingale inequality, also known as Kolmogorov’s submartingale inequality is a result in the study of stochastic processes. It gives a bound on the probability that a submartingale exceeds any given value over a given interval of time. As the name suggests, the result is usually given in the case that the process is a martingale, but the result is also valid for submartingales.

In probability theory, Bernstein inequalities give bounds on the probability that the sum of random variables deviates from its mean. In the simplest case, let X1, ..., Xn be independent Bernoulli random variables taking values +1 and −1 with probability 1/2, then for every positive ,

In mathematics, the Grothendieck inequality states that there is a universal constant with the following property. If Mij is an n × n matrix with

<span class="mw-page-title-main">Anatoly Karatsuba</span> Russian mathematician (1937–2008)

Anatoly Alexeyevich Karatsuba was a Russian mathematician working in the field of analytic number theory, p-adic numbers and Dirichlet series.

In probability theory, concentration inequalities provide mathematical bounds on the probability of a random variable deviating from some value.

In additive number theory, an area of mathematics, the Erdős–Tetali theorem is an existence theorem concerning economical additive bases of every order. More specifically, it states that for every fixed integer , there exists a subset of the natural numbers satisfying

References

  1. McDiarmid, Colin (1989). "On the method of bounded differences". Surveys in Combinatorics, 1989: Invited Papers at the Twelfth British Combinatorial Conference: 148–188.
  2. 1 2 Doob, J. L. (1940). "Regularity properties of certain families of chance variables" (PDF). Transactions of the American Mathematical Society. 47 (3): 455–486. doi: 10.2307/1989964 . JSTOR   1989964.
  3. Chou, Chi-Ning; Love, Peter J.; Sandhu, Juspreet Singh; Shi, Jonathan (2022). "Limitations of Local Quantum Algorithms on Random Max-k-XOR and Beyond". 49th International Colloquium on Automata, Languages, and Programming (ICALP 2022). 229: 41:13. arXiv: 2108.06049 . doi:10.4230/LIPIcs.ICALP.2022.41 . Retrieved 8 July 2022.
  4. 1 2 3 4 Ying, Yiming (2004). "McDiarmid's inequalities of Bernstein and Bennett forms" (PDF). City University of Hong Kong. Retrieved 10 July 2022.
  5. Combes, Richard (2015). "An extension of McDiarmid's inequality". arXiv: 1511.05240 [cs.LG].
  6. Wu, Xinxing; Zhang, Junping (April 2018). "Distribution-dependent concentration inequalities for tighter generalization bounds". Science China Information Sciences. 61 (4): 048105:1–048105:3. arXiv: 1607.05506 . doi:10.1007/s11432-017-9225-2. S2CID   255199895 . Retrieved 10 July 2022.
  7. Kontorovich, Aryeh (22 June 2014). "Concentration in unbounded metric spaces and algorithmic stability". Proceedings of the 31st International Conference on Machine Learning. 32 (2): 28–36. arXiv: 1309.1007 . Retrieved 10 July 2022.
  8. 1 2 Maurer, Andreas; Pontil, Pontil (2021). "Concentration inequalities under sub-Gaussian and sub-exponential conditions" (PDF). Advances in Neural Information Processing Systems. 34: 7588–7597. Retrieved 10 July 2022.