Subadditivity

Last updated

In mathematics, subadditivity is a property of a function that states, roughly, that evaluating the function for the sum of two elements of the domain always returns something less than or equal to the sum of the function's values at each element. There are numerous examples of subadditive functions in various areas of mathematics, particularly norms and square roots. Additive maps are special cases of subadditive functions.

Contents

Definitions

A subadditive function is a function , having a domain A and an ordered codomain B that are both closed under addition, with the following property:

An example is the square root function, having the non-negative real numbers as domain and codomain, since we have:

A sequence , is called subadditive if it satisfies the inequality

for all m and n. This is a special case of subadditive function, if a sequence is interpreted as a function on the set of natural numbers.

Note that while a concave sequence is subadditive, the converse is false. For example, randomly assign with values in , then the sequence is subadditive but not concave.

Properties

Sequences

A useful result pertaining to subadditive sequences is the following lemma due to Michael Fekete. [1]

Fekete's Subadditive Lemma  For every subadditive sequence , the limit exists and is equal to the infimum . (The limit may be .)

Proof

Let .

By definition, . So it suffices to show .

If not, then there exists a subsequence , and an , such that for all .

Since , there exists an such that .

By infinitary pigeonhole principle, there exists a sub-subsequence , whose indices all belong to the same residue class modulo , and so they advance by multiples of . This sequence, continued for long enough, would be forced by subadditivity to dip below the slope line, a contradiction.

In more detail, by subadditivity, we have

which implies

The analogue of Fekete's lemma holds for superadditive sequences as well, that is: (The limit then may be positive infinity: consider the sequence .)

There are extensions of Fekete's lemma that do not require the inequality to hold for all m and n, but only for m and n such that

Proof

Continue the proof as before, until we have just used the infinite pigeonhole principle.

Consider the sequence . Since , we have . Similarly, we have , etc.

By the assumption, for any , we can use subadditivity on them if

If we were dealing with continuous variables, then we can use subadditivity to go from to , then to , and so on, which covers the entire interval .

Though we don't have continuous variables, we can still cover enough integers to complete the proof. Let be large enough, such that

then let be the smallest number in the intersection . By the assumption on , it's easy to see (draw a picture) that the intervals and touch in the middle. Thus, by repeating this process, we cover the entirety of .

With that, all are forced down as in the previous proof.

Moreover, the condition may be weakened as follows: provided that is an increasing function such that the integral converges (near the infinity). [2]

There are also results that allow one to deduce the rate of convergence to the limit whose existence is stated in Fekete's lemma if some kind of both superadditivity and subadditivity is present. [3] [4]

Besides, analogues of Fekete's lemma have been proved for subadditive real maps (with additional assumptions) from finite subsets of an amenable group [5] [6] , [7] and further, of a cancellative left-amenable semigroup. [8]

Functions

Theorem: [9]   For every measurable subadditive function the limit exists and is equal to (The limit may be )

If f is a subadditive function, and if 0 is in its domain, then f(0) ≥ 0. To see this, take the inequality at the top. . Hence

A concave function with is also subadditive. To see this, one first observes that . Then looking at the sum of this bound for and , will finally verify that f is subadditive. [10]

The negative of a subadditive function is superadditive.


Examples in various domains

Entropy

Entropy plays a fundamental role in information theory and statistical physics, as well as in quantum mechanics in a generalized formulation due to von Neumann. Entropy appears always as a subadditive quantity in all of its formulations, meaning the entropy of a supersystem or a set union of random variables is always less or equal than the sum of the entropies of its individual components. Additionally, entropy in physics satisfies several more strict inequalities such as the Strong Subadditivity of Entropy in classical statistical mechanics and its quantum analog.

Economics

Subadditivity is an essential property of some particular cost functions. It is, generally, a necessary and sufficient condition for the verification of a natural monopoly. It implies that production from only one firm is socially less expensive (in terms of average costs) than production of a fraction of the original quantity by an equal number of firms.

Economies of scale are represented by subadditive average cost functions.

Except in the case of complementary goods, the price of goods (as a function of quantity) must be subadditive. Otherwise, if the sum of the cost of two items is cheaper than the cost of the bundle of two of them together, then nobody would ever buy the bundle, effectively causing the price of the bundle to "become" the sum of the prices of the two separate items. Thus proving that it is not a sufficient condition for a natural monopoly; since the unit of exchange may not be the actual cost of an item. This situation is familiar to everyone in the political arena where some minority asserts that the loss of some particular freedom at some particular level of government means that many governments are better; whereas the majority assert that there is some other correct unit of cost.[ citation needed ]

Finance

Subadditivity is one of the desirable properties of coherent risk measures in risk management. [11] The economic intuition behind risk measure subadditivity is that a portfolio risk exposure should, at worst, simply equal the sum of the risk exposures of the individual positions that compose the portfolio. In any other case the effects of diversification would result in a portfolio exposure that is lower than the sum of the individual risk exposures. The lack of subadditivity is one of the main critiques of VaR models which do not rely on the assumption of normality of risk factors. The Gaussian VaR ensures subadditivity: for example, the Gaussian VaR of a two unitary long positions portfolio at the confidence level is, assuming that the mean portfolio value variation is zero and the VaR is defined as a negative loss,

where is the inverse of the normal cumulative distribution function at probability level , are the individual positions returns variances and is the linear correlation measure between the two individual positions returns. Since variance is always positive,

Thus the Gaussian VaR is subadditive for any value of and, in particular, it equals the sum of the individual risk exposures when which is the case of no diversification effects on portfolio risk.

Thermodynamics

Subadditivity occurs in the thermodynamic properties of non-ideal solutions and mixtures like the excess molar volume and heat of mixing or excess enthalpy.

Combinatorics on words

A factorial language is one where if a word is in , then all factors of that word are also in . In combinatorics on words, a common problem is to determine the number of length- words in a factorial language. Clearly , so is subadditive, and hence Fekete's lemma can be used to estimate the growth of . [12]

For every , sample two strings of length uniformly at random on the alphabet . The expected length of the longest common subsequence is a super-additive function of , and thus there exists a number , such that the expected length grows as . By checking the case with , we easily have . The exact value of even , however, is only known to be between 0.788 and 0.827. [13]

See also

Notes

  1. Fekete, M. (1923). "Über die Verteilung der Wurzeln bei gewissen algebraischen Gleichungen mit ganzzahligen Koeffizienten". Mathematische Zeitschrift. 17 (1): 228–249. doi:10.1007/BF01504345. S2CID   186223729.
  2. de Bruijn, N.G.; Erdös, P. (1952). "Some linear and some quadratic recursion formulas. II". Nederl. Akad. Wetensch. Proc. Ser. A. 55: 152–163. doi:10.1016/S1385-7258(52)50021-0. (The same as Indagationes Math.14.) See also Steele 1997, Theorem 1.9.2.
  3. Michael J. Steele. "Probability theory and combinatorial optimization". SIAM, Philadelphia (1997). ISBN   0-89871-380-3.
  4. Michael J. Steele (2011). CBMS Lectures on Probability Theory and Combinatorial Optimization. University of Cambridge.
  5. Lindenstrauss, Elon; Weiss, Benjamin (2000). "Mean topological dimension". Israel Journal of Mathematics . 115 (1): 1–24. CiteSeerX   10.1.1.30.3552 . doi: 10.1007/BF02810577 . ISSN   0021-2172. Theorem 6.1
  6. Ornstein, Donald S.; Weiss, Benjamin (1987). "Entropy and isomorphism theorems for actions of amenable groups". Journal d'Analyse Mathématique . 48 (1): 1–141. doi: 10.1007/BF02790325 . ISSN   0021-7670.
  7. Gromov, Misha (1999). "Topological Invariants of Dynamical Systems and Spaces of Holomorphic Maps: I". Mathematical Physics, Analysis and Geometry. 2 (4): 323–415. doi:10.1023/A:1009841100168. ISSN   1385-0172. S2CID   117100302.
  8. Ceccherini-Silberstein, Tullio; Krieger, Fabrice; Coornaert, Michel (2014). "An analogue of Fekete's lemma for subadditive functions on cancellative amenable semigroups". Journal d'Analyse Mathématique . 124: 59–81. arXiv: 1209.6179 . doi: 10.1007/s11854-014-0027-4 . Theorem 1.1
  9. Hille 1948, Theorem 6.6.1. (Measurability is stipulated in Sect. 6.2 "Preliminaries".)
  10. Schechter, Eric (1997). Handbook of Analysis and its Foundations. San Diego: Academic Press. ISBN   978-0-12-622760-4., p.314,12.25
  11. Rau-Bredow, H. (2019). "Bigger Is Not Always Safer: A Critical Analysis of the Subadditivity Assumption for Coherent Risk Measures". Risks. 7 (3): 91. doi: 10.3390/risks7030091 . hdl: 10419/257929 .
  12. Shur, Arseny (2012). "Growth properties of power-free languages". Computer Science Review. 6 (5–6): 187–208. doi:10.1016/j.cosrev.2012.09.001.
  13. Lueker, George S. (May 2009). "Improved bounds on the average length of longest common subsequences". Journal of the ACM. 56 (3): 1–38. doi:10.1145/1516512.1516519. ISSN   0004-5411. S2CID   7232681.

Related Research Articles

In complex analysis, an entire function, also called an integral function, is a complex-valued function that is holomorphic on the whole complex plane. Typical examples of entire functions are polynomials and the exponential function, and any finite sums, products and compositions of these, such as the trigonometric functions sine and cosine and their hyperbolic counterparts sinh and cosh, as well as derivatives and integrals of entire functions such as the error function. If an entire function has a root at , then , taking the limit value at , is an entire function. On the other hand, the natural logarithm, the reciprocal function, and the square root are all not entire functions, nor can they be continued analytically to an entire function.

<span class="mw-page-title-main">Entropy (information theory)</span> Expected amount of information needed to specify the output of a stochastic data source

In information theory, the entropy of a random variable is the average level of "information", "surprise", or "uncertainty" inherent to the variable's possible outcomes. Given a discrete random variable , which takes values in the alphabet and is distributed according to :

<span class="mw-page-title-main">Uncertainty principle</span> Foundational principle in quantum physics

The uncertainty principle, also known as Heisenberg's indeterminacy principle, is a fundamental concept in quantum mechanics. It states that there is a limit to the precision with which certain pairs of physical properties, such as position and momentum, can be simultaneously known. In other words, the more accurately one property is measured, the less accurately the other property can be known.

In mathematics, an infinite series of numbers is said to converge absolutely if the sum of the absolute values of the summands is finite. More precisely, a real or complex series is said to converge absolutely if for some real number Similarly, an improper integral of a function, is said to converge absolutely if the integral of the absolute value of the integrand is finite—that is, if

In the mathematical field of real analysis, the monotone convergence theorem is any of a number of related theorems proving the convergence of monotonic sequences that are also bounded. Informally, the theorems state that if a sequence is increasing and bounded above by a supremum, then the sequence will converge to the supremum; in the same way, if a sequence is decreasing and is bounded below by an infimum, it will converge to the infimum.

In mathematics, Fatou's lemma establishes an inequality relating the Lebesgue integral of the limit inferior of a sequence of functions to the limit inferior of integrals of these functions. The lemma is named after Pierre Fatou.

In mathematics, the limit of a sequence of sets is a set whose elements are determined by the sequence in either of two equivalent ways: (1) by upper and lower bounds on the sequence that converge monotonically to the same set and (2) by convergence of a sequence of indicator functions which are themselves real-valued. As is the case with sequences of other objects, convergence is not necessary or even usual.

In mathematics, a Dirichlet series is any series of the form

In probability theory and statistics, the generalized extreme value (GEV) distribution is a family of continuous probability distributions developed within extreme value theory to combine the Gumbel, Fréchet and Weibull families also known as type I, II and III extreme value distributions. By the extreme value theorem the GEV distribution is the only possible limit distribution of properly normalized maxima of a sequence of independent and identically distributed random variables. Note that a limit distribution needs to exist, which requires regularity conditions on the tail of the distribution. Despite this, the GEV distribution is often used as an approximation to model the maxima of long (finite) sequences of random variables.

Differential entropy is a concept in information theory that began as an attempt by Claude Shannon to extend the idea of (Shannon) entropy, a measure of average (surprisal) of a random variable, to continuous probability distributions. Unfortunately, Shannon did not derive this formula, and rather just assumed it was the correct continuous analogue of discrete entropy, but it is not. The actual continuous version of discrete entropy is the limiting density of discrete points (LDDP). Differential entropy is commonly encountered in the literature, but it is a limiting case of the LDDP, and one that loses its fundamental association with discrete entropy.

In probability theory, the theory of large deviations concerns the asymptotic behaviour of remote tails of sequences of probability distributions. While some basic ideas of the theory can be traced to Laplace, the formalization started with insurance mathematics, namely ruin theory with Cramér and Lundberg. A unified formalization of large deviation theory was developed in 1966, in a paper by Varadhan. Large deviations theory formalizes the heuristic ideas of concentration of measures and widely generalizes the notion of convergence of probability measures.

In mathematics, the Stirling polynomials are a family of polynomials that generalize important sequences of numbers appearing in combinatorics and analysis, which are closely related to the Stirling numbers, the Bernoulli numbers, and the generalized Bernoulli polynomials. There are multiple variants of the Stirling polynomial sequence considered below most notably including the Sheffer sequence form of the sequence, , defined characteristically through the special form of its exponential generating function, and the Stirling (convolution) polynomials, , which also satisfy a characteristic ordinary generating function and that are of use in generalizing the Stirling numbers to arbitrary complex-valued inputs. We consider the "convolution polynomial" variant of this sequence and its properties second in the last subsection of the article. Still other variants of the Stirling polynomials are studied in the supplementary links to the articles given in the references.

In mathematics, Fejér's theorem, named after Hungarian mathematician Lipót Fejér, states the following:

In mathematics, the theory of optimal stopping or early stopping is concerned with the problem of choosing a time to take a particular action, in order to maximise an expected reward or minimise an expected cost. Optimal stopping problems can be found in areas of statistics, economics, and mathematical finance. A key example of an optimal stopping problem is the secretary problem. Optimal stopping problems can often be written in the form of a Bellman equation, and are therefore often solved using dynamic programming.

Expected shortfall (ES) is a risk measure—a concept used in the field of financial risk measurement to evaluate the market risk or credit risk of a portfolio. The "expected shortfall at q% level" is the expected return on the portfolio in the worst of cases. ES is an alternative to value at risk that is more sensitive to the shape of the tail of the loss distribution.

<span class="mw-page-title-main">Generalized Pareto distribution</span> Family of probability distributions often used to model tails or extreme values

In statistics, the generalized Pareto distribution (GPD) is a family of continuous probability distributions. It is often used to model the tails of another distribution. It is specified by three parameters: location , scale , and shape . Sometimes it is specified by only scale and shape and sometimes only by its shape parameter. Some references give the shape parameter as .

In the field of mathematical analysis, a general Dirichlet series is an infinite series that takes the form of

Uncertainty theory is a branch of mathematics based on normality, monotonicity, self-duality, countable subadditivity, and product measure axioms.

Generalized relative entropy is a measure of dissimilarity between two quantum states. It is a "one-shot" analogue of quantum relative entropy and shares many properties of the latter quantity.

In mathematics, Kingman's subadditive ergodic theorem is one of several ergodic theorems. It can be seen as a generalization of Birkhoff's ergodic theorem. Intuitively, the subadditive ergodic theorem is a kind of random variable version of Fekete's lemma. As a result, it can be rephrased in the language of probability, e.g. using a sequence of random variables and expected values. The theorem is named after John Kingman.

References

This article incorporates material from subadditivity on PlanetMath, which is licensed under the Creative Commons Attribution/Share-Alike License.