In mathematics, Karamata's inequality, [1] named after Jovan Karamata, [2] also known as the majorization inequality, is a theorem in elementary algebra for convex and concave real-valued functions, defined on an interval of the real line. It generalizes the discrete form of Jensen's inequality, and generalizes in turn to the concept of Schur-convex functions.
Let I be an interval of the real line and let f denote a real-valued, convex function defined on I. If x1, …, xn and y1, …, yn are numbers in I such that (x1, …, xn) majorizes (y1, …, yn), then
| (1) |
Here majorization means that x1, …, xn and y1, …, yn satisfies
and |
| (2) |
and we have the inequalities
for all i ∈ {1, …, n − 1}. |
| (3) |
and the equality
| (4) |
If f is a strictly convex function, then the inequality ( 1 ) holds with equality if and only if we have xi = yi for all i ∈ {1, …, n}.
| (5) |
The finite form of Jensen's inequality is a special case of this result. Consider the real numbers x1, …, xn ∈ I and let
denote their arithmetic mean. Then (x1, …, xn) majorizes the n-tuple (a, a, …, a), since the arithmetic mean of the i largest numbers of (x1, …, xn) is at least as large as the arithmetic mean a of all the n numbers, for every i ∈ {1, …, n − 1}. By Karamata's inequality ( 1 ) for the convex function f,
Dividing by n gives Jensen's inequality. The sign is reversed if f is concave.
We may assume that the numbers are in decreasing order as specified in ( 2 ).
If xi = yi for all i ∈ {1, …, n}, then the inequality ( 1 ) holds with equality, hence we may assume in the following that xi ≠ yi for at least one i.
If xi = yi for an i ∈ {1, …, n − 1}, then the inequality ( 1 ) and the majorization properties ( 3 ) and ( 4 ) are not affected if we remove xi and yi. Hence we may assume that xi ≠ yi for all i ∈ {1, …, n − 1}.
It is a property of convex functions that for two numbers x ≠ y in the interval I the slope
of the secant line through the points (x, f (x)) and (y, f (y)) of the graph of f is a monotonically non-decreasing function in x for y fixed (and vice versa). This implies that
| (6) |
for all i ∈ {1, …, n − 1}. Define A0 = B0 = 0 and
for all i ∈ {1, …, n}. By the majorization property ( 3 ), Ai ≥ Bi for all i ∈ {1, …, n − 1} and by ( 4 ), An = Bn. Hence,
| (7) |
which proves Karamata's inequality ( 1 ).
To discuss the case of equality in ( 1 ), note that x1 > y1 by ( 3 ) and our assumption xi ≠ yi for all i ∈ {1, …, n − 1}. Let i be the smallest index such that (xi, yi) ≠ (xi+1, yi+1), which exists due to ( 4 ). Then Ai > Bi. If f is strictly convex, then there is strict inequality in ( 6 ), meaning that ci+1 < ci. Hence there is a strictly positive term in the sum on the right hand side of ( 7 ) and equality in ( 1 ) cannot hold.
If the convex function f is non-decreasing, then cn ≥ 0. The relaxed condition ( 5 ) means that An ≥ Bn, which is enough to conclude that cn(An−Bn) ≥ 0 in the last step of ( 7 ).
If the function f is strictly convex and non-decreasing, then cn > 0. It only remains to discuss the case An > Bn. However, then there is a strictly positive term on the right hand side of ( 7 ) and equality in ( 1 ) cannot hold.
In mathematics, generalized means are a family of functions for aggregating sets of numbers. These include as special cases the Pythagorean means.
In mathematics, the limit inferior and limit superior of a sequence can be thought of as limiting bounds on the sequence. They can be thought of in a similar fashion for a function. For a set, they are the infimum and supremum of the set's limit points, respectively. In general, when there are multiple objects around which a sequence, function, or set accumulates, the inferior and superior limits extract the smallest and largest of them; the type of object and the measure of size is context-dependent, but the notion of extreme limits is invariant. Limit inferior is also called infimum limit, limit infimum, liminf, inferior limit, lower limit, or inner limit; limit superior is also known as supremum limit, limit supremum, limsup, superior limit, upper limit, or outer limit.
In probability theory, there exist several different notions of convergence of random variables. The convergence of sequences of random variables to some limit random variable is an important concept in probability theory, and its applications to statistics and stochastic processes. The same concepts are known in more general mathematics as stochastic convergence and they formalize the idea that a sequence of essentially random or unpredictable events can sometimes be expected to settle down into a behavior that is essentially unchanging when items far enough into the sequence are studied. The different possible notions of convergence relate to how such a behavior can be characterized: two readily understood behaviors are that the sequence eventually takes a constant value, and that values in the sequence continue to change but can be described by an unchanging probability distribution.
In mathematics, the triangle inequality states that for any triangle, the sum of the lengths of any two sides must be greater than or equal to the length of the remaining side. This statement permits the inclusion of degenerate triangles, but some authors, especially those writing about elementary geometry, will exclude this possibility, thus leaving out the possibility of equality. If x, y, and z are the lengths of the sides of the triangle, with no side being greater than z, then the triangle inequality states that
In mathematics, an inequality is a relation which makes a non-equal comparison between two numbers or other mathematical expressions. It is used most often to compare two numbers on the number line by their size. There are several different notations used to represent different kinds of inequalities:
In mathematical analysis, the Minkowski inequality establishes that the Lp spaces are normed vector spaces. Let S be a measure space, let 1 ≤ p < ∞ and let f and g be elements of Lp(S). Then f + g is in Lp(S), and we have the triangle inequality
In mathematics, a real-valued function is called convex if the line segment between any two points on the graph of the function does not lie below the graph between the two points. Equivalently, a function is convex if its epigraph is a convex set. A twice-differentiable function of a single variable is convex if and only if its second derivative is nonnegative on its entire domain. Well-known examples of convex functions of a single variable include the quadratic function and the exponential function . In simple terms, a convex function refers to a function whose graph is shaped like a cup , while a concave function's graph is shaped like a cap .
In mathematics, a concave function is the negative of a convex function. A concave function is also synonymously called concave downwards, concave down, convex upwards, convex cap, or upper convex.
In mathematics, Jensen's inequality, named after the Danish mathematician Johan Jensen, relates the value of a convex function of an integral to the integral of the convex function. It was proved by Jensen in 1906. Given its generality, the inequality appears in many forms depending on the context, some of which are presented below. In its simplest form the inequality states that the convex transformation of a mean is less than or equal to the mean applied after convex transformation; it is a simple corollary that the opposite is true of concave transformations.
In mathematics, the inequality of arithmetic and geometric means, or more briefly the AM–GM inequality, states that the arithmetic mean of a list of non-negative real numbers is greater than or equal to the geometric mean of the same list; and further, that the two means are equal if and only if every number in the list is the same.
In mathematics, the rearrangement inequality states that
In mathematics, Muirhead's inequality, named after Robert Franklin Muirhead, also known as the "bunching" method, generalizes the inequality of arithmetic and geometric means.
In mathematics, a function f is logarithmically convex or superconvex if , the composition of the logarithm with f, is itself a convex function.
In mathematics, there are two different results that share the common name of the Ky Fan inequality. One is an inequality involving the geometric mean and arithmetic mean of two sets of real numbers of the unit interval. The result was published on page 5 of the book Inequalities by Edwin F. Beckenbach and Richard E. Bellman (1961), who refer to an unpublished result of Ky Fan. They mention the result in connection with the inequality of arithmetic and geometric means and Augustin Louis Cauchy's proof of this inequality by forward-backward-induction; a method which can also be used to prove the Ky Fan inequality.
In mathematics, majorization is a preorder on vectors of real numbers. For a vector , we denote by the vector with the same components, but sorted in descending order. Given , we say that weakly majorizesfrom below written as iff
In mathematics, the three classical Pythagorean means are the arithmetic mean (AM), the geometric mean (GM), and the harmonic mean (HM). These means were studied with proportions by Pythagoreans and later generations of Greek mathematicians because of their importance in geometry and music.
In mathematics, a Schur-convex function, also known as S-convex, isotonic function and order-preserving function is a function that for all such that is majorized by , one has that . Named after Issai Schur, Schur-convex functions are used in the study of majorization. Every function that is convex and symmetric is also Schur-convex. The opposite implication is not true, but all Schur-convex functions are symmetric.
In functional analysis, compact operators are linear operators on Banach spaces that map bounded sets to relatively compact sets. In the case of a Hilbert space H, the compact operators are the closure of the finite rank operators in the uniform operator topology. In general, operators on infinite-dimensional spaces feature properties that do not appear in the finite-dimensional case, i.e. for matrices. The compact operators are notable in that they share as much similarity with matrices as one can expect from a general operator. In particular, the spectral properties of compact operators resemble those of square matrices.
In mathematics, there are many kinds of inequalities involving matrices and linear operators on Hilbert spaces. This article covers some important operator inequalities connected with traces of matrices.
In mathematics, Young's inequality for products is a mathematical inequality about the product of two numbers. The inequality is named after William Henry Young and should not be confused with Young's convolution inequality.
An explanation of Karamata's inequality and majorization theory can be found here.