Convex function

Last updated
Convex function on an interval. ConvexFunction.svg
Convex function on an interval.

A function (in black) is convex if and only if the region above its graph (in green) is a convex set. Epigraph convex.svg
A function (in black) is convex if and only if the region above its graph (in green) is a convex set.
A graph of the bivariate convex function x + xy + y. Grafico 3d x2+xy+y2.png
A graph of the bivariate convex function x + xy + y.
Convex vs. Not convex Convex vs. Not-convex.jpg
Convex vs. Not convex

In mathematics, a real-valued function is called convex if the line segment between any two distinct points on the graph of the function lies above the graph between the two points. Equivalently, a function is convex if its epigraph (the set of points on or above the graph of the function) is a convex set. In simple terms, a convex function graph is shaped like a cup (or a straight line like a linear function), while a concave function's graph is shaped like a cap .

Contents

A twice-differentiable function of a single variable is convex if and only if its second derivative is nonnegative on its entire domain. [1] Well-known examples of convex functions of a single variable include a linear function (where is a real number), a quadratic function ( as a nonnegative real number) and an exponential function ( as a nonnegative real number).

Convex functions play an important role in many areas of mathematics. They are especially important in the study of optimization problems where they are distinguished by a number of convenient properties. For instance, a strictly convex function on an open set has no more than one minimum. Even in infinite-dimensional spaces, under suitable additional hypotheses, convex functions continue to satisfy such properties and as a result, they are the most well-understood functionals in the calculus of variations. In probability theory, a convex function applied to the expected value of a random variable is always bounded above by the expected value of the convex function of the random variable. This result, known as Jensen's inequality, can be used to deduce inequalities such as the arithmeticgeometric mean inequality and Hölder's inequality.

Definition

Visualizing a convex function and Jensen's Inequality

Let be a convex subset of a real vector space and let be a function.

Then is called convex if and only if any of the following equivalent conditions hold:

  1. For all and all :

    The right hand side represents the straight line between and in the graph of as a function of increasing from to or decreasing from to sweeps this line. Similarly, the argument of the function in the left hand side represents the straight line between and in or the -axis of the graph of So, this condition requires that the straight line between any pair of points on the curve of to be above or just meets the graph. [2]

  2. For all and all such that :
    The difference of this second condition with respect to the first condition above is that this condition does not include the intersection points (for example, and ) between the straight line passing through a pair of points on the curve of (the straight line is represented by the right hand side of this condition) and the curve of the first condition includes the intersection points as it becomes or at or or In fact, the intersection points do not need to be considered in a condition of convex using
    because and are always true (so not useful to be a part of a condition).

The second statement characterizing convex functions that are valued in the real line is also the statement used to define convex functions that are valued in the extended real number line where such a function is allowed to take as a value. The first statement is not used because it permits to take or as a value, in which case, if or respectively, then would be undefined (because the multiplications and are undefined). The sum is also undefined so a convex extended real-valued function is typically only allowed to take exactly one of and as a value.

The second statement can also be modified to get the definition of strict convexity, where the latter is obtained by replacing with the strict inequality Explicitly, the map is called strictly convex if and only if for all real and all such that :

A strictly convex function is a function that the straight line between any pair of points on the curve is above the curve except for the intersection points between the straight line and the curve. An example of a function which is convex but not strictly convex is . This function is not strictly convex because any two points sharing an x coordinate will have a straight line between them, while any two points NOT sharing an x coordinate will have a greater value of the function than the points between them.

The function is said to be concave (resp. strictly concave) if ( multiplied by −1) is convex (resp. strictly convex).

Alternative naming

The term convex is often referred to as convex down or concave upward, and the term concave is often referred as concave down or convex upward. [3] [4] [5] If the term "convex" is used without an "up" or "down" keyword, then it refers strictly to a cup shaped graph . As an example, Jensen's inequality refers to an inequality involving a convex or convex-(down), function. [6]

Properties

Many properties of convex functions have the same simple formulation for functions of many variables as for functions of one variable. See below the properties for the case of many variables, as some of them are not listed for functions of one variable.

Functions of one variable

Proof Since is convex, by using one of the convex function definitions above and letting it follows that for all real
From , it follows that

Namely, .

Functions of several variables

Operations that preserve convexity

Strongly convex functions

The concept of strong convexity extends and parametrizes the notion of strict convexity. Intuitively, a strongly-convex function is a function that grows as fast as a quadratic function. [11] A strongly convex function is also strictly convex, but not vice versa. If a one-dimensional function is twice continuously differentiable and the domain is the real line, then we can characterize it as follows:

For example, let be strictly convex, and suppose there is a sequence of points such that . Even though , the function is not strongly convex because will become arbitrarily small.

More generally, a differentiable function is called strongly convex with parameter if the following inequality holds for all points in its domain: [12]

or, more generally,

where is any inner product, and is the corresponding norm. Some authors, such as [13] refer to functions satisfying this inequality as elliptic functions.

An equivalent condition is the following: [14]

It is not necessary for a function to be differentiable in order to be strongly convex. A third definition [14] for a strongly convex function, with parameter is that, for all in the domain and

Notice that this definition approaches the definition for strict convexity as and is identical to the definition of a convex function when Despite this, functions exist that are strictly convex but are not strongly convex for any (see example below).

If the function is twice continuously differentiable, then it is strongly convex with parameter if and only if for all in the domain, where is the identity and is the Hessian matrix, and the inequality means that is positive semi-definite. This is equivalent to requiring that the minimum eigenvalue of be at least for all If the domain is just the real line, then is just the second derivative so the condition becomes . If then this means the Hessian is positive semidefinite (or if the domain is the real line, it means that ), which implies the function is convex, and perhaps strictly convex, but not strongly convex.

Assuming still that the function is twice continuously differentiable, one can show that the lower bound of implies that it is strongly convex. Using Taylor's Theorem there exists

such that

Then

by the assumption about the eigenvalues, and hence we recover the second strong convexity equation above.

A function is strongly convex with parameter m if and only if the function

is convex.

A twice continuously differentiable function on a compact domain that satisfies for all is strongly convex. The proof of this statement follows from the extreme value theorem, which states that a continuous function on a compact set has a maximum and minimum.

Strongly convex functions are in general easier to work with than convex or strictly convex functions, since they are a smaller class. Like strictly convex functions, strongly convex functions have unique minima on compact sets.

Properties of strongly-convex functions

If f is a strongly-convex function with parameter m, then: [15] :Prop.6.1.4

Uniformly convex functions

A uniformly convex function, [16] [17] with modulus , is a function that, for all in the domain and satisfies

where is a function that is non-negative and vanishes only at 0. This is a generalization of the concept of strongly convex function; by taking we recover the definition of strong convexity.

It is worth noting that some authors require the modulus to be an increasing function, [17] but this condition is not required by all authors. [16]

Examples

Functions of one variable

Functions of n variables

See also

Notes

  1. "Lecture Notes 2" (PDF). www.stat.cmu.edu. Retrieved 3 March 2017.
  2. "Concave Upward and Downward". Archived from the original on 2013-12-18.
  3. Stewart, James (2015). Calculus (8th ed.). Cengage Learning. pp. 223–224. ISBN   978-1305266643.
  4. W. Hamming, Richard (2012). Methods of Mathematics Applied to Calculus, Probability, and Statistics (illustrated ed.). Courier Corporation. p. 227. ISBN   978-0-486-13887-9. Extract of page 227
  5. Uvarov, Vasiliĭ Borisovich (1988). Mathematical Analysis. Mir Publishers. p. 126-127. ISBN   978-5-03-000500-3.
  6. Prügel-Bennett, Adam (2020). The Probability Companion for Engineering and Computer Science (illustrated ed.). Cambridge University Press. p. 160. ISBN   978-1-108-48053-6. Extract of page 160
  7. 1 2 Boyd, Stephen P.; Vandenberghe, Lieven (2004). Convex Optimization (pdf). Cambridge University Press. ISBN   978-0-521-83378-3 . Retrieved October 15, 2011.
  8. Donoghue, William F. (1969). Distributions and Fourier Transforms. Academic Press. p. 12. ISBN   9780122206504 . Retrieved August 29, 2012.
  9. "If f is strictly convex in a convex set, show it has no more than 1 minimum". Math StackExchange. 21 Mar 2013. Retrieved 14 May 2016.
  10. Altenberg, L., 2012. Resolvent positive linear operators exhibit the reduction phenomenon. Proceedings of the National Academy of Sciences, 109(10), pp.3705-3710.
  11. "Strong convexity · Xingyu Zhou's blog". xingyuzhou.org. Retrieved 2023-09-27.
  12. Dimitri Bertsekas (2003). Convex Analysis and Optimization . Contributors: Angelia Nedic and Asuman E. Ozdaglar. Athena Scientific. p.  72. ISBN   9781886529458.
  13. Philippe G. Ciarlet (1989). Introduction to numerical linear algebra and optimisation. Cambridge University Press. ISBN   9780521339841.
  14. 1 2 Yurii Nesterov (2004). Introductory Lectures on Convex Optimization: A Basic Course . Kluwer Academic Publishers. pp.  63–64. ISBN   9781402075537.
  15. Nemirovsky and Ben-Tal (2023). "Optimization III: Convex Optimization" (PDF).
  16. 1 2 C. Zalinescu (2002). Convex Analysis in General Vector Spaces. World Scientific. ISBN   9812380671.
  17. 1 2 H. Bauschke and P. L. Combettes (2011). Convex Analysis and Monotone Operator Theory in Hilbert Spaces . Springer. p.  144. ISBN   978-1-4419-9467-7.
  18. Kingman, J. F. C. (1961). "A Convexity Property of Positive Matrices". The Quarterly Journal of Mathematics. 12: 283–284. doi:10.1093/qmath/12.1.283.
  19. Cohen, J.E., 1981. Convexity of the dominant eigenvalue of an essentially nonnegative matrix. Proceedings of the American Mathematical Society, 81(4), pp.657-658.

Related Research Articles

In mathematical analysis, the Minkowski inequality establishes that the Lp spaces are normed vector spaces. Let be a measure space, let and let and be elements of Then is in and we have the triangle inequality

In mathematics, a concave function is one for which the value at any convex combination of elements in the domain is greater than or equal to the convex combination of the values at the endpoints. Equivalently, a concave function is any function for which the hypograph is convex. The class of concave functions is in a sense the opposite of the class of convex functions. A concave function is also synonymously called concave downwards, concave down, convex upwards, convex cap, or upper convex.

<span class="mw-page-title-main">Jensen's inequality</span> Theorem of convex functions

In mathematics, Jensen's inequality, named after the Danish mathematician Johan Jensen, relates the value of a convex function of an integral to the integral of the convex function. It was proved by Jensen in 1906, building on an earlier proof of the same inequality for doubly-differentiable functions by Otto Hölder in 1889. Given its generality, the inequality appears in many forms depending on the context, some of which are presented below. In its simplest form the inequality states that the convex transformation of a mean is less than or equal to the mean applied after convex transformation; it is a simple corollary that the opposite is true of concave transformations.

In mathematics and mathematical optimization, the convex conjugate of a function is a generalization of the Legendre transformation which applies to non-convex functions. It is also known as Legendre–Fenchel transformation, Fenchel transformation, or Fenchel conjugate. It allows in particular for a far reaching generalization of Lagrangian duality.

In mathematics, a function f is logarithmically convex or superconvex if , the composition of the logarithm with f, is itself a convex function.

In convex analysis, a non-negative function f : RnR+ is logarithmically concave if its domain is a convex set, and if it satisfies the inequality

<span class="mw-page-title-main">Convex analysis</span>

Convex analysis is the branch of mathematics devoted to the study of properties of convex functions and convex sets, often with applications in convex minimization, a subdomain of optimization theory.

In mathematics, subharmonic and superharmonic functions are important classes of functions used extensively in partial differential equations, complex analysis and potential theory.

In mathematical optimization theory, duality or the duality principle is the principle that optimization problems may be viewed from either of two perspectives, the primal problem or the dual problem. If the primal is a minimization problem then the dual is a maximization problem. Any feasible solution to the primal (minimization) problem is at least as large as any feasible solution to the dual (maximization) problem. Therefore, the solution to the primal is an upper bound to the solution of the dual, and the solution of the dual is a lower bound to the solution of the primal. This fact is called weak duality.

In mathematics, specifically statistics and information geometry, a Bregman divergence or Bregman distance is a measure of difference between two points, defined in terms of a strictly convex function; they form an important class of divergences. When the points are interpreted as probability distributions – notably as either values of the parameter of a parametric model or as a data set of observed values – the resulting distance is a statistical distance. The most basic Bregman divergence is the squared Euclidean distance.

<span class="mw-page-title-main">Quasiconvex function</span> Mathematical function with convex lower level sets

In mathematics, a quasiconvex function is a real-valued function defined on an interval or on a convex subset of a real vector space such that the inverse image of any set of the form is a convex set. For a function of a single variable, along any stretch of the curve the highest point is one of the endpoints. The negative of a quasiconvex function is said to be quasiconcave.

Hardy's inequality is an inequality in mathematics, named after G. H. Hardy. It states that if is a sequence of non-negative real numbers, then for every real number p > 1 one has

Stochastic approximation methods are a family of iterative methods typically used for root-finding problems or for optimization problems. The recursive update rules of stochastic approximation methods can be used, among other things, for solving linear systems when the collected data is corrupted by noise, or for approximating extreme values of functions which cannot be computed directly, but only estimated via noisy observations.

In mathematics, uniformly convex spaces are common examples of reflexive Banach spaces. The concept of uniform convexity was first introduced by James A. Clarkson in 1936.

In mathematics, the Brascamp–Lieb inequality is either of two inequalities. The first is a result in geometry concerning integrable functions on n-dimensional Euclidean space . It generalizes the Loomis–Whitney inequality and Hölder's inequality. The second is a result of probability theory which gives a concentration inequality for log-concave probability distributions. Both are named after Herm Jan Brascamp and Elliott H. Lieb.

In probability theory and statistics, a stochastic order quantifies the concept of one random variable being "bigger" than another. These are usually partial orders, so that one random variable may be neither stochastically greater than, less than, nor equal to another random variable . Many different orders exist, which have different applications.

In mathematics, there are many kinds of inequalities involving matrices and linear operators on Hilbert spaces. This article covers some important operator inequalities connected with traces of matrices.

K-convex functions, first introduced by Scarf, are a special weakening of the concept of convex function which is crucial in the proof of the optimality of the policy in inventory control theory. The policy is characterized by two numbers s and S, , such that when the inventory level falls below level s, an order is issued for a quantity that brings the inventory up to level S, and nothing is ordered otherwise. Gallego and Sethi have generalized the concept of K-convexity to higher dimensional Euclidean spaces.

<span class="mw-page-title-main">Young's inequality for products</span> Mathematical concept

In mathematics, Young's inequality for products is a mathematical inequality about the product of two numbers. The inequality is named after William Henry Young and should not be confused with Young's convolution inequality.

In measure and probability theory in mathematics, a convex measure is a probability measure that — loosely put — does not assign more mass to any intermediate set "between" two measurable sets A and B than it does to A or B individually. There are multiple ways in which the comparison between the probabilities of A and B and the intermediate set can be made, leading to multiple definitions of convexity, such as log-concavity, harmonic convexity, and so on. The mathematician Christer Borell was a pioneer of the detailed study of convex measures on locally convex spaces in the 1970s.

References