Convex analysis is the branch of mathematics devoted to the study of properties of convex functions and convex sets, often with applications in convex minimization, a subdomain of optimization theory.
A subset of some vector space is convex if it satisfies any of the following equivalent conditions:
Throughout, will be a map valued in the extended real numbers with a domain that is a convex subset of some vector space. The map is a convex function if
| (Convexity ≤) |
holds for any real and any with If this remains true of when the defining inequality ( Convexity ≤ ) is replaced by the strict inequality
| (Convexity <) |
then is called strictly convex. [1]
Convex functions are related to convex sets. Specifically, the function is convex if and only if its epigraph
| (Epigraph def.) |
is a convex set. [2] The epigraphs of extended real-valued functions play a role in convex analysis that is analogous to the role played by graphs of real-valued function in real analysis. Specifically, the epigraph of an extended real-valued function provides geometric intuition that can be used to help formula or prove conjectures.
The domain of a function is denoted by while its effective domain is the set [2]
| (dom f def.) |
The function is called proper if and for all [2] Alternatively, this means that there exists some in the domain of at which and is also never equal to In words, a function is proper if its domain is not empty, it never takes on the value and it also is not identically equal to If is a proper convex function then there exist some vector and some such that
where denotes the dot product of these vectors.
The convex conjugate of an extended real-valued function (not necessarily convex) is the function from the (continuous) dual space of and [3]
where the brackets denote the canonical duality The biconjugate of is the map defined by for every If denotes the set of -valued functions on then the map defined by is called the Legendre-Fenchel transform.
If and then the subdifferential set is
For example, in the important special case where is a norm on , it can be shown [proof 1] that if then this definition reduces down to:
For any and which is called the Fenchel-Young inequality. This inequality is an equality (i.e. ) if and only if It is in this way that the subdifferential set is directly related to the convex conjugate
The biconjugate of a function is the conjugate of the conjugate, typically written as The biconjugate is useful for showing when strong or weak duality hold (via the perturbation function).
For any the inequality follows from the Fenchel–Young inequality. For proper functions, if and only if is convex and lower semi-continuous by Fenchel–Moreau theorem. [3] [4]
A convex minimization (primal) problem is one of the form
In optimization theory, the duality principle states that optimization problems may be viewed from either of two perspectives, the primal problem or the dual problem.
In general given two dual pairs separated locally convex spaces and Then given the function we can define the primal problem as finding such that
If there are constraint conditions, these can be built into the function by letting where is the indicator function. Then let be a perturbation function such that [5]
The dual problem with respect to the chosen perturbation function is given by
where is the convex conjugate in both variables of
The duality gap is the difference of the right and left hand sides of the inequality [6] [5] [7]
This principle is the same as weak duality. If the two sides are equal to each other, then the problem is said to satisfy strong duality.
There are many conditions for strong duality to hold such as:
For a convex minimization problem with inequality constraints,
the Lagrangian dual problem is
where the objective function is the Lagrange dual function defined as follows:
The Cauchy–Schwarz inequality is considered one of the most important and widely used inequalities in mathematics.
In mathematics, the Lp spaces are function spaces defined using a natural generalization of the p-norm for finite-dimensional vector spaces. They are sometimes called Lebesgue spaces, named after Henri Lebesgue, although according to the Bourbaki group they were first introduced by Frigyes Riesz. Lp spaces form an important class of Banach spaces in functional analysis, and of topological vector spaces. Because of their key role in the mathematical analysis of measure and probability spaces, Lebesgue spaces are used also in the theoretical discussion of problems in physics, statistics, economics, finance, engineering, and other disciplines.
Distributions, also known as Schwartz distributions or generalized functions, are objects that generalize the classical notion of functions in mathematical analysis. Distributions make it possible to differentiate functions whose derivatives do not exist in the classical sense. In particular, any locally integrable function has a distributional derivative.
In mathematics, a linear form is a linear map from a vector space to its field of scalars.
In mathematics, a real-valued function is called convex if the line segment between any two points on the graph of the function lies above the graph between the two points. Equivalently, a function is convex if its epigraph is a convex set. A twice-differentiable function of a single variable is convex if and only if its second derivative is nonnegative on its entire domain. Well-known examples of convex functions of a single variable include the quadratic function and the exponential function . In simple terms, a convex function refers to a function whose graph is shaped like a cup , while a concave function's graph is shaped like a cap .
In mathematics, Jensen's inequality, named after the Danish mathematician Johan Jensen, relates the value of a convex function of an integral to the integral of the convex function. It was proved by Jensen in 1906, building on an earlier proof of the same inequality for doubly-differentiable functions by Otto Hölder in 1889. Given its generality, the inequality appears in many forms depending on the context, some of which are presented below. In its simplest form the inequality states that the convex transformation of a mean is less than or equal to the mean applied after convex transformation; it is a simple corollary that the opposite is true of concave transformations.
In mathematics, the Legendre transformation, named after Adrien-Marie Legendre, is an involutive transformation on real-valued convex functions of one real variable. In physical problems, it is used to convert functions of one quantity into functions of the conjugate quantity. In this way, it is commonly used in classical mechanics to derive the Hamiltonian formalism out of the Lagrangian formalism and in thermodynamics to derive the thermodynamic potentials, as well as in the solution of differential equations of several variables.
In functional analysis and related branches of mathematics, the Banach–Alaoglu theorem states that the closed unit ball of the dual space of a normed vector space is compact in the weak* topology. A common proof identifies the unit ball with the weak-* topology as a closed subset of a product of compact sets with the product topology. As a consequence of Tychonoff's theorem, this product, and hence the unit ball within, is compact.
In mathematics, a norm is a function from a real or complex vector space to the non-negative real numbers that behaves in certain ways like the distance from the origin: it commutes with scaling, obeys a form of the triangle inequality, and is zero only at the origin. In particular, the Euclidean distance of a vector from the origin is a norm, called the Euclidean norm, or 2-norm, which may also be defined as the square root of the inner product of a vector with itself.
In mathematics and mathematical optimization, the convex conjugate of a function is a generalization of the Legendre transformation which applies to non-convex functions. It is also known as Legendre–Fenchel transformation, Fenchel transformation, or Fenchel conjugate. It allows in particular for a far reaching generalization of Lagrangian duality.
In functional and convex analysis, and related disciplines of mathematics, the polar set is a special convex set associated to any subset of a vector space lying in the dual space The bipolar of a subset is the polar of but lies in .
In mathematics, Fenchel's duality theorem is a result in the theory of convex functions named after Werner Fenchel.
In mathematics, specifically statistics and information geometry, a Bregman divergence or Bregman distance is a measure of difference between two points, defined in terms of a strictly convex function; they form an important class of divergences. When the points are interpreted as probability distributions – notably as either values of the parameter of a parametric model or as a data set of observed values – the resulting distance is a statistical distance. The most basic Bregman divergence is the squared Euclidean distance.
In geometry, the hyperplane separation theorem is a theorem about disjoint convex sets in n-dimensional Euclidean space. There are several rather similar versions. In one version of the theorem, if both these sets are closed and at least one of them is compact, then there is a hyperplane in between them and even two parallel hyperplanes in between them separated by a gap. In another version, if both disjoint convex sets are open, then there is a hyperplane in between them, but not necessarily any gap. An axis which is orthogonal to a separating hyperplane is a separating axis, because the orthogonal projections of the convex bodies onto the axis are disjoint.
In mathematics, the Hilbert projection theorem is a famous result of convex analysis that says that for every vector in a Hilbert space and every nonempty closed convex there exists a unique vector for which is minimized over the vectors ; that is, such that for every
In functional analysis, the dual norm is a measure of size for a continuous linear function defined on a normed vector space.
In mathematics, Hilbert spaces allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. A Hilbert space is a vector space equipped with an inner product which defines a distance function for which it is a complete metric space. Hilbert spaces arise naturally and frequently in mathematics and physics, typically as function spaces.
In mathematical analysis, Ekeland's variational principle, discovered by Ivar Ekeland, is a theorem that asserts that there exist nearly optimal solutions to some optimization problems.
In mathematical optimization, the perturbation function is any function which relates to primal and dual problems. The name comes from the fact that any such function defines a perturbation of the initial problem. In many cases this takes the form of shifting the constraints.
In mathematics, the bipolar theorem is a theorem in functional analysis that characterizes the bipolar of a set. In convex analysis, the bipolar theorem refers to a necessary and sufficient conditions for a cone to be equal to its bipolar. The bipolar theorem can be seen as a special case of the Fenchel–Moreau theorem.