Fenchel's duality theorem

Last updated

In mathematics, Fenchel's duality theorem is a result in the theory of convex functions named after Werner Fenchel.

Contents

Let ƒ be a proper convex function on Rn and let g be a proper concave function on Rn. Then, if regularity conditions are satisfied,

where ƒ * is the convex conjugate of ƒ (also referred to as the FenchelLegendre transform) and g * is the concave conjugate of g. That is,

Mathematical theorem

Let X and Y be Banach spaces, and be convex functions and be a bounded linear map. Then the Fenchel problems:

satisfy weak duality, i.e. . Note that are the convex conjugates of f,g respectively, and is the adjoint operator. The perturbation function for this dual problem is given by .

Suppose that f,g, and A satisfy either

  1. f and g are lower semi-continuous and where is the algebraic interior and , where h is some function, is the set , or
  2. where are the points where the function is continuous.

Then strong duality holds, i.e. . If then supremum is attained. [1]

One-dimensional illustration

In the following figure, the minimization problem on the left side of the equation is illustrated. One seeks to vary x such that the vertical distance between the convex and concave curves at x is as small as possible. The position of the vertical line in the figure is the (approximate) optimum.

FencheDual02.png

The next figure illustrates the maximization problem on the right hand side of the above equation. Tangents are drawn to each of the two curves such that both tangents have the same slope p. The problem is to adjust p in such a way that the two tangents are as far away from each other as possible (more precisely, such that the points where they intersect the y-axis are as far from each other as possible). Imagine the two tangents as metal bars with vertical springs between them that push them apart and against the two parabolas that are fixed in place.

FenchelDual01.png

Fenchel's theorem states that the two problems have the same solution. The points having the minimum vertical separation are also the tangency points for the maximally separated parallel tangents.

See also

Related Research Articles

Distributions, also known as Schwartz distributions or generalized functions, are objects that generalize the classical notion of functions in mathematical analysis. Distributions make it possible to differentiate functions whose derivatives do not exist in the classical sense. In particular, any locally integrable function has a distributional derivative. Distributions are widely used in the theory of partial differential equations, where it may be easier to establish the existence of distributional solutions than classical solutions, or appropriate classical solutions may not exist. Distributions are also important in physics and engineering where many problems naturally lead to differential equations whose solutions or initial conditions are distributions, such as the Dirac delta function.

Convex function real function with secant line between points above the graph itself

In mathematics, a real-valued function defined on an n-dimensional interval is called convex if the line segment between any two points on the graph of the function lies above the graph between the two points. Equivalently, a function is convex if its epigraph is a convex set. A twice-differentiable function of a single variable is convex if and only if its second derivative is nonnegative on its entire domain. Well-known examples of convex functions of a single variable include the squaring function and the exponential function . In simple terms, a convex function refers to a function that is in the shape of a cup , and a concave function is in the shape of a cap .

Jensens inequality Theorem of convex functions

In mathematics, Jensen's inequality, named after the Danish mathematician Johan Jensen, relates the value of a convex function of an integral to the integral of the convex function. It was proved by Jensen in 1906. Given its generality, the inequality appears in many forms depending on the context, some of which are presented below. In its simplest form the inequality states that the convex transformation of a mean is less than or equal to the mean applied after convex transformation; it is a simple corollary that the opposite is true of concave transformations.

Legendre transformation

In mathematics and physics, the Legendre transformation, named after Adrien-Marie Legendre, is an involutive transformation on the real-valued convex functions of one real variable. In physical problems, it is used to convert functions of one quantity into functions of the conjugate quantity. In this way, it is commonly used in classical mechanics to derive the Hamiltonian formalism out of the Lagrangian formalism and in thermodynamics to derive the thermodynamic potentials, as well as in the solution of differential equations of several variables.

In mathematics and mathematical optimization, the convex conjugate of a function is a generalization of the Legendre transformation which applies to non-convex functions. It is also known as Legendre–Fenchel transformation, Fenchel transformation, or Fenchel conjugate. It allows in particular for a far reaching generalization of Lagrangian duality.

Convex optimization is a subfield of mathematical optimization that studies the problem of minimizing convex functions over convex sets. Many classes of convex optimization problems admit polynomial-time algorithms, whereas mathematical optimization is in general NP-hard.

In mathematics, there are usually many different ways to construct a topological tensor product of two topological vector spaces. For Hilbert spaces or nuclear spaces there is a simple well-behaved theory of tensor products, but for general Banach spaces or locally convex topological vector spaces the theory is notoriously subtle.

Convex analysis

Convex analysis is the branch of mathematics devoted to the study of properties of convex functions and convex sets, often with applications in convex minimization, a subdomain of optimization theory.

In mathematical optimization theory, duality or the duality principle is the principle that optimization problems may be viewed from either of two perspectives, the primal problem or the dual problem. The solution to the dual problem provides a lower bound to the solution of the primal (minimization) problem. However in general the optimal values of the primal and dual problems need not be equal. Their difference is called the duality gap. For convex optimization problems, the duality gap is zero under a constraint qualification condition.

In mathematics, the concepts of essential supremum and essential infimum are related to the notions of supremum and infimum, but adapted to measure theory and functional analysis, where one often deals with statements that are not valid for all elements in a set, but rather almost everywhere, i.e., except on a set of measure zero.

Ordered vector space

In mathematics, an ordered vector space or partially ordered vector space is a vector space equipped with a partial order that is compatible with the vector space operations.

In functional analysis, the dual norm is a measure of the "size" of each continuous linear functional defined on a normed vector space.

In mathematical analysis, Ekeland's variational principle, discovered by Ivar Ekeland, is a theorem that asserts that there exists nearly optimal solutions to some optimization problems.

In optimization problems in applied mathematics, the duality gap is the difference between the primal and dual solutions. If is the optimal dual value and is the optimal primal value then the duality gap is equal to . This value is always greater than or equal to 0. The duality gap is zero if and only if strong duality holds. Otherwise the gap is strictly positive and weak duality holds.

In mathematical optimization, the perturbation function is any function which relates to primal and dual problems. The name comes from the fact that any such function defines a perturbation of the initial problem. In many cases this takes the form of shifting the constraints.

In convex analysis, a branch of mathematics, the effective domain is an extension of the domain of a function.

In mathematics, the bipolar theorem is a theorem in functional analysis that characterizes the bipolar of a set. In convex analysis, the bipolar theorem refers to a necessary and sufficient conditions for a cone to be equal to its bipolar. The bipolar theorem can be seen as a special case of the Fenchel–Moreau theorem.

Proximal gradient methods are a generalized form of projection used to solve non-differentiable convex optimization problems.

In mathematical analysis, epi-convergence is a type of convergence for real-valued and extended real-valued functions.

In the field of Functional Analysis, it is possible to generalize the notion of derivative to infinite dimensional topological vector spaces (TVSs) in multiple ways. But when the domain of TVS-value functions is a subset of finite-dimensional Euclidean space then the number of generalizations of the derivative is much more limited and derivatives are more well behaved. This article presents the theory of k-times continuously differentiable functions on an open subset of Euclidean space , which is an important special case of differentiation between arbitrary TVSs. All vector spaces will be assumed to be over the field where is either the real numbers or the complex numbers

References

  1. Borwein, Jonathan; Zhu, Qiji (2005). Techniques of Variational Analysis . Springer. pp.  135–137. ISBN   978-1-4419-2026-3.