Second-order cone programming

Last updated

A second-order cone program (SOCP) is a convex optimization problem of the form

Contents

minimize
subject to

where the problem parameters are , and . is the optimization variable. is the Euclidean norm and indicates transpose. [1] The "second-order cone" in SOCP arises from the constraints, which are equivalent to requiring the affine function to lie in the second-order cone in . [1]

SOCPs can be solved by interior point methods [2] and in general, can be solved more efficiently than semidefinite programming (SDP) problems. [3] Some engineering applications of SOCP include filter design, antenna array weight design, truss design, and grasping force optimization in robotics. [4] Applications in quantitative finance include portfolio optimization; some market impact constraints, because they are not linear, cannot be solved by quadratic programming but can be formulated as SOCP problems. [5] [6] [7]

Second-order cone

The standard or unit second-order cone of dimension is defined as

.

The second-order cone is also known by quadratic cone or ice-cream cone or Lorentz cone. The standard second-order cone in is .

The set of points satisfying a second-order cone constraint is the inverse image of the unit second-order cone under an affine mapping:

and hence is convex.

The second-order cone can be embedded in the cone of the positive semidefinite matrices since

i.e., a second-order cone constraint is equivalent to a linear matrix inequality (Here means is semidefinite matrix). Similarly, we also have,

.

Relation with other optimization problems

A hierarchy of convex optimization problems. (LP: linear program, QP: quadratic program, SOCP second-order cone program, SDP: semidefinite program, CP: cone program.) Hierarchy compact convex.png
A hierarchy of convex optimization problems. (LP: linear program, QP: quadratic program, SOCP second-order cone program, SDP: semidefinite program, CP: cone program.)

When for , the SOCP reduces to a linear program. When for , the SOCP is equivalent to a convex quadratically constrained linear program.

Convex quadratically constrained quadratic programs can also be formulated as SOCPs by reformulating the objective function as a constraint. [4] Semidefinite programming subsumes SOCPs as the SOCP constraints can be written as linear matrix inequalities (LMI) and can be reformulated as an instance of semidefinite program. [4] The converse, however, is not valid: there are positive semidefinite cones that do not admit any second-order cone representation. [3] In fact, while any closed convex semialgebraic set in the plane can be written as a feasible region of a SOCP, [8] it is known that there exist convex semialgebraic sets that are not representable by SDPs, that is, there exist convex semialgebraic sets that can not be written as a feasible region of a SDP. [9]

Examples

Quadratic constraint

Consider a convex quadratic constraint of the form

This is equivalent to the SOCP constraint

Stochastic linear programming

Consider a stochastic linear program in inequality form

minimize
subject to

where the parameters are independent Gaussian random vectors with mean and covariance and . This problem can be expressed as the SOCP

minimize
subject to

where is the inverse normal cumulative distribution function. [1]

Stochastic second-order cone programming

We refer to second-order cone programs as deterministic second-order cone programs since data defining them are deterministic. Stochastic second-order cone programs are a class of optimization problems that are defined to handle uncertainty in data defining deterministic second-order cone programs. [10]

Other examples

Other modeling examples are available at the MOSEK modeling cookbook. [11]

Solvers and scripting (programming) languages

NameLicenseBrief info
AMPL commercialAn algebraic modeling language with SOCP support
Artelys Knitro commercial
CPLEX commercial
CVXPY open sourcePython modeling language with support for SOCP. Supports open source and commercial solvers.
CVXOPT open sourceConvex solver with support for SOCP
ECOS open sourceSOCP solver optimized for embedded applications
FICO Xpress commercial
Gurobi Optimizer commercial
MATLAB commercialThe coneprog function solves SOCP problems [12] using an interior-point algorithm [13]
MOSEK commercialparallel interior-point algorithm
NAG Numerical Library commercialGeneral purpose numerical library with SOCP solver
SCS open sourceSCS (Splitting Conic Solver) is a numerical optimization package for solving large-scale convex quadratic cone problems.

See also

Related Research Articles

Quadratic programming (QP) is the process of solving certain mathematical optimization problems involving quadratic functions. Specifically, one seeks to optimize a multivariate quadratic function subject to linear constraints on the variables. Quadratic programming is a type of nonlinear programming.

In mathematics, a symmetric matrix with real entries is positive-definite if the real number is positive for every nonzero real column vector where is the transpose of . More generally, a Hermitian matrix is positive-definite if the real number is positive for every nonzero complex column vector where denotes the conjugate transpose of

<span class="mw-page-title-main">Linear programming</span> Method to solve some optimization problems

Linear programming (LP), also called linear optimization, is a method to achieve the best outcome in a mathematical model whose requirements are represented by linear relationships. Linear programming is a special case of mathematical programming.

<span class="mw-page-title-main">Mathematical optimization</span> Study of mathematical algorithms for optimization problems

Mathematical optimization or mathematical programming is the selection of a best element, with regard to some criterion, from some set of available alternatives. It is generally divided into two subfields: discrete optimization and continuous optimization. Optimization problems arise in all quantitative disciplines from computer science and engineering to operations research and economics, and the development of solution methods has been of interest in mathematics for centuries.

In mathematics, the Hessian matrix, Hessian or Hesse matrix is a square matrix of second-order partial derivatives of a scalar-valued function, or scalar field. It describes the local curvature of a function of many variables. The Hessian matrix was developed in the 19th century by the German mathematician Ludwig Otto Hesse and later named after him. Hesse originally used the term "functional determinants". The Hessian is sometimes denoted by H or, ambiguously, by ∇2.

Convex optimization is a subfield of mathematical optimization that studies the problem of minimizing convex functions over convex sets. Many classes of convex optimization problems admit polynomial-time algorithms, whereas mathematical optimization is in general NP-hard.

In mathematical optimization, the Karush–Kuhn–Tucker (KKT) conditions, also known as the Kuhn–Tucker conditions, are first derivative tests for a solution in nonlinear programming to be optimal, provided that some regularity conditions are satisfied.

In mathematics, Farkas' lemma is a solvability theorem for a finite system of linear inequalities. It was originally proven by the Hungarian mathematician Gyula Farkas. Farkas' lemma is the key result underpinning the linear programming duality and has played a central role in the development of mathematical optimization. It is used amongst other things in the proof of the Karush–Kuhn–Tucker theorem in nonlinear programming. Remarkably, in the area of the foundations of quantum theory, the lemma also underlies the complete set of Bell inequalities in the form of necessary and sufficient conditions for the existence of a local hidden-variable theory, given data from any specific set of measurements.

<span class="mw-page-title-main">Convex cone</span> Mathematical set closed under positive linear combinations

In linear algebra, a cone—sometimes called a linear cone for distinguishing it from other sorts of cones—is a subset of a vector space that is closed under positive scalar multiplication; that is, C is a cone if implies for every positive scalar s.

Semidefinite programming (SDP) is a subfield of convex optimization concerned with the optimization of a linear objective function over the intersection of the cone of positive semidefinite matrices with an affine space, i.e., a spectrahedron.

In convex optimization, a linear matrix inequality (LMI) is an expression of the form

In mathematics, the Grothendieck inequality states that there is a universal constant with the following property. If Mij is an n × n matrix with

In mathematical optimization, a quadratically constrained quadratic program (QCQP) is an optimization problem in which both the objective function and the constraints are quadratic functions. It has the form

Conic optimization is a subfield of convex optimization that studies problems consisting of minimizing a convex function over the intersection of an affine subspace and a convex cone.

In mathematics, every analytic function can be used for defining a matrix function that maps square matrices with complex entries to square matrices of the same size.

A self-concordant function is a function satisfying a certain differential inequality, which makes it particularly easy for optimizaiton using Newton's method

Sparse principal component analysis is a technique used in statistical analysis and, in particular, in the analysis of multivariate data sets. It extends the classic method of principal component analysis (PCA) for the reduction of dimensionality of data by introducing sparsity structures to the input variables.

A sum-of-squares optimization program is an optimization problem with a linear cost function and a particular type of constraint on the decision variables. These constraints are of the form that when the decision variables are used as coefficients in certain polynomials, those polynomials should have the polynomial SOS property. When fixing the maximum degree of the polynomials involved, sum-of-squares optimization is also known as the Lasserre hierarchy of relaxations in semidefinite programming.

<span class="mw-page-title-main">Matrix completion</span>

Matrix completion is the task of filling in the missing entries of a partially observed matrix, which is equivalent to performing data imputation in statistics. A wide range of datasets are naturally organized in matrix form. One example is the movie-ratings matrix, as appears in the Netflix problem: Given a ratings matrix in which each entry represents the rating of movie by customer , if customer has watched movie and is otherwise missing, we would like to predict the remaining entries in order to make good recommendations to customers on what to watch next. Another example is the document-term matrix: The frequencies of words used in a collection of documents can be represented as a matrix, where each entry corresponds to the number of times the associated term appears in the indicated document.

Quantum optimization algorithms are quantum algorithms that are used to solve optimization problems. Mathematical optimization deals with finding the best solution to a problem from a set of possible solutions. Mostly, the optimization problem is formulated as a minimization problem, where one tries to minimize an error which depends on the solution: the optimal solution has the minimal error. Different optimization techniques are applied in various fields such as mechanics, economics and engineering, and as the complexity and amount of data involved rise, more efficient ways of solving optimization problems are needed. Quantum computing may allow problems which are not practically feasible on classical computers to be solved, or suggest a considerable speed up with respect to the best known classical algorithm.

References

  1. 1 2 3 Boyd, Stephen; Vandenberghe, Lieven (2004). Convex Optimization (PDF). Cambridge University Press. ISBN   978-0-521-83378-3 . Retrieved July 15, 2019.
  2. Potra, lorian A.; Wright, Stephen J. (1 December 2000). "Interior-point methods". Journal of Computational and Applied Mathematics. 124 (1–2): 281–302. Bibcode:2000JCoAM.124..281P. doi:10.1016/S0377-0427(00)00433-7.
  3. 1 2 Fawzi, Hamza (2019). "On representing the positive semidefinite cone using the second-order cone". Mathematical Programming. 175 (1–2): 109–118. arXiv: 1610.04901 . doi:10.1007/s10107-018-1233-0. ISSN   0025-5610. S2CID   119324071.
  4. 1 2 3 Lobo, Miguel Sousa; Vandenberghe, Lieven; Boyd, Stephen; Lebret, Hervé (1998). "Applications of second-order cone programming". Linear Algebra and Its Applications. 284 (1–3): 193–228. doi: 10.1016/S0024-3795(98)10032-0 .
  5. "Solving SOCP" (PDF).
  6. "portfolio optimization" (PDF).
  7. Li, Haksun (16 January 2022). Numerical Methods Using Java: For Data Science, Analysis, and Engineering. APress. pp. Chapter 10. ISBN   978-1484267967.
  8. Scheiderer, Claus (2020-04-08). "Second-order cone representation for convex subsets of the plane". arXiv: 2004.04196 [math.OC].
  9. Scheiderer, Claus (2018). "Spectrahedral Shadows". SIAM Journal on Applied Algebra and Geometry. 2 (1): 26–44. doi: 10.1137/17M1118981 . ISSN   2470-6566.
  10. Alzalg, Baha M. (2012-10-01). "Stochastic second-order cone programming: Applications models". Applied Mathematical Modelling. 36 (10): 5122–5134. doi:10.1016/j.apm.2011.12.053. ISSN   0307-904X.
  11. "MOSEK Modeling Cookbook - Conic Quadratic Optimization".
  12. "Second-order cone programming solver - MATLAB coneprog". MathWorks. 2021-03-01. Retrieved 2021-07-15.
  13. "Second-Order Cone Programming Algorithm - MATLAB & Simulink". MathWorks. 2021-03-01. Retrieved 2021-07-15.
  14. "MOSEK Modeling Cookbook - the Power Cones".