Finsler's lemma is a mathematical result named after Paul Finsler. It states equivalent ways to express the positive definiteness of a quadratic form Q constrained by a linear form L. Since it is equivalent to another lemmas used in optimization and control theory, such as Yakubovich's S-lemma, [1] Finsler's lemma has been given many proofs and has been widely used, particularly in results related to robust optimization and linear matrix inequalities.
Let x ∈ Rn, and Q ∈ Rn x n and L ∈ Rn x n be symmetric matrices. The following statements are equivalent: [2]
When the matrix L is indefinite, replacing strict inequalities with non-strict ones still maintains the equivalence between the statements of Finsler's lemma. However, if L is not indefinite, additional assumptions are necessary to ensure equivalence between the statements. [3]
In the particular case that L is positive semi-definite, it is possible to decompose it as L = BTB. The following statements, which are also referred as Finsler's lemma in the literature, are equivalent: [4]
There is also a variant of Finsler's lemma for quadratic matrix inequalities, known as matrix Finsler's lemma, which states that the following statements are equivalent for symmetric matrices Q and L belonging to R(l+k)x(l+k): [5] [6]
under the assumption that
and
satisfy the following assumptions:
The equivalence between the following statements is also common on the literature of linear matrix inequalities, and is known as the Projection Lemma (or also as Elimination Lemma): [7]
This lemma generalizes one of the Finsler's lemma variants by including an extra matrix C and an extra constraint involving this extra matrix.
It is interesting to note that if the strict inequalities are changed to non-strict inequalities, the equivalence does not hold anymore: only the second statement imply the first statement. Nevertheless, it still possible to obtain the equivalence between the statements under extra assumptions. [8]
Finsler's lemma also generalizes for matrices Q and B depending on a parameter s within a set S. In this case, it is natural to ask if the same variable μ (respectively X) can satisfy for all (respectively, ). If Q and B depends continuously on the parameter s, and S is compact, then this is true. If S is not compact, but Q and B are still continuous matrix-valued functions, then μ and X can be guaranteed to be at least continuous functions. [9]
The matrix variant of Finsler lemma has been applied to the data-driven control of Lur'e systems [5] and in a data-driven robust linear matrix inequality-based model predictive control scheme. [10]
Finsler's lemma can be used to give novel linear matrix inequality (LMI) characterizations to stability and control problems. [4] The set of LMIs stemmed from this procedure yields less conservative results when applied to control problems where the system matrices has dependence on a parameter, such as robust control problems and control of linear-parameter varying systems. [11] This approach has recently been called as S-variable approach [12] [13] and the LMIs stemming from this approach are known as SV-LMIs (also known as dilated LMIs [14] ).
A nonlinear system has the universal stabilizability property if every forward-complete solution of a system can be globally stabilized. By the use of Finsler's lemma, it is possible to derive a sufficient condition for universal stabilizability in terms of a differential linear matrix inequality. [15]
Autocorrelation, sometimes known as serial correlation in the discrete time case, is the correlation of a signal with a delayed copy of itself as a function of delay. Informally, it is the similarity between observations of a random variable as a function of the time lag between them. The analysis of autocorrelation is a mathematical tool for finding repeating patterns, such as the presence of a periodic signal obscured by noise, or identifying the missing fundamental frequency in a signal implied by its harmonic frequencies. It is often used in signal processing for analyzing functions or series of values, such as time domain signals.
In mathematical physics and mathematics, the Pauli matrices are a set of three 2 × 2 complex matrices that are traceless, Hermitian, involutory and unitary. Usually indicated by the Greek letter sigma, they are occasionally denoted by tau when used in connection with isospin symmetries.
In mathematics, the Lp spaces are function spaces defined using a natural generalization of the p-norm for finite-dimensional vector spaces. They are sometimes called Lebesgue spaces, named after Henri Lebesgue, although according to the Bourbaki group they were first introduced by Frigyes Riesz.
In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One definition is that a random vector is said to be k-variate normally distributed if every linear combination of its k components has a univariate normal distribution. Its importance derives mainly from the multivariate central limit theorem. The multivariate normal distribution is often used to describe, at least approximately, any set of (possibly) correlated real-valued random variables, each of which clusters around a mean value.
In the mathematical field of real analysis, the monotone convergence theorem is any of a number of related theorems proving the good convergence behaviour of monotonic sequences, i.e. sequences that are non-increasing, or non-decreasing. In its simplest form, it says that a non-decreasing bounded-above sequence of real numbers converges to its smallest upper bound, its supremum. Likewise, a non-increasing bounded-below sequence converges to its largest lower bound, its infimum. In particular, infinite sums of non-negative numbers converge to the supremum of the partial sums if and only if the partial sums are bounded.
In group theory, the quaternion group Q8 (sometimes just denoted by Q) is a non-abelian group of order eight, isomorphic to the eight-element subset of the quaternions under multiplication. It is given by the group presentation
In mathematical analysis, Hölder's inequality, named after Otto Hölder, is a fundamental inequality between integrals and an indispensable tool for the study of Lp spaces.
In mathematics, particularly linear algebra, an orthonormal basis for an inner product space with finite dimension is a basis for whose vectors are orthonormal, that is, they are all unit vectors and orthogonal to each other. For example, the standard basis for a Euclidean space is an orthonormal basis, where the relevant inner product is the dot product of vectors. The image of the standard basis under a rotation or reflection is also orthonormal, and every orthonormal basis for arises in this fashion. An orthonormal basis can be derived from an orthogonal basis via normalization. The choice of an origin and an orthonormal basis forms a coordinate frame known as an orthonormal frame.
In mathematics, Fatou's lemma establishes an inequality relating the Lebesgue integral of the limit inferior of a sequence of functions to the limit inferior of integrals of these functions. The lemma is named after Pierre Fatou.
In mathematics, the conformal group of an inner product space is the group of transformations from the space to itself that preserve angles. More formally, it is the group of transformations that preserve the conformal geometry of the space.
In mathematics, the Riesz–Thorin theorem, often referred to as the Riesz–Thorin interpolation theorem or the Riesz–Thorin convexity theorem, is a result about interpolation of operators. It is named after Marcel Riesz and his student G. Olof Thorin.
In mathematics, a norm is a function from a real or complex vector space to the non-negative real numbers that behaves in certain ways like the distance from the origin: it commutes with scaling, obeys a form of the triangle inequality, and is zero only at the origin. In particular, the Euclidean distance in a Euclidean space is defined by a norm on the associated Euclidean vector space, called the Euclidean norm, the 2-norm, or, sometimes, the magnitude or length of the vector. This norm can be defined as the square root of the inner product of a vector with itself.
Semidefinite programming (SDP) is a subfield of mathematical programming concerned with the optimization of a linear objective function over the intersection of the cone of positive semidefinite matrices with an affine space, i.e., a spectrahedron.
In control theory, the linear–quadratic–Gaussian (LQG) control problem is one of the most fundamental optimal control problems, and it can also be operated repeatedly for model predictive control. It concerns linear systems driven by additive white Gaussian noise. The problem is to determine an output feedback law that is optimal in the sense of minimizing the expected value of a quadratic cost criterion. Output measurements are assumed to be corrupted by Gaussian noise and the initial state, likewise, is assumed to be a Gaussian random vector.
In convex optimization, a linear matrix inequality (LMI) is an expression of the form
In mathematics, ergodicity expresses the idea that a point of a moving system, either a dynamical system or a stochastic process, will eventually visit all parts of the space that the system moves in, in a uniform and random sense. This implies that the average behavior of the system can be deduced from the trajectory of a "typical" point. Equivalently, a sufficiently large collection of random samples from a process can represent the average statistical properties of the entire process. Ergodicity is a property of the system; it is a statement that the system cannot be reduced or factored into smaller components. Ergodic theory is the study of systems possessing ergodicity.
The Kalman–Yakubovich–Popov lemma is a result in system analysis and control theory which states: Given a number , two n-vectors B, C and an n x n Hurwitz matrix A, if the pair is completely controllable, then a symmetric matrix P and a vector Q satisfying
In mathematics and mathematical physics, raising and lowering indices are operations on tensors which change their type. Raising and lowering indices are a form of index manipulation in tensor expressions.
In functional analysis, the dual norm is a measure of size for a continuous linear function defined on a normed vector space.
Baranyi and Yam proposed the TP model transformation as a new concept in quasi-LPV (qLPV) based control, which plays a central role in the highly desirable bridging between identification and polytopic systems theories. It is also used as a TS (Takagi-Sugeno) fuzzy model transformation. It is uniquely effective in manipulating the convex hull of polytopic forms, and, hence, has revealed and proved the fact that convex hull manipulation is a necessary and crucial step in achieving optimal solutions and decreasing conservativeness in modern linear matrix inequality based control theory. Thus, although it is a transformation in a mathematical sense, it has established a conceptually new direction in control theory and has laid the ground for further new approaches towards optimality.