Generalized inverse

Last updated

In mathematics, and in particular, algebra, a generalized inverse (or, g-inverse) of an element x is an element y that has some properties of an inverse element but not necessarily all of them. The purpose of constructing a generalized inverse of a matrix is to obtain a matrix that can serve as an inverse in some sense for a wider class of matrices than invertible matrices. Generalized inverses can be defined in any mathematical structure that involves associative multiplication, that is, in a semigroup. This article describes generalized inverses of a matrix .

Contents

A matrix is a generalized inverse of a matrix if [1] [2] [3] A generalized inverse exists for an arbitrary matrix, and when a matrix has a regular inverse, this inverse is its unique generalized inverse. [1]

Motivation

Consider the linear system

where is an matrix and the column space of . If and is nonsingular then will be the solution of the system. Note that, if is nonsingular, then

Now suppose is rectangular (), or square and singular. Then we need a right candidate of order such that for all

[4]

That is, is a solution of the linear system . Equivalently, we need a matrix of order such that

Hence we can define the generalized inverse as follows: Given an matrix , an matrix is said to be a generalized inverse of if [1] [2] [3] The matrix has been termed a regular inverse of by some authors. [5]

Types

Important types of generalized inverse include:

Some generalized inverses are defined and classified based on the Penrose conditions:

where denotes conjugate transpose. If satisfies the first condition, then it is a generalized inverse of . If it satisfies the first two conditions, then it is a reflexive generalized inverse of . If it satisfies all four conditions, then it is the pseudoinverse of , which is denoted by and also known as the Moore–Penrose inverse, after the pioneering works by E. H. Moore and Roger Penrose. [2] [7] [8] [9] [10] [11] It is convenient to define an -inverse of as an inverse that satisfies the subset of the Penrose conditions listed above. Relations, such as , can be established between these different classes of -inverses. [1]

When is non-singular, any generalized inverse and is therefore unique. For a singular , some generalised inverses, such as the Drazin inverse and the Moore–Penrose inverse, are unique, while others are not necessarily uniquely defined.

Examples

Reflexive generalized inverse

Let

Since , is singular and has no regular inverse. However, and satisfy Penrose conditions (1) and (2), but not (3) or (4). Hence, is a reflexive generalized inverse of .

One-sided inverse

Let

Since is not square, has no regular inverse. However, is a right inverse of . The matrix has no left inverse.

Inverse of other semigroups (or rings)

The element b is a generalized inverse of an element a if and only if , in any semigroup (or ring, since the multiplication function in any ring is a semigroup).

The generalized inverses of the element 3 in the ring are 3, 7, and 11, since in the ring :

The generalized inverses of the element 4 in the ring are 1, 4, 7, and 10, since in the ring :

If an element a in a semigroup (or ring) has an inverse, the inverse must be the only generalized inverse of this element, like the elements 1, 5, 7, and 11 in the ring .

In the ring , any element is a generalized inverse of 0, however, 2 has no generalized inverse, since there is no b in such that .

Construction

The following characterizations are easy to verify:

Uses

Any generalized inverse can be used to determine whether a system of linear equations has any solutions, and if so to give all of them. If any solutions exist for the n × m linear system

,

with vector of unknowns and vector of constants, all solutions are given by

,

parametric on the arbitrary vector , where is any generalized inverse of . Solutions exist if and only if is a solution, that is, if and only if . If A has full column rank, the bracketed expression in this equation is the zero matrix and so the solution is unique. [12]

Generalized inverses of matrices

The generalized inverses of matrices can be characterized as follows. Let , and

be its singular-value decomposition. Then for any generalized inverse , there exist [1] matrices , , and such that

Conversely, any choice of , , and for matrix of this form is a generalized inverse of . [1] The -inverses are exactly those for which , the -inverses are exactly those for which , and the -inverses are exactly those for which . In particular, the pseudoinverse is given by :

Transformation consistency properties

In practical applications it is necessary to identify the class of matrix transformations that must be preserved by a generalized inverse. For example, the Moore–Penrose inverse, satisfies the following definition of consistency with respect to transformations involving unitary matrices U and V:

.

The Drazin inverse, satisfies the following definition of consistency with respect to similarity transformations involving a nonsingular matrix S:

.

The unit-consistent (UC) inverse, [13] satisfies the following definition of consistency with respect to transformations involving nonsingular diagonal matrices D and E:

.

The fact that the Moore–Penrose inverse provides consistency with respect to rotations (which are orthonormal transformations) explains its widespread use in physics and other applications in which Euclidean distances must be preserved. The UC inverse, by contrast, is applicable when system behavior is expected to be invariant with respect to the choice of units on different state variables, e.g., miles versus kilometers.

See also

Citations

Sources

Textbook

Publication

Related Research Articles

<span class="mw-page-title-main">Pauli matrices</span> Matrices important in quantum mechanics and the study of spin

In mathematical physics and mathematics, the Pauli matrices are a set of three 2 × 2 complex matrices that are traceless, Hermitian, involutory and unitary. Usually indicated by the Greek letter sigma, they are occasionally denoted by tau when used in connection with isospin symmetries.

In mathematics, a symmetric matrix with real entries is positive-definite if the real number is positive for every nonzero real column vector where is the row vector transpose of More generally, a Hermitian matrix is positive-definite if the real number is positive for every nonzero complex column vector where denotes the conjugate transpose of

In mathematics, the concept of an inverse element generalises the concepts of opposite and reciprocal of numbers.

In linear algebra, an orthogonal matrix, or orthonormal matrix, is a real square matrix whose columns and rows are orthonormal vectors.

In mechanics and geometry, the 3D rotation group, often denoted SO(3), is the group of all rotations about the origin of three-dimensional Euclidean space under the operation of composition.

In linear algebra, an n-by-n square matrix A is called invertible if there exists an n-by-n square matrix B such thatwhere In denotes the n-by-n identity matrix and the multiplication used is ordinary matrix multiplication. If this is the case, then the matrix B is uniquely determined by A, and is called the (multiplicative) inverse of A, denoted by A−1. Matrix inversion is the process of finding the matrix which when multiplied by the original matrix gives the identity matrix.

In mathematics, the conjugate transpose, also known as the Hermitian transpose, of an complex matrix is an matrix obtained by transposing and applying complex conjugation to each entry. There are several notations, such as or , , or .

In linear algebra, a square matrix  is called diagonalizable or non-defective if it is similar to a diagonal matrix. That is, if there exists an invertible matrix  and a diagonal matrix such that . This is equivalent to . This property exists for any linear map: for a finite-dimensional vector space , a linear map  is called diagonalizable if there exists an ordered basis of  consisting of eigenvectors of . These definitions are equivalent: if  has a matrix representation as above, then the column vectors of  form a basis consisting of eigenvectors of , and the diagonal entries of  are the corresponding eigenvalues of ; with respect to this eigenvector basis,  is represented by .

In mathematics, and in particular linear algebra, the Moore–Penrose inverse of a matrix , often called the pseudoinverse, is the most widely known generalization of the inverse matrix. It was independently described by E. H. Moore in 1920, Arne Bjerhammar in 1951, and Roger Penrose in 1955. Earlier, Erik Ivar Fredholm had introduced the concept of a pseudoinverse of integral operators in 1903. The terms pseudoinverse and generalized inverse are sometimes used as synonyms for the Moore–Penrose inverse of a matrix, but sometimes applied to other elements of algebraic structures which share some but not all properties expected for an inverse element.

The Schur complement of a block matrix, encountered in linear algebra and the theory of matrices, is defined as follows.

In multivariable calculus, the implicit function theorem is a tool that allows relations to be converted to functions of several real variables. It does so by representing the relation as the graph of a function. There may not be a single function whose graph can represent the entire relation, but there may be such a function on a restriction of the domain of the relation. The implicit function theorem gives a sufficient condition to ensure that there is such a function.

In mathematics, the matrix exponential is a matrix function on square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exponential gives the exponential map between a matrix Lie algebra and the corresponding Lie group.

In linear algebra, linear transformations can be represented by matrices. If is a linear transformation mapping to and is a column vector with entries, then for some matrix , called the transformation matrix of . Note that has rows and columns, whereas the transformation is from to . There are alternative expressions of transformation matrices involving row vectors that are preferred by some authors.

<span class="mw-page-title-main">Pushforward (differential)</span> Linear approximation of smooth maps on tangent spaces

In differential geometry, pushforward is a linear approximation of smooth maps on tangent spaces. Suppose that is a smooth map between smooth manifolds; then the differential of at a point , denoted , is, in some sense, the best linear approximation of near . It can be viewed as a generalization of the total derivative of ordinary calculus. Explicitly, the differential is a linear map from the tangent space of at to the tangent space of at , . Hence it can be used to push tangent vectors on forward to tangent vectors on . The differential of a map is also called, by various authors, the derivative or total derivative of .

In mathematics, a logarithm of a matrix is another matrix such that the matrix exponential of the latter matrix equals the original matrix. It is thus a generalization of the scalar logarithm and in some sense an inverse function of the matrix exponential. Not all matrices have a logarithm and those matrices that do have a logarithm may have more than one logarithm. The study of logarithms of matrices leads to Lie theory since when a matrix has a logarithm then it is in an element of a Lie group and the logarithm is the corresponding element of the vector space of the Lie algebra.

In mathematics and physics, in particular quantum information, the term generalized Pauli matrices refers to families of matrices which generalize the properties of the Pauli matrices. Here, a few classes of such matrices are summarized.

In mathematics, every analytic function can be used for defining a matrix function that maps square matrices with complex entries to square matrices of the same size.

In linear algebra, eigendecomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in this way. When the matrix being factorized is a normal or real symmetric matrix, the decomposition is called "spectral decomposition", derived from the spectral theorem.

<span class="mw-page-title-main">Stokes' theorem</span> Theorem in vector calculus

Stokes' theorem, also known as the Kelvin–Stokes theorem after Lord Kelvin and George Stokes, the fundamental theorem for curls or simply the curl theorem, is a theorem in vector calculus on . Given a vector field, the theorem relates the integral of the curl of the vector field over some surface, to the line integral of the vector field around the boundary of the surface. The classical theorem of Stokes can be stated in one sentence:

<span class="mw-page-title-main">Generalized pencil-of-function method</span>

Generalized pencil-of-function method (GPOF), also known as matrix pencil method, is a signal processing technique for estimating a signal or extracting information with complex exponentials. Being similar to Prony and original pencil-of-function methods, it is generally preferred to those for its robustness and computational efficiency.