Generalized inverse

Last updated

In mathematics, and in particular, algebra, a generalized inverse (or, g-inverse) of an element x is an element y that has some properties of an inverse element but not necessarily all of them. The purpose of constructing a generalized inverse of a matrix is to obtain a matrix that can serve as an inverse in some sense for a wider class of matrices than invertible matrices. Generalized inverses can be defined in any mathematical structure that involves associative multiplication, that is, in a semigroup. This article describes generalized inverses of a matrix .

Contents

A matrix is a generalized inverse of a matrix if [1] [2] [3] A generalized inverse exists for an arbitrary matrix, and when a matrix has a regular inverse, this inverse is its unique generalized inverse. [1]

Motivation

Consider the linear system

where is an matrix and the column space of . If is nonsingular (which implies ) then will be the solution of the system. Note that, if is nonsingular, then

Now suppose is rectangular (), or square and singular. Then we need a right candidate of order such that for all

[4]

That is, is a solution of the linear system . Equivalently, we need a matrix of order such that

Hence we can define the generalized inverse as follows: Given an matrix , an matrix is said to be a generalized inverse of if [1] [2] [3] The matrix has been termed a regular inverse of by some authors. [5]

Types

Important types of generalized inverse include:

Some generalized inverses are defined and classified based on the Penrose conditions:

where denotes conjugate transpose. If satisfies the first condition, then it is a generalized inverse of . If it satisfies the first two conditions, then it is a reflexive generalized inverse of . If it satisfies all four conditions, then it is the pseudoinverse of , which is denoted by and also known as the Moore–Penrose inverse, after the pioneering works by E. H. Moore and Roger Penrose. [2] [7] [8] [9] [10] [11] It is convenient to define an -inverse of as an inverse that satisfies the subset of the Penrose conditions listed above. Relations, such as , can be established between these different classes of -inverses. [1]

When is non-singular, any generalized inverse and is therefore unique. For a singular , some generalised inverses, such as the Drazin inverse and the Moore–Penrose inverse, are unique, while others are not necessarily uniquely defined.

Examples

Reflexive generalized inverse

Let

Since , is singular and has no regular inverse. However, and satisfy Penrose conditions (1) and (2), but not (3) or (4). Hence, is a reflexive generalized inverse of .

One-sided inverse

Let

Since is not square, has no regular inverse. However, is a right inverse of . The matrix has no left inverse.

Inverse of other semigroups (or rings)

The element b is a generalized inverse of an element a if and only if , in any semigroup (or ring, since the multiplication function in any ring is a semigroup).

The generalized inverses of the element 3 in the ring are 3, 7, and 11, since in the ring :

The generalized inverses of the element 4 in the ring are 1, 4, 7, and 10, since in the ring :

If an element a in a semigroup (or ring) has an inverse, the inverse must be the only generalized inverse of this element, like the elements 1, 5, 7, and 11 in the ring .

In the ring , any element is a generalized inverse of 0, however, 2 has no generalized inverse, since there is no b in such that .

Construction

The following characterizations are easy to verify:

Uses

Any generalized inverse can be used to determine whether a system of linear equations has any solutions, and if so to give all of them. If any solutions exist for the n × m linear system

,

with vector of unknowns and vector of constants, all solutions are given by

,

parametric on the arbitrary vector , where is any generalized inverse of . Solutions exist if and only if is a solution, that is, if and only if . If A has full column rank, the bracketed expression in this equation is the zero matrix and so the solution is unique. [12]

Generalized inverses of matrices

The generalized inverses of matrices can be characterized as follows. Let , and

be its singular-value decomposition. Then for any generalized inverse , there exist [1] matrices , , and such that

Conversely, any choice of , , and for matrix of this form is a generalized inverse of . [1] The -inverses are exactly those for which , the -inverses are exactly those for which , and the -inverses are exactly those for which . In particular, the pseudoinverse is given by :

Transformation consistency properties

In practical applications it is necessary to identify the class of matrix transformations that must be preserved by a generalized inverse. For example, the Moore–Penrose inverse, satisfies the following definition of consistency with respect to transformations involving unitary matrices U and V:

.

The Drazin inverse, satisfies the following definition of consistency with respect to similarity transformations involving a nonsingular matrix S:

.

The unit-consistent (UC) inverse, [13] satisfies the following definition of consistency with respect to transformations involving nonsingular diagonal matrices D and E:

.

The fact that the Moore–Penrose inverse provides consistency with respect to rotations (which are orthonormal transformations) explains its widespread use in physics and other applications in which Euclidean distances must be preserved. The UC inverse, by contrast, is applicable when system behavior is expected to be invariant with respect to the choice of units on different state variables, e.g., miles versus kilometers.

See also

Citations

Sources

Textbook

Publication

Related Research Articles

<span class="mw-page-title-main">Pauli matrices</span> Matrices important in quantum mechanics and the study of spin

In mathematical physics and mathematics, the Pauli matrices are a set of three 2 × 2 complex matrices that are Hermitian, involutory and unitary. Usually indicated by the Greek letter sigma, they are occasionally denoted by tau when used in connection with isospin symmetries.

In mathematics, a symmetric matrix with real entries is positive-definite if the real number is positive for every nonzero real column vector where is the transpose of . More generally, a Hermitian matrix is positive-definite if the real number is positive for every nonzero complex column vector where denotes the conjugate transpose of

In mathematics, the concept of an inverse element generalises the concepts of opposite and reciprocal of numbers.

In linear algebra, an orthogonal matrix, or orthonormal matrix, is a real square matrix whose columns and rows are orthonormal vectors.

<span class="mw-page-title-main">Orthogonal group</span> Type of group in mathematics

In mathematics, the orthogonal group in dimension , denoted , is the group of distance-preserving transformations of a Euclidean space of dimension that preserve a fixed point, where the group operation is given by composing transformations. The orthogonal group is sometimes called the general orthogonal group, by analogy with the general linear group. Equivalently, it is the group of orthogonal matrices, where the group operation is given by matrix multiplication. The orthogonal group is an algebraic group and a Lie group. It is compact.

In mechanics and geometry, the 3D rotation group, often denoted SO(3), is the group of all rotations about the origin of three-dimensional Euclidean space under the operation of composition.

In linear algebra, an n-by-n square matrix A is called invertible if there exists an n-by-n square matrix B such that

In mathematics, the conjugate transpose, also known as the Hermitian transpose, of an complex matrix is an matrix obtained by transposing and applying complex conjugation to each entry. There are several notations, such as or , , or .

In mathematics, and in particular linear algebra, the Moore–Penrose inverse of a matrix is the most widely known generalization of the inverse matrix. It was independently described by E. H. Moore in 1920, Arne Bjerhammar in 1951, and Roger Penrose in 1955. Earlier, Erik Ivar Fredholm had introduced the concept of a pseudoinverse of integral operators in 1903. When referring to a matrix, the term pseudoinverse, without further specification, is often used to indicate the Moore–Penrose inverse. The term generalized inverse is sometimes used as a synonym for pseudoinverse.

In linear algebra and the theory of matrices, the Schur complement of a block matrix is defined as follows.

In multivariable calculus, the implicit function theorem is a tool that allows relations to be converted to functions of several real variables. It does so by representing the relation as the graph of a function. There may not be a single function whose graph can represent the entire relation, but there may be such a function on a restriction of the domain of the relation. The implicit function theorem gives a sufficient condition to ensure that there is such a function.

In mathematics, the matrix exponential is a matrix function on square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exponential gives the exponential map between a matrix Lie algebra and the corresponding Lie group.

In linear algebra, linear transformations can be represented by matrices. If is a linear transformation mapping to and is a column vector with entries, then

<span class="mw-page-title-main">Pushforward (differential)</span> Linear approximation of smooth maps on tangent spaces

In differential geometry, pushforward is a linear approximation of smooth maps on tangent spaces. Suppose that is a smooth map between smooth manifolds; then the differential of at a point , denoted , is, in some sense, the best linear approximation of near . It can be viewed as a generalization of the total derivative of ordinary calculus. Explicitly, the differential is a linear map from the tangent space of at to the tangent space of at , . Hence it can be used to push tangent vectors on forward to tangent vectors on . The differential of a map is also called, by various authors, the derivative or total derivative of .

In the mathematical discipline of matrix theory, a Jordan matrix, named after Camille Jordan, is a block diagonal matrix over a ring R, where each block along the diagonal, called a Jordan block, has the following form:

In mathematics and physics, in particular quantum information, the term generalized Pauli matrices refers to families of matrices which generalize the properties of the Pauli matrices. Here, a few classes of such matrices are summarized.

In mathematics, every analytic function can be used for defining a matrix function that maps square matrices with complex entries to square matrices of the same size.

In linear algebra, eigendecomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in this way. When the matrix being factorized is a normal or real symmetric matrix, the decomposition is called "spectral decomposition", derived from the spectral theorem.

<span class="mw-page-title-main">Stokes' theorem</span> Theorem in vector calculus

Stokes' theorem, also known as the Kelvin–Stokes theorem after Lord Kelvin and George Stokes, the fundamental theorem for curls or simply the curl theorem, is a theorem in vector calculus on . Given a vector field, the theorem relates the integral of the curl of the vector field over some surface, to the line integral of the vector field around the boundary of the surface. The classical theorem of Stokes can be stated in one sentence: The line integral of a vector field over a loop is equal to its curl through the enclosed surface. It is illustrated in the figure, where the direction of positive circulation of the bounding contour ∂Σ, and the direction n of positive flux through the surface Σ, are related by a right-hand-rule. For the right hand the fingers circulate along ∂Σ and the thumb is directed along n.

<span class="mw-page-title-main">Generalized pencil-of-function method</span>

Generalized pencil-of-function method (GPOF), also known as matrix pencil method, is a signal processing technique for estimating a signal or extracting information with complex exponentials. Being similar to Prony and original pencil-of-function methods, it is generally preferred to those for its robustness and computational efficiency.