Partial inverse of a matrix

Last updated

In linear algebra and statistics, the partial inverse of a matrix is an operation related to Gaussian elimination which has applications in numerical analysis and statistics. It is also known by various authors as the principal pivot transform, or as the sweep, gyration, or exchange operator.

Given an matrix over a vector space partitioned into blocks:

If is invertible, then the partial inverse of around the pivot block is created by inverting , putting the Schur complement in place of , and adjusting the off-diagonal elements accordingly: [1]

Conceptually, partial inversion corresponds to a rotation [2] of the graph of the matrix , such that, for conformally-partitioned column matrices and : [1]

As defined this way, this operator is its own inverse: , and if the pivot block is chosen to be the entire matrix, then the transform simply gives the matrix inverse . Note that some authors define a related operation (under one of the other names) which is not an inverse per se; particularly, one common definition instead has .

The transform is often presented as a pivot around a single non-zero element , in which case one has

Partial inverses obey a number of nice properties: [3]

Use of the partial inverse in numerical analysis is due to the fact that there is some flexibility in the choices of pivots, allowing for non-invertible elements to be avoided, and because the operation of rotation (of the graph of the pivoted matrix) has better numerical stability than the shearing operation which is implicitly performed by Gaussian elimination. [2] Use in statistics is due to the fact that the resulting matrix nicely decomposes into blocks which have useful meanings in the context of linear regression. [3]

Related Research Articles

In linear algebra, the determinant is a scalar value that can be computed from the elements of a square matrix and encodes certain properties of the linear transformation described by the matrix. The determinant of a matrix A is denoted det(A), det A, or |A|. Geometrically, it can be viewed as the volume scaling factor of the linear transformation described by the matrix. This is also the signed volume of the n-dimensional parallelepiped spanned by the column or row vectors of the matrix. The determinant is positive or negative according to whether the linear mapping preserves or reverses the orientation of n-space.

Diffeomorphism Isomorphism of smooth manifolds; a smooth bijection with a smooth inverse

In mathematics, a diffeomorphism is an isomorphism of smooth manifolds. It is an invertible function that maps one differentiable manifold to another such that both the function and its inverse are smooth.

Inverse function function that "reverses" another function: if the function f applied to an input x gives a result of y, then applying its inverse function g to y gives the result x, and vice versa. i.e., f(x) = y if and only if g(y) = x

In mathematics, an inverse function is a function that "reverses" another function: if the function f applied to an input x gives a result of y, then applying its inverse function g to y gives the result x, and vice versa, i.e., f(x) = y if and only if g(y) = x.

In mathematics, when X is a finite set with at least two elements, the permutations of X fall into two classes of equal size: the even permutations and the odd permutations. If any total ordering of X is fixed, the parity of a permutation of X can be defined as the parity of the number of inversions for σ, i.e., of pairs of elements x, y of X such that x < y and σ(x) > σ(y).

Matrix multiplication Mathematical operation in linear algebra

In mathematics, matrix multiplication is a binary operation that produces a matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the number of rows in the second matrix. The result matrix, known as the matrix product, has the number of rows of the first and the number of columns of the second matrix.

In linear algebra, the adjugate, classical adjoint, or adjunct of a square matrix is the transpose of its cofactor matrix.

In linear algebra, an n-by-n square matrix A is called invertible if there exists an n-by-n square matrix B such that

In differential geometry, the Ricci curvature tensor, named after Gregorio Ricci-Curbastro, is a geometric object which is determined by a choice of Riemannian or pseudo-Riemannian metric on a manifold. It can be considered, broadly, as a measure of the degree to which the geometry of a given metric differs from that of ordinary Euclidean space or pseudo-Euclidean space.

In geometry, inversive geometry is the study of those properties of figures that are preserved by a generalization of a type of transformation of the Euclidean plane, called inversion. These transformations preserve angles and map generalized circles into generalized circles, where a generalized circle means either a circle or a line. Many difficult problems in geometry become much more tractable when an inversion is applied.

In mathematics, the affine group or general affine group of any affine space over a field K is the group of all invertible affine transformations from the space into itself.

In linear algebra, a QR decomposition, also known as a QR factorization or QU factorization is a decomposition of a matrix A into a product A = QR of an orthogonal matrix Q and an upper triangular matrix R. QR decomposition is often used to solve the linear least squares problem and is the basis for a particular eigenvalue algorithm, the QR algorithm.

In geometry and complex analysis, a Möbius transformation of the complex plane is a rational function of the form

In mathematics, and in particular linear algebra, the Moore–Penrose inverse of a matrix is the most widely known generalization of the inverse matrix. It was independently described by E. H. Moore in 1920, Arne Bjerhammar in 1951, and Roger Penrose in 1955. Earlier, Erik Ivar Fredholm had introduced the concept of a pseudoinverse of integral operators in 1903. When referring to a matrix, the term pseudoinverse, without further specification, is often used to indicate the Moore–Penrose inverse. The term generalized inverse is sometimes used as a synonym for pseudoinverse.

In mathematics, the determinant of a skew-symmetric matrix can always be written as the square of a polynomial in the matrix entries, a polynomial with integer coefficients that only depend on the size of the matrix. The value of this polynomial, when applied to the coefficients of a skew-symmetric matrix, is called the Pfaffian of that matrix. The term Pfaffian was introduced by Cayley (1852) who indirectly named them after Johann Friedrich Pfaff. The Pfaffian is nonvanishing only for 2n × 2n skew-symmetric matrices, in which case it is a polynomial of degree n.

In mathematics, more specifically in multivariable calculus, the implicit function theorem is a tool that allows relations to be converted to functions of several real variables. It does so by representing the relation as the graph of a function. There may not be a single function whose graph can represent the entire relation, but there may be such a function on a restriction of the domain of the relation. The implicit function theorem gives a sufficient condition to ensure that there is such a function.

In mathematics, the square root of a matrix extends the notion of square root from numbers to matrices. A matrix B is said to be a square root of A if the matrix product BB is equal to A.

In mathematics, a logarithm of a matrix is another matrix such that the matrix exponential of the latter matrix equals the original matrix. It is thus a generalization of the scalar logarithm and in some sense an inverse function of the matrix exponential. Not all matrices have a logarithm and those matrices that do have a logarithm may have more than one logarithm. The study of logarithms of matrices leads to Lie theory since when a matrix has a logarithm then it is in a Lie group and the logarithm is the corresponding element of the vector space of the Lie algebra.

In mathematics, the converse relation, or transpose, of a binary relation is the relation that occurs when the order of the elements is switched in the relation. For example, the converse of the relation 'child of' is the relation 'parent of'. In formal terms, if X and Y are sets and LX × Y is a relation from X to Y, then LT is the relation defined so that y LTx if and only if x L y. In set-builder notation, LT = {(y, x) ∈ Y × X | ∈ L}.

In numerical analysis and linear algebra, lower–upper (LU) decomposition or factorization factors a matrix as the product of a lower triangular matrix and an upper triangular matrix. The product sometimes includes a permutation matrix as well. LU decomposition can be viewed as the matrix form of Gaussian elimination. Computers usually solve square systems of linear equations using LU decomposition, and it is also a key step when inverting a matrix or computing the determinant of a matrix. LU decomposition was introduced by Polish mathematician Tadeusz Banachiewicz in 1938.

The purpose of this page is to provide supplementary materials for the ordinary least squares article, reducing the load of the main article with mathematics and improving its accessibility, while at the same time retaining the completeness of exposition.

References

  1. 1 2 Tsatsomeros, M. J. (2000). Principal pivot transforms: properties and applications. Linear Algebra and its Applications, 307 (1-3), 151–165.
  2. 1 2 Sweeping a matrix rotates its graph,
  3. 1 2 Exceedingly Simple Principal Pivot Transforms