Adjugate matrix

Last updated

In linear algebra, the adjugate or classical adjoint of a square matrix A, adj(A), is the transpose of its cofactor matrix. [1] [2] It is occasionally known as adjunct matrix, [3] [4] or "adjoint", [5] though that normally refers to a different concept, the adjoint operator which for a matrix is the conjugate transpose.

Contents

The product of a matrix with its adjugate gives a diagonal matrix (entries not on the main diagonal are zero) whose diagonal entries are the determinant of the original matrix:

where I is the identity matrix of the same size as A. Consequently, the multiplicative inverse of an invertible matrix can be found by dividing its adjugate by its determinant.

Definition

The adjugate of A is the transpose of the cofactor matrix C of A,

In more detail, suppose R is a unital+ commutative ring and A is an n×n matrix with entries from R. The (i, j)- minor of A, denoted Mij, is the determinant of the (n  1)×(n  1) matrix that results from deleting row i and column j of A. The cofactor matrix of A is the n×n matrix C whose (i, j) entry is the (i, j) cofactor of A, which is the (i, j)-minor times a sign factor:

The adjugate of A is the transpose of C, that is, the n×n matrix whose (i, j) entry is the (j,i) cofactor of A,

Important consequence

The adjugate is defined so that the product of A with its adjugate yields a diagonal matrix whose diagonal entries are the determinant det(A). That is,

where I is the n×n identity matrix. This is a consequence of the Laplace expansion of the determinant.

The above formula implies one of the fundamental results in matrix algebra, that A is invertible if and only if det(A) is an invertible element of R. When this holds, the equation above yields

Examples

1×1 generic matrix

Since the determinant of a 0 × 0 matrix is 1, the adjugate of any 1×1 matrix (complex scalar) is . Observe that

2×2 generic matrix

The adjugate of the 2×2 matrix

is

By direct computation,

In this case, it is also true that det(adj(A)) = det(A) and hence that adj(adj(A)) = A.

3×3 generic matrix

Consider a 3×3 matrix

Its cofactor matrix is

where

Its adjugate is the transpose of its cofactor matrix,

3×3 numeric matrix

As a specific example, we have

It is easy to check the adjugate is the inverse times the determinant, −6.

The −1 in the second row, third column of the adjugate was computed as follows. The (2,3) entry of the adjugate is the (3,2) cofactor of A. This cofactor is computed using the submatrix obtained by deleting the third row and second column of the original matrix A,

The (3,2) cofactor is a sign times the determinant of this submatrix:

and this is the (2,3) entry of the adjugate.

Properties

For any n×n matrix A, elementary computations show that adjugates have the following properties:

Over the complex numbers,

Suppose that B is another n×n matrix. Then

This can be proved in three ways. One way, valid for any commutative ring, is a direct computation using the Cauchy–Binet formula. The second way, valid for the real or complex numbers, is to first observe that for invertible matrices A and B,

Because every non-invertible matrix is the limit of invertible matrices, continuity of the adjugate then implies that the formula remains true when one of A or B is not invertible.

A corollary of the previous formula is that, for any non-negative integer k,

If A is invertible, then the above formula also holds for negative k.

From the identity

we deduce

Suppose that A commutes with B. Multiplying the identity AB = BA on the left and right by adj(A) proves that

If A is invertible, this implies that adj(A) also commutes with B. Over the real or complex numbers, continuity implies that adj(A) commutes with B even when A is not invertible.

Finally, there is a more general proof than the second proof, which only requires that an n×n matrix has entries over a field with at least 2n +1 elements (e.g. a 5×5 matrix over the integers modulo 11). det(A+tI) is a polynomial in t with degree at most n, so it has at most n roots. Note that the ijth entry of adj((A+tI)(B)) is a polynomial of at most order n, and likewise for adj(A+tI)adj(B). These two polynomials at the ijth entry agree on at least n +1 points, as we have at least n +1 elements of the field where A+tI is invertible, and we have proven the identity for invertible matrices. Polynomials of degree n which agree on n +1 points must be identical (subtract them from each other and you have n +1 roots for a polynomial of degree at most n – a contradiction unless their difference is identically zero). As the two polynomials are identical, they take the same value for every value of t. Thus, they take the same value when t = 0.

Using the above properties and other elementary computations, it is straightforward to show that if A has one of the following properties, then adjA does as well:

If A is skew-symmetric, then adj(A) is skew-symmetric for even n and symmetric for odd n. Similarly, if A is skew-Hermitian, then adj(A) is skew-Hermitian for even n and Hermitian for odd n.

If A is invertible, then, as noted above, there is a formula for adj(A) in terms of the determinant and inverse of A. When A is not invertible, the adjugate satisfies different but closely related formulas.

Column substitution and Cramer's rule

Partition A into column vectors:

Let b be a column vector of size n. Fix 1in and consider the matrix formed by replacing column i of A by b:

Laplace expand the determinant of this matrix along column i. The result is entry i of the product adj(A)b. Collecting these determinants for the different possible i yields an equality of column vectors

This formula has the following concrete consequence. Consider the linear system of equations

Assume that A is non-singular. Multiplying this system on the left by adj(A) and dividing by the determinant yields

Applying the previous formula to this situation yields Cramer's rule,

where xi is the ith entry of x.

Characteristic polynomial

Let the characteristic polynomial of A be

The first divided difference of p is a symmetric polynomial of degree n 1,

Multiply sIA by its adjugate. Since p(A) = 0 by the Cayley–Hamilton theorem, some elementary manipulations reveal

In particular, the resolvent of A is defined to be

and by the above formula, this is equal to

Jacobi's formula

The adjugate also appears in Jacobi's formula for the derivative of the determinant. If A(t) is continuously differentiable, then

It follows that the total derivative of the determinant is the transpose of the adjugate:

Cayley–Hamilton formula

Let pA(t) be the characteristic polynomial of A. The Cayley–Hamilton theorem states that

Separating the constant term and multiplying the equation by adj(A) gives an expression for the adjugate that depends only on A and the coefficients of pA(t). These coefficients can be explicitly represented in terms of traces of powers of A using complete exponential Bell polynomials. The resulting formula is

where n is the dimension of A, and the sum is taken over s and all sequences of kl ≥ 0 satisfying the linear Diophantine equation

For the 2×2 case, this gives

For the 3×3 case, this gives

For the 4×4 case, this gives

The same formula follows directly from the terminating step of the Faddeev–LeVerrier algorithm, which efficiently determines the characteristic polynomial of A.

In general, adjugate matrix of arbitrary dimension N matrix can be computed by Einstein's convention.

Relation to exterior algebras

The adjugate can be viewed in abstract terms using exterior algebras. Let V be an n-dimensional vector space. The exterior product defines a bilinear pairing

Abstractly, is isomorphic to R, and under any such isomorphism the exterior product is a perfect pairing. Therefore, it yields an isomorphism

Explicitly, this pairing sends vV to , where

Suppose that T : VV is a linear transformation. Pullback by the (n 1)st exterior power of T induces a morphism of Hom spaces. The adjugate of T is the composite

If V = Rn is endowed with its canonical basis e1, …, en, and if the matrix of T in this basis is A, then the adjugate of T is the adjugate of A. To see why, give the basis

Fix a basis vector ei of Rn. The image of ei under is determined by where it sends basis vectors:

On basis vectors, the (n 1)st exterior power of T is

Each of these terms maps to zero under except the k = i term. Therefore, the pullback of is the linear transformation for which

that is, it equals

Applying the inverse of shows that the adjugate of T is the linear transformation for which

Consequently, its matrix representation is the adjugate of A.

If V is endowed with an inner product and a volume form, then the map φ can be decomposed further. In this case, φ can be understood as the composite of the Hodge star operator and dualization. Specifically, if ω is the volume form, then it, together with the inner product, determines an isomorphism

This induces an isomorphism

A vector v in Rn corresponds to the linear functional

By the definition of the Hodge star operator, this linear functional is dual to *v. That is, ωφ equals v ↦ *v.

Higher adjugates

Let A be an n×n matrix, and fix r 0. The rth higher adjugate of A is an matrix, denoted adjrA, whose entries are indexed by size r subsets I and J of {1, ..., m}[ citation needed ]. Let Ic and Jc denote the complements of I and J, respectively. Also let denote the submatrix of A containing those rows and columns whose indices are in Ic and Jc, respectively. Then the (I, J) entry of adjrA is

where σ(I) and σ(J) are the sum of the elements of I and J, respectively.

Basic properties of higher adjugates include [ citation needed ]:

Higher adjugates may be defined in abstract algebraic terms in a similar fashion to the usual adjugate, substituting and for and , respectively.

Iterated adjugates

Iteratively taking the adjugate of an invertible matrix Ak times yields

For example,

See also

Related Research Articles

In mathematics, the determinant is a scalar-valued function of the entries of a square matrix. The determinant of a matrix A is commonly denoted det(A), det A, or |A|. Its value characterizes some properties of the matrix and the linear map represented, on a given basis, by the matrix. In particular, the determinant is nonzero if and only if the matrix is invertible and the corresponding linear map is an isomorphism.

<span class="mw-page-title-main">Pauli matrices</span> Matrices important in quantum mechanics and the study of spin

In mathematical physics and mathematics, the Pauli matrices are a set of three 2 × 2 complex matrices that are traceless, Hermitian, involutory and unitary. Usually indicated by the Greek letter sigma, they are occasionally denoted by tau when used in connection with isospin symmetries.

In linear algebra, the trace of a square matrix A, denoted tr(A), is the sum of the elements on its main diagonal, . It is only defined for a square matrix.

In linear algebra, Cramer's rule is an explicit formula for the solution of a system of linear equations with as many equations as unknowns, valid whenever the system has a unique solution. It expresses the solution in terms of the determinants of the (square) coefficient matrix and of matrices obtained from it by replacing one column by the column vector of right-sides of the equations. It is named after Gabriel Cramer, who published the rule for an arbitrary number of unknowns in 1750, although Colin Maclaurin also published special cases of the rule in 1748, and possibly knew of it as early as 1729.

<span class="mw-page-title-main">Square matrix</span> Matrix with the same number of rows and columns

In mathematics, a square matrix is a matrix with the same number of rows and columns. An n-by-n matrix is known as a square matrix of order . Any two square matrices of the same order can be added and multiplied.

Ray transfer matrix analysis is a mathematical form for performing ray tracing calculations in sufficiently simple problems which can be solved considering only paraxial rays. Each optical element is described by a 2 × 2ray transfer matrix which operates on a vector describing an incoming light ray to calculate the outgoing ray. Multiplication of the successive matrices thus yields a concise ray transfer matrix describing the entire optical system. The same mathematics is also used in accelerator physics to track particles through the magnet installations of a particle accelerator, see electron optics.

<span class="mw-page-title-main">Cayley–Hamilton theorem</span> Every square matrix over a commutative ring satisfies its own characteristic equation

In linear algebra, the Cayley–Hamilton theorem states that every square matrix over a commutative ring satisfies its own characteristic equation.

<span class="mw-page-title-main">Transpose</span> Matrix operation which flips a matrix over its diagonal

In linear algebra, the transpose of a matrix is an operator which flips a matrix over its diagonal; that is, it switches the row and column indices of the matrix A by producing another matrix, often denoted by AT.

In linear algebra, an invertible matrix is a square matrix which has an inverse. In other words, if some other matrix is multiplied by the invertible matrix, the result can be multiplied by an inverse to undo the operation. Invertible matrices are the same size as their inverse.

In mathematics, the conjugate transpose, also known as the Hermitian transpose, of an complex matrix is an matrix obtained by transposing and applying complex conjugation to each entry. There are several notations, such as or , , or .

In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The characteristic polynomial of an endomorphism of a finite-dimensional vector space is the characteristic polynomial of the matrix of that endomorphism over any basis. The characteristic equation, also known as the determinantal equation, is the equation obtained by equating the characteristic polynomial to zero.

In linear algebra, a minor of a matrix A is the determinant of some smaller square matrix generated from A by removing one or more of its rows and columns. Minors obtained by removing just one row and one column from square matrices are required for calculating matrix cofactors, which are useful for computing both the determinant and inverse of square matrices. The requirement that the square matrix be smaller than the original matrix is often omitted in the definition.

<span class="mw-page-title-main">Block matrix</span> Matrix defined using smaller matrices called blocks

In mathematics, a block matrix or a partitioned matrix is a matrix that is interpreted as having been broken into sections called blocks or submatrices.

In mathematics, the determinant of an m-by-m skew-symmetric matrix can always be written as the square of a polynomial in the matrix entries, a polynomial with integer coefficients that only depends on m. When m is odd, the polynomial is zero, and when m is even, it is a nonzero polynomial of degree m/2, and is unique up to multiplication by ±1. The convention on skew-symmetric tridiagonal matrices, given below in the examples, then determines one specific polynomial, called the Pfaffian polynomial. The value of this polynomial, when applied to the entries of a skew-symmetric matrix, is called the Pfaffian of that matrix. The term Pfaffian was introduced by Cayley, who indirectly named them after Johann Friedrich Pfaff.

In mathematics, Dodgson condensation or method of contractants is a method of computing the determinants of square matrices. It is named for its inventor, Charles Lutwidge Dodgson (better known by his pseudonym, as Lewis Carroll, the popular author), who discovered it in 1866. The method in the case of an n × n matrix is to construct an (n − 1) × (n − 1) matrix, an (n − 2) × (n − 2), and so on, finishing with a 1 × 1 matrix, which has one entry, the determinant of the original matrix.

In matrix calculus, Jacobi's formula expresses the derivative of the determinant of a matrix A in terms of the adjugate of A and the derivative of A.

<span class="mw-page-title-main">Matrix (mathematics)</span> Array of numbers

In mathematics, a matrix is a rectangular array or table of numbers, symbols, or expressions, with elements or entries arranged in rows and columns, which is used to represent a mathematical object or property of such an object.

Geometric algebra is an extension of vector algebra, providing additional algebraic structures on vector spaces, with geometric interpretations.

In linear algebra, a branch of mathematics, a (multiplicative) compound matrix is a matrix whose entries are all minors, of a given size, of another matrix. Compound matrices are closely related to exterior algebras, and their computation appears in a wide array of problems, such as in the analysis of nonlinear time-varying dynamical systems and generalizations of positive systems, cooperative systems and contracting systems.

<span class="mw-page-title-main">Faddeev–LeVerrier algorithm</span>

In mathematics, the Faddeev–LeVerrier algorithm is a recursive method to calculate the coefficients of the characteristic polynomial of a square matrix, A, named after Dmitry Konstantinovich Faddeev and Urbain Le Verrier. Calculation of this polynomial yields the eigenvalues of A as its roots; as a matrix polynomial in the matrix A itself, it vanishes by the Cayley–Hamilton theorem. Computing the characteristic polynomial directly from the definition of the determinant is computationally cumbersome insofar as it introduces a new symbolic quantity ; by contrast, the Faddeev-Le Verrier algorithm works directly with coefficients of matrix .

References

  1. Gantmacher, F. R. (1960). The Theory of Matrices. Vol. 1. New York: Chelsea. pp. 76–89. ISBN   0-8218-1376-5.
  2. Strang, Gilbert (1988). "Section 4.4: Applications of determinants" . Linear Algebra and its Applications (3rd ed.). Harcourt Brace Jovanovich. pp.  231–232. ISBN   0-15-551005-3.
  3. Claeyssen, J.C.R. (1990). "On predicting the response of non-conservative linear vibrating systems by using dynamical matrix solutions". Journal of Sound and Vibration. 140 (1): 73–84. Bibcode:1990JSV...140...73C. doi:10.1016/0022-460X(90)90907-H.
  4. Chen, W.; Chen, W.; Chen, Y.J. (2004). "A characteristic matrix approach for analyzing resonant ring lattice devices". IEEE Photonics Technology Letters. 16 (2): 458–460. Bibcode:2004IPTL...16..458C. doi:10.1109/LPT.2003.823104.
  5. Householder, Alston S. (2006). The Theory of Matrices in Numerical Analysis. Dover Books on Mathematics. pp. 166–168. ISBN   0-486-44972-6.

Bibliography