Minor (linear algebra)

Last updated

In linear algebra, a minor of a matrix A is the determinant of some smaller square matrix, cut down from A by removing one or more of its rows and columns. Minors obtained by removing just one row and one column from square matrices (first minors) are required for calculating matrix cofactors, which in turn are useful for computing both the determinant and inverse of square matrices. The requirement that the square matrix be smaller than the original matrix is often omitted in the definition.

Contents

Definition and illustration

First minors

If A is a square matrix, then the minor of the entry in the ith row and jth column (also called the (i, j) minor, or a first minor [1] ) is the determinant of the submatrix formed by deleting the ith row and jth column. This number is often denoted Mi,j. The (i, j) cofactor is obtained by multiplying the minor by .

To illustrate these definitions, consider the following 3 by 3 matrix,

To compute the minor M2,3 and the cofactor C2,3, we find the determinant of the above matrix with row 2 and column 3 removed.

So the cofactor of the (2,3) entry is

General definition

Let A be an m×n matrix and k an integer with 0 < km, and kn. A k×kminor of A, also called minor determinant of order k of A or, if m = n, (nk)th minor determinant of A (the word "determinant" is often omitted, and the word "degree" is sometimes used instead of "order") is the determinant of a k×k matrix obtained from A by deleting mk rows and nk columns. Sometimes the term is used to refer to the k×k matrix obtained from A as above (by deleting mk rows and nk columns), but this matrix should be referred to as a (square) submatrix of A, leaving the term "minor" to refer to the determinant of this matrix. For a matrix A as above, there are a total of minors of size k×k. The minor of order zero is often defined to be 1. For a square matrix, the zeroth minor is just the determinant of the matrix. [2] [3]

Let and be ordered sequences (in natural order, as it is always assumed when talking about minors unless otherwise stated) of indexes, call them I and J, respectively. The minor corresponding to these choices of indexes is denoted or or or or or (where the denotes the sequence of indexes I, etc.), depending on the source. Also, there are two types of denotations in use in literature: by the minor associated to ordered sequences of indexes I and J, some authors [4] mean the determinant of the matrix that is formed as above, by taking the elements of the original matrix from the rows whose indexes are in I and columns whose indexes are in J, whereas some other authors mean by a minor associated to I and J the determinant of the matrix formed from the original matrix by deleting the rows in I and columns in J. [2] Which notation is used should always be checked from the source in question. In this article, we use the inclusive definition of choosing the elements from rows of I and columns of J. The exceptional case is the case of the first minor or the (i, j)-minor described above; in that case, the exclusive meaning is standard everywhere in the literature and is used in this article also.

Complement

The complement, Bijk...,pqr..., of a minor, Mijk...,pqr..., of a square matrix, A, is formed by the determinant of the matrix A from which all the rows (ijk...) and columns (pqr...) associated with Mijk...,pqr... have been removed. The complement of the first minor of an element aij is merely that element. [5]

Applications of minors and cofactors

Cofactor expansion of the determinant

The cofactors feature prominently in Laplace's formula for the expansion of determinants, which is a method of computing larger determinants in terms of smaller ones. Given an n×n matrix , the determinant of A, denoted det(A), can be written as the sum of the cofactors of any row or column of the matrix multiplied by the entries that generated them. In other words, defining then the cofactor expansion along the jth column gives:

The cofactor expansion along the ith row gives:

Inverse of a matrix

One can write down the inverse of an invertible matrix by computing its cofactors by using Cramer's rule, as follows. The matrix formed by all of the cofactors of a square matrix A is called the cofactor matrix (also called the matrix of cofactors or, sometimes, comatrix):

Then the inverse of A is the transpose of the cofactor matrix times the reciprocal of the determinant of A:

The transpose of the cofactor matrix is called the adjugate matrix (also called the classical adjoint) of A.

The above formula can be generalized as follows: Let and be ordered sequences (in natural order) of indexes (here A is an n×n matrix). Then [6]

where I′, J′ denote the ordered sequences of indices (the indices are in natural order of magnitude, as above) complementary to I, J, so that every index 1, ..., n appears exactly once in either I or I′, but not in both (similarly for the J and J′) and denotes the determinant of the submatrix of A formed by choosing the rows of the index set I and columns of index set J. Also, . A simple proof can be given using wedge product. Indeed,

where are the basis vectors. Acting by A on both sides, one gets

The sign can be worked out to be , so the sign is determined by the sums of elements in I and J.

Other applications

Given an m×n matrix with real entries (or entries from any other field) and rank r, then there exists at least one non-zero r×r minor, while all larger minors are zero.

We will use the following notation for minors: if A is an m×n matrix, I is a subset of {1,...,m} with k elements, and J is a subset of {1,...,n} with k elements, then we write [A]I,J for the k×k minor of A that corresponds to the rows with index in I and the columns with index in J.

Both the formula for ordinary matrix multiplication and the Cauchy–Binet formula for the determinant of the product of two matrices are special cases of the following general statement about the minors of a product of two matrices. Suppose that A is an m×n matrix, B is an n×p matrix, I is a subset of {1,...,m} with k elements and J is a subset of {1,...,p} with k elements. Then

where the sum extends over all subsets K of {1,...,n} with k elements. This formula is a straightforward extension of the Cauchy–Binet formula.

Multilinear algebra approach

A more systematic, algebraic treatment of minors is given in multilinear algebra, using the wedge product: the k-minors of a matrix are the entries in the kth exterior power map.

If the columns of a matrix are wedged together k at a time, the k×k minors appear as the components of the resulting k-vectors. For example, the 2×2 minors of the matrix

are −13 (from the first two rows), −7 (from the first and last row), and 5 (from the last two rows). Now consider the wedge product

where the two expressions correspond to the two columns of our matrix. Using the properties of the wedge product, namely that it is bilinear and alternating,

and antisymmetric,

we can simplify this expression to

where the coefficients agree with the minors computed earlier.

A remark about different notation

In some books, instead of cofactor the term adjunct is used. [7] Moreover, it is denoted as Aij and defined in the same way as cofactor:

Using this notation the inverse matrix is written this way:

Keep in mind that adjunct is not adjugate or adjoint. In modern terminology, the "adjoint" of a matrix most often refers to the corresponding adjoint operator.

See also

Related Research Articles

In mathematics, the determinant is a scalar value that is a function of the entries of a square matrix. The determinant of a matrix A is commonly denoted det(A), det A, or |A|. Its value characterizes some properties of the matrix and the linear map represented by the matrix. In particular, the determinant is nonzero if and only if the matrix is invertible and the linear map represented by the matrix is an isomorphism. The determinant of a product of matrices is the product of their determinants (which follows directly from the above properties).

In linear algebra, the rank of a matrix A is the dimension of the vector space generated by its columns. This corresponds to the maximal number of linearly independent columns of A. This, in turn, is identical to the dimension of the vector space spanned by its rows. Rank is thus a measure of the "nondegenerateness" of the system of linear equations and linear transformation encoded by A. There are multiple equivalent definitions of rank. A matrix's rank is one of its most fundamental characteristics.

In mathematics, a symmetric matrix with real entries is positive-definite if the real number is positive for every nonzero real column vector where is the transpose of . More generally, a Hermitian matrix is positive-definite if the real number is positive for every nonzero complex column vector where denotes the conjugate transpose of

In linear algebra, Cramer's rule is an explicit formula for the solution of a system of linear equations with as many equations as unknowns, valid whenever the system has a unique solution. It expresses the solution in terms of the determinants of the (square) coefficient matrix and of matrices obtained from it by replacing one column by the column vector of right-sides of the equations. It is named after Gabriel Cramer (1704–1752), who published the rule for an arbitrary number of unknowns in 1750, although Colin Maclaurin also published special cases of the rule in 1748.

<span class="mw-page-title-main">Square matrix</span> Matrix with the same number of rows and columns

In mathematics, a square matrix is a matrix with the same number of rows and columns. An n-by-n matrix is known as a square matrix of order . Any two square matrices of the same order can be added and multiplied.

In mathematics, a Hermitian matrix is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the i-th row and j-th column is equal to the complex conjugate of the element in the j-th row and i-th column, for all indices i and j:

In linear algebra, the adjugate or classical adjoint of a square matrix A is the transpose of its cofactor matrix and is denoted by adj(A). It is also occasionally known as adjunct matrix, or "adjoint", though the latter term today normally refers to a different concept, the adjoint operator which for a matrix is the conjugate transpose.

In linear algebra, an n-by-n square matrix A is called invertible, if there exists an n-by-n square matrix B such that

<span class="mw-page-title-main">Exterior algebra</span> Algebra of exterior/ wedge products

In mathematics, the exterior algebra of a vector space V is a graded associative algebra

In mathematics, particularly in matrix theory, a permutation matrix is a square binary matrix that has exactly one entry of 1 in each row and each column and 0s elsewhere. Each such matrix, say P, represents a permutation of m elements and, when used to multiply another matrix, say A, results in permuting the rows or columns of the matrix A.

In linear algebra, a Vandermonde matrix, named after Alexandre-Théophile Vandermonde, is a matrix with the terms of a geometric progression in each row: an matrix

In linear algebra, a Hankel matrix, named after Hermann Hankel, is a square matrix in which each ascending skew-diagonal from left to right is constant, e.g.:

In mathematics, a block matrix or a partitioned matrix is a matrix that is interpreted as having been broken into sections called blocks or submatrices. Intuitively, a matrix interpreted as a block matrix can be visualized as the original matrix with a collection of horizontal and vertical lines, which break it up, or partition it, into a collection of smaller matrices. Any matrix may be interpreted as a block matrix in one or more ways, with each interpretation defined by how its rows and columns are partitioned.

In mathematics, the determinant of an m×m skew-symmetric matrix can always be written as the square of a polynomial in the matrix entries, a polynomial with integer coefficients that only depends on m. When m is odd, the polynomial is zero. When m=2n is even, it is a nonzero polynomial of degree n, and is unique up to multiplication by ±1. The convention on skew-symmetric tridiagonal matrices, given below in the examples, then determines one specific polynomial, called the Pfaffian polynomial. The value of this polynomial, when applied to the entries of a skew-symmetric matrix, is called the Pfaffian of that matrix. The term Pfaffian was introduced by Cayley (1852) who indirectly named them after Johann Friedrich Pfaff.

In linear algebra, a circulant matrix is a square matrix in which all row vectors are composed of the same elements and each row vector is rotated one element to the right relative to the preceding row vector. It is a particular kind of Toeplitz matrix.

In mathematics, Dodgson condensation or method of contractants is a method of computing the determinants of square matrices. It is named for its inventor, Charles Lutwidge Dodgson (better known by his pseudonym, as Lewis Carroll, the popular author), who discovered it in 1866. The method in the case of an n × n matrix is to construct an (n − 1) × (n − 1) matrix, an (n − 2) × (n − 2), and so on, finishing with a 1 × 1 matrix, which has one entry, the determinant of the original matrix.

In mathematics, a totally positive matrix is a square matrix in which all the minors are positive: that is, the determinant of every square submatrix is a positive number. A totally positive matrix has all entries positive, so it is also a positive matrix; and it has all principal minors positive. A symmetric totally positive matrix is therefore also positive-definite. A totally non-negative matrix is defined similarly, except that all the minors must be non-negative. Some authors use "totally positive" to include all totally non-negative matrices.

In linear algebra, the Laplace expansion, named after Pierre-Simon Laplace, also called cofactor expansion, is an expression of the determinant of an n × n matrix B as a weighted sum of minors, which are the determinants of some (n − 1) × (n − 1) submatrices of B. Specifically, for every i, the Laplace expansion along the ith row is the equality

<span class="mw-page-title-main">Matrix (mathematics)</span> Array of numbers

In mathematics, a matrix is a rectangular array or table of numbers, symbols, or expressions, arranged in rows and columns, which is used to represent a mathematical object or a property of such an object.

In linear algebra, a branch of mathematics, a (multiplicative) compound matrix is a matrix whose entries are all minors, of a given size, of another matrix. Compound matrices are closely related to exterior algebras, and their computation appears in a wide array of problems, such as in the analysis of nonlinear time-varying dynamical systems and generalizations of positive systems, cooperative systems and contracting systems.

References

  1. Burnside, William Snow & Panton, Arthur William (1886) Theory of Equations: with an Introduction to the Theory of Binary Algebraic Form .
  2. 1 2 Elementary Matrix Algebra (Third edition), Franz E. Hohn, The Macmillan Company, 1973, ISBN   978-0-02-355950-1
  3. 1 2 3 "Minor". Encyclopedia of Mathematics.
  4. Linear Algebra and Geometry, Igor R. Shafarevich, Alexey O. Remizov, Springer-Verlag Berlin Heidelberg, 2013, ISBN   978-3-642-30993-9
  5. Bertha Jeffreys, Methods of Mathematical Physics, p.135, Cambridge University Press, 1999 ISBN   0-521-66402-0.
  6. Viktor Vasil_evich Prasolov (13 June 1994). Problems and Theorems in Linear Algebra. American Mathematical Soc. pp. 15–. ISBN   978-0-8218-0236-6.
  7. Felix Gantmacher, Theory of matrices (1st ed., original language is Russian), Moscow: State Publishing House of technical and theoretical literature, 1953, p.491,