Weyr canonical form

Last updated
The image shows an example of a general Weyr matrix consisting of two blocks each of which is a basic Weyr matrix. The basic Weyr matrix in the top-left corner has the structure (4,2,1) and the other one has the structure (2,2,1,1). WeyrMatrixExample.jpg
The image shows an example of a general Weyr matrix consisting of two blocks each of which is a basic Weyr matrix. The basic Weyr matrix in the top-left corner has the structure (4,2,1) and the other one has the structure (2,2,1,1).

In mathematics, in linear algebra, a Weyr canonical form (or, Weyr form or Weyr matrix) is a square matrix which (in some sense) induces "nice" properties with matrices it commutes with. It also has a particularly simple structure and the conditions for possessing a Weyr form are fairly weak, making it a suitable tool for studying classes of commuting matrices. A square matrix is said to be in the Weyr canonical form if the matrix has the structure defining the Weyr canonical form. The Weyr form was discovered by the Czech mathematician Eduard Weyr in 1885. [1] [2] [3] The Weyr form did not become popular among mathematicians and it was overshadowed by the closely related, but distinct, canonical form known by the name Jordan canonical form. [3] The Weyr form has been rediscovered several times since Weyr’s original discovery in 1885. [4] This form has been variously called as modified Jordan form,reordered Jordan form,second Jordan form, and H-form. [4] The current terminology is credited to Shapiro who introduced it in a paper published in the American Mathematical Monthly in 1999. [4] [5]

Contents

Recently several applications have been found for the Weyr matrix. Of particular interest is an application of the Weyr matrix in the study of phylogenetic invariants in biomathematics.

Definitions

Basic Weyr matrix

Definition

A basic Weyr matrix with eigenvalue is an matrix of the following form: There is an integer partition

of with

such that, when is viewed as an block matrix , where the block is an matrix, the following three features are present:

  1. The main diagonal blocks are the scalar matrices for .
  2. The first superdiagonal blocks are full column rank matrices in reduced row-echelon form (that is, an identity matrix followed by zero rows) for .
  3. All other blocks of W are zero (that is, when ).

In this case, we say that has Weyr structure .

Example

The following is an example of a basic Weyr matrix.

BasicWeyrMatrix.jpg

In this matrix, and . So has the Weyr structure . Also,

and

General Weyr matrix

Definition

Let be a square matrix and let be the distinct eigenvalues of . We say that is in Weyr form (or is a Weyr matrix) if has the following form:

where is a basic Weyr matrix with eigenvalue for .

Example

The following image shows an example of a general Weyr matrix consisting of three basic Weyr matrix blocks. The basic Weyr matrix in the top-left corner has the structure (4,2,1) with eigenvalue 4, the middle block has structure (2,2,1,1) with eigenvalue -3 and the one in the lower-right corner has the structure (3, 2) with eigenvalue 0.

WeyrMatrixExample02.jpg

Relation between Weyr and Jordan forms

The Weyr canonical form is related to the Jordan form by a simple permutation for each Weyr basic block as follows: The first index of each Weyr subblock forms the largest Jordan chain. After crossing out these rows and columns, the first index of each new subblock forms the second largest Jordan chain, and so forth. [6]

The Weyr form is canonical

That the Weyr form is a canonical form of a matrix is a consequence of the following result: [3] Each square matrix over an algebraically closed field is similar to a Weyr matrix which is unique up to permutation of its basic blocks. The matrix is called the Weyr (canonical) form of .

Computation of the Weyr canonical form

Reduction to the nilpotent case

Let be a square matrix of order over an algebraically closed field and let the distinct eigenvalues of be . The Jordan–Chevalley decomposition theorem states that is similar to a block diagonal matrix of the form

where is a diagonal matrix, is a nilpotent matrix, and , justifying the reduction of into subblocks . So the problem of reducing to the Weyr form reduces to the problem of reducing the nilpotent matrices to the Weyr form. This is leads to the generalized eigenspace decomposition theorem.

Reduction of a nilpotent matrix to the Weyr form

Given a nilpotent square matrix of order over an algebraically closed field , the following algorithm produces an invertible matrix and a Weyr matrix such that .

Step 1

Let

Step 2

  1. Compute a basis for the null space of .
  2. Extend the basis for the null space of to a basis for the -dimensional vector space .
  3. Form the matrix consisting of these basis vectors.
  4. Compute . is a square matrix of size nullity .

Step 3

If is nonzero, repeat Step 2 on .

  1. Compute a basis for the null space of .
  2. Extend the basis for the null space of to a basis for the vector space having dimension nullity .
  3. Form the matrix consisting of these basis vectors.
  4. Compute . is a square matrix of size nullity nullity.

Step 4

Continue the processes of Steps 1 and 2 to obtain increasingly smaller square matrices and associated invertible matrices until the first zero matrix is obtained.

Step 5

The Weyr structure of is where = nullity.

Step 6

  1. Compute the matrix (here the 's are appropriately sized identity matrices).
  2. Compute . is a matrix of the following form:
.

Step 7

Use elementary row operations to find an invertible matrix of appropriate size such that the product is a matrix of the form .

Step 8

Set diag and compute . In this matrix, the -block is .

Step 9

Find a matrix formed as a product of elementary matrices such that is a matrix in which all the blocks above the block contain only 's.

Step 10

Repeat Steps 8 and 9 on column converting -block to via conjugation by some invertible matrix . Use this block to clear out the blocks above, via conjugation by a product of elementary matrices.

Step 11

Repeat these processes on columns, using conjugations by . The resulting matrix is now in Weyr form.

Step 12

Let . Then .

Applications of the Weyr form

Some well-known applications of the Weyr form are listed below: [3]

  1. The Weyr form can be used to simplify the proof of Gerstenhaber’s Theorem which asserts that the subalgebra generated by two commuting matrices has dimension at most .
  2. A set of finite matrices is said to be approximately simultaneously diagonalizable if they can be perturbed to simultaneously diagonalizable matrices. The Weyr form is used to prove approximate simultaneous diagonalizability of various classes of matrices. The approximate simultaneous diagonalizability property has applications in the study of phylogenetic invariants in biomathematics.
  3. The Weyr form can be used to simplify the proofs of the irreducibility of the variety of all k-tuples of commuting complex matrices.

Related Research Articles

Ray transfer matrix analysis is a mathematical form for performing ray tracing calculations in sufficiently simple problems which can be solved considering only paraxial rays. Each optical element is described by a 2×2 ray transfer matrix which operates on a vector describing an incoming light ray to calculate the outgoing ray. Multiplication of the successive matrices thus yields a concise ray transfer matrix describing the entire optical system. The same mathematics is also used in accelerator physics to track particles through the magnet installations of a particle accelerator, see electron optics.

<span class="mw-page-title-main">Cayley–Hamilton theorem</span> Every square matrix over a commutative ring satisfies its own characteristic equation

In linear algebra, the Cayley–Hamilton theorem states that every square matrix over a commutative ring satisfies its own characteristic equation.

In mathematics, particularly in linear algebra, a skew-symmetricmatrix is a square matrix whose transpose equals its negative. That is, it satisfies the condition

In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagonal matrix is , while an example of a 3×3 diagonal matrix is. An identity matrix of any size, or any multiple of it is a diagonal matrix called scalar matrix, for example, . In geometry, a diagonal matrix may be used as a scaling matrix, since matrix multiplication with it results in changing scale (size) and possibly also shape; only a scalar matrix results in uniform change in scale.

In linear algebra, a square matrix  is called diagonalizable or non-defective if it is similar to a diagonal matrix. That is, if there exists an invertible matrix  and a diagonal matrix such that . This is equivalent to . This property exists for any linear map: for a finite-dimensional vector space , a linear map  is called diagonalizable if there exists an ordered basis of  consisting of eigenvectors of . These definitions are equivalent: if  has a matrix representation as above, then the column vectors of  form a basis consisting of eigenvectors of , and the diagonal entries of  are the corresponding eigenvalues of ; with respect to this eigenvector basis,  is represented by .

<span class="mw-page-title-main">Jordan normal form</span> Form of a matrix indicating its eigenvalues and their algebraic multiplicities

In linear algebra, a Jordan normal form, also known as a Jordan canonical form (JCF), is an upper triangular matrix of a particular form called a Jordan matrix representing a linear operator on a finite-dimensional vector space with respect to some basis. Such a matrix has each non-zero off-diagonal entry equal to 1, immediately above the main diagonal, and with identical diagonal entries to the left and below them.

In linear algebra, a Householder transformation is a linear transformation that describes a reflection about a plane or hyperplane containing the origin. The Householder transformation was used in a 1958 paper by Alston Scott Householder.

In numerical analysis, one of the most important problems is designing efficient and stable algorithms for finding the eigenvalues of a matrix. These eigenvalue algorithms may also find eigenvectors.

In linear algebra, the Frobenius companion matrix of the monic polynomial

In mathematics, the matrix exponential is a matrix function on square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exponential gives the exponential map between a matrix Lie algebra and the corresponding Lie group.

In linear algebra, a nilpotent matrix is a square matrix N such that

In linear algebra, it is often important to know which vectors have their directions unchanged by a given linear transformation. An eigenvector or characteristic vector is such a vector. Thus an eigenvector of a linear transformation is scaled by a constant factor when the linear transformation is applied to it: . The corresponding eigenvalue, characteristic value, or characteristic root is the multiplying factor .

In mathematics, a logarithm of a matrix is another matrix such that the matrix exponential of the latter matrix equals the original matrix. It is thus a generalization of the scalar logarithm and in some sense an inverse function of the matrix exponential. Not all matrices have a logarithm and those matrices that do have a logarithm may have more than one logarithm. The study of logarithms of matrices leads to Lie theory since when a matrix has a logarithm then it is in an element of a Lie group and the logarithm is the corresponding element of the vector space of the Lie algebra.

In the mathematical discipline of matrix theory, a Jordan matrix, named after Camille Jordan, is a block diagonal matrix over a ring R, where each block along the diagonal, called a Jordan block, has the following form:

In linear algebra, Weyl's inequality is a theorem about the changes to eigenvalues of an Hermitian matrix that is perturbed. It can be used to estimate the eigenvalues of a perturbed Hermitian matrix.

In linear algebra, a defective matrix is a square matrix that does not have a complete basis of eigenvectors, and is therefore not diagonalizable. In particular, an matrix is defective if and only if it does not have linearly independent eigenvectors. A complete basis is formed by augmenting the eigenvectors with generalized eigenvectors, which are necessary for solving defective systems of ordinary differential equations and other problems.

In mathematics, every analytic function can be used for defining a matrix function that maps square matrices with complex entries to square matrices of the same size.

In linear algebra, eigendecomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in this way. When the matrix being factorized is a normal or real symmetric matrix, the decomposition is called "spectral decomposition", derived from the spectral theorem.

In mathematics, the quadratic eigenvalue problem (QEP), is to find scalar eigenvalues , left eigenvectors and right eigenvectors such that

In linear algebra, two matrices and are said to commute if , or equivalently if their commutator is zero. A set of matrices is said to commute if they commute pairwise, meaning that every pair of matrices in the set commute with each other.

References

  1. Eduard Weyr (1885). "Répartition des matrices en espèces et formation de toutes les espèces" (PDF). Comptes Rendus de l'Académie des Sciences de Paris. 100: 966–969. Retrieved 10 December 2013.
  2. Eduard Weyr (1890). "Zur Theorie der bilinearen Formen". Monatshefte für Mathematik und Physik. 1: 163–236.
  3. 1 2 3 4 Kevin C. Meara; John Clark; Charles I. Vinsonhaler (2011). Advanced Topics in Linear Algebra: Weaving Matrix Problems through the Weyr Form. Oxford University Press.
  4. 1 2 3 Kevin C. Meara; John Clark; Charles I. Vinsonhaler (2011). Advanced Topics in Linear Algebra: Weaving Matrix Problems through the Weyr Form. Oxford University Press. pp. 44, 81–82.
  5. Shapiro, H. (1999). "The Weyr characteristic" (PDF). The American Mathematical Monthly. 106 (10): 919–929. doi:10.2307/2589746. JSTOR   2589746. S2CID   56072601.
  6. Sergeichuk, "Canonical matrices for linear matrix problems", Arxiv:0709.2485 [math.RT], 2007