In mathematics, in linear algebra, a Weyr canonical form (or, Weyr form or Weyr matrix) is a square matrix which (in some sense) induces "nice" properties with matrices it commutes with. It also has a particularly simple structure and the conditions for possessing a Weyr form are fairly weak, making it a suitable tool for studying classes of commuting matrices. A square matrix is said to be in the Weyr canonical form if the matrix has the structure defining the Weyr canonical form. The Weyr form was discovered by the Czech mathematician Eduard Weyr in 1885. [1] [2] [3] The Weyr form did not become popular among mathematicians and it was overshadowed by the closely related, but distinct, canonical form known by the name Jordan canonical form. [3] The Weyr form has been rediscovered several times since Weyr’s original discovery in 1885. [4] This form has been variously called as modified Jordan form,reordered Jordan form,second Jordan form, and H-form. [4] The current terminology is credited to Shapiro who introduced it in a paper published in the American Mathematical Monthly in 1999. [4] [5]
Recently several applications have been found for the Weyr matrix. Of particular interest is an application of the Weyr matrix in the study of phylogenetic invariants in biomathematics.
A basic Weyr matrix with eigenvalue is an matrix of the following form: There is an integer partition
such that, when is viewed as an block matrix , where the block is an matrix, the following three features are present:
In this case, we say that has Weyr structure .
The following is an example of a basic Weyr matrix.
In this matrix, and . So has the Weyr structure . Also,
and
Let be a square matrix and let be the distinct eigenvalues of . We say that is in Weyr form (or is a Weyr matrix) if has the following form:
where is a basic Weyr matrix with eigenvalue for .
The following image shows an example of a general Weyr matrix consisting of three basic Weyr matrix blocks. The basic Weyr matrix in the top-left corner has the structure (4,2,1) with eigenvalue 4, the middle block has structure (2,2,1,1) with eigenvalue -3 and the one in the lower-right corner has the structure (3, 2) with eigenvalue 0.
The Weyr canonical form is related to the Jordan form by a simple permutation for each Weyr basic block as follows: The first index of each Weyr subblock forms the largest Jordan chain. After crossing out these rows and columns, the first index of each new subblock forms the second largest Jordan chain, and so forth. [6]
That the Weyr form is a canonical form of a matrix is a consequence of the following result: [3] Each square matrix over an algebraically closed field is similar to a Weyr matrix which is unique up to permutation of its basic blocks. The matrix is called the Weyr (canonical) form of .
Let be a square matrix of order over an algebraically closed field and let the distinct eigenvalues of be . The Jordan–Chevalley decomposition theorem states that is similar to a block diagonal matrix of the form
where is a diagonal matrix, is a nilpotent matrix, and , justifying the reduction of into subblocks . So the problem of reducing to the Weyr form reduces to the problem of reducing the nilpotent matrices to the Weyr form. This is leads to the generalized eigenspace decomposition theorem.
Given a nilpotent square matrix of order over an algebraically closed field , the following algorithm produces an invertible matrix and a Weyr matrix such that .
Step 1
Let
Step 2
Step 3
If is nonzero, repeat Step 2 on .
Step 4
Continue the processes of Steps 1 and 2 to obtain increasingly smaller square matrices and associated invertible matrices until the first zero matrix is obtained.
Step 5
The Weyr structure of is where = nullity.
Step 6
Step 7
Use elementary row operations to find an invertible matrix of appropriate size such that the product is a matrix of the form .
Step 8
Set diag and compute . In this matrix, the -block is .
Step 9
Find a matrix formed as a product of elementary matrices such that is a matrix in which all the blocks above the block contain only 's.
Step 10
Repeat Steps 8 and 9 on column converting -block to via conjugation by some invertible matrix . Use this block to clear out the blocks above, via conjugation by a product of elementary matrices.
Step 11
Repeat these processes on columns, using conjugations by . The resulting matrix is now in Weyr form.
Step 12
Let . Then .
Some well-known applications of the Weyr form are listed below: [3]
Ray transfer matrix analysis is a mathematical form for performing ray tracing calculations in sufficiently simple problems which can be solved considering only paraxial rays. Each optical element is described by a 2×2 ray transfer matrix which operates on a vector describing an incoming light ray to calculate the outgoing ray. Multiplication of the successive matrices thus yields a concise ray transfer matrix describing the entire optical system. The same mathematics is also used in accelerator physics to track particles through the magnet installations of a particle accelerator, see electron optics.
In linear algebra, the Cayley–Hamilton theorem states that every square matrix over a commutative ring satisfies its own characteristic equation.
In mathematics, particularly in linear algebra, a skew-symmetricmatrix is a square matrix whose transpose equals its negative. That is, it satisfies the condition
In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagonal matrix is , while an example of a 3×3 diagonal matrix is. An identity matrix of any size, or any multiple of it is a diagonal matrix called scalar matrix, for example, . In geometry, a diagonal matrix may be used as a scaling matrix, since matrix multiplication with it results in changing scale (size) and possibly also shape; only a scalar matrix results in uniform change in scale.
In linear algebra, a square matrix is called diagonalizable or non-defective if it is similar to a diagonal matrix. That is, if there exists an invertible matrix and a diagonal matrix such that . This is equivalent to . This property exists for any linear map: for a finite-dimensional vector space , a linear map is called diagonalizable if there exists an ordered basis of consisting of eigenvectors of . These definitions are equivalent: if has a matrix representation as above, then the column vectors of form a basis consisting of eigenvectors of , and the diagonal entries of are the corresponding eigenvalues of ; with respect to this eigenvector basis, is represented by .
In linear algebra, a Jordan normal form, also known as a Jordan canonical form (JCF), is an upper triangular matrix of a particular form called a Jordan matrix representing a linear operator on a finite-dimensional vector space with respect to some basis. Such a matrix has each non-zero off-diagonal entry equal to 1, immediately above the main diagonal, and with identical diagonal entries to the left and below them.
In linear algebra, a Householder transformation is a linear transformation that describes a reflection about a plane or hyperplane containing the origin. The Householder transformation was used in a 1958 paper by Alston Scott Householder.
In numerical analysis, one of the most important problems is designing efficient and stable algorithms for finding the eigenvalues of a matrix. These eigenvalue algorithms may also find eigenvectors.
In linear algebra, the Frobenius companion matrix of the monic polynomial
In mathematics, the matrix exponential is a matrix function on square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exponential gives the exponential map between a matrix Lie algebra and the corresponding Lie group.
In linear algebra, a nilpotent matrix is a square matrix N such that
In linear algebra, it is often important to know which vectors have their directions unchanged by a given linear transformation. An eigenvector or characteristic vector is such a vector. Thus an eigenvector of a linear transformation is scaled by a constant factor when the linear transformation is applied to it: . The corresponding eigenvalue, characteristic value, or characteristic root is the multiplying factor .
In mathematics, a logarithm of a matrix is another matrix such that the matrix exponential of the latter matrix equals the original matrix. It is thus a generalization of the scalar logarithm and in some sense an inverse function of the matrix exponential. Not all matrices have a logarithm and those matrices that do have a logarithm may have more than one logarithm. The study of logarithms of matrices leads to Lie theory since when a matrix has a logarithm then it is in an element of a Lie group and the logarithm is the corresponding element of the vector space of the Lie algebra.
In the mathematical discipline of matrix theory, a Jordan matrix, named after Camille Jordan, is a block diagonal matrix over a ring R, where each block along the diagonal, called a Jordan block, has the following form:
In linear algebra, Weyl's inequality is a theorem about the changes to eigenvalues of an Hermitian matrix that is perturbed. It can be used to estimate the eigenvalues of a perturbed Hermitian matrix.
In linear algebra, a defective matrix is a square matrix that does not have a complete basis of eigenvectors, and is therefore not diagonalizable. In particular, an matrix is defective if and only if it does not have linearly independent eigenvectors. A complete basis is formed by augmenting the eigenvectors with generalized eigenvectors, which are necessary for solving defective systems of ordinary differential equations and other problems.
In mathematics, every analytic function can be used for defining a matrix function that maps square matrices with complex entries to square matrices of the same size.
In linear algebra, eigendecomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in this way. When the matrix being factorized is a normal or real symmetric matrix, the decomposition is called "spectral decomposition", derived from the spectral theorem.
In mathematics, the quadratic eigenvalue problem (QEP), is to find scalar eigenvalues , left eigenvectors and right eigenvectors such that
In linear algebra, two matrices and are said to commute if , or equivalently if their commutator is zero. A set of matrices is said to commute if they commute pairwise, meaning that every pair of matrices in the set commute with each other.