Corner transfer matrix

Last updated

In statistical mechanics, the corner transfer matrix describes the effect of adding a quadrant to a lattice. Introduced by Rodney Baxter in 1968 as an extension of the Kramers-Wannier row-to-row transfer matrix, it provides a powerful method of studying lattice models. Calculations with corner transfer matrices led Baxter to the exact solution of the hard hexagon model in 1980.

Contents

Definition

Consider an IRF (interaction-round-a-face) model, i.e. a square lattice model with a spin σi assigned to each site i and interactions limited to spins around a common face. Let the total energy be given by

where for each face the surrounding sites i, j, k and l are arranged as follows:

CTM - IRF model.png

For a lattice with N sites, the partition function is

where the sum is over all possible spin configurations and w is the Boltzmann weight

To simplify the notation, we use a ferromagnetic Ising-type lattice where each spin has the value +1 or −1, and the ground state is given by all spins up (i.e. the total energy is minimised when all spins on the lattice have the value +1). We also assume the lattice has 4-fold rotational symmetry (up to boundary conditions) and is reflection-invariant. These simplifying assumptions are not crucial, and extending the definition to the general case is relatively straightforward.

Now consider the lattice quadrant shown below:

CTM - lattice quadrant.png

The outer boundary sites, marked by triangles, are assigned their ground state spins (+1 in this case). The sites marked by open circles form the inner boundaries of the quadrant; their associated spin sets are labelled {σ1,...,σm} and {σ'1,...,σ'm}, where σ1 = σ'1. There are 2m possible configurations for each inner boundary, so we define a 2m×2m matrix entry-wise by

The matrix A, then, is the corner transfer matrix for the given lattice quadrant. Since the outer boundary spins are fixed and the sum is over all interior spins, each entry of A is a function of the inner boundary spins. The Kronecker delta in the expression ensures that σ1 = σ'1, so by ordering the configurations appropriately we may cast A as a block diagonal matrix:

Corner transfer matrices are related to the partition function in a simple way. In our simplified example, we construct the full lattice from four rotated copies of the lattice quadrant, where the inner boundary spin sets σ, σ', σ" and σ'" are allowed to differ:

CTM - full lattice.png

The partition function is then written in terms of the corner transfer matrix A as

Discussion

Recursion relation

A corner transfer matrix A2m (defined for an m×m quadrant) may be expressed in terms of smaller corner transfer matrices A2m-1 and A2m-2 (defined for reduced (m-1)×(m-1) and (m-2)×(m-2) quadrants respectively). This recursion relation allows, in principle, the iterative calculation of the corner transfer matrix for any lattice quadrant of finite size.

Like their row-to-row counterparts, corner transfer matrices may be factored into face transfer matrices, which correspond to adding a single face to the lattice. For the lattice quadrant given earlier, the face transfer matrices are of size 2m×2m and defined entry-wise by

where 2 ≤ im+1. Near the outer boundary, specifically, we have

So the corner transfer matrix A factorises as

where

Graphically, this corresponds to:

CTM - FTM factorisation.png

We also require the 2m×2m matrices A* and A**, defined entry-wise by

where the A matrices whose entries appear on the RHS are of size 2m-1×2m-1 and 2m-2×2m-2 respectively. This is more clearly written as

Now from the definitions of A, A*, A**, Ui and Fj, we have

which gives the recursion relation for A2m in terms of A2m-1 and A2m-2.

Diagonal form

When using corner transfer matrices to perform calculations, it is both analytically and numerically convenient to work with their diagonal forms instead. To facilitate this, the recursion relation may be rewritten directly in terms of the diagonal forms and eigenvector matrices of A, A* and A**.

Recalling that the lattice in our example is reflection-invariant, in the sense that

we see that A is a symmetric matrix (i.e. it is diagonalisable by an orthogonal matrix). So we write

where Ad is a diagonal matrix (normalised such that its numerically largest entry is 1), αm is the largest eigenvalue of A, and PTP = I. Likewise for A* and A**, we have

where Ad*, Ad**, P* and P** are defined in an analogous fashion to A* and A**, i.e. in terms of the smaller (normalised) diagonal forms and (orthogonal) eigenvector matrices of A2m-1 and A2m-2.

By substituting these diagonalisations into the recursion relation, we obtain

where

Now At is also symmetric, and may be calculated if Ad*, Ad** and R* are known; diagonalising At then yields its normalised diagonal form Ad, its largest eigenvalue κ, and its orthogonal eigenvector matrix R.

Applications

Spin expectation value

Corner transfer matrices (or their diagonal forms) may be used to calculate quantities such as the spin expectation value at a particular site deep inside the lattice. For the full lattice given earlier, the spin expectation value at the central site is given by

With the configurations ordered such that A is block diagonal as before, we may define a 2m×2m diagonal matrix

such that

Partition function per site

Another important quantity for lattice models is the partition function per site, evaluated in the thermodynamic limit and written as

In our example, this reduces to

since tr Ad4 is a convergent sum as m → ∞ and Ad becomes infinite-dimensional. Furthermore, the number of faces 2m(m+1) approaches the number of sites N in the thermodynamic limit, so we have

which is consistent with the earlier equation giving κ as the largest eigenvalue for At. In other words, the partition function per site is given exactly by the diagonalised recursion relation for corner transfer matrices in the thermodynamic limit; this allows κ to be approximated via the iterative process of calculating Ad for a large lattice.

The matrices involved grow exponentially in size, however, and in actual numerical calculations they must be truncated at each step. One way of doing this is to keep the n largest eigenvalues at each step, for some fixed n. In most cases, the sequence of approximations obtained by taking n = 1,2,3,... converges rapidly, and to the exact value (for an exactly solvable model).

See also

Related Research Articles

<span class="mw-page-title-main">Pauli matrices</span> Matrices important in quantum mechanics and the study of spin

In mathematical physics and mathematics, the Pauli matrices are a set of three 2 × 2 complex matrices that are traceless, Hermitian, involutory and unitary. Usually indicated by the Greek letter sigma, they are occasionally denoted by tau when used in connection with isospin symmetries.

<span class="mw-page-title-main">Spinor</span> Non-tensorial representation of the spin group

In geometry and physics, spinors are elements of a complex number-based vector space that can be associated with Euclidean space. A spinor transforms linearly when the Euclidean space is subjected to a slight (infinitesimal) rotation, but unlike geometric vectors and tensors, a spinor transforms to its negative when the space rotates through 360°. It takes a rotation of 720° for a spinor to go back to its original state. This property characterizes spinors: spinors can be viewed as the "square roots" of vectors.

In particle physics, the Dirac equation is a relativistic wave equation derived by British physicist Paul Dirac in 1928. In its free form, or including electromagnetic interactions, it describes all spin-1/2 massive particles, called "Dirac particles", such as electrons and quarks for which parity is a symmetry. It is consistent with both the principles of quantum mechanics and the theory of special relativity, and was the first theory to account fully for special relativity in the context of quantum mechanics. It was validated by accounting for the fine structure of the hydrogen spectrum in a completely rigorous way. It has become vital in the building of the Standard Model.

<span class="mw-page-title-main">Lorentz group</span> Lie group of Lorentz transformations

In physics and mathematics, the Lorentz group is the group of all Lorentz transformations of Minkowski spacetime, the classical and quantum setting for all (non-gravitational) physical phenomena. The Lorentz group is named for the Dutch physicist Hendrik Lorentz.

The Ising model, named after the physicists Ernst Ising and Wilhelm Lenz, is a mathematical model of ferromagnetism in statistical mechanics. The model consists of discrete variables that represent magnetic dipole moments of atomic "spins" that can be in one of two states. The spins are arranged in a graph, usually a lattice, allowing each spin to interact with its neighbors. Neighboring spins that agree have a lower energy than those that disagree; the system tends to the lowest energy but heat disturbs this tendency, thus creating the possibility of different structural phases. The model allows the identification of phase transitions as a simplified model of reality. The two-dimensional square-lattice Ising model is one of the simplest statistical models to show a phase transition.

In mathematics, the field trace is a particular function defined with respect to a finite field extension L/K, which is a K-linear map from L onto K.

In mathematics, the determinant of an m-by-m skew-symmetric matrix can always be written as the square of a polynomial in the matrix entries, a polynomial with integer coefficients that only depends on m. When m is odd, the polynomial is zero, and when m is even, it is a nonzero polynomial of degree m/2, and is unique up to multiplication by ±1. The convention on skew-symmetric tridiagonal matrices, given below in the examples, then determines one specific polynomial, called the Pfaffian polynomial. The value of this polynomial, when applied to the entries of a skew-symmetric matrix, is called the Pfaffian of that matrix. The term Pfaffian was introduced by Cayley, who indirectly named them after Johann Friedrich Pfaff.

In linear algebra, the generalized singular value decomposition (GSVD) is the name of two different techniques based on the singular value decomposition (SVD). The two versions differ because one version decomposes two matrices and the other version uses a set of constraints imposed on the left and right singular vectors of a single-matrix SVD.

<span class="mw-page-title-main">Two-state quantum system</span> Simple quantum mechanical system

In quantum mechanics, a two-state system is a quantum system that can exist in any quantum superposition of two independent quantum states. The Hilbert space describing such a system is two-dimensional. Therefore, a complete basis spanning the space will consist of two independent states. Any two-state system can also be seen as a qubit.

In mathematical physics, the gamma matrices, also called the Dirac matrices, are a set of conventional matrices with specific anticommutation relations that ensure they generate a matrix representation of the Clifford algebra It is also possible to define higher-dimensional gamma matrices. When interpreted as the matrices of the action of a set of orthogonal basis vectors for contravariant vectors in Minkowski space, the column vectors on which the matrices act become a space of spinors, on which the Clifford algebra of spacetime acts. This in turn makes it possible to represent infinitesimal spatial rotations and Lorentz boosts. Spinors facilitate spacetime computations in general, and in particular are fundamental to the Dirac equation for relativistic spin particles. Gamma matrices were introduced by Paul Dirac in 1928.

In physics, the Majorana equation is a relativistic wave equation. It is named after the Italian physicist Ettore Majorana, who proposed it in 1937 as a means of describing fermions that are their own antiparticle. Particles corresponding to this equation are termed Majorana particles, although that term now has a more expansive meaning, referring to any fermionic particle that is its own anti-particle.

The quantum Heisenberg model, developed by Werner Heisenberg, is a statistical mechanical model used in the study of critical points and phase transitions of magnetic systems, in which the spins of the magnetic systems are treated quantum mechanically. It is related to the prototypical Ising model, where at each site of a lattice, a spin represents a microscopic magnetic dipole to which the magnetic moment is either up or down. Except the coupling between magnetic dipole moments, there is also a multipolar version of Heisenberg model called the multipolar exchange interaction.

In mathematics and physics, in particular quantum information, the term generalized Pauli matrices refers to families of matrices which generalize the properties of the Pauli matrices. Here, a few classes of such matrices are summarized.

In numerical linear algebra, the alternating-direction implicit (ADI) method is an iterative method used to solve Sylvester matrix equations. It is a popular method for solving the large matrix equations that arise in systems theory and control, and can be formulated to construct solutions in a memory-efficient, factored form. It is also used to numerically solve parabolic and elliptic partial differential equations, and is a classic method used for modeling heat conduction and solving the diffusion equation in two or more dimensions. It is an example of an operator splitting method.

In statistical mechanics, the eight-vertex model is a generalization of the ice-type (six-vertex) models. It was discussed by Sutherland and Fan & Wu, and solved by Rodney Baxter in the zero-field case.

The Luttinger–Kohn model is a flavor of the k·p perturbation theory used for calculating the structure of multiple, degenerate electronic bands in bulk and quantum well semiconductors. The method is a generalization of the single band k·p theory.

In the geometry of numbers, the Klein polyhedron, named after Felix Klein, is used to generalize the concept of simple continued fractions to higher dimensions.

For certain applications in linear algebra, it is useful to know properties of the probability distribution of the largest eigenvalue of a finite sum of random matrices. Suppose is a finite sequence of random matrices. Analogous to the well-known Chernoff bound for sums of scalars, a bound on the following is sought for a given parameter t:

In statistics, the matrix t-distribution is the generalization of the multivariate t-distribution from vectors to matrices.

<span class="mw-page-title-main">Symmetry in quantum mechanics</span> Properties underlying modern physics

Symmetries in quantum mechanics describe features of spacetime and particles which are unchanged under some transformation, in the context of quantum mechanics, relativistic quantum mechanics and quantum field theory, and with applications in the mathematical formulation of the standard model and condensed matter physics. In general, symmetry in physics, invariance, and conservation laws, are fundamentally important constraints for formulating physical theories and models. In practice, they are powerful methods for solving problems and predicting what can happen. While conservation laws do not always give the answer to the problem directly, they form the correct constraints and the first steps to solving a multitude of problems. In application, understanding symmetries can also provide insights on the eigenstates that can be expected. For example, the existence of degenerate states can be inferred by the presence of non commuting symmetry operators or that the non degenerate states are also eigenvectors of symmetry operators.

References