Density matrix renormalization group

Last updated

The density matrix renormalization group (DMRG) is a numerical variational technique devised to obtain the low-energy physics of quantum many-body systems with high accuracy. As a variational method, DMRG is an efficient algorithm that attempts to find the lowest-energy matrix product state wavefunction of a Hamiltonian. It was invented in 1992 by Steven R. White and it is nowadays the most efficient method for 1-dimensional systems. [1]

Contents

History

The first application of the DMRG, by Steven R. White and Reinhard Noack, was a toy model: to find the spectrum of a spin 0 particle in a 1D box.[ when? ] This model had been proposed by Kenneth G. Wilson as a test for any new renormalization group method, because they all happened to fail with this simple problem.[ when? ] The DMRG overcame the problems of previous renormalization group methods by connecting two blocks with the two sites in the middle rather than just adding a single site to a block at each step as well as by using the density matrix to identify the most important states to be kept at the end of each step. After succeeding with the toy model, the DMRG method was tried with success on the quantum Heisenberg model.

Principle

The main problem of quantum many-body physics is the fact that the Hilbert space grows exponentially with size. In other words if one considers a lattice, with some Hilbert space of dimension on each site of the lattice, then the total Hilbert space would have dimension , where is the number of sites on the lattice. For example, a spin-1/2 chain of length L has 2L degrees of freedom. The DMRG is an iterative, variational method that reduces effective degrees of freedom to those most important for a target state. The state one is most often interested in is the ground state.

After a warmup cycle[ definition needed ], the method splits the system into two subsystems, or blocks, which need not have equal sizes, and two sites in between. A set of representative states has been chosen for the block during the warmup. This set of left blocks + two sites + right blocks is known as the superblock. Now a candidate for the ground state of the superblock, which is a reduced version of the full system, may be found. It may have a rather poor accuracy, but the method is iterative and improves with the steps below.

Decomposition of the system into left and right blocks, according to DMRG. Dmrg1.png
Decomposition of the system into left and right blocks, according to DMRG.

The candidate ground state that has been found is projected into the Hilbert subspace for each block using a density matrix, hence the name. Thus, the relevant states for each block are updated. [ further explanation needed ]

Now one of the blocks grows at the expense of the other and the procedure is repeated. When the growing block reaches maximum size, the other starts to grow in its place. Each time we return to the original (equal sizes) situation, we say that a sweep has been completed. Normally, a few sweeps are enough to get a precision of a part in 1010 for a 1D lattice.

The DMRG sweep. Dmrg2.png
The DMRG sweep.

Implementation guide

A practical implementation of the DMRG algorithm is a lengthy work[ opinion ]. A few of the main computational tricks are these:

Applications

The DMRG has been successfully applied to get the low energy properties of spin chains: Ising model in a transverse field, Heisenberg model, etc., fermionic systems, such as the Hubbard model, problems with impurities such as the Kondo effect, boson systems, and the physics of quantum dots joined with quantum wires. It has been also extended to work on tree graphs, and has found applications in the study of dendrimers. For 2D systems with one of the dimensions much larger than the other DMRG is also accurate, and has proved useful in the study of ladders.

The method has been extended to study equilibrium statistical physics in 2D, and to analyze non-equilibrium phenomena in 1D.

The DMRG has also been applied to the field of quantum chemistry to study strongly correlated systems.

Example: Quantum Heisenberg model

Let us consider an "infinite" DMRG algorithm for the antiferromagnetic quantum Heisenberg chain. The recipe can be applied for every translationally invariant one-dimensional lattice.

DMRG is a renormalization-group technique because it offers an efficient truncation of the Hilbert space of one-dimensional quantum systems.

Starting point

To simulate an infinite chain, starting with four sites. The first is the block site, the last the universe-block site and the remaining are the added sites, the right one is added to the universe-block site and the other to the block site.

The Hilbert space for the single site is with the base . With this base the spin operators are , and for the single site. For every block, the two blocks and the two sites, there is its own Hilbert space , its base ()and its own operators

where

At the starting point all four Hilbert spaces are equivalent to , all spin operators are equivalent to , and and . In the following iterations, this is only true for the left and right sites.

Step 1: Form the Hamiltonian matrix for the superblock

The ingredients are the four block operators and the four universe-block operators, which at the first iteration are matrices, the three left-site spin operators and the three right-site spin operators, which are always matrices. The Hamiltonian matrix of the superblock (the chain), which at the first iteration has only four sites, is formed by these operators. In the Heisenberg antiferromagnetic S=1 model the Hamiltonian is:

These operators live in the superblock state space: , the base is . For example: (convention):

The Hamiltonian in the DMRG form is (we set ):

The operators are matrices, , for example:

Step 2: Diagonalize the superblock Hamiltonian

At this point you must choose the eigenstate of the Hamiltonian for which some observables is calculated, this is the target state . At the beginning you can choose the ground state and use some advanced algorithm to find it, one of these is described in:

This step is the most time-consuming part of the algorithm.

If is the target state, expectation value of various operators can be measured at this point using .

Step 3: Reduce density matrix

Form the reduced density matrix for the first two block system, the block and the left-site. By definition it is the matrix:

Diagonalize and form the matrix , which rows are the eigenvectors associated with the largest eigenvalues of . So is formed by the most significant eigenstates of the reduced density matrix. You choose looking to the parameter : .

Step 4: New block and universe-block operators

Form the matrix representation of operators for the system composite of the block and left-site, and for the system composite of right-site and universe-block, for example:

Now, form the matrix representations of the new block and universe-block operators, form a new block by changing basis with the transformation , for example:

At this point the iteration is ended and the algorithm goes back to step 1.

The algorithm stops successfully when the observable converges to some value.

Matrix product ansatz

The success of the DMRG for 1D systems is related to the fact that it is a variational method within the space of matrix product states (MPS). These are states of the form

where are the values of the e.g. z-component of the spin in a spin chain, and the Asi are matrices of arbitrary dimension m. As m  ∞, the representation becomes exact. This theory was exposed by S. Rommer and S. Ostlund in .

In quantum chemistry application, stands for the four possibilities of the projection of the spin quantum number of the two electrons that can occupy a single orbital, thus , where the first (second) entry of these kets corresponds to the spin-up(down) electron. In quantum chemistry, (for a given ) and (for a given ) are traditionally chosen to be row and column matrices, respectively. This way, the result of is a scalar value and the trace operation is unnecessary. is the number of sites (the orbitals basically) used in the simulation.

The matrices in the MPS ansatz are not unique, one can, for instance, insert in the middle of , then define and , and the state will stay unchanged. Such gauge freedom is employed to transform the matrices into a canonical form. Three types of canonical form exist: (1) left-normalized form, when

for all , (2) right-normalized form, when

for all , and (3) mixed-canonical form when both left- and right-normalized matrices exist among the matrices in the above MPS ansatz.

The goal of the DMRG calculation is then to solve for the elements of each of the matrices. The so-called one-site and two-site algorithms have been devised for this purpose. In the one-site algorithm, only one matrix (one site) whose elements are solved for at a time. Two-site just means that two matrices are first contracted (multiplied) into a single matrix, and then its elements are solved. The two-site algorithm is proposed because the one-site algorithm is much more prone to getting trapped at a local minimum. Having the MPS in one of the above canonical forms has the advantage of making the computation more favorable - it leads to the ordinary eigenvalue problem. Without canonicalization, one will be dealing with a generalized eigenvalue problem.

Extensions

In 2004 the time-evolving block decimation method was developed to implement real-time evolution of matrix product states. The idea is based on the classical simulation of a quantum computer. Subsequently, a new method was devised to compute real-time evolution within the DMRG formalism - See the paper by A. Feiguin and S.R. White .

In recent years, some proposals to extend the method to 2D and 3D have been put forward, extending the definition of the matrix product states. See this paper by F. Verstraete and I. Cirac, .

Further reading

See also

Related Research Articles

<span class="mw-page-title-main">Lie algebra</span> Algebraic structure used in analysis

In mathematics, a Lie algebra is a vector space together with an operation called the Lie bracket, an alternating bilinear map , that satisfies the Jacobi identity. Otherwise said, a Lie algebra is an algebra over a field where the multiplication operation is now called Lie bracket and has two additional properties: it is alternating and satisfies the Jacobi identity. The Lie bracket of two vectors and is denoted . The Lie bracket does not need to be associative, meaning that the Lie algebra can be non associative. Given an associative algebra, a Lie bracket can be and is often defined through the commutator, namely defining correctly defines a Lie bracket in addition to the already existing multiplication operation.

<span class="mw-page-title-main">Pauli matrices</span> Matrices important in quantum mechanics and the study of spin

In mathematical physics and mathematics, the Pauli matrices are a set of three 2 × 2 complex matrices which are Hermitian, involutory and unitary. Usually indicated by the Greek letter sigma, they are occasionally denoted by tau when used in connection with isospin symmetries.

In mathematics, a symmetric matrix with real entries is positive-definite if the real number is positive for every nonzero real column vector where is the transpose of . More generally, a Hermitian matrix is positive-definite if the real number is positive for every nonzero complex column vector where denotes the conjugate transpose of

In linear algebra, the trace of a square matrix A, denoted tr(A), is defined to be the sum of elements on the main diagonal of A. The trace is only defined for a square matrix.

In mechanics and geometry, the 3D rotation group, often denoted SO(3), is the group of all rotations about the origin of three-dimensional Euclidean space under the operation of composition.

<span class="mw-page-title-main">Georgi–Glashow model</span> Grand Unified Theory proposed in 1974

In particle physics, the Georgi–Glashow model is a particular Grand Unified Theory (GUT) proposed by Howard Georgi and Sheldon Glashow in 1974. In this model, the Standard Model gauge groups SU(3) × SU(2) × U(1) are combined into a single simple gauge group SU(5). The unified group SU(5) is then thought to be spontaneously broken into the Standard Model subgroup below a very high energy scale called the grand unification scale.

In mathematics, a Lie superalgebra is a generalisation of a Lie algebra to include a Z2‑grading. Lie superalgebras are important in theoretical physics where they are used to describe the mathematics of supersymmetry. In most of these theories, the even elements of the superalgebra correspond to bosons and odd elements to fermions.

In mathematics, the Heisenberg group, named after Werner Heisenberg, is the group of 3×3 upper triangular matrices of the form

In mathematics, the Kronecker product, sometimes denoted by ⊗, is an operation on two matrices of arbitrary size resulting in a block matrix. It is a specialization of the tensor product from vectors to matrices and gives the matrix of the tensor product linear map with respect to a standard choice of basis. The Kronecker product is to be distinguished from the usual matrix multiplication, which is an entirely different operation. The Kronecker product is also sometimes called matrix direct product.

In quantum computing and specifically the quantum circuit model of computation, a quantum logic gate is a basic quantum circuit operating on a small number of qubits. They are the building blocks of quantum circuits, like classical logic gates are for conventional digital circuits.

<span class="mw-page-title-main">Representation theory of the Lorentz group</span> Representation of the symmetry group of spacetime in special relativity

The Lorentz group is a Lie group of symmetries of the spacetime of special relativity. This group can be realized as a collection of matrices, linear transformations, or unitary operators on some Hilbert space; it has a variety of representations. This group is significant because special relativity together with quantum mechanics are the two physical theories that are most thoroughly established, and the conjunction of these two theories is the study of the infinite-dimensional unitary representations of the Lorentz group. These have both historical importance in mainstream physics, as well as connections to more speculative present-day theories.

The quantum Heisenberg model, developed by Werner Heisenberg, is a statistical mechanical model used in the study of critical points and phase transitions of magnetic systems, in which the spins of the magnetic systems are treated quantum mechanically. It is related to the prototypical Ising model, where at each site of a lattice, a spin represents a microscopic magnetic dipole to which the magnetic moment is either up or down. Except the coupling between magnetic dipole moments, there is also a multipolar version of Heisenberg model called the multipolar exchange interaction.

Quantum walks are quantum analogues of classical random walks. In contrast to the classical random walk, where the walker occupies definite states and the randomness arises due to stochastic transitions between states, in quantum walks randomness arises through: (1) quantum superposition of states, (2) non-random, reversible unitary evolution and (3) collapse of the wave function due to state measurements.

In multilinear algebra, the tensor rank decomposition or the decomposition of a tensor is the decomposition of a tensor in terms of a sum of minimum tensors. This is an open problem.

In discrete mathematics, ideal lattices are a special class of lattices and a generalization of cyclic lattices. Ideal lattices naturally occur in many parts of number theory, but also in other areas. In particular, they have a significant place in cryptography. Micciancio defined a generalization of cyclic lattices as ideal lattices. They can be used in cryptosystems to decrease by a square root the number of parameters necessary to describe a lattice, making them more efficient. Ideal lattices are a new concept, but similar lattice classes have been used for a long time. For example, cyclic lattices, a special case of ideal lattices, are used in NTRUEncrypt and NTRUSign.

In algebra, the Nichols algebra of a braided vector space is a braided Hopf algebra which is denoted by and named after the mathematician Warren Nichols. It takes the role of quantum Borel part of a pointed Hopf algebra such as a quantum groups and their well known finite-dimensional truncations. Nichols algebras can immediately be used to write down new such quantum groups by using the Radford biproduct.

Given a Hilbert space with a tensor product structure a product numerical range is defined as a numerical range with respect to the subset of product vectors. In some situations, especially in the context of quantum mechanics product numerical range is known as local numerical range

<span class="mw-page-title-main">Lie algebra extension</span> Creating a "larger" Lie algebra from a smaller one, in one of several ways

In the theory of Lie groups, Lie algebras and their representation theory, a Lie algebra extensione is an enlargement of a given Lie algebra g by another Lie algebra h. Extensions arise in several ways. There is the trivial extension obtained by taking a direct sum of two Lie algebras. Other types are the split extension and the central extension. Extensions may arise naturally, for instance, when forming a Lie algebra from projective group representations. Such a Lie algebra will contain central charges.

In quantum computing and quantum information theory, the Clifford gates are the elements of the Clifford group, a set of mathematical transformations which normalize the n-qubit Pauli group, i.e., map tensor products of Pauli matrices to tensor products of Pauli matrices through conjugation. The notion was introduced by Daniel Gottesman and is named after the mathematician William Kingdon Clifford. Quantum circuits that consist of only Clifford gates can be efficiently simulated with a classical computer due to the Gottesman–Knill theorem.

<span class="mw-page-title-main">Tensor sketch</span> Algorithm for reducing the dimension of tensors

In statistics, machine learning and algorithms, a tensor sketch is a type of dimensionality reduction that is particularly efficient when applied to vectors that have tensor structure. Such a sketch can be used to speed up explicit kernel methods, bilinear pooling in neural networks and is a cornerstone in many numerical linear algebra algorithms.

References

  1. Nakatani, Naoki (2018), "Matrix Product States and Density Matrix Renormalization Group Algorithm", Reference Module in Chemistry, Molecular Sciences and Chemical Engineering, Elsevier, doi:10.1016/b978-0-12-409547-2.11473-8, ISBN   978-0-12-409547-2 , retrieved 2021-04-21