Trace distance

Last updated

In quantum mechanics, and especially quantum information and the study of open quantum systems, the trace distanceT is a metric on the space of density matrices and gives a measure of the distinguishability between two states. It is the quantum generalization of the Kolmogorov distance for classical probability distributions.

Contents

Definition

The trace distance is defined as half of the trace norm of the difference of the matrices:

where is the trace norm of , and is the unique positive semidefinite such that (which is always defined for positive semidefinite ). This can be thought of as the matrix obtained from taking the algebraic square roots of its eigenvalues. For the trace distance, we more specifically have an expression of the form where is Hermitian. This quantity equals the sum of the singular values of , which being Hermitian, equals the sum of the absolute values of its eigenvalues. More explicitly,

where is the -th eigenvalue of , and is its rank.

The factor of two ensures that the trace distance between normalized density matrices takes values in the range .

Connection with the total variation distance

The trace distance can be seen as a direct quantum generalization of the total variation distance between probability distributions. Given a pair of probability distributions , their total variation distance is

Attempting to directly apply this definition to quantum states raises the problem that quantum states can result in different probability distributions depending on how they are measured. A natural choice is then to consider the total variation distance between the classical probability distribution obtained measuring the two states, maximized over the possible choices of measurement, which results precisely in the trace distance between the quantum states. More explicitly, this is the quantity

with the maximization performed with respect to all possible POVMs . To see why this is the case, we start observing that there is a unique decomposition with positive semidefinite matrices with orthogonal support. With these operators we can write concisely . Furthermore , and thus . We thus have

This shows that

where denotes the classical probability distribution resulting from measuring with the POVM , , and the maximum is performed over all POVMs . To conclude that the inequality is saturated by some POVM, we need only consider the projective measurement with elements corresponding to the eigenvectors of . With this choice,

where are the eigenvalues of .

Physical interpretation

By using the Hölder duality for Schatten norms, the trace distance can be written in variational form as [1]

As for its classical counterpart, the trace distance can be related to the maximum probability of distinguishing between two quantum states:

For example, suppose Alice prepares a system in either the state or , each with probability and sends it to Bob who has to discriminate between the two states using a binary measurement. Let Bob assign the measurement outcome and a POVM element such as the outcome and a POVM element to identify the state or , respectively. His expected probability of correctly identifying the incoming state is then given by

Therefore, when applying an optimal measurement, Bob has the maximal probability

of correctly identifying in which state Alice prepared the system. [2]

Properties

The trace distance has the following properties [1]

For qubits, the trace distance is equal to half the Euclidean distance in the Bloch representation.

Relationship to other distance measures

Fidelity

The fidelity of two quantum states is related to the trace distance by the inequalities

The upper bound inequality becomes an equality when and are pure states. [Note that the definition for Fidelity used here is the square of that used in Nielsen and Chuang]

Total variation distance

The trace distance is a generalization of the total variation distance, and for two commuting density matrices, has the same value as the total variation distance of the two corresponding probability distributions.

Related Research Articles

<span class="mw-page-title-main">Multivariate normal distribution</span> Generalization of the one-dimensional normal distribution to higher dimensions

In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One definition is that a random vector is said to be k-variate normally distributed if every linear combination of its k components has a univariate normal distribution. Its importance derives mainly from the multivariate central limit theorem. The multivariate normal distribution is often used to describe, at least approximately, any set of (possibly) correlated real-valued random variables each of which clusters around a mean value.

In quantum mechanics, a density matrix is a matrix that describes the quantum state of a physical system. It allows for the calculation of the probabilities of the outcomes of any measurement performed upon this system, using the Born rule. It is a generalization of the more usual state vectors or wavefunctions: while those can only represent pure states, density matrices can also represent mixed states. Mixed states arise in quantum mechanics in two different situations: first when the preparation of the system is not fully known, and thus one must deal with a statistical ensemble of possible preparations, and second when one wants to describe a physical system which is entangled with another, without describing their combined state.

In physics, the CHSH inequality can be used in the proof of Bell's theorem, which states that certain consequences of entanglement in quantum mechanics can not be reproduced by local hidden-variable theories. Experimental verification of the inequality being violated is seen as confirmation that nature cannot be described by such theories. CHSH stands for John Clauser, Michael Horne, Abner Shimony, and Richard Holt, who described it in a much-cited paper published in 1969. They derived the CHSH inequality, which, as with John Stewart Bell's original inequality, is a constraint on the statistical occurrence of "coincidences" in a Bell test which is necessarily true if there exist underlying local hidden variables, an assumption that is sometimes termed local realism. It is in fact the case that the inequality is routinely violated by modern experiments in quantum mechanics.

In probability theory, the central limit theorem states that, under certain circumstances, the probability distribution of the scaled mean of a random sample converges to a normal distribution as the sample size increases to infinity. Under stronger assumptions, the Berry–Esseen theorem, or Berry–Esseen inequality, gives a more quantitative result, because it also specifies the rate at which this convergence takes place by giving a bound on the maximal error of approximation between the normal distribution and the true distribution of the scaled sample mean. The approximation is measured by the Kolmogorov–Smirnov distance. In the case of independent samples, the convergence rate is n−1/2, where n is the sample size, and the constant is estimated in terms of the third absolute normalized moments.

In quantum physics, a measurement is the testing or manipulation of a physical system to yield a numerical result. The predictions that quantum physics makes are in general probabilistic. The mathematical tools for making predictions about what measurement outcomes may occur were developed during the 20th century and make use of linear algebra and functional analysis.

In mathematics, the explicit formulae for L-functions are relations between sums over the complex number zeroes of an L-function and sums over prime powers, introduced by Riemann (1859) for the Riemann zeta function. Such explicit formulae have been applied also to questions on bounding the discriminant of an algebraic number field, and the conductor of a number field.

In probability theory, calculation of the sum of normally distributed random variables is an instance of the arithmetic of random variables, which can be quite complex based on the probability distributions of the random variables involved and their relationships.

In functional analysis and quantum measurement theory, a positive operator-valued measure (POVM) is a measure whose values are positive semi-definite operators on a Hilbert space. POVMs are a generalisation of projection-valued measures (PVM) and, correspondingly, quantum measurements described by POVMs are a generalisation of quantum measurement described by PVMs.

In probability theory and statistics, the Jensen–Shannon divergence is a method of measuring the similarity between two probability distributions. It is also known as information radius (IRad) or total divergence to the average. It is based on the Kullback–Leibler divergence, with some notable differences, including that it is symmetric and it always has a finite value. The square root of the Jensen–Shannon divergence is a metric often referred to as Jensen–Shannon distance.

In mathematics, a π-system on a set is a collection of certain subsets of such that

Holevo's theorem is an important limitative theorem in quantum computing, an interdisciplinary field of physics and computer science. It is sometimes called Holevo's bound, since it establishes an upper bound to the amount of information that can be known about a quantum state. It was published by Alexander Holevo in 1973.

In quantum mechanics, notably in quantum information theory, fidelity is a measure of the "closeness" of two quantum states. It expresses the probability that one state will pass a test to identify as the other. The fidelity is not a metric on the space of density matrices, but it can be used to define the Bures metric on this space.

In quantum information theory, quantum relative entropy is a measure of distinguishability between two quantum states. It is the quantum mechanical analog of relative entropy.

<span class="mw-page-title-main">SIC-POVM</span>

A symmetric, informationally complete, positive operator-valued measure (SIC-POVM) is a special case of a generalized measurement on a Hilbert space, used in the field of quantum mechanics. A measurement of the prescribed form satisfies certain defining qualities that makes it an interesting candidate for a "standard quantum measurement", utilized in the study of foundational quantum mechanics, most notably in QBism. Furthermore, it has been shown that applications exist in quantum state tomography and quantum cryptography, and a possible connection has been discovered with Hilbert's twelfth problem.

The Fannes–Audenaert inequality is a mathematical bound on the difference between the von Neumann entropies of two density matrices as a function of their trace distance. It was proved by Koenraad M. R. Audenaert in 2007 as an optimal refinement of Mark Fannes' original inequality, which was published in 1973. Mark Fannes is a Belgian physicist specialised in mathematical quantum mechanics. He works at the KU Leuven. Koenraad M. R. Audenaert is a Belgian physicist and civil engineer. He currently works at Royal Holloway, University of London.

A product distribution is a probability distribution constructed as the distribution of the product of random variables having two other known distributions. Given two statistically independent random variables X and Y, the distribution of the random variable Z that is formed as the product is a product distribution.

In quantum information theory, the classical capacity of a quantum channel is the maximum rate at which classical data can be sent over it error-free in the limit of many uses of the channel. Holevo, Schumacher, and Westmoreland proved the following least upper bound on the classical capacity of any quantum channel :

The min-entropy, in information theory, is the smallest of the Rényi family of entropies, corresponding to the most conservative way of measuring the unpredictability of a set of outcomes, as the negative logarithm of the probability of the most likely outcome. The various Rényi entropies are all equal for a uniform distribution, but measure the unpredictability of a nonuniform distribution in different ways. The min-entropy is never greater than the ordinary or Shannon entropy and that in turn is never greater than the Hartley or max-entropy, defined as the logarithm of the number of outcomes with nonzero probability.

Generalized relative entropy is a measure of dissimilarity between two quantum states. It is a "one-shot" analogue of quantum relative entropy and shares many properties of the latter quantity.

The term quantum state discrimination collectively refers to quantum-informatics techniques, with the help of which, by performing a small number of measurements on a physical system, its specific quantum state can be identified. And this is provided that the set of states in which the system can be is known in advance, and we only need to determine which one it is. This assumption distinguishes such techniques from quantum tomography, which does not impose additional requirements on the state of the system, but requires many times more measurements.

References

  1. 1 2 Nielsen, Michael A.; Chuang, Isaac L. (2010). "9. Distance measures for quantum information". Quantum Computation and Quantum Information (2nd ed.). Cambridge: Cambridge University Press. ISBN   978-1-107-00217-3. OCLC   844974180.
  2. S. M. Barnett, "Quantum Information", Oxford University Press, 2009, Chapter 4
  3. Wilde, Mark (2017). "From Classical to Quantum Shannon Theory". arXiv. 1106.1445.