Polarization identity

Last updated
Vectors involved in the polarization identity
2
||
x
||
2
+
2
||
y
||
2
=
||
x
+
y
||
2
+
||
x
-
y
||
2
.
{\displaystyle 2\|x\|^{2}+2\|y\|^{2}=\|x+y\|^{2}+\|x-y\|^{2}.} Parallelogram law.svg
Vectors involved in the polarization identity

In linear algebra, a branch of mathematics, the polarization identity is any one of a family of formulas that express the inner product of two vectors in terms of the norm of a normed vector space. If a norm arises from an inner product then the polarization identity can be used to express this inner product entirely in terms of the norm. The polarization identity shows that a norm can arise from at most one inner product; however, there exist norms that do not arise from any inner product.

Contents

The norm associated with any inner product space satisfies the parallelogram law: In fact, as observed by John von Neumann, [1] the parallelogram law characterizes those norms that arise from inner products. Given a normed space , the parallelogram law holds for if and only if there exists an inner product on such that for all in which case this inner product is uniquely determined by the norm via the polarization identity. [2] [3]

Polarization identities

Any inner product on a vector space induces a norm by the equation

The polarization identities reverse this relationship, recovering the inner product from the norm. Every inner product satisfies:

Solving for gives the formula If the inner product is real then and this formula becomes a polarization identity for real inner products.

Real vector spaces

If the vector space is over the real numbers then the polarization identities are: [4]


These various forms are all equivalent by the parallelogram law: [proof 1]


This further implies that class is not a Hilbert space whenever , as the parallelogram law is not satisfied. For the sake of counterexample, consider and for any two disjoint subsets of general domain and compute the measure of both sets under parallelogram law.

Complex vector spaces

For vector spaces over the complex numbers, the above formulas are not quite correct because they do not describe the imaginary part of the (complex) inner product. However, an analogous expression does ensure that both real and imaginary parts are retained. The complex part of the inner product depends on whether it is antilinear in the first or the second argument. The notation which is commonly used in physics will be assumed to be antilinear in the first argument while which is commonly used in mathematics, will be assumed to be antilinear its the second argument. They are related by the formula:

The real part of any inner product (no matter which argument is antilinear and no matter if it is real or complex) is a symmetric bilinear map that for any is always equal to: [4] [proof 1]

It is always a symmetric map, meaning that [proof 1]

and it also satisfies: [proof 1]

Thus which in plain English says that to move a factor of to the other argument, introduce a negative sign.

Proof of properties of

Let

Then implies

and

Moreover,

which proves that

From it follows that and so that

which proves that

Unlike its real part, the imaginary part of a complex inner product depends on which argument is antilinear.

Antilinear in first argument

The polarization identities for the inner product which is antilinear in the first argument, are

where The second to last equality is similar to the formula expressing a linear functional in terms of its real part:

Antilinear in second argument

The polarization identities for the inner product which is antilinear in the second argument, follows from that of by the relationship: So for any [4]

This expression can be phrased symmetrically as: [5]

Summary of both cases

Thus if denotes the real and imaginary parts of some inner product's value at the point of its domain, then its imaginary part will be:

where the scalar is always located in the same argument that the inner product is antilinear in.

Using the above formula for the imaginary part becomes:

Reconstructing the inner product

In a normed space if the parallelogram law

holds, then there exists a unique inner product on such that for all [4] [1]

Proof

We will only give the real case here; the proof for complex vector spaces is analogous.

By the above formulas, if the norm is described by an inner product (as we hope), then it must satisfy

which may serve as a definition of the unique candidate for the role of a suitable inner product. Thus, the uniqueness is guaranteed.

It remains to prove that this formula indeed defines an inner product and that this inner product induces the norm Explicitly, the following will be shown:

(This axiomatization omits positivity, which is implied by (1) and the fact that is a norm.)

For properties (1) and (2), substitute: and

For property (3), it is convenient to work in reverse. It remains to show that

or equivalently,

Now apply the parallelogram identity:

Thus it remains to verify:

But the latter claim can be verified by subtracting the following two further applications of the parallelogram identity:

Thus (3) holds.

It can be verified by induction that (3) implies (4), as long as But "(4) when " implies "(4) when ". And any positive-definite, real-valued, -bilinear form satisfies the Cauchy–Schwarz inequality, so that is continuous. Thus must be -linear as well.

Another necessary and sufficient condition for there to exist an inner product that induces a given norm is for the norm to satisfy Ptolemy's inequality, which is: [6]

Applications and consequences

If is a complex Hilbert space then is real if and only if its imaginary part is which happens if and only if Similarly, is (purely) imaginary if and only if For example, from it can be concluded that is real and that is purely imaginary.

Isometries

If is a linear isometry between two Hilbert spaces (so for all ) then

that is, linear isometries preserve inner products.

If is instead an antilinear isometry then

Relation to the law of cosines

The second form of the polarization identity can be written as

This is essentially a vector form of the law of cosines for the triangle formed by the vectors and In particular,

where is the angle between the vectors and

The equation is numerically unstable if u and v are similar because of catastrophic cancellation and should be avoided for numeric computation.

Derivation

The basic relation between the norm and the dot product is given by the equation

Then

and similarly

Forms (1) and (2) of the polarization identity now follow by solving these equations for while form (3) follows from subtracting these two equations. (Adding these two equations together gives the parallelogram law.)

Generalizations

Symmetric bilinear forms

The polarization identities are not restricted to inner products. If is any symmetric bilinear form on a vector space, and is the quadratic form defined by

then

The so-called symmetrization map generalizes the latter formula, replacing by a homogeneous polynomial of degree defined by where is a symmetric -linear map. [7]

The formulas above even apply in the case where the field of scalars has characteristic two, though the left-hand sides are all zero in this case. Consequently, in characteristic two there is no formula for a symmetric bilinear form in terms of a quadratic form, and they are in fact distinct notions, a fact which has important consequences in L-theory; for brevity, in this context "symmetric bilinear forms" are often referred to as "symmetric forms".

These formulas also apply to bilinear forms on modules over a commutative ring, though again one can only solve for if 2 is invertible in the ring, and otherwise these are distinct notions. For example, over the integers, one distinguishes integral quadratic forms from integral symmetric forms, which are a narrower notion.

More generally, in the presence of a ring involution or where 2 is not invertible, one distinguishes -quadratic forms and -symmetric forms; a symmetric form defines a quadratic form, and the polarization identity (without a factor of 2) from a quadratic form to a symmetric form is called the "symmetrization map", and is not in general an isomorphism. This has historically been a subtle distinction: over the integers it was not until the 1950s that relation between "twos out" (integral quadratic form) and "twos in" (integral symmetric form) was understood – see discussion at integral quadratic form; and in the algebraization of surgery theory, Mishchenko originally used symmetricL-groups, rather than the correct quadraticL-groups (as in Wall and Ranicki) – see discussion at L-theory.

Homogeneous polynomials of higher degree

Finally, in any of these contexts these identities may be extended to homogeneous polynomials (that is, algebraic forms) of arbitrary degree, where it is known as the polarization formula, and is reviewed in greater detail in the article on the polarization of an algebraic form.

See also

Notes and references

  1. 1 2 Lax 2002, p. 53.
  2. Philippe Blanchard, Erwin Brüning (2003). "Proposition 14.1.2 (Fréchet–von Neumann–Jordan)". Mathematical methods in physics: distributions, Hilbert space operators, and variational methods. Birkhäuser. p. 192. ISBN   0817642285.
  3. Gerald Teschl (2009). "Theorem 0.19 (Jordan–von Neumann)". Mathematical methods in quantum mechanics: with applications to Schrödinger operators. American Mathematical Society Bookstore. p. 19. ISBN   978-0-8218-4660-5.
  4. 1 2 3 4 Schechter 1996, pp. 601–603.
  5. Butler, Jon (20 June 2013). "norm - Derivation of the polarization identities?". Mathematics Stack Exchange. Archived from the original on 14 October 2020. Retrieved 2020-10-14. See Harald Hanche-Olson's answer.
  6. Apostol, Tom M. (1967). "Ptolemy's Inequality and the Chordal Metric". Mathematics Magazine. 40 (5): 233–235. doi:10.2307/2688275. JSTOR   2688275.
  7. Butler 2013. See Keith Conrad (KCd)'s answer.
    1. 1 2 3 4 A proof can be found here.

    Bibliography

    Related Research Articles

    In mathematics, more specifically in functional analysis, a Banach space is a complete normed vector space. Thus, a Banach space is a vector space with a metric that allows the computation of vector length and distance between vectors and is complete in the sense that a Cauchy sequence of vectors always converges to a well-defined limit that is within the space.

    Bra–ket notation, also called Dirac notation, is a notation for linear algebra and linear operators on complex vector spaces together with their dual space both in the finite-dimensional and infinite-dimensional case. It is specifically designed to ease the types of calculations that frequently come up in quantum mechanics. Its use in quantum mechanics is quite widespread.

    <span class="mw-page-title-main">Inner product space</span> Generalization of the dot product; used to define Hilbert spaces

    In mathematics, an inner product space is a real vector space or a complex vector space with an operation called an inner product. The inner product of two vectors in the space is a scalar, often denoted with angle brackets such as in . Inner products allow formal definitions of intuitive geometric notions, such as lengths, angles, and orthogonality of vectors. Inner product spaces generalize Euclidean vector spaces, in which the inner product is the dot product or scalar product of Cartesian coordinates. Inner product spaces of infinite dimension are widely used in functional analysis. Inner product spaces over the field of complex numbers are sometimes referred to as unitary spaces. The first usage of the concept of a vector space with an inner product is due to Giuseppe Peano, in 1898.

    The Riesz representation theorem, sometimes called the Riesz–Fréchet representation theorem after Frigyes Riesz and Maurice René Fréchet, establishes an important connection between a Hilbert space and its continuous dual space. If the underlying field is the real numbers, the two are isometrically isomorphic; if the underlying field is the complex numbers, the two are isometrically anti-isomorphic. The (anti-) isomorphism is a particular natural isomorphism.

    The Cauchy–Schwarz inequality is an upper bound on the inner product between two vectors in an inner product space in terms of the product of the vector norms. It is considered one of the most important and widely used inequalities in mathematics.

    In linear algebra, two vectors in an inner product space are orthonormal if they are orthogonal unit vectors. A unit vector means that the vector has a length of 1, which is also known as normalized. Orthogonal means that the vectors are all perpendicular to each other. A set of vectors form an orthonormal set if all vectors in the set are mutually orthogonal and all of unit length. An orthonormal set which forms a basis is called an orthonormal basis.

    In mathematics, particularly linear algebra, an orthonormal basis for an inner product space V with finite dimension is a basis for whose vectors are orthonormal, that is, they are all unit vectors and orthogonal to each other. For example, the standard basis for a Euclidean space is an orthonormal basis, where the relevant inner product is the dot product of vectors. The image of the standard basis under a rotation or reflection is also orthonormal, and every orthonormal basis for arises in this fashion.

    In mathematics, a function between two complex vector spaces is said to be antilinear or conjugate-linear if

    In mathematics, the Hodge star operator or Hodge star is a linear map defined on the exterior algebra of a finite-dimensional oriented vector space endowed with a nondegenerate symmetric bilinear form. Applying the operator to an element of the algebra produces the Hodge dual of the element. This map was introduced by W. V. D. Hodge.

    <span class="mw-page-title-main">Parallelogram law</span> The sum of the squares of the 4 sides of a parallelogram equals that of the 2 diagonals

    In mathematics, the simplest form of the parallelogram law belongs to elementary geometry. It states that the sum of the squares of the lengths of the four sides of a parallelogram equals the sum of the squares of the lengths of the two diagonals. We use these notations for the sides: AB, BC, CD, DA. But since in Euclidean geometry a parallelogram necessarily has opposite sides equal, that is, AB = CD and BC = DA, the law can be stated as

    In mathematics as well as physics, a linear operator acting on an inner product space is called positive-semidefinite if, for every , and , where is the domain of . Positive-semidefinite operators are denoted as . The operator is said to be positive-definite, and written , if for all .

    In mathematics, specifically in operator theory, each linear operator on an inner product space defines a Hermitian adjoint operator on that space according to the rule

    In linear algebra, the Gram matrix of a set of vectors in an inner product space is the Hermitian matrix of inner products, whose entries are given by the inner product . If the vectors are the columns of matrix then the Gram matrix is in the general case that the vector coordinates are complex numbers, which simplifies to for the case that the vector coordinates are real numbers.

    In mathematics, a norm is a function from a real or complex vector space to the non-negative real numbers that behaves in certain ways like the distance from the origin: it commutes with scaling, obeys a form of the triangle inequality, and is zero only at the origin. In particular, the Euclidean distance in a Euclidean space is defined by a norm on the associated Euclidean vector space, called the Euclidean norm, the 2-norm, or, sometimes, the magnitude of the vector. This norm can be defined as the square root of the inner product of a vector with itself.

    In the mathematical fields of linear algebra and functional analysis, the orthogonal complement of a subspace W of a vector space V equipped with a bilinear form B is the set W of all vectors in V that are orthogonal to every vector in W. Informally, it is called the perp, short for perpendicular complement. It is a subspace of V.

    <span class="mw-page-title-main">LSZ reduction formula</span> Connection between correlation functions and the S-matrix

    In quantum field theory, the Lehmann–Symanzik–Zimmermann (LSZ) reduction formula is a method to calculate S-matrix elements from the time-ordered correlation functions of a quantum field theory. It is a step of the path that starts from the Lagrangian of some quantum field theory and leads to prediction of measurable quantities. It is named after the three German physicists Harry Lehmann, Kurt Symanzik and Wolfhart Zimmermann.

    <span class="mw-page-title-main">Ptolemy's inequality</span>

    In Euclidean geometry, Ptolemy's inequality relates the six distances determined by four points in the plane or in a higher-dimensional space. It states that, for any four points A, B, C, and D, the following inequality holds:

    In functional analysis, the dual norm is a measure of size for a continuous linear function defined on a normed vector space.

    <span class="mw-page-title-main">Hilbert space</span> Type of topological vector space

    In mathematics, Hilbert spaces allow the methods of linear algebra and calculus to be generalized from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise naturally and frequently in mathematics and physics, typically as function spaces. Formally, a Hilbert space is a vector space equipped with an inner product that induces a distance function for which the space is a complete metric space.

    In discrete mathematics, ideal lattices are a special class of lattices and a generalization of cyclic lattices. Ideal lattices naturally occur in many parts of number theory, but also in other areas. In particular, they have a significant place in cryptography. Micciancio defined a generalization of cyclic lattices as ideal lattices. They can be used in cryptosystems to decrease by a square root the number of parameters necessary to describe a lattice, making them more efficient. Ideal lattices are a new concept, but similar lattice classes have been used for a long time. For example, cyclic lattices, a special case of ideal lattices, are used in NTRUEncrypt and NTRUSign.