Overcompleteness

Last updated

Overcompleteness is a concept from linear algebra that is widely used in mathematics, computer science, engineering, and statistics (usually in the form of overcomplete frames). It was introduced by R. J. Duffin and A. C. Schaeffer in 1952. [1]

Contents

Formally, a subset of the vectors of a Banach space , sometimes called a "system", is complete if every element in can be approximated arbitrarily well in norm by finite linear combinations of elements in . [2] A system is called overcomplete if it contains more vectors than necessary to be complete, i.e., there exist that can be removed from the system such that remains complete. In research areas such as signal processing and function approximation, overcompleteness can help researchers to achieve a more stable, more robust, or more compact decomposition than using a basis. [3]

Relation between overcompleteness and frames

The theory of frames originates in a paper by Duffin and Schaeffer on non-harmonic Fourier series. [1] A frame is defined to be a set of non-zero vectors such that for an arbitrary ,

where denotes the inner product, and are positive constants called bounds of the frame. When and can be chosen such that , the frame is called a tight frame. [4]

It can be seen that . An example of frame can be given as follows. Let each of and be an orthonormal basis of , then

is a frame of with bounds .

Let be the frame operator,

A frame that is not a Riesz basis, in which case it consists of a set of functions more than a basis, is said to be overcomplete or redundant. [5] In this case, given , it can have different decompositions based on the frame. The frame given in the example above is an overcomplete frame.

When frames are used for function estimation, one may want to compare the performance of different frames. The parsimony of the approximating functions by different frames may be considered as one way to compare their performances. [6]

Given a tolerance and a frame in , for any function , define the set of all approximating functions that satisfy

Then let

indicates the parsimony of utilizing frame to approximate . Different may have different based on the hardness to be approximated with elements in the frame. The worst case to estimate a function in is defined as

For another frame , if , then frame is better than frame at level . And if there exists a that for each , we have , then is better than broadly.

Overcomplete frames are usually constructed in three ways.

  1. Combine a set of bases, such as wavelet basis and Fourier basis, to obtain an overcomplete frame.
  2. Enlarge the range of parameters in some frame, such as in Gabor frame and wavelet frame, to have an overcomplete frame.
  3. Add some other functions to an existing complete basis to achieve an overcomplete frame.

An example of an overcomplete frame is shown below. The collected data is in a two-dimensional space, and in this case a basis with two elements should be able to explain all the data. However, when noise is included in the data, a basis may not be able to express the properties of the data. If an overcomplete frame with four elements corresponding to the four axes in the figure is used to express the data, each point would be able to have a good expression by the overcomplete frame.

The flexibility of the overcomplete frame is one of its key advantages when used in expressing a signal or approximating a function. However, because of this redundancy, a function can have multiple expressions under an overcomplete frame. [7] When the frame is finite, the decomposition can be expressed as

where is the function one wants to approximate, is the matrix containing all the elements in the frame, and is the coefficients of under the representation of . Without any other constraint, the frame will choose to give with minimal norm in . Based on this, some other properties may also be considered when solving the equation, such as sparsity. So different researchers have been working on solving this equation by adding other constraints in the objective function. For example, a constraint minimizing 's norm in may be used in solving this equation. This should be equivalent to the Lasso regression in statistics community. Bayesian approach is also used to eliminate the redundancy in an overcomplete frame. Lweicki and Sejnowski proposed an algorithm for overcomplete frame by viewing it as a probabilistic model of the observed data. [7] Recently, the overcomplete Gabor frame has been combined with bayesian variable selection method to achieve both small norm expansion coefficients in and sparsity in elements. [8]

Examples of overcomplete frames

In modern analysis in signal processing and other engineering field, various overcomplete frames are proposed and used. Here two common used frames, Gabor frames and wavelet frames, are introduced and discussed.

Gabor frames

In usual Fourier transformation, the function in time domain is transformed to the frequency domain. However, the transformation only shows the frequency property of this function and loses its information in the time domain. If a window function , which only has nonzero value in a small interval, is multiplied with the original function before operating the Fourier transformation, both the information in time and frequency domains may remain at the chosen interval. When a sequence of translation of is used in the transformation, the information of the function in time domain are kept after the transformation.

Let operators

A Gabor frame (named after Dennis Gabor and also called Weyl-Heisenberg frame) in is defined as the form , where and is a fixed function. [5] However, not for every and forms a frame on . For example, when , it is not a frame for . When , is possible to be a frame, in which case it is a Riesz basis. So the possible situation for being an overcomplete frame is . The Gabor family is also a frame and sharing the same frame bounds as

Different kinds of window function may be used in Gabor frame. Here examples of three window functions are shown, and the condition for the corresponding Gabor system being a frame is shown as follows.

(1) , is a frame when

(2) , is a frame when

(3) , where is the indicator function. The situation for to be a frame stands as follows.

1) or , not a frame

2) and , not a frame

3) , is a frame

4) and is an irrational, and , is a frame

5) , and are relatively primes, , not a frame

6) and , where and be a natural number, not a frame

7) , , , where is the biggest integer not exceeding , is a frame.

The above discussion is a summary of chapter 8 in. [5]

Wavelet frames

A collection of wavelet usually refers to a set of functions based on

This forms an orthonormal basis for . However, when can take values in , the set represents an overcomplete frame and called undecimated wavelet basis. In general case, a wavelet frame is defined as a frame for of the form

where , , and . The upper and lower bound of this frame can be computed as follows. Let be the Fourier transform for

When are fixed, define

Then

Furthermore, when

, for all odd integers

the generated frame is a tight frame.

The discussion in this section is based on chapter 11 in. [5]

Applications

Overcomplete Gabor frames and Wavelet frames have been used in various research area including signal detection, image representation, object recognition, noise reduction, sampling theory, operator theory, harmonic analysis, nonlinear sparse approximation, pseudodifferential operators, wireless communications, geophysics, quantum computing, and filter banks. [3] [5]

Related Research Articles

In particle physics, the Dirac equation is a relativistic wave equation derived by British physicist Paul Dirac in 1928. In its free form, or including electromagnetic interactions, it describes all spin-12 massive particles, called "Dirac particles", such as electrons and quarks for which parity is a symmetry. It is consistent with both the principles of quantum mechanics and the theory of special relativity, and was the first theory to account fully for special relativity in the context of quantum mechanics. It was validated by accounting for the fine structure of the hydrogen spectrum in a completely rigorous way.

<span class="mw-page-title-main">Quantum decoherence</span> Loss of quantum coherence

Quantum decoherence is the loss of quantum coherence, the process in which a system's behaviour changes from that which can be explained by quantum mechanics to that which can be explained by classical mechanics. In quantum mechanics, particles such as electrons are described by a wave function, a mathematical representation of the quantum state of a system; a probabilistic interpretation of the wave function is used to explain various quantum effects. As long as there exists a definite phase relation between different states, the system is said to be coherent. A definite phase relationship is necessary to perform quantum computing on quantum information encoded in quantum states. Coherence is preserved under the laws of quantum physics.

In physics, an operator is a function over a space of physical states onto another space of physical states. The simplest example of the utility of operators is the study of symmetry. Because of this, they are useful tools in classical mechanics. Operators are even more important in quantum mechanics, where they form an intrinsic part of the formulation of the theory.

In mathematics, the covariant derivative is a way of specifying a derivative along tangent vectors of a manifold. Alternatively, the covariant derivative is a way of introducing and working with a connection on a manifold by means of a differential operator, to be contrasted with the approach given by a principal connection on the frame bundle – see affine connection. In the special case of a manifold isometrically embedded into a higher-dimensional Euclidean space, the covariant derivative can be viewed as the orthogonal projection of the Euclidean directional derivative onto the manifold's tangent space. In this case the Euclidean derivative is broken into two parts, the extrinsic normal component and the intrinsic covariant derivative component.

The Gram–Charlier A series, and the Edgeworth series are series that approximate a probability distribution in terms of its cumulants. The series are the same; but, the arrangement of terms differ. The key idea of these expansions is to write the characteristic function of the distribution whose probability density function f is to be approximated in terms of the characteristic function of a distribution with known and suitable properties, and to recover f through the inverse Fourier transform.

<span class="mw-page-title-main">Electromagnetic tensor</span> Mathematical object that describes the electromagnetic field in spacetime

In electromagnetism, the electromagnetic tensor or electromagnetic field tensor is a mathematical object that describes the electromagnetic field in spacetime. The field tensor was first used after the four-dimensional tensor formulation of special relativity was introduced by Hermann Minkowski. The tensor allows related physical laws to be written very concisely, and allows for the quantization of the electromagnetic field by Lagrangian formulation described below.

In theoretical physics, the Wess–Zumino model has become the first known example of an interacting four-dimensional quantum field theory with linearly realised supersymmetry. In 1974, Julius Wess and Bruno Zumino studied, using modern terminology, dynamics of a single chiral superfield whose cubic superpotential leads to a renormalizable theory.

The time-evolving block decimation (TEBD) algorithm is a numerical scheme used to simulate one-dimensional quantum many-body systems, characterized by at most nearest-neighbour interactions. It is dubbed Time-evolving Block Decimation because it dynamically identifies the relevant low-dimensional Hilbert subspaces of an exponentially larger original Hilbert space. The algorithm, based on the Matrix Product States formalism, is highly efficient when the amount of entanglement in the system is limited, a requirement fulfilled by a large class of quantum many-body systems in one dimension.

The Newman–Penrose (NP) formalism is a set of notation developed by Ezra T. Newman and Roger Penrose for general relativity (GR). Their notation is an effort to treat general relativity in terms of spinor notation, which introduces complex forms of the usual variables used in GR. The NP formalism is itself a special case of the tetrad formalism, where the tensors of the theory are projected onto a complete vector basis at each point in spacetime. Usually this vector basis is chosen to reflect some symmetry of the spacetime, leading to simplified expressions for physical observables. In the case of the NP formalism, the vector basis chosen is a null tetrad: a set of four null vectors—two real, and a complex-conjugate pair. The two real members often asymptotically point radially inward and radially outward, and the formalism is well adapted to treatment of the propagation of radiation in curved spacetime. The Weyl scalars, derived from the Weyl tensor, are often used. In particular, it can be shown that one of these scalars— in the appropriate frame—encodes the outgoing gravitational radiation of an asymptotically flat system.

The theoretical and experimental justification for the Schrödinger equation motivates the discovery of the Schrödinger equation, the equation that describes the dynamics of nonrelativistic particles. The motivation uses photons, which are relativistic particles with dynamics described by Maxwell's equations, as an analogue for all types of particles.

<span class="mw-page-title-main">Mathematical descriptions of the electromagnetic field</span> Formulations of electromagnetism

There are various mathematical descriptions of the electromagnetic field that are used in the study of electromagnetism, one of the four fundamental interactions of nature. In this article, several approaches are discussed, although the equations are in terms of electric and magnetic fields, potentials, and charges with currents, generally speaking.

In mathematical physics, spacetime algebra (STA) is a name for the Clifford algebra Cl1,3(R), or equivalently the geometric algebra G(M4). According to David Hestenes, spacetime algebra can be particularly closely associated with the geometry of special relativity and relativistic spacetime.

A hydrogen-like atom (or hydrogenic atom) is any atom or ion with a single valence electron. These atoms are isoelectronic with hydrogen. Examples of hydrogen-like atoms include, but are not limited to, hydrogen itself, all alkali metals such as Rb and Cs, singly ionized alkaline earth metals such as Ca+ and Sr+ and other ions such as He+, Li2+, and Be3+ and isotopes of any of the above. A hydrogen-like atom includes a positively charged core consisting of the atomic nucleus and any core electrons as well as a single valence electron. Because helium is common in the universe, the spectroscopy of singly ionized helium is important in EUV astronomy, for example, of DO white dwarf stars.

In 1927, a year after the publication of the Schrödinger equation, Hartree formulated what are now known as the Hartree equations for atoms, using the concept of self-consistency that Lindsay had introduced in his study of many electron systems in the context of Bohr theory. Hartree assumed that the nucleus together with the electrons formed a spherically symmetric field. The charge distribution of each electron was the solution of the Schrödinger equation for an electron in a potential , derived from the field. Self-consistency required that the final field, computed from the solutions, was self-consistent with the initial field, and he thus called his method the self-consistent field method.

<span class="mw-page-title-main">Gravitational lensing formalism</span>

In general relativity, a point mass deflects a light ray with impact parameter by an angle approximately equal to

Coherent states have been introduced in a physical context, first as quasi-classical states in quantum mechanics, then as the backbone of quantum optics and they are described in that spirit in the article Coherent states. However, they have generated a huge variety of generalizations, which have led to a tremendous amount of literature in mathematical physics. In this article, we sketch the main directions of research on this line. For further details, we refer to several existing surveys.

In the ADM formulation of general relativity one splits spacetime into spatial slices and time, the basic variables are taken to be the induced metric, , on the spatial slice, and its conjugate momentum variable related to the extrinsic curvature, ,. These are the metric canonical coordinates.

<span class="mw-page-title-main">Loop representation in gauge theories and quantum gravity</span> Description of gauge theories using loop operators

Attempts have been made to describe gauge theories in terms of extended objects such as Wilson loops and holonomies. The loop representation is a quantum hamiltonian representation of gauge theories in terms of loops. The aim of the loop representation in the context of Yang–Mills theories is to avoid the redundancy introduced by Gauss gauge symmetries allowing to work directly in the space of physical states. The idea is well known in the context of lattice Yang–Mills theory. Attempts to explore the continuous loop representation was made by Gambini and Trias for canonical Yang–Mills theory, however there were difficulties as they represented singular objects. As we shall see the loop formalism goes far beyond a simple gauge invariant description, in fact it is the natural geometrical framework to treat gauge theories and quantum gravity in terms of their fundamental physical excitations.

In set theory and logic, Buchholz's ID hierarchy is a hierarchy of subsystems of first-order arithmetic. The systems/theories are referred to as "the formal theories of ν-times iterated inductive definitions". IDν extends PA by ν iterated least fixed points of monotone operators.

In theoretical physics, more specifically in quantum field theory and supersymmetry, supersymmetric Yang–Mills, also known as super Yang–Mills and abbreviated to SYM, is a supersymmetric generalization of Yang–Mills theory, which is a gauge theory that plays an important part in the mathematical formulation of forces in particle physics.

References

  1. 1 2 R. J. Duffin and A. C. Schaeffer, A class of nonharmonic Fourier series, Transactions of the American Mathematical Society, vol. 72, no. 2, pp. 341{366, 1952. [Online]. Available: https://www.jstor.org/stable/1990760
  2. C. Heil, A Basis Theory Primer: Expanded Edition. Boston, MA: Birkhauser, 2010.
  3. 1 2 R. Balan, P. Casazza, C. Heil, and Z. Landau, Density, overcompleteness, and localization of frames. I. theory, Journal of Fourier Analysis and Applications, vol. 12, no. 2, 2006.
  4. K. Grochenig, Foundations of time-frequency analysis. Boston, MA: Birkhauser, 2000.
  5. 1 2 3 4 5 O. Christensen, An Introduction to Frames and Riesz Bases. Boston, MA: Birkhauser, 2003.
  6. , STA218, Data Mining Class Note at Duke University
  7. 1 2 M. S. Lewicki and T. J. Sejnowski, Learning overcomplete representations, Neural Computation, vol. 12, no. 2, pp. 337{365, 2000.
  8. P. Wolfe, S. Godsill, and W. Ng, Bayesian variable selection and regularization for time-frequency surface estimation, J. R. Statist. Soc. B, vol. 66, no. 3, 2004.