Parseval's theorem

Last updated

In mathematics, Parseval's theorem [1] usually refers to the result that the Fourier transform is unitary; loosely, that the sum (or integral) of the square of a function is equal to the sum (or integral) of the square of its transform. It originates from a 1799 theorem about series by Marc-Antoine Parseval, which was later applied to the Fourier series. It is also known as Rayleigh's energy theorem, or Rayleigh's identity, after John William Strutt, Lord Rayleigh. [2]

Contents

Although the term "Parseval's theorem" is often used to describe the unitarity of any Fourier transform, especially in physics, the most general form of this property is more properly called the Plancherel theorem. [3]

Statement of Parseval's theorem

Suppose that and are two complex-valued functions on of period that are square integrable (with respect to the Lebesgue measure) over intervals of period length, with Fourier series

and

respectively. Then

 

 

 

 

(Eq.1)

where is the imaginary unit and horizontal bars indicate complex conjugation. Substituting and :

As is the case with the middle terms in this example, many terms will integrate to over a full period of length (see harmonics):

More generally, if and are instead two complex-valued functions on of period that are square integrable (with respect to the Lebesgue measure) over intervals of period length, with Fourier series

and

respectively. Then

 

 

 

 

(Eq.2)

Even more generally, given an abelian locally compact group G with Pontryagin dual G^, Parseval's theorem says the Pontryagin–Fourier transform is a unitary operator between Hilbert spaces L2(G) and L2(G^) (with integration being against the appropriately scaled Haar measures on the two groups.) When G is the unit circle T, G^ is the integers and this is the case discussed above. When G is the real line , G^ is also and the unitary transform is the Fourier transform on the real line. When G is the cyclic group Zn, again it is self-dual and the Pontryagin–Fourier transform is what is called discrete Fourier transform in applied contexts.

Parseval's theorem can also be expressed as follows: Suppose is a square-integrable function over (i.e., and are integrable on that interval), with the Fourier series

Then [4] [5] [6]

Notation used in engineering

In electrical engineering, Parseval's theorem is often written as:

where represents the continuous Fourier transform (in normalized, unitary form) of , and is frequency in radians per second.

The interpretation of this form of the theorem is that the total energy of a signal can be calculated by summing power-per-sample across time or spectral power across frequency.

For discrete time signals, the theorem becomes:

where is the discrete-time Fourier transform (DTFT) of and represents the angular frequency (in radians per sample) of .

Alternatively, for the discrete Fourier transform (DFT), the relation becomes:

where is the DFT of , both of length .

We show the DFT case below. For the other cases, the proof is similar. By using the definition of inverse DFT of , we can derive

where represents complex conjugate.

See also

Parseval's theorem is closely related to other mathematical results involving unitary transformations:

Notes

  1. Parseval des Chênes, Marc-Antoine Mémoire sur les séries et sur l'intégration complète d'une équation aux différences partielles linéaire du second ordre, à coefficients constants" presented before the Académie des Sciences (Paris) on 5 April 1799. This article was published in Mémoires présentés à l’Institut des Sciences, Lettres et Arts, par divers savants, et lus dans ses assemblées. Sciences, mathématiques et physiques. (Savants étrangers.), vol. 1, pages 638–648 (1806).
  2. Rayleigh, J.W.S. (1889) "On the character of the complete radiation at a given temperature," Philosophical Magazine, vol. 27, pages 460–469. Available on-line here.
  3. Plancherel, Michel (1910) "Contribution à l'etude de la representation d'une fonction arbitraire par les integrales définies," Rendiconti del Circolo Matematico di Palermo, vol. 30, pages 298–335.
  4. Arthur E. Danese (1965). Advanced Calculus. Vol. 1. Boston, MA: Allyn and Bacon, Inc. p. 439.
  5. Wilfred Kaplan (1991). Advanced Calculus (4th ed.). Reading, MA: Addison Wesley. p.  519. ISBN   0-201-57888-3.
  6. Georgi P. Tolstov (1962). Fourier Series . Translated by Silverman, Richard. Englewood Cliffs, NJ: Prentice-Hall, Inc. p.  119.

Related Research Articles

<span class="mw-page-title-main">Fourier transform</span> Mathematical transform that expresses a function of time as a function of frequency

In physics, engineering and mathematics, the Fourier transform (FT) is an integral transform that converts a function into a form that describes the frequencies present in the original function. The output of the transform is a complex-valued function of frequency. The term Fourier transform refers to both this complex-valued function and the mathematical operation. When a distinction needs to be made the Fourier transform is sometimes called the frequency domain representation of the original function. The Fourier transform is analogous to decomposing the sound of a musical chord into the intensities of its constituent pitches.

<span class="mw-page-title-main">Fourier series</span> Decomposition of periodic functions into sums of simpler sinusoidal forms

A Fourier series is an expansion of a periodic function into a sum of trigonometric functions. The Fourier series is an example of a trigonometric series, but not all trigonometric series are Fourier series. By expressing a function as a sum of sines and cosines, many problems involving the function become easier to analyze because trigonometric functions are well understood. For example, Fourier series were first used by Joseph Fourier to find solutions to the heat equation. This application is possible because the derivatives of trigonometric functions fall into simple patterns. Fourier series cannot be used to approximate arbitrary functions, because most functions have infinitely many terms in their Fourier series, and the series do not always converge. Well-behaved functions, for example smooth functions, have Fourier series that converge to the original function. The coefficients of the Fourier series are determined by integrals of the function multiplied by trigonometric functions, described in Common forms of the Fourier series below.

<span class="mw-page-title-main">Residue theorem</span> Concept of complex analysis

In complex analysis, the residue theorem, sometimes called Cauchy's residue theorem, is a powerful tool to evaluate line integrals of analytic functions over closed curves; it can often be used to compute real integrals and infinite series as well. It generalizes the Cauchy integral theorem and Cauchy's integral formula. The residue theorem should not be confused with special cases of the generalized Stokes' theorem; however, the latter can be used as an ingredient of its proof.

In calculus, and more generally in mathematical analysis, integration by parts or partial integration is a process that finds the integral of a product of functions in terms of the integral of the product of their derivative and antiderivative. It is frequently used to transform the antiderivative of a product of functions into an antiderivative for which a solution can be more easily found. The rule can be thought of as an integral version of the product rule of differentiation; it is indeed derived using the product rule.

In vector calculus, the divergence theorem, also known as Gauss's theorem or Ostrogradsky's theorem, is a theorem relating the flux of a vector field through a closed surface to the divergence of the field in the volume enclosed.

<span class="mw-page-title-main">Error function</span> Sigmoid shape special function

In mathematics, the error function, often denoted by erf, is a function defined as:

In mathematics, the Fourier inversion theorem says that for many types of functions it is possible to recover a function from its Fourier transform. Intuitively it may be viewed as the statement that if we know all frequency and phase information about a wave then we may reconstruct the original wave precisely.

In mathematical analysis, Parseval's identity, named after Marc-Antoine Parseval, is a fundamental result on the summability of the Fourier series of a function. The identity asserts the equality of the energy of a periodic signal and the energy of its frequency domain representation. Geometrically, it is a generalized Pythagorean theorem for inner-product spaces.

The Basel problem is a problem in mathematical analysis with relevance to number theory, concerning an infinite sum of inverse squares. It was first posed by Pietro Mengoli in 1650 and solved by Leonhard Euler in 1734, and read on 5 December 1735 in The Saint Petersburg Academy of Sciences. Since the problem had withstood the attacks of the leading mathematicians of the day, Euler's solution brought him immediate fame when he was twenty-eight. Euler generalised the problem considerably, and his ideas were taken up more than a century later by Bernhard Riemann in his seminal 1859 paper "On the Number of Primes Less Than a Given Magnitude", in which he defined his zeta function and proved its basic properties. The problem is named after Basel, hometown of Euler as well as of the Bernoulli family who unsuccessfully attacked the problem.

In mathematics, the Poisson summation formula is an equation that relates the Fourier series coefficients of the periodic summation of a function to values of the function's continuous Fourier transform. Consequently, the periodic summation of a function is completely defined by discrete samples of the original function's Fourier transform. And conversely, the periodic summation of a function's Fourier transform is completely defined by discrete samples of the original function. The Poisson summation formula was discovered by Siméon Denis Poisson and is sometimes called Poisson resummation.

<span class="mw-page-title-main">Gaussian integral</span> Integral of the Gaussian function, equal to sqrt(π)

The Gaussian integral, also known as the Euler–Poisson integral, is the integral of the Gaussian function over the entire real line. Named after the German mathematician Carl Friedrich Gauss, the integral is

In mathematics and signal processing, the Hilbert transform is a specific singular integral that takes a function, u(t) of a real variable and produces another function of a real variable H(u)(t). The Hilbert transform is given by the Cauchy principal value of the convolution with the function (see § Definition). The Hilbert transform has a particularly simple representation in the frequency domain: It imparts a phase shift of ±90° (π/2 radians) to every frequency component of a function, the sign of the shift depending on the sign of the frequency (see § Relationship with the Fourier transform). The Hilbert transform is important in signal processing, where it is a component of the analytic representation of a real-valued signal u(t). The Hilbert transform was first introduced by David Hilbert in this setting, to solve a special case of the Riemann–Hilbert problem for analytic functions.

In the mathematical field of complex analysis, contour integration is a method of evaluating certain integrals along paths in the complex plane.

<span class="mw-page-title-main">Dirichlet integral</span> Integral of sin(x)/x from 0 to infinity.

In mathematics, there are several integrals known as the Dirichlet integral, after the German mathematician Peter Gustav Lejeune Dirichlet, one of which is the improper integral of the sinc function over the positive real line:

<span class="mw-page-title-main">Wallis product</span> Infinite product for pi

In mathematics, the Wallis product for π, published in 1656 by John Wallis, states that

In mathematics, the Hankel transform expresses any given function f(r) as the weighted sum of an infinite number of Bessel functions of the first kind Jν(kr). The Bessel functions in the sum are all of the same order ν, but differ in a scaling factor k along the r axis. The necessary coefficient Fν of each Bessel function in the sum, as a function of the scaling factor k constitutes the transformed function. The Hankel transform is an integral transform and was first developed by the mathematician Hermann Hankel. It is also known as the Fourier–Bessel transform. Just as the Fourier transform for an infinite interval is related to the Fourier series over a finite interval, so the Hankel transform over an infinite interval is related to the Fourier–Bessel series over a finite interval.

In mathematics, the Riesz–Fischer theorem in real analysis is any of a number of closely related results concerning the properties of the space L2 of square integrable functions. The theorem was proven independently in 1907 by Frigyes Riesz and Ernst Sigismund Fischer.

In mathematics, Maass forms or Maass wave forms are studied in the theory of automorphic forms. Maass forms are complex-valued smooth functions of the upper half plane, which transform in a similar way under the operation of a discrete subgroup of as modular forms. They are eigenforms of the hyperbolic Laplace operator defined on and satisfy certain growth conditions at the cusps of a fundamental domain of . In contrast to modular forms, Maass forms need not be holomorphic. They were studied first by Hans Maass in 1949.

In mathematics, the Parseval–Gutzmer formula states that, if is an analytic function on a closed disk of radius r with Taylor series

In mathematical analysis and applications, multidimensional transforms are used to analyze the frequency content of signals in a domain of two or more dimensions.