Trigonometric series

Last updated

In mathematics, a trigonometric series is an infinite series of the form

Contents

where is the variable and and are coefficients. It is an infinite version of a trigonometric polynomial.

A trigonometric series is called the Fourier series of the integrable function if the coefficients have the form:

Examples

The Fourier series for the identity function suffers from the Gibbs phenomenon near the ends of the periodic interval. Fourier series for the identity function.png
The Fourier series for the identity function suffers from the Gibbs phenomenon near the ends of the periodic interval.

Every Fourier series gives an example of a trigonometric series. Let the function on be extended periodically (see sawtooth wave). Then its Fourier coefficients are:

Which gives an example of a trigonometric series:

The trigonometric series sin 2x / log 2 + sin 3x / log 3 + sin 4x / log 4 + ... is not a Fourier series. Trigonometric series with wiggles and spike.png
The trigonometric series sin 2x/log 2+ sin 3x/log 3+ sin 4x/log 4 + ... is not a Fourier series.

The converse is false however, not every trigonometric series is a Fourier series. The series

is a trigonometric series which converges for all but is not a Fourier series. [1] Here for and all other coefficients are zero.

Uniqueness of Trigonometric series

The uniqueness and the zeros of trigonometric series was an active area of research in 19th century Europe. First, Georg Cantor proved that if a trigonometric series is convergent to a function on the interval , which is identically zero, or more generally, is nonzero on at most finitely many points, then the coefficients of the series are all zero. [2]

Later Cantor proved that even if the set S on which is nonzero is infinite, but the derived set S' of S is finite, then the coefficients are all zero. In fact, he proved a more general result. Let S0 = S and let Sk+1 be the derived set of Sk. If there is a finite number n for which Sn is finite, then all the coefficients are zero. Later, Lebesgue proved that if there is a countably infinite ordinal α such that Sα is finite, then the coefficients of the series are all zero. Cantor's work on the uniqueness problem famously led him to invent transfinite ordinal numbers, which appeared as the subscripts α in Sα . [3]

Notes

  1. Hardy, Godfrey Harold; Rogosinski, Werner Wolfgang (1956) [1st ed. 1944]. Fourier Series (3rd ed.). Cambridge University Press. pp. 4–5.
  2. http://www.math.caltech.edu/papers/uniqueness.pdf [ bare URL PDF ]
  3. Cooke, Roger (1993), "Uniqueness of trigonometric series and descriptive set theory, 1870–1985", Archive for History of Exact Sciences, 45 (4): 281–334, doi:10.1007/BF01886630, S2CID   122744778.

Related Research Articles

<span class="mw-page-title-main">Taylor series</span> Mathematical approximation of a function

In mathematics, the Taylor series or Taylor expansion of a function is an infinite sum of terms that are expressed in terms of the function's derivatives at a single point. For most common functions, the function and the sum of its Taylor series are equal near this point. Taylor series are named after Brook Taylor, who introduced them in 1715. A Taylor series is also called a Maclaurin series when 0 is the point where the derivatives are considered, after Colin Maclaurin, who made extensive use of this special case of Taylor series in the 18th century.

<span class="mw-page-title-main">Fourier series</span> Decomposition of periodic functions into sums of simpler sinusoidal forms

A Fourier series is an expansion of a periodic function into a sum of trigonometric functions. The Fourier series is an example of a trigonometric series, but not all trigonometric series are Fourier series. By expressing a function as a sum of sines and cosines, many problems involving the function become easier to analyze because trigonometric functions are well understood. For example, Fourier series were first used by Joseph Fourier to find solutions to the heat equation. This application is possible because the derivatives of trigonometric functions fall into simple patterns. Fourier series cannot be used to approximate arbitrary functions, because most functions have infinitely many terms in their Fourier series, and the series do not always converge. Well-behaved functions, for example smooth functions, have Fourier series that converge to the original function. The coefficients of the Fourier series are determined by integrals of the function multiplied by trigonometric functions, described in Common forms of the Fourier series below.

In linear algebra, two vectors in an inner product space are orthonormal if they are orthogonal unit vectors. A unit vector means that the vector has a length of 1, which is also known as normalized. Orthogonal means that the vectors are all perpendicular to each other. A set of vectors form an orthonormal set if all vectors in the set are mutually orthogonal and all of unit length. An orthonormal set which forms a basis is called an orthonormal basis.

Integration is the basic operation in integral calculus. While differentiation has straightforward rules by which the derivative of a complicated function can be found by differentiating its simpler component functions, integration does not, so tables of known integrals are often useful. This page lists some of the most common antiderivatives.

<span class="mw-page-title-main">Improper integral</span> Concept in mathematical analysis

In mathematical analysis, an improper integral is an extension of the notion of a definite integral to cases that violate the usual assumptions for that kind of integral. In the context of Riemann integrals, this typically involves unboundedness, either of the set over which the integral is taken or of the integrand, or both. It may also involve bounded but not closed sets or bounded but not continuous functions. While an improper integral is typically written symbolically just like a standard definite integral, it actually represents a limit of a definite integral or a sum of such limits; thus improper integrals are said to converge or diverge. If a regular definite integral is worked out as if it is improper, the same answer will result.

In mathematics, the Gibbs phenomenon is the oscillatory behavior of the Fourier series of a piecewise continuously differentiable periodic function around a jump discontinuity. The th partial Fourier series of the function produces large peaks around the jump which overshoot and undershoot the function values. As more sinusoids are used, this approximation error approaches a limit of about 9% of the jump, though the infinite Fourier series sum does eventually converge almost everywhere except points of discontinuity.

In mathematics, an almost periodic function is, loosely speaking, a function of a real number that is periodic to within any desired level of accuracy, given suitably long, well-distributed "almost-periods". The concept was first studied by Harald Bohr and later generalized by Vyacheslav Stepanov, Hermann Weyl and Abram Samoilovitch Besicovitch, amongst others. There is also a notion of almost periodic functions on locally compact abelian groups, first studied by John von Neumann.

<span class="mw-page-title-main">Clausen function</span> Transcendental single-variable function

In mathematics, the Clausen function, introduced by Thomas Clausen, is a transcendental, special function of a single variable. It can variously be expressed in the form of a definite integral, a trigonometric series, and various other forms. It is intimately connected with the polylogarithm, inverse tangent integral, polygamma function, Riemann zeta function, Dirichlet eta function, and Dirichlet beta function.

In signal processing, a finite impulse response (FIR) filter is a filter whose impulse response is of finite duration, because it settles to zero in finite time. This is in contrast to infinite impulse response (IIR) filters, which may have internal feedback and may continue to respond indefinitely.

In mathematics, orthogonal functions belong to a function space that is a vector space equipped with a bilinear form. When the function space has an interval as the domain, the bilinear form may be the integral of the product of functions over the interval:

The Basel problem is a problem in mathematical analysis with relevance to number theory, concerning an infinite sum of inverse squares. It was first posed by Pietro Mengoli in 1650 and solved by Leonhard Euler in 1734, and read on 5 December 1735 in The Saint Petersburg Academy of Sciences. Since the problem had withstood the attacks of the leading mathematicians of the day, Euler's solution brought him immediate fame when he was twenty-eight. Euler generalised the problem considerably, and his ideas were taken up more than a century later by Bernhard Riemann in his seminal 1859 paper "On the Number of Primes Less Than a Given Magnitude", in which he defined his zeta function and proved its basic properties. The problem is named after Basel, hometown of Euler as well as of the Bernoulli family who unsuccessfully attacked the problem.

In the mathematical subfields of numerical analysis and mathematical analysis, a trigonometric polynomial is a finite linear combination of functions sin(nx) and cos(nx) with n taking on the values of one or more natural numbers. The coefficients may be taken as real numbers, for real-valued functions. For complex coefficients, there is no difference between such a function and a finite Fourier series.

<span class="mw-page-title-main">Sinc function</span> Special mathematical function defined as sin(x)/x

In mathematics, physics and engineering, the sinc function, denoted by sinc(x), has two forms, normalized and unnormalized.

<span class="mw-page-title-main">Fejér kernel</span>

In mathematics, the Fejér kernel is a summability kernel used to express the effect of Cesàro summation on Fourier series. It is a non-negative kernel, giving rise to an approximate identity. It is named after the Hungarian mathematician Lipót Fejér (1880–1959).

In mathematics, Parseval's theorem usually refers to the result that the Fourier transform is unitary; loosely, that the sum of the square of a function is equal to the sum of the square of its transform. It originates from a 1799 theorem about series by Marc-Antoine Parseval, which was later applied to the Fourier series. It is also known as Rayleigh's energy theorem, or Rayleigh's identity, after John William Strutt, Lord Rayleigh.

In mathematics, a half range Fourier series is a Fourier series defined on an interval instead of the more common , with the implication that the analyzed function should be extended to as either an even or odd function. This allows the expansion of the function in a series solely of sines (odd) or cosines (even). The choice between odd and even is typically motivated by boundary conditions associated with a differential equation satisfied by .

In mathematics, convergence tests are methods of testing for the convergence, conditional convergence, absolute convergence, interval of convergence or divergence of an infinite series .

Clenshaw–Curtis quadrature and Fejér quadrature are methods for numerical integration, or "quadrature", that are based on an expansion of the integrand in terms of Chebyshev polynomials. Equivalently, they employ a change of variables and use a discrete cosine transform (DCT) approximation for the cosine series. Besides having fast-converging accuracy comparable to Gaussian quadrature rules, Clenshaw–Curtis quadrature naturally leads to nested quadrature rules, which is important for both adaptive quadrature and multidimensional quadrature (cubature).

In mathematics, particularly the field of calculus and Fourier analysis, the Fourier sine and cosine series are two mathematical series named after Joseph Fourier.

<span class="mw-page-title-main">Dirichlet kernel</span>

In mathematical analysis, the Dirichlet kernel, named after the German mathematician Peter Gustav Lejeune Dirichlet, is the collection of periodic functions defined as

References

See also