Trigonometric series

Last updated

In mathematics, a trigonometric series is an infinite series of the form

Contents

where is the variable and and are coefficients. It is an infinite version of a trigonometric polynomial.

A trigonometric series is called the Fourier series of the integrable function if the coefficients have the form:

Examples

The Fourier series for the identity function suffers from the Gibbs phenomenon near the ends of the periodic interval. Fourier series for the identity function.png
The Fourier series for the identity function suffers from the Gibbs phenomenon near the ends of the periodic interval.

Every Fourier series gives an example of a trigonometric series. Let the function on be extended periodically (see sawtooth wave). Then its Fourier coefficients are:

Which gives an example of a trigonometric series:

The trigonometric series sin 2x / log 2 + sin 3x / log 3 + sin 4x / log 4 + ... is not a Fourier series. Trigonometric series with wiggles and spike.png
The trigonometric series sin 2x/log 2+ sin 3x/log 3+ sin 4x/log 4 + ... is not a Fourier series.

The converse is false however, not every trigonometric series is a Fourier series. The series

is a trigonometric series which converges for all but is not a Fourier series. [1] Here for and all other coefficients are zero.

Uniqueness of Trigonometric series

The uniqueness and the zeros of trigonometric series was an active area of research in 19th century Europe. First, Georg Cantor proved that if a trigonometric series is convergent to a function on the interval , which is identically zero, or more generally, is nonzero on at most finitely many points, then the coefficients of the series are all zero. [2]

Later Cantor proved that even if the set S on which is nonzero is infinite, but the derived set S' of S is finite, then the coefficients are all zero. In fact, he proved a more general result. Let S0 = S and let Sk+1 be the derived set of Sk. If there is a finite number n for which Sn is finite, then all the coefficients are zero. Later, Lebesgue proved that if there is a countably infinite ordinal α such that Sα is finite, then the coefficients of the series are all zero. Cantor's work on the uniqueness problem famously led him to invent transfinite ordinal numbers, which appeared as the subscripts α in Sα . [3]

Notes

  1. Hardy, Godfrey Harold; Rogosinski, Werner Wolfgang (1956) [1st ed. 1944]. Fourier Series (3rd ed.). Cambridge University Press. pp. 4–5.
  2. Kechris, Alexander S. (1997). "Set theory and uniqueness for trigonometric series" (PDF). Caltech.
  3. Cooke, Roger (1993). "Uniqueness of trigonometric series and descriptive set theory, 1870–1985". Archive for History of Exact Sciences. 45 (4): 281–334. doi:10.1007/BF01886630. S2CID   122744778.{{cite journal}}: CS1 maint: postscript (link)

Related Research Articles

<span class="mw-page-title-main">Trigonometric functions</span> Functions of an angle

In mathematics, the trigonometric functions are real functions which relate an angle of a right-angled triangle to ratios of two side lengths. They are widely used in all sciences that are related to geometry, such as navigation, solid mechanics, celestial mechanics, geodesy, and many others. They are among the simplest periodic functions, and as such are also widely used for studying periodic phenomena through Fourier analysis.

<span class="mw-page-title-main">Fourier series</span> Decomposition of periodic functions into sums of simpler sinusoidal forms

A Fourier series is an expansion of a periodic function into a sum of trigonometric functions. The Fourier series is an example of a trigonometric series, but not all trigonometric series are Fourier series. By expressing a function as a sum of sines and cosines, many problems involving the function become easier to analyze because trigonometric functions are well understood. For example, Fourier series were first used by Joseph Fourier to find solutions to the heat equation. This application is possible because the derivatives of trigonometric functions fall into simple patterns. Fourier series cannot be used to approximate arbitrary functions, because most functions have infinitely many terms in their Fourier series, and the series do not always converge. Well-behaved functions, for example smooth functions, have Fourier series that converge to the original function. The coefficients of the Fourier series are determined by integrals of the function multiplied by trigonometric functions, described in Common forms of the Fourier series below.

In linear algebra, two vectors in an inner product space are orthonormal if they are orthogonal unit vectors. A unit vector means that the vector has a length of 1, which is also known as normalized. Orthogonal means that the vectors are all perpendicular to each other. A set of vectors form an orthonormal set if all vectors in the set are mutually orthogonal and all of unit length. An orthonormal set which forms a basis is called an orthonormal basis.

Integration is the basic operation in integral calculus. While differentiation has straightforward rules by which the derivative of a complicated function can be found by differentiating its simpler component functions, integration does not, so tables of known integrals are often useful. This page lists some of the most common antiderivatives.

A generalized Fourier series is the expansion of a square integrable function into a sum of square integrable orthogonal basis functions. The standard Fourier series uses an orthonormal basis of trigonometric functions, and the series expansion is applied to periodic functions. In contrast, a generalized Fourier series uses any set of orthogonal basis functions and can apply to any square integrable function.

In mathematics, the Gibbs phenomenon is the oscillatory behavior of the Fourier series of a piecewise continuously differentiable periodic function around a jump discontinuity. The th partial Fourier series of the function produces large peaks around the jump which overshoot and undershoot the function values. As more sinusoids are used, this approximation error approaches a limit of about 9% of the jump, though the infinite Fourier series sum does eventually converge almost everywhere except points of discontinuity.

In mathematics, an almost periodic function is, loosely speaking, a function of a real variable that is periodic to within any desired level of accuracy, given suitably long, well-distributed "almost-periods". The concept was first studied by Harald Bohr and later generalized by Vyacheslav Stepanov, Hermann Weyl and Abram Samoilovitch Besicovitch, amongst others. There is also a notion of almost periodic functions on locally compact abelian groups, first studied by John von Neumann.

<span class="mw-page-title-main">Clausen function</span> Transcendental single-variable function

In mathematics, the Clausen function, introduced by Thomas Clausen, is a transcendental, special function of a single variable. It can variously be expressed in the form of a definite integral, a trigonometric series, and various other forms. It is intimately connected with the polylogarithm, inverse tangent integral, polygamma function, Riemann zeta function, Dirichlet eta function, and Dirichlet beta function.

In signal processing, a finite impulse response (FIR) filter is a filter whose impulse response is of finite duration, because it settles to zero in finite time. This is in contrast to infinite impulse response (IIR) filters, which may have internal feedback and may continue to respond indefinitely.

In mathematics, orthogonal functions belong to a function space that is a vector space equipped with a bilinear form. When the function space has an interval as the domain, the bilinear form may be the integral of the product of functions over the interval:

The Basel problem is a problem in mathematical analysis with relevance to number theory, concerning an infinite sum of inverse squares. It was first posed by Pietro Mengoli in 1650 and solved by Leonhard Euler in 1734, and read on 5 December 1735 in The Saint Petersburg Academy of Sciences. Since the problem had withstood the attacks of the leading mathematicians of the day, Euler's solution brought him immediate fame when he was twenty-eight. Euler generalised the problem considerably, and his ideas were taken up more than a century later by Bernhard Riemann in his seminal 1859 paper "On the Number of Primes Less Than a Given Magnitude", in which he defined his zeta function and proved its basic properties. The problem is named after Basel, hometown of Euler as well as of the Bernoulli family who unsuccessfully attacked the problem.

In the mathematical subfields of numerical analysis and mathematical analysis, a trigonometric polynomial is a finite linear combination of functions sin(nx) and cos(nx) with n taking on the values of one or more natural numbers. The coefficients may be taken as real numbers, for real-valued functions. For complex coefficients, there is no difference between such a function and a finite Fourier series.

<span class="mw-page-title-main">Sinc function</span> Special mathematical function defined as sin(x)/x

In mathematics, physics and engineering, the sinc function, denoted by sinc(x), has two forms, normalized and unnormalized.

<span class="mw-page-title-main">Fejér kernel</span>

In mathematics, the Fejér kernel is a summability kernel used to express the effect of Cesàro summation on Fourier series. It is a non-negative kernel, giving rise to an approximate identity. It is named after the Hungarian mathematician Lipót Fejér (1880–1959).

<span class="mw-page-title-main">Dirichlet integral</span> Integral of sin(x)/x from 0 to infinity.

In mathematics, there are several integrals known as the Dirichlet integral, after the German mathematician Peter Gustav Lejeune Dirichlet, one of which is the improper integral of the sinc function over the positive real line:

<span class="mw-page-title-main">Series expansion</span> Expression of a function as an infinite sum of simpler functions

In mathematics, a series expansion is a technique that expresses a function as an infinite sum, or series, of simpler functions. It is a method for calculating a function that cannot be expressed by just elementary operators.

<span class="mw-page-title-main">Sine and cosine</span> Fundamental trigonometric functions

In mathematics, sine and cosine are trigonometric functions of an angle. The sine and cosine of an acute angle are defined in the context of a right triangle: for the specified angle, its sine is the ratio of the length of the side that is opposite that angle to the length of the longest side of the triangle, and the cosine is the ratio of the length of the adjacent leg to that of the hypotenuse. For an angle , the sine and cosine functions are denoted as and .

Clenshaw–Curtis quadrature and Fejér quadrature are methods for numerical integration, or "quadrature", that are based on an expansion of the integrand in terms of Chebyshev polynomials. Equivalently, they employ a change of variables and use a discrete cosine transform (DCT) approximation for the cosine series. Besides having fast-converging accuracy comparable to Gaussian quadrature rules, Clenshaw–Curtis quadrature naturally leads to nested quadrature rules, which is important for both adaptive quadrature and multidimensional quadrature (cubature).

In mathematics, particularly the field of calculus and Fourier analysis, the Fourier sine and cosine series are two mathematical series named after Joseph Fourier.

<span class="mw-page-title-main">Dirichlet kernel</span> Concept in mathematical analysis

In mathematical analysis, the Dirichlet kernel, named after the German mathematician Peter Gustav Lejeune Dirichlet, is the collection of periodic functions defined as

References

See also