Power series

Last updated

In mathematics, a power series (in one variable) is an infinite series of the form

Contents

where an represents the coefficient of the nth term and c is a constant. Power series are useful in mathematical analysis, where they arise as Taylor series of infinitely differentiable functions. In fact, Borel's theorem implies that every power series is the Taylor series of some smooth function.

In many situations, c (the center of the series) is equal to zero, for instance when considering a Maclaurin series. In such cases, the power series takes the simpler form

Beyond their role in mathematical analysis, power series also occur in combinatorics as generating functions (a kind of formal power series) and in electronic engineering (under the name of the Z-transform). The familiar decimal notation for real numbers can also be viewed as an example of a power series, with integer coefficients, but with the argument x fixed at 110. In number theory, the concept of p-adic numbers is also closely related to that of a power series.

Examples

Polynomial

The exponential function (in blue), and its improving approximation by the sum of the first n + 1 terms of its Maclaurin power series (in red). So
n=0 gives
f
(
x
)
=
1
{\displaystyle f(x)=1}
,
n=1
f
(
x
)
=
1
+
x
{\displaystyle f(x)=1+x}
,
n=2
f
(
x
)
=
1
+
x
+
x
2
/
2
{\displaystyle f(x)=1+x+x^{2}/2}
,
n=3
f
(
x
)
=
1
+
x
+
x
2
/
2
+
x
3
/
6
{\displaystyle f(x)=1+x+x^{2}/2+x^{3}/6}
etcetera. Exp series.gif
The exponential function (in blue), and its improving approximation by the sum of the first n +1 terms of its Maclaurin power series (in red). So
n=0 gives ,
n=1 ,
n=2 ,
n=3 etcetera.

A polynomial of degree d can be expressed as a power series around any center c, where all terms of degree higher than d have a zero coefficient. For instance, the polynomial can be written as a power series around the center as

or around the center as

This is because of the Taylor series expansion of f(x) around is

as and the non-zero derivatives are , so and , a constant.

Or indeed the expansion is possible around any other center c. [1] One can view power series as being like "polynomials of infinite degree," although power series are not polynomials.

Geometric series, exponential function and sine

The geometric series formula

which is valid for , is one of the most important examples of a power series, as are the exponential function formula

and the sine formula

valid for all real x.

These power series are also examples of Taylor series.

On the set of exponents

Negative powers are not permitted in a power series; for instance, is not considered a power series (although it is a Laurent series). Similarly, fractional powers such as are not permitted (but see Puiseux series). The coefficients are not allowed to depend on , thus for instance:

is not a power series.

Radius of convergence

A power series is convergent for some values of the variable x, which will always include x = c (as usual, evaluates as 1 and the sum of the series is thus for x = c). The series may diverge for other values of x. If c is not the only point of convergence, then there is always a number r with 0 < r ≤ ∞ such that the series converges whenever |xc| < r and diverges whenever |xc| > r. The number r is called the radius of convergence of the power series; in general it is given as

or, equivalently,

(this is the Cauchy–Hadamard theorem; see limit superior and limit inferior for an explanation of the notation). The relation

is also satisfied, if this limit exists.

The set of the complex numbers such that |xc| < r is called the disc of convergence of the series. The series converges absolutely inside its disc of convergence, and converges uniformly on every compact subset of the disc of convergence.

For |xc| = r, there is no general statement on the convergence of the series. However, Abel's theorem states that if the series is convergent for some value z such that |zc| = r, then the sum of the series for x = z is the limit of the sum of the series for x = c + t (zc) where t is a real variable less than 1 that tends to 1.

Operations on power series

Addition and subtraction

When two functions f and g are decomposed into power series around the same center c, the power series of the sum or difference of the functions can be obtained by termwise addition and subtraction. That is, if

and

then

It is not true that if two power series and have the same radius of convergence, then also has this radius of convergence. If and , then both series have the same radius of convergence of 1, but the series has a radius of convergence of 3.

The sum of two power series will have, at minimum, a radius of convergence of the smaller of the two radii of convergence of the two series (and it may be higher than either, as seen in the example above). [2]

Multiplication and division

With the same definitions for and , the power series of the product and quotient of the functions can be obtained as follows:

The sequence is known as the convolution of the sequences and .

For division, if one defines the sequence by

then

and one can solve recursively for the terms by comparing coefficients.

Solving the corresponding equations yields the formulae based on determinants of certain matrices of the coefficients of and

Differentiation and integration

Once a function is given as a power series as above, it is differentiable on the interior of the domain of convergence. It can be differentiated and integrated quite easily, by treating every term separately:

Both of these series have the same radius of convergence as the original one.

Analytic functions

A function f defined on some open subset U of R or C is called analytic if it is locally given by a convergent power series. This means that every aU has an open neighborhood VU, such that there exists a power series with center a that converges to f(x) for every xV.

Every power series with a positive radius of convergence is analytic on the interior of its region of convergence. All holomorphic functions are complex-analytic. Sums and products of analytic functions are analytic, as are quotients as long as the denominator is non-zero.

If a function is analytic, then it is infinitely differentiable, but in the real case the converse is not generally true. For an analytic function, the coefficients an can be computed as

where denotes the nth derivative of f at c, and . This means that every analytic function is locally represented by its Taylor series.

The global form of an analytic function is completely determined by its local behavior in the following sense: if f and g are two analytic functions defined on the same connected open set U, and if there exists an element cU such that f(n)(c) = g(n)(c) for all n ≥ 0, then f(x) = g(x) for all xU.

If a power series with radius of convergence r is given, one can consider analytic continuations of the series, i.e. analytic functions f which are defined on larger sets than { x | |xc| < r } and agree with the given power series on this set. The number r is maximal in the following sense: there always exists a complex number x with |xc| = r such that no analytic continuation of the series can be defined at x.

The power series expansion of the inverse function of an analytic function can be determined using the Lagrange inversion theorem.

Behavior near the boundary

The sum of a power series with a positive radius of convergence is an analytic function at every point in the interior of the disc of convergence. However, different behavior can occur at points on the boundary of that disc. For example:

  1. Divergence while the sum extends to an analytic function: has radius of convergence equal to and diverges at every point of . Nevertheless, the sum in is , which is analytic at every point of the plane except for .
  2. Convergent at some points divergent at others: has radius of convergence . It converges for , while it diverges for .
  3. Absolute convergence at every point of the boundary: has radius of convergence , while it converges absolutely, and uniformly, at every point of due to Weierstrass M-test applied with the hyper-harmonic convergent series .
  4. Convergent on the closure of the disc of convergence but not continuous sum: Sierpiński gave an example [3] of a power series with radius of convergence , convergent at all points with , but the sum is an unbounded function and, in particular, discontinuous. A sufficient condition for one-sided continuity at a boundary point is given by Abel's theorem.

Formal power series

In abstract algebra, one attempts to capture the essence of power series without being restricted to the fields of real and complex numbers, and without the need to talk about convergence. This leads to the concept of formal power series, a concept of great utility in algebraic combinatorics.

Power series in several variables

An extension of the theory is necessary for the purposes of multivariable calculus. A power series is here defined to be an infinite series of the form

where j = (j1, …, jn) is a vector of natural numbers, the coefficients a(j1, …, jn) are usually real or complex numbers, and the center c = (c1, …, cn) and argument x = (x1, …, xn) are usually real or complex vectors. The symbol is the product symbol, denoting multiplication. In the more convenient multi-index notation this can be written

where is the set of natural numbers, and so is the set of ordered n-tuples of natural numbers.

The theory of such series is trickier than for single-variable series, with more complicated regions of convergence. For instance, the power series is absolutely convergent in the set between two hyperbolas. (This is an example of a log-convex set, in the sense that the set of points , where lies in the above region, is a convex set. More generally, one can show that when c=0, the interior of the region of absolute convergence is always a log-convex set in this sense.) On the other hand, in the interior of this region of convergence one may differentiate and integrate under the series sign, just as one may with ordinary power series. [4]

Order of a power series

Let α be a multi-index for a power series f(x1, x2, …, xn). The order of the power series f is defined to be the least value such that there is aα ≠ 0 with , or if f ≡ 0. In particular, for a power series f(x) in a single variable x, the order of f is the smallest power of x with a nonzero coefficient. This definition readily extends to Laurent series.

Notes

  1. Howard Levi (1967). Polynomials, Power Series, and Calculus. Van Nostrand. p. 24.
  2. Erwin Kreyszig, Advanced Engineering Mathematics, 8th ed, page 747
  3. Wacław Sierpiński (1916). "Sur une série potentielle qui, étant convergente en tout point de son cercle de convergence, représente sur ce cercle une fonction discontinue. (French)". Rendiconti del Circolo Matematico di Palermo. Palermo Rend. 41: 187–190. doi:10.1007/BF03018294. JFM   46.1466.03. S2CID   121218640.
  4. Beckenbach, E. F. (1948). "Convex functions". Bulletin of the American Mathematical Society. 54 (5): 439–460. doi: 10.1090/S0002-9904-1948-08994-7 .

Related Research Articles

In mathematics, a series is, roughly speaking, the operation of adding infinitely many quantities, one after the other, to a given starting quantity. The study of series is a major part of calculus and its generalization, mathematical analysis. Series are used in most areas of mathematics, even for studying finite structures through generating functions. In addition to their ubiquity in mathematics, infinite series are also widely used in other quantitative disciplines such as physics, computer science, statistics and finance.

<span class="mw-page-title-main">Taylor series</span> Mathematical approximation of a function

In mathematics, the Taylor series or Taylor expansion of a function is an infinite sum of terms that are expressed in terms of the function's derivatives at a single point. For most common functions, the function and the sum of its Taylor series are equal near this point. Taylor series are named after Brook Taylor, who introduced them in 1715. A Taylor series is also called a Maclaurin series when 0 is the point where the derivatives are considered, after Colin Maclaurin, who made extensive use of this special case of Taylor series in the 18th century.

<span class="mw-page-title-main">Taylor's theorem</span> Approximation of a function by a truncated power series

In calculus, Taylor's theorem gives an approximation of a -times differentiable function around a given point by a polynomial of degree , called the -th-order Taylor polynomial. For a smooth function, the Taylor polynomial is the truncation at the order of the Taylor series of the function. The first-order Taylor polynomial is the linear approximation of the function, and the second-order Taylor polynomial is often referred to as the quadratic approximation. There are several versions of Taylor's theorem, some giving explicit estimates of the approximation error of the function by its Taylor polynomial.

<span class="mw-page-title-main">Fourier series</span> Decomposition of periodic functions into sums of simpler sinusoidal forms

A Fourier series is an expansion of a periodic function into a sum of trigonometric functions. The Fourier series is an example of a trigonometric series, but not all trigonometric series are Fourier series. By expressing a function as a sum of sines and cosines, many problems involving the function become easier to analyze because trigonometric functions are well understood. For example, Fourier series were first used by Joseph Fourier to find solutions to the heat equation. This application is possible because the derivatives of trigonometric functions fall into simple patterns. Fourier series cannot be used to approximate arbitrary functions, because most functions have infinitely many terms in their Fourier series, and the series do not always converge. Well-behaved functions, for example smooth functions, have Fourier series that converge to the original function. The coefficients of the Fourier series are determined by integrals of the function multiplied by trigonometric functions, described in Common forms of the Fourier series below.

<span class="mw-page-title-main">Laurent series</span> Power series with negative powers

In mathematics, the Laurent series of a complex function is a representation of that function as a power series which includes terms of negative degree. It may be used to express complex functions in cases where a Taylor series expansion cannot be applied. The Laurent series was named after and first published by Pierre Alphonse Laurent in 1843. Karl Weierstrass may have discovered it first in a paper written in 1841, but it was not published until after his death.

In mathematics, the radius of convergence of a power series is the radius of the largest disk at the center of the series in which the series converges. It is either a non-negative real number or . When it is positive, the power series converges absolutely and uniformly on compact sets inside the open disk of radius equal to the radius of convergence, and it is the Taylor series of the analytic function to which it converges. In case of multiple singularities of a function, the radius of convergence is the shortest or minimum of all the respective distances calculated from the center of the disk of convergence to the respective singularities of the function.

In mathematics, an infinite series of numbers is said to converge absolutely if the sum of the absolute values of the summands is finite. More precisely, a real or complex series is said to converge absolutely if for some real number Similarly, an improper integral of a function, is said to converge absolutely if the integral of the absolute value of the integrand is finite—that is, if

In mathematical analysis, the Lagrange inversion theorem, also known as the Lagrange–Bürmann formula, gives the Taylor series expansion of the inverse function of an analytic function. Lagrange inversion is a special case of the inverse function theorem.

In mathematics, a generating function is a representation of an infinite sequence of numbers as the coefficients of a formal power series. Unlike an ordinary series, the formal power series is not required to converge: in fact, the generating function is not actually regarded as a function, and the "variable" remains an indeterminate. Generating functions were first introduced by Abraham de Moivre in 1730, in order to solve the general linear recurrence problem. One can generalize to formal power series in more than one indeterminate, to encode information about infinite multi-dimensional arrays of numbers.

In mathematics, a Dirichlet series is any series of the form

In mathematics, the binomial series is a generalization of the polynomial that comes from a binomial formula expression like for a nonnegative integer . Specifically, the binomial series is the MacLaurin series for the function , where and . Explicitly,

In mathematics, more specifically in mathematical analysis, the Cauchy product is the discrete convolution of two infinite series. It is named after the French mathematician Augustin-Louis Cauchy.

In mathematics, a divergent series is an infinite series that is not convergent, meaning that the infinite sequence of the partial sums of the series does not have a finite limit.

In mathematics, a series is the sum of the terms of an infinite sequence of numbers. More precisely, an infinite sequence defines a series S that is denoted

In mathematics, Borel summation is a summation method for divergent series, introduced by Émile Borel. It is particularly useful for summing divergent asymptotic series, and in some sense gives the best possible sum for such series. There are several variations of this method that are also called Borel summation, and a generalization of it called Mittag-Leffler summation.

<span class="mw-page-title-main">Puiseux series</span> Power series with rational exponents

In mathematics, Puiseux series are a generalization of power series that allow for negative and fractional exponents of the indeterminate. For example, the series

In real analysis and complex analysis, branches of mathematics, the identity theorem for analytic functions states: given functions f and g analytic on a domain D, if f = g on some , where has an accumulation point in D, then f = g on D.

In mathematics, Abel's test is a method of testing for the convergence of an infinite series. The test is named after mathematician Niels Henrik Abel, who proved it in 1826. There are two slightly different versions of Abel's test – one is used with series of real numbers, and the other is used with power series in complex analysis. Abel's uniform convergence test is a criterion for the uniform convergence of a series of functions dependent on parameters.

In mathematics, for a sequence of complex numbers a1, a2, a3, ... the infinite product

In analytic number theory, a Dirichlet series, or Dirichlet generating function (DGF), of a sequence is a common way of understanding and summing arithmetic functions in a meaningful way. A little known, or at least often forgotten about, way of expressing formulas for arithmetic functions and their summatory functions is to perform an integral transform that inverts the operation of forming the DGF of a sequence. This inversion is analogous to performing an inverse Z-transform to the generating function of a sequence to express formulas for the series coefficients of a given ordinary generating function.

References