Integration by parts

Last updated

In calculus, and more generally in mathematical analysis, integration by parts or partial integration is a process that finds the integral of a product of functions in terms of the integral of the product of their derivative and antiderivative. It is frequently used to transform the antiderivative of a product of functions into an antiderivative for which a solution can be more easily found. The rule can be thought of as an integral version of the product rule of differentiation; it is indeed derived using the product rule.

Contents

The integration by parts formula states:

Or, letting and while and the formula can be written more compactly:

The former expression is written as a definite integral and the latter is written as an indefinite integral. Applying the appropriate limits to the latter expression should yield the former, but the latter is not necessarily equivalent to the former.

Mathematician Brook Taylor discovered integration by parts, first publishing the idea in 1715. [1] [2] More general formulations of integration by parts exist for the Riemann–Stieltjes and Lebesgue–Stieltjes integrals. The discrete analogue for sequences is called summation by parts.

Theorem

Product of two functions

The theorem can be derived as follows. For two continuously differentiable functions and , the product rule states:

Integrating both sides with respect to ,

and noting that an indefinite integral is an antiderivative gives

where we neglect writing the constant of integration. This yields the formula for integration by parts:

or in terms of the differentials ,

This is to be understood as an equality of functions with an unspecified constant added to each side. Taking the difference of each side between two values and and applying the fundamental theorem of calculus gives the definite integral version: The original integral contains the derivative v'; to apply the theorem, one must find v, the antiderivative of v', then evaluate the resulting integral

Validity for less smooth functions

It is not necessary for and to be continuously differentiable. Integration by parts works if is absolutely continuous and the function designated is Lebesgue integrable (but not necessarily continuous). [3] (If has a point of discontinuity then its antiderivative may not have a derivative at that point.)

If the interval of integration is not compact, then it is not necessary for to be absolutely continuous in the whole interval or for to be Lebesgue integrable in the interval, as a couple of examples (in which and are continuous and continuously differentiable) will show. For instance, if

is not absolutely continuous on the interval [1, ∞), but nevertheless:

so long as is taken to mean the limit of as and so long as the two terms on the right-hand side are finite. This is only true if we choose Similarly, if

is not Lebesgue integrable on the interval [1, ∞), but nevertheless

with the same interpretation.

One can also easily come up with similar examples in which and are not continuously differentiable.

Further, if is a function of bounded variation on the segment and is differentiable on then

where denotes the signed measure corresponding to the function of bounded variation , and functions are extensions of to which are respectively of bounded variation and differentiable.[ citation needed ]

Product of many functions

Integrating the product rule for three multiplied functions, , , , gives a similar result:

In general, for factors

which leads to

Visualization

Graphical interpretation of the theorem. The pictured curve is parametrized by the variable t. Integration by parts v2.svg
Graphical interpretation of the theorem. The pictured curve is parametrized by the variable t.

Consider a parametric curve . Assuming that the curve is locally one-to-one and integrable, we can define

The area of the blue region is

Similarly, the area of the red region is

The total area A1 + A2 is equal to the area of the bigger rectangle, x2y2, minus the area of the smaller one, x1y1:

Or, in terms of t, Or, in terms of indefinite integrals, this can be written as Rearranging: Thus integration by parts may be thought of as deriving the area of the blue region from the area of rectangles and that of the red region.

This visualization also explains why integration by parts may help find the integral of an inverse function f−1(x) when the integral of the function f(x) is known. Indeed, the functions x(y) and y(x) are inverses, and the integral ∫ x dy may be calculated as above from knowing the integral ∫ y dx. In particular, this explains use of integration by parts to integrate logarithm and inverse trigonometric functions. In fact, if is a differentiable one-to-one function on an interval, then integration by parts can be used to derive a formula for the integral of in terms of the integral of . This is demonstrated in the article, Integral of inverse functions.

Applications

Finding antiderivatives

Integration by parts is a heuristic rather than a purely mechanical process for solving integrals; given a single function to integrate, the typical strategy is to carefully separate this single function into a product of two functions u(x)v(x) such that the residual integral from the integration by parts formula is easier to evaluate than the single function. The following form is useful in illustrating the best strategy to take:

On the right-hand side, u is differentiated and v is integrated; consequently it is useful to choose u as a function that simplifies when differentiated, or to choose v as a function that simplifies when integrated. As a simple example, consider:

Since the derivative of ln(x) is 1/x, one makes (ln(x)) part u; since the antiderivative of 1/x2 is −1/x, one makes 1/x2 part v. The formula now yields:

The antiderivative of −1/x2 can be found with the power rule and is 1/x.

Alternatively, one may choose u and v such that the product u′ (∫v dx) simplifies due to cancellation. For example, suppose one wishes to integrate:

If we choose u(x) = ln(|sin(x)|) and v(x) = sec2x, then u differentiates to using the chain rule and v integrates to tan x; so the formula gives:

The integrand simplifies to 1, so the antiderivative is x. Finding a simplifying combination frequently involves experimentation.

In some applications, it may not be necessary to ensure that the integral produced by integration by parts has a simple form; for example, in numerical analysis, it may suffice that it has small magnitude and so contributes only a small error term. Some other special techniques are demonstrated in the examples below.

Polynomials and trigonometric functions

In order to calculate

let:

then:

where C is a constant of integration.

For higher powers of in the form

repeatedly using integration by parts can evaluate integrals such as these; each application of the theorem lowers the power of by one.

Exponentials and trigonometric functions

An example commonly used to examine the workings of integration by parts is

Here, integration by parts is performed twice. First let

then:

Now, to evaluate the remaining integral, we use integration by parts again, with:

Then:

Putting these together,

The same integral shows up on both sides of this equation. The integral can simply be added to both sides to get

which rearranges to

where again (and ) is a constant of integration.

A similar method is used to find the integral of secant cubed.

Functions multiplied by unity

Two other well-known examples are when integration by parts is applied to a function expressed as a product of 1 and itself. This works if the derivative of the function is known, and the integral of this derivative times is also known.

The first example is . We write this as:

Let:

then:

where is the constant of integration.

The second example is the inverse tangent function :

Rewrite this as

Now let:

then

using a combination of the inverse chain rule method and the natural logarithm integral condition.

LIATE rule

The LIATE rule is a rule of thumb for integration by parts. It involves choosing as u the function that comes first in the following list: [4]

The function which is to be dv is whichever comes last in the list. The reason is that functions lower on the list generally have simpler antiderivatives than the functions above them. The rule is sometimes written as "DETAIL", where D stands for dv and the top of the list is the function chosen to be dv. An alternative to this rule is the ILATE rule, where inverse trigonometric functions come before logarithmic functions.

To demonstrate the LIATE rule, consider the integral

Following the LIATE rule, u = x, and dv = cos(x) dx, hence du = dx, and v = sin(x), which makes the integral become which equals

In general, one tries to choose u and dv such that du is simpler than u and dv is easy to integrate. If instead cos(x) was chosen as u, and x dx as dv, we would have the integral

which, after recursive application of the integration by parts formula, would clearly result in an infinite recursion and lead nowhere.

Although a useful rule of thumb, there are exceptions to the LIATE rule. A common alternative is to consider the rules in the "ILATE" order instead. Also, in some cases, polynomial terms need to be split in non-trivial ways. For example, to integrate

one would set

so that

Then

Finally, this results in

Integration by parts is often used as a tool to prove theorems in mathematical analysis.

Wallis product

The Wallis infinite product for

may be derived using integration by parts.

Gamma function identity

The gamma function is an example of a special function, defined as an improper integral for . Integration by parts illustrates it to be an extension of the factorial function:

Since

when is a natural number, that is, , applying this formula repeatedly gives the factorial:

Use in harmonic analysis

Integration by parts is often used in harmonic analysis, particularly Fourier analysis, to show that quickly oscillating integrals with sufficiently smooth integrands decay quickly. The most common example of this is its use in showing that the decay of function's Fourier transform depends on the smoothness of that function, as described below.

Fourier transform of derivative

If is a -times continuously differentiable function and all derivatives up to the th one decay to zero at infinity, then its Fourier transform satisfies

where is the th derivative of . (The exact constant on the right depends on the convention of the Fourier transform used.) This is proved by noting that

so using integration by parts on the Fourier transform of the derivative we get

Applying this inductively gives the result for general . A similar method can be used to find the Laplace transform of a derivative of a function.

Decay of Fourier transform

The above result tells us about the decay of the Fourier transform, since it follows that if and are integrable then

In other words, if satisfies these conditions then its Fourier transform decays at infinity at least as quickly as 1/|ξ|k. In particular, if then the Fourier transform is integrable.

The proof uses the fact, which is immediate from the definition of the Fourier transform, that

Using the same idea on the equality stated at the start of this subsection gives

Summing these two inequalities and then dividing by 1 + |2πξk| gives the stated inequality.

Use in operator theory

One use of integration by parts in operator theory is that it shows that the −∆ (where ∆ is the Laplace operator) is a positive operator on (see Lp space). If is smooth and compactly supported then, using integration by parts, we have

Other applications

Repeated integration by parts

Considering a second derivative of in the integral on the LHS of the formula for partial integration suggests a repeated application to the integral on the RHS:

Extending this concept of repeated partial integration to derivatives of degree n leads to

This concept may be useful when the successive integrals of are readily available (e.g., plain exponentials or sine and cosine, as in Laplace or Fourier transforms), and when the nth derivative of vanishes (e.g., as a polynomial function with degree ). The latter condition stops the repeating of partial integration, because the RHS-integral vanishes.

In the course of the above repetition of partial integrations the integrals and and get related. This may be interpreted as arbitrarily "shifting" derivatives between and within the integrand, and proves useful, too (see Rodrigues' formula).

Tabular integration by parts

The essential process of the above formula can be summarized in a table; the resulting method is called "tabular integration" [5] and was featured in the film Stand and Deliver (1988). [6]

For example, consider the integral

and take

Begin to list in column A the function and its subsequent derivatives until zero is reached. Then list in column B the function and its subsequent integrals until the size of column B is the same as that of column A. The result is as follows:

# iSignA: derivatives B: integrals
0+
1
2+
3
4+

The product of the entries in row i of columns A and B together with the respective sign give the relevant integrals in step i in the course of repeated integration by parts. Step i = 0 yields the original integral. For the complete result in step i > 0 the ith integral must be added to all the previous products (0 ≤ j < i) of the jth entry of column A and the (j + 1)st entry of column B (i.e., multiply the 1st entry of column A with the 2nd entry of column B, the 2nd entry of column A with the 3rd entry of column B, etc. ...) with the given jth sign. This process comes to a natural halt, when the product, which yields the integral, is zero (i = 4 in the example). The complete result is the following (with the alternating signs in each term):

This yields

The repeated partial integration also turns out useful, when in the course of respectively differentiating and integrating the functions and their product results in a multiple of the original integrand. In this case the repetition may also be terminated with this index i.This can happen, expectably, with exponentials and trigonometric functions. As an example consider

# iSignA: derivatives B: integrals
0+
1
2+

In this case the product of the terms in columns A and B with the appropriate sign for index i = 2 yields the negative of the original integrand (compare rows i = 0and i = 2).

Observing that the integral on the RHS can have its own constant of integration , and bringing the abstract integral to the other side, gives:

and finally:

where .

Higher dimensions

Integration by parts can be extended to functions of several variables by applying a version of the fundamental theorem of calculus to an appropriate product rule. There are several such pairings possible in multivariate calculus, involving a scalar-valued function u and vector-valued function (vector field) V. [7]

The product rule for divergence states:

Suppose is an open bounded subset of with a piecewise smooth boundary . Integrating over with respect to the standard volume form , and applying the divergence theorem, gives:

where is the outward unit normal vector to the boundary, integrated with respect to its standard Riemannian volume form . Rearranging gives:

or in other words The regularity requirements of the theorem can be relaxed. For instance, the boundary need only be Lipschitz continuous, and the functions u, v need only lie in the Sobolev space .

Green's first identity

Consider the continuously differentiable vector fields and , where is the i-th standard basis vector for . Now apply the above integration by parts to each times the vector field :

Summing over i gives a new integration by parts formula:

The case , where , is known as the first of Green's identities:

See also

Notes

  1. "Brook Taylor". History.MCS.St-Andrews.ac.uk. Retrieved May 25, 2018.
  2. "Brook Taylor". Stetson.edu. Archived from the original on January 3, 2018. Retrieved May 25, 2018.
  3. "Integration by parts". Encyclopedia of Mathematics.
  4. Kasube, Herbert E. (1983). "A Technique for Integration by Parts". The American Mathematical Monthly . 90 (3): 210–211. doi:10.2307/2975556. JSTOR   2975556.
  5. Thomas, G. B.; Finney, R. L. (1988). Calculus and Analytic Geometry (7th ed.). Reading, MA: Addison-Wesley. ISBN   0-201-17069-8.
  6. Horowitz, David (1990). "Tabular Integration by Parts" (PDF). The College Mathematics Journal . 21 (4): 307–311. doi:10.2307/2686368. JSTOR   2686368.
  7. Rogers, Robert C. (September 29, 2011). "The Calculus of Several Variables" (PDF).

Further reading

Related Research Articles

In mathematics, the Laplace transform, named after Pierre-Simon Laplace, is an integral transform that converts a function of a real variable to a function of a complex variable .

<span class="mw-page-title-main">Wave equation</span> Differential equation important in physics

The wave equation is a second-order linear partial differential equation for the description of waves or standing wave fields such as mechanical waves or electromagnetic waves. It arises in fields like acoustics, electromagnetism, and fluid dynamics.

<span class="mw-page-title-main">Hyperbolic functions</span> Collective name of 6 mathematical functions

In mathematics, hyperbolic functions are analogues of the ordinary trigonometric functions, but defined using the hyperbola rather than the circle. Just as the points (cos t, sin t) form a circle with a unit radius, the points (cosh t, sinh t) form the right half of the unit hyperbola. Also, similarly to how the derivatives of sin(t) and cos(t) are cos(t) and –sin(t) respectively, the derivatives of sinh(t) and cosh(t) are cosh(t) and +sinh(t) respectively.

<span class="mw-page-title-main">Fourier series</span> Decomposition of periodic functions into sums of simpler sinusoidal forms

A Fourier series is an expansion of a periodic function into a sum of trigonometric functions. The Fourier series is an example of a trigonometric series. By expressing a function as a sum of sines and cosines, many problems involving the function become easier to analyze because trigonometric functions are well understood. For example, Fourier series were first used by Joseph Fourier to find solutions to the heat equation. This application is possible because the derivatives of trigonometric functions fall into simple patterns. Fourier series cannot be used to approximate arbitrary functions, because most functions have infinitely many terms in their Fourier series, and the series do not always converge. Well-behaved functions, for example smooth functions, have Fourier series that converge to the original function. The coefficients of the Fourier series are determined by integrals of the function multiplied by trigonometric functions, described in Common forms of the Fourier series below.

In vector calculus, the divergence theorem, also known as Gauss's theorem or Ostrogradsky's theorem, is a theorem relating the flux of a vector field through a closed surface to the divergence of the field in the volume enclosed.

In calculus, integration by substitution, also known as u-substitution, reverse chain rule or change of variables, is a method for evaluating integrals and antiderivatives. It is the counterpart to the chain rule for differentiation, and can loosely be thought of as using the chain rule "backwards."

In mathematics, differential forms provide a unified approach to define integrands over curves, surfaces, solids, and higher-dimensional manifolds. The modern notion of differential forms was pioneered by Élie Cartan. It has many applications, especially in geometry, topology and physics.

Integration is the basic operation in integral calculus. While differentiation has straightforward rules by which the derivative of a complicated function can be found by differentiating its simpler component functions, integration does not, so tables of known integrals are often useful. This page lists some of the most common antiderivatives.

In mathematics, a Gaussian function, often simply referred to as a Gaussian, is a function of the base form and with parametric extension for arbitrary real constants a, b and non-zero c. It is named after the mathematician Carl Friedrich Gauss. The graph of a Gaussian is a characteristic symmetric "bell curve" shape. The parameter a is the height of the curve's peak, b is the position of the center of the peak, and c controls the width of the "bell".

<span class="mw-page-title-main">Green's function</span> Impulse response of an inhomogeneous linear differential operator

In mathematics, a Green's function is the impulse response of an inhomogeneous linear differential operator defined on a domain with specified initial conditions or boundary conditions.

<span class="mw-page-title-main">Path integral formulation</span> Formulation of quantum mechanics

The path integral formulation is a description in quantum mechanics that generalizes the stationary action principle of classical mechanics. It replaces the classical notion of a single, unique classical trajectory for a system with a sum, or functional integral, over an infinity of quantum-mechanically possible trajectories to compute a quantum amplitude.

<span class="mw-page-title-main">Radon transform</span> Integral transform

In mathematics, the Radon transform is the integral transform which takes a function f defined on the plane to a function Rf defined on the (two-dimensional) space of lines in the plane, whose value at a particular line is equal to the line integral of the function over that line. The transform was introduced in 1917 by Johann Radon, who also provided a formula for the inverse transform. Radon further included formulas for the transform in three dimensions, in which the integral is taken over planes. It was later generalized to higher-dimensional Euclidean spaces and more broadly in the context of integral geometry. The complex analogue of the Radon transform is known as the Penrose transform. The Radon transform is widely applicable to tomography, the creation of an image from the projection data associated with cross-sectional scans of an object.

<span class="mw-page-title-main">Sinc function</span> Special mathematical function defined as sin(x)/x

In mathematics, physics and engineering, the sinc function, denoted by sinc(x), has two forms, normalized and unnormalized.

In the mathematical field of complex analysis, contour integration is a method of evaluating certain integrals along paths in the complex plane.

<span class="mw-page-title-main">Dirichlet integral</span> Integral of sin(x)/x from 0 to infinity.

In mathematics, there are several integrals known as the Dirichlet integral, after the German mathematician Peter Gustav Lejeune Dirichlet, one of which is the improper integral of the sinc function over the positive real number line.

In calculus, the Leibniz integral rule for differentiation under the integral sign, named after Gottfried Wilhelm Leibniz, states that for an integral of the form where and the integrands are functions dependent on the derivative of this integral is expressible as where the partial derivative indicates that inside the integral, only the variation of with is considered in taking the derivative.

<span class="mw-page-title-main">Multiple integral</span> Generalization of definite integrals to functions of multiple variables

In mathematics (specifically multivariable calculus), a multiple integral is a definite integral of a function of several real variables, for instance, f(x, y) or f(x, y, z).

The gradient theorem, also known as the fundamental theorem of calculus for line integrals, says that a line integral through a gradient field can be evaluated by evaluating the original scalar field at the endpoints of the curve. The theorem is a generalization of the second fundamental theorem of calculus to any curve in a plane or space rather than just the real line.

Common integrals in quantum field theory are all variations and generalizations of Gaussian integrals to the complex plane and to multiple dimensions. Other integrals can be approximated by versions of the Gaussian integral. Fourier integrals are also considered.