Power rule

Last updated

In calculus, the power rule is used to differentiate functions of the form , whenever is a real number. Since differentiation is a linear operation on the space of differentiable functions, polynomials can also be differentiated using this rule. The power rule underlies the Taylor series as it relates a power series with a function's derivatives.

Contents

Statement of the power rule

Let be a function satisfying for all , where . [lower-alpha 1] Then,

The power rule for integration states that

for any real number . It can be derived by inverting the power rule for differentiation. In this equation C is any constant.

Proofs

Proof for real exponents

To start, we should choose a working definition of the value of , where is any real number. Although it is feasible to define the value as the limit of a sequence of rational powers that approach the irrational power whenever we encounter such a power, or as the least upper bound of a set of rational powers less than the given power, this type of definition is not amenable to differentiation. It is therefore preferable to use a functional definition, which is usually taken to be for all values of , where is the natural exponential function and is Euler's number. [1] [2] First, we may demonstrate that the derivative of is .

If , then , where is the natural logarithm function, the inverse function of the exponential function, as demonstrated by Euler. [3] Since the latter two functions are equal for all values of , their derivatives are also equal, whenever either derivative exists, so we have, by the chain rule, or , as was required. Therefore, applying the chain rule to , we see that which simplifies to .

When , we may use the same definition with , where we now have . This necessarily leads to the same result. Note that because does not have a conventional definition when is not a rational number, irrational power functions are not well defined for negative bases. In addition, as rational powers of −1 with even denominators (in lowest terms) are not real numbers, these expressions are only real valued for rational powers with odd denominators (in lowest terms).

Finally, whenever the function is differentiable at , the defining limit for the derivative is: which yields 0 only when is a rational number with odd denominator (in lowest terms) and , and 1 when . For all other values of , the expression is not well-defined for , as was covered above, or is not a real number, so the limit does not exist as a real-valued derivative. For the two cases that do exist, the values agree with the value of the existing power rule at 0, so no exception need be made.

The exclusion of the expression (the case ) from our scheme of exponentiation is due to the fact that the function has no limit at (0,0), since approaches 1 as x approaches 0, while approaches 0 as y approaches 0. Thus, it would be problematic to ascribe any particular value to it, as the value would contradict one of the two cases, dependent on the application. It is traditionally left undefined.

Proofs for integer exponents

Proof by induction (natural numbers)

Let . It is required to prove that The base case may be when or , depending on how the set of natural numbers is defined.

When ,

When ,

Therefore, the base case holds either way.

Suppose the statement holds for some natural number k, i.e.

When ,By the principle of mathematical induction, the statement is true for all natural numbers n.

Proof by binomial theorem (natural number)

Let , where .

Then,

Since n choose 1 is equal to n, and the rest of the terms all contain h, which is 0, the rest of the terms cancel. This proof only works for natural numbers as the binomial theorem only works for natural numbers.

Generalization to negative integer exponents

For a negative integer n, let so that m is a positive integer. Using the reciprocal rule,In conclusion, for any integer ,

Generalization to rational exponents

Upon proving that the power rule holds for integer exponents, the rule can be extended to rational exponents.

Proof by chain rule

This proof is composed of two steps that involve the use of the chain rule for differentiation.

  1. Let , where . Then . By the chain rule, . Solving for , Thus, the power rule applies for rational exponents of the form , where is a nonzero natural number. This can be generalized to rational exponents of the form by applying the power rule for integer exponents using the chain rule, as shown in the next step.
  2. Let , where so that . By the chain rule,

From the above results, we can conclude that when is a rational number,

Proof by implicit differentiation

A more straightforward generalization of the power rule to rational exponents makes use of implicit differentiation.

Let , where so that .

Then,Differentiating both sides of the equation with respect to ,Solving for ,Since ,Applying laws of exponents,Thus, letting , we can conclude that when is a rational number.

History

The power rule for integrals was first demonstrated in a geometric form by Italian mathematician Bonaventura Cavalieri in the early 17th century for all positive integer values of , and during the mid 17th century for all rational powers by the mathematicians Pierre de Fermat, Evangelista Torricelli, Gilles de Roberval, John Wallis, and Blaise Pascal, each working independently. At the time, they were treatises on determining the area between the graph of a rational power function and the horizontal axis. With hindsight, however, it is considered the first general theorem of calculus to be discovered. [4] The power rule for differentiation was derived by Isaac Newton and Gottfried Wilhelm Leibniz, each independently, for rational power functions in the mid 17th century, who both then used it to derive the power rule for integrals as the inverse operation. This mirrors the conventional way the related theorems are presented in modern basic calculus textbooks, where differentiation rules usually precede integration rules. [5]

Although both men stated that their rules, demonstrated only for rational quantities, worked for all real powers, neither sought a proof of such, as at the time the applications of the theory were not concerned with such exotic power functions, and questions of convergence of infinite series were still ambiguous.

The unique case of was resolved by Flemish Jesuit and mathematician Grégoire de Saint-Vincent and his student Alphonse Antonio de Sarasa in the mid 17th century, who demonstrated that the associated definite integral,

representing the area between the rectangular hyperbola and the x-axis, was a logarithmic function, whose base was eventually discovered to be the transcendental number e. The modern notation for the value of this definite integral is , the natural logarithm.

Generalizations

Complex power functions

If we consider functions of the form where is any complex number and is a complex number in a slit complex plane that excludes the branch point of 0 and any branch cut connected to it, and we use the conventional multivalued definition , then it is straightforward to show that, on each branch of the complex logarithm, the same argument used above yields a similar result: . [6]

In addition, if is a positive integer, then there is no need for a branch cut: one may define , or define positive integral complex powers through complex multiplication, and show that for all complex , from the definition of the derivative and the binomial theorem.

However, due to the multivalued nature of complex power functions for non-integer exponents, one must be careful to specify the branch of the complex logarithm being used. In addition, no matter which branch is used, if is not a positive integer, then the function is not differentiable at 0.

See also

Related Research Articles

In calculus, the chain rule is a formula that expresses the derivative of the composition of two differentiable functions f and g in terms of the derivatives of f and g. More precisely, if is the function such that for every x, then the chain rule is, in Lagrange's notation, or, equivalently,

In mathematics, the derivative is a fundamental tool that quantifies the sensitivity of change of a function's output with respect to its input. The derivative of a function of a single variable at a chosen input value, when it exists, is the slope of the tangent line to the graph of the function at that point. The tangent line is the best linear approximation of the function near that input value. For this reason, the derivative is often described as the instantaneous rate of change, the ratio of the instantaneous change in the dependent variable to that of the independent variable. The process of finding a derivative is called differentiation.

<span class="mw-page-title-main">Exponential function</span> Mathematical function, denoted exp(x) or e^x

The exponential function is a mathematical function denoted by or . Unless otherwise specified, the term generally refers to the positive-valued function of a real variable, although it can be extended to the complex numbers or generalized to other mathematical objects like matrices or Lie algebras. The exponential function originated from the operation of taking powers of a number, but various modern definitions allow it to be rigorously extended to all real arguments , including irrational numbers. Its ubiquitous occurrence in pure and applied mathematics led mathematician Walter Rudin to consider the exponential function to be "the most important function in mathematics".

<span class="mw-page-title-main">Natural logarithm</span> Logarithm to the base of the mathematical constant e

The natural logarithm of a number is its logarithm to the base of the mathematical constant e, which is an irrational and transcendental number approximately equal to 2.718281828459. The natural logarithm of x is generally written as ln x, logex, or sometimes, if the base e is implicit, simply log x. Parentheses are sometimes added for clarity, giving ln(x), loge(x), or log(x). This is done particularly when the argument to the logarithm is not a single symbol, so as to prevent ambiguity.

<span class="mw-page-title-main">Taylor's theorem</span> Approximation of a function by a truncated power series

In calculus, Taylor's theorem gives an approximation of a -times differentiable function around a given point by a polynomial of degree , called the -th-order Taylor polynomial. For a smooth function, the Taylor polynomial is the truncation at the order of the Taylor series of the function. The first-order Taylor polynomial is the linear approximation of the function, and the second-order Taylor polynomial is often referred to as the quadratic approximation. There are several versions of Taylor's theorem, some giving explicit estimates of the approximation error of the function by its Taylor polynomial.

<span class="mw-page-title-main">Exponentiation</span> Arithmetic operation

In mathematics, exponentiation is an operation involving two numbers: the base and the exponent or power. Exponentiation is written as bn, where b is the base and n is the power; this is pronounced as "b (raised) to the n". When n is a positive integer, exponentiation corresponds to repeated multiplication of the base: that is, bn is the product of multiplying n bases:

<span class="mw-page-title-main">Euler's constant</span> Constant value used in mathematics

Euler's constant is a mathematical constant, usually denoted by the lowercase Greek letter gamma, defined as the limiting difference between the harmonic series and the natural logarithm, denoted here by log:

In calculus, the quotient rule is a method of finding the derivative of a function that is the ratio of two differentiable functions. Let , where both f and g are differentiable and The quotient rule states that the derivative of h(x) is

<span class="mw-page-title-main">Product rule</span> Formula for the derivative of a product

In calculus, the product rule is a formula used to find the derivatives of products of two or more functions. For two functions, it may be stated in Lagrange's notation as or in Leibniz's notation as

<span class="mw-page-title-main">Integral test for convergence</span> Test for infinite series of monotonous terms for convergence

In mathematics, the integral test for convergence is a method used to test infinite series of monotonic terms for convergence. It was developed by Colin Maclaurin and Augustin-Louis Cauchy and is sometimes known as the Maclaurin–Cauchy test.

The Basel problem is a problem in mathematical analysis with relevance to number theory, concerning an infinite sum of inverse squares. It was first posed by Pietro Mengoli in 1650 and solved by Leonhard Euler in 1734, and read on 5 December 1735 in The Saint Petersburg Academy of Sciences. Since the problem had withstood the attacks of the leading mathematicians of the day, Euler's solution brought him immediate fame when he was twenty-eight. Euler generalised the problem considerably, and his ideas were taken up more than a century later by Bernhard Riemann in his seminal 1859 paper "On the Number of Primes Less Than a Given Magnitude", in which he defined his zeta function and proved its basic properties. The problem is named after Basel, hometown of Euler as well as of the Bernoulli family who unsuccessfully attacked the problem.

In mathematics, the exponential function can be characterized in many ways. This article presents some common characterizations, discusses why each makes sense, and proves that they are all equivalent.

In mathematics, the total derivative of a function f at a point is the best linear approximation near this point of the function with respect to its arguments. Unlike partial derivatives, the total derivative approximates the function with respect to all of its arguments, not just a single one. In many situations, this is the same as considering all partial derivatives simultaneously. The term "total derivative" is primarily used when f is a function of several variables, because when f is a function of a single variable, the total derivative is the same as the ordinary derivative of the function.

In mathematical analysis, and applications in geometry, applied mathematics, engineering, and natural sciences, a function of a real variable is a function whose domain is the real numbers , or a subset of that contains an interval of positive length. Most real functions that are considered and studied are differentiable in some interval. The most widely considered such functions are the real functions, which are the real-valued functions of a real variable, that is, the functions of a real variable whose codomain is the set of real numbers.

In mathematics, the derivative is a fundamental construction of differential calculus and admits many possible generalizations within the fields of mathematical analysis, combinatorics, algebra, geometry, etc.

In calculus, the Leibniz integral rule for differentiation under the integral sign, named after Gottfried Wilhelm Leibniz, states that for an integral of the form where and the integrands are functions dependent on the derivative of this integral is expressible as where the partial derivative indicates that inside the integral, only the variation of with is considered in taking the derivative.

This is a summary of differentiation rules, that is, rules for computing the derivative of a function in calculus.

A product integral is any product-based counterpart of the usual sum-based integral of calculus. The product integral was developed by the mathematician Vito Volterra in 1887 to solve systems of linear differential equations.

In mathematical analysis and its applications, a function of several real variables or real multivariate function is a function with more than one argument, with all arguments being real variables. This concept extends the idea of a function of a real variable to several variables. The "input" variables take real values, while the "output", also called the "value of the function", may be real or complex. However, the study of the complex-valued functions may be easily reduced to the study of the real-valued functions, by considering the real and imaginary parts of the complex function; therefore, unless explicitly specified, only real-valued functions will be considered in this article.

In mathematics, calculus on Euclidean space is a generalization of calculus of functions in one or several variables to calculus of functions on Euclidean space as well as a finite-dimensional real vector space. This calculus is also known as advanced calculus, especially in the United States. It is similar to multivariable calculus but is somewhat more sophisticated in that it uses linear algebra more extensively and covers some concepts from differential geometry such as differential forms and Stokes' formula in terms of differential forms. This extensive use of linear algebra also allows a natural generalization of multivariable calculus to calculus on Banach spaces or topological vector spaces.

References

Notes

  1. If is a rational number whose lowest terms representation has an odd denominator, then the domain of is understood to be . Otherwise, the domain is .

Citations

  1. Landau, Edmund (1951). Differential and Integral Calculus. New York: Chelsea Publishing Company. p. 45. ISBN   978-0821828304.
  2. Spivak, Michael (1994). Calculus (3 ed.). Texas: Publish or Perish, Inc. pp. 336–342. ISBN   0-914098-89-6.
  3. Maor, Eli (1994). e: The Story of a Number . New Jersey: Princeton University Press. p.  156. ISBN   0-691-05854-7.
  4. Boyer, Carl (1959). The History of the Calculus and its Conceptual Development. New York: Dover. p.  127. ISBN   0-486-60509-4.
  5. Boyer, Carl (1959). The History of the Calculus and its Conceptual Development. New York: Dover. pp.  191, 205. ISBN   0-486-60509-4.
  6. Freitag, Eberhard; Busam, Rolf (2009). Complex Analysis (2 ed.). Heidelberg: Springer-Verlag. p. 46. ISBN   978-3-540-93982-5.

Further reading