Derivative

Last updated

In mathematics, the derivative is a fundamental tool that quantifies the sensitivity of change of a function's output with respect to its input. The derivative of a function of a single variable at a chosen input value, when it exists, is the slope of the tangent line to the graph of the function at that point. The tangent line is the best linear approximation of the function near that input value. For this reason, the derivative is often described as the instantaneous rate of change, the ratio of the instantaneous change in the dependent variable to that of the independent variable. [1] The process of finding a derivative is called differentiation.

Contents

There are multiple different notations for differentiation, two of the most commonly used being Leibniz notation and prime notation. Leibniz notation, named after Gottfried Wilhelm Leibniz, is represented as the ratio of two differentials, whereas prime notation is written by adding a prime mark. Higher order notations represent repeated differentiation, and they are usually denoted in Leibniz notation by adding superscripts to the differentials, and in prime notation by adding additional prime marks. The higher order derivatives can be applied in physics; for example, while the first derivative of the position of a moving object with respect to time is the object's velocity, how the position changes as time advances, the second derivative is the object's acceleration, how the velocity changes as time advances.

Derivatives can be generalized to functions of several real variables. In this generalization, the derivative is reinterpreted as a linear transformation whose graph is (after an appropriate translation) the best linear approximation to the graph of the original function. The Jacobian matrix is the matrix that represents this linear transformation with respect to the basis given by the choice of independent and dependent variables. It can be calculated in terms of the partial derivatives with respect to the independent variables. For a real-valued function of several variables, the Jacobian matrix reduces to the gradient vector.

Definition

As a limit

A function of a real variable is differentiable at a point of its domain, if its domain contains an open interval containing , and the limit exists. [2] This means that, for every positive real number , there exists a positive real number such that, for every such that and then is defined, and where the vertical bars denote the absolute value. This is an example of the (ε, δ)-definition of limit. [3]

If the function is differentiable at , that is if the limit exists, then this limit is called the derivative of at . Multiple notations for the derivative exist. [4] The derivative of at can be denoted , read as " prime of "; or it can be denoted , read as "the derivative of with respect to at " or " by (or over) at ". See § Notation below. If is a function that has a derivative at every point in its domain, then a function can be defined by mapping every point to the value of the derivative of at . This function is written and is called the derivative function or the derivative of. The function sometimes has a derivative at most, but not all, points of its domain. The function whose value at equals whenever is defined and elsewhere is undefined is also called the derivative of . It is still a function, but its domain may be smaller than the domain of . [5]

For example, let be the squaring function: . Then the quotient in the definition of the derivative is [6] The division in the last step is valid as long as . The closer is to , the closer this expression becomes to the value . The limit exists, and for every input the limit is . So, the derivative of the squaring function is the doubling function: .

Tangent to a curve.svg
The graph of a function, drawn in black, and a tangent line to that graph, drawn in red. The slope of the tangent line is equal to the derivative of the function at the marked point.
Tangent function animation.gif
The derivative at different points of a differentiable function. In this case, the derivative is equal to .

The ratio in the definition of the derivative is the slope of the line through two points on the graph of the function , specifically the points and . As is made smaller, these points grow closer together, and the slope of this line approaches the limiting value, the slope of the tangent to the graph of at . In other words, the derivative is the slope of the tangent. [7]

Using infinitesimals

One way to think of the derivative is as the ratio of an infinitesimal change in the output of the function to an infinitesimal change in its input. [8] In order to make this intuition rigorous, a system of rules for manipulating infinitesimal quantities is required. [9] The system of hyperreal numbers is a way of treating infinite and infinitesimal quantities. The hyperreals are an extension of the real numbers that contain numbers greater than anything of the form for any finite number of terms. Such numbers are infinite, and their reciprocals are infinitesimals. The application of hyperreal numbers to the foundations of calculus is called nonstandard analysis. This provides a way to define the basic concepts of calculus such as the derivative and integral in terms of infinitesimals, thereby giving a precise meaning to the in the Leibniz notation. Thus, the derivative of becomes for an arbitrary infinitesimal , where denotes the standard part function, which "rounds off" each finite hyperreal to the nearest real. [10] Taking the squaring function as an example again,

Continuity and differentiability

Right-continuous.svg
This function does not have a derivative at the marked point, as the function is not continuous there (specifically, it has a jump discontinuity).
Absolute value.svg
The absolute value function is continuous but fails to be differentiable at x = 0 since the tangent slopes do not approach the same value from the left as they do from the right.

If is differentiable at , then must also be continuous at . [11] As an example, choose a point and let be the step function that returns the value 1 for all less than , and returns a different value 10 for all greater than or equal to . The function cannot have a derivative at . If is negative, then is on the low part of the step, so the secant line from to is very steep; as tends to zero, the slope tends to infinity. If is positive, then is on the high part of the step, so the secant line from to has slope zero. Consequently, the secant lines do not approach any single slope, so the limit of the difference quotient does not exist. However, even if a function is continuous at a point, it may not be differentiable there. For example, the absolute value function given by is continuous at , but it is not differentiable there. If is positive, then the slope of the secant line from 0 to is one; if is negative, then the slope of the secant line from to is . [12] This can be seen graphically as a "kink" or a "cusp" in the graph at . Even a function with a smooth graph is not differentiable at a point where its tangent is vertical: For instance, the function given by is not differentiable at . In summary, a function that has a derivative is continuous, but there are continuous functions that do not have a derivative. [13]

Most functions that occur in practice have derivatives at all points or almost every point. Early in the history of calculus, many mathematicians assumed that a continuous function was differentiable at most points. [14] Under mild conditions (for example, if the function is a monotone or a Lipschitz function), this is true. However, in 1872, Weierstrass found the first example of a function that is continuous everywhere but differentiable nowhere. This example is now known as the Weierstrass function. [15] In 1931, Stefan Banach proved that the set of functions that have a derivative at some point is a meager set in the space of all continuous functions. Informally, this means that hardly any random continuous functions have a derivative at even one point. [16]

Notation

One common way of writing the derivative of a function is Leibniz notation, introduced by Gottfried Wilhelm Leibniz in 1675, which denotes a derivative as the quotient of two differentials, such as and . [17] [18] It is still commonly used when the equation is viewed as a functional relationship between dependent and independent variables. The first derivative is denoted by , read as "the derivative of with respect to ". [19] This derivative can alternately be treated as the application of a differential operator to a function, Higher derivatives are expressed using the notation for the -th derivative of . These are abbreviations for multiple applications of the derivative operator; for example, [20] Unlike some alternatives, Leibniz notation involves explicit specification of the variable for differentiation, in the denominator, which removes ambiguity when working with multiple interrelated quantities. The derivative of a composed function can be expressed using the chain rule: if and then [21]

Another common notation for differentiation is by using the prime mark in the symbol of a function . This is known as prime notation, due to Joseph-Louis Lagrange. [22] The first derivative is written as , read as " prime of , or , read as " prime". [23] Similarly, the second and the third derivatives can be written as and , respectively. [24] For denoting the number of higher derivatives beyond this point, some authors use Roman numerals in superscript, whereas others place the number in parentheses, such as or . [25] The latter notation generalizes to yield the notation for the th derivative of . [20]

In Newton's notation or the dot notation, a dot is placed over a symbol to represent a time derivative. If is a function of , then the first and second derivatives can be written as and , respectively. This notation is used exclusively for derivatives with respect to time or arc length. It is typically used in differential equations in physics and differential geometry. [26] However, the dot notation becomes unmanageable for high-order derivatives (of order 4 or more) and cannot deal with multiple independent variables.

Another notation is D-notation, which represents the differential operator by the symbol . [20] The first derivative is written and higher derivatives are written with a superscript, so the -th derivative is . This notation is sometimes called Euler notation, although it seems that Leonhard Euler did not use it, and the notation was introduced by Louis François Antoine Arbogast. [27] To indicate a partial derivative, the variable differentiated by is indicated with a subscript, for example given the function , its partial derivative with respect to can be written or . Higher partial derivatives can be indicated by superscripts or multiple subscripts, e.g. and . [28]

Rules of computation

In principle, the derivative of a function can be computed from the definition by considering the difference quotient and computing its limit. Once the derivatives of a few simple functions are known, the derivatives of other functions are more easily computed using rules for obtaining derivatives of more complicated functions from simpler ones. This process of finding a derivative is known as differentiation. [29]

Rules for basic functions

The following are the rules for the derivatives of the most common basic functions. Here, is a real number, and is the base of the natural logarithm, approximately 2.71828. [30]

Rules for combined functions

Given that the and are the functions. The following are some of the most basic rules for deducing the derivative of functions from derivatives of basic functions. [31]

Computation example

The derivative of the function given by is Here the second term was computed using the chain rule and the third term using the product rule. The known derivatives of the elementary functions , , , , and , as well as the constant , were also used.

Higher-order derivatives

Higher order derivatives are the result of differentiating a function repeatedly. Given that is a differentiable function, the derivative of is the first derivative, denoted as . The derivative of is the second derivative, denoted as , and the derivative of is the third derivative, denoted as . By continuing this process, if it exists, the th derivative is the derivative of the th derivative or the derivative of order . As has been discussed above, the generalization of derivative of a function may be denoted as . [32] A function that has successive derivatives is called times differentiable. If the -th derivative is continuous, then the function is said to be of differentiability class . [33] A function that has infinitely many derivatives is called infinitely differentiable or smooth . [34] Any polynomial function is infinitely differentiable; taking derivatives repeatedly will eventually result in a constant function, and all subsequent derivatives of that function are zero. [35]

One application of higher-order derivatives is in physics. Suppose that a function represents the position of an object at the time. The first derivative of that function is the velocity of an object with respect to time, the second derivative of the function is the acceleration of an object with respect to time, [29] and the third derivative is the jerk. [36]

In other dimensions

Vector-valued functions

A vector-valued function of a real variable sends real numbers to vectors in some vector space . A vector-valued function can be split up into its coordinate functions , meaning that . This includes, for example, parametric curves in or . The coordinate functions are real-valued functions, so the above definition of derivative applies to them. The derivative of is defined to be the vector, called the tangent vector, whose coordinates are the derivatives of the coordinate functions. That is, [37] if the limit exists. The subtraction in the numerator is the subtraction of vectors, not scalars. If the derivative of exists for every value of , then is another vector-valued function. [37]

Partial derivatives

Functions can depend upon more than one variable. A partial derivative of a function of several variables is its derivative with respect to one of those variables, with the others held constant. Partial derivatives are used in vector calculus and differential geometry. As with ordinary derivatives, multiple notations exist: the partial derivative of a function with respect to the variable is variously denoted by

, , , , or ,

among other possibilities. [38] It can be thought of as the rate of change of the function in the -direction. [39] Here is a rounded d called the partial derivative symbol. To distinguish it from the letter d, ∂ is sometimes pronounced "der", "del", or "partial" instead of "dee". [40] For example, let , then the partial derivative of function with respect to both variables and are, respectively: In general, the partial derivative of a function in the direction at the point is defined to be: [41]

This is fundamental for the study of the functions of several real variables. Let be such a real-valued function. If all partial derivatives with respect to are defined at the point , these partial derivatives define the vector which is called the gradient of at . If is differentiable at every point in some domain, then the gradient is a vector-valued function that maps the point to the vector . Consequently, the gradient determines a vector field. [42]

Directional derivatives

If is a real-valued function on , then the partial derivatives of measure its variation in the direction of the coordinate axes. For example, if is a function of and , then its partial derivatives measure the variation in in the and direction. However, they do not directly measure the variation of in any other direction, such as along the diagonal line . These are measured using directional derivatives. Given a vector , then the directional derivative of in the direction of at the point is: [43]

If all the partial derivatives of exist and are continuous at , then they determine the directional derivative of in the direction by the formula: [44]

Total derivative, total differential and Jacobian matrix

When is a function from an open subset of to , then the directional derivative of in a chosen direction is the best linear approximation to at that point and in that direction. However, when , no single directional derivative can give a complete picture of the behavior of . The total derivative gives a complete picture by considering all directions at once. That is, for any vector starting at , the linear approximation formula holds: [45] Similarly with the single-variable derivative, is chosen so that the error in this approximation is as small as possible. The total derivative of at is the unique linear transformation such that [45] Here is a vector in , so the norm in the denominator is the standard length on . However, is a vector in , and the norm in the numerator is the standard length on . [45] If is a vector starting at , then is called the pushforward of by . [46]

If the total derivative exists at , then all the partial derivatives and directional derivatives of exist at , and for all , is the directional derivative of in the direction . If is written using coordinate functions, so that , then the total derivative can be expressed using the partial derivatives as a matrix. This matrix is called the Jacobian matrix of at : [47]

Generalizations

The concept of a derivative can be extended to many other settings. The common thread is that the derivative of a function at a point serves as a linear approximation of the function at that point.

See also

Notes

  1. Apostol 1967, p. 160; Stewart 2002, p. 129130; Strang et al. 2023, p.  224.
  2. Apostol 1967, p. 160; Stewart 2002, p. 127; Strang et al. 2023, p.  220.
  3. Gonick 2012, p. 83; Thomas et al. 2014, p. 60.
  4. Gonick 2012, p. 88; Strang et al. 2023, p.  234.
  5. Gonick 2012, p. 83; Strang et al. 2023, p.  232.
  6. Gonick 2012, pp. 77–80.
  7. Thompson 1998, pp. 34, 104; Stewart 2002, p. 128.
  8. Thompson 1998, pp. 84–85.
  9. Keisler 2012, pp. 902–904.
  10. Keisler 2012, p. 45; Henle & Kleinberg 2003, p. 66.
  11. Gonick 2012, p. 156; Thomas et al. 2014, p. 114; Strang et al. 2023, p.  237.
  12. Gonick 2012, p. 149; Thomas et al. 2014, p. 113; Strang et al. 2023, p.  237.
  13. Gonick 2012, p. 156; Thomas et al. 2014, p. 114; Strang et al. 2023, pp.  237–238.
  14. Jašek 1922; Jarník 1922; Rychlík 1923.
  15. David 2018.
  16. Banach 1931, cited in Hewitt & Stromberg 1965.
  17. Apostol 1967, p. 172.
  18. Cajori 2007, p. 204.
  19. Moore & Siegel 2013, p. 110.
  20. 1 2 3 Varberg, Purcell & Rigdon 2007, p. 125126.
  21. In the formulation of calculus in terms of limits, various authors have assigned the symbol various meanings. Some authors such as Varberg, Purcell & Rigdon 2007, p. 119 and Stewart 2002, p. 177 do not assign a meaning to by itself, but only as part of the symbol . Others define as an independent variable, and define by . In non-standard analysis is defined as an infinitesimal. It is also interpreted as the exterior derivative of a function . See differential (infinitesimal) for further information.
  22. Schwartzman 1994, p.  171.
  23. Moore & Siegel 2013, p. 110; Goodman 1963, p. 7879.
  24. Varberg, Purcell & Rigdon 2007, p. 125126; Cajori 2007, p. 228.
  25. Choudary & Niculescu 2014, p.  222; Apostol 1967, p. 171.
  26. Evans 1999, p. 63; Kreyszig 1991, p. 1.
  27. Cajori 1923.
  28. Apostol 1967, p. 172; Varberg, Purcell & Rigdon 2007, p. 125126.
  29. 1 2 Apostol 1967, p. 160.
  30. Varberg, Purcell & Rigdon 2007. See p. 133 for the power rule, p. 115116 for the trigonometric functions, p. 326 for the natural logarithm, p. 338339 for exponential with base , p. 343 for the exponential with base , p. 344 for the logarithm with base , and p. 369 for the inverse of trigonometric functions.
  31. For constant rule and sum rule, see Apostol 1967 , p. 161, 164, respectively. For the product rule, quotient rule, and chain rule, see Varberg, Purcell & Rigdon 2007 , p. 111112, 119, respectively. For the special case of the product rule, that is, the product of a constant and a function, see Varberg, Purcell & Rigdon 2007 , p. 108109.
  32. Apostol 1967, p. 160; Varberg, Purcell & Rigdon 2007, p. 125126.
  33. Warner 1983, p. 5.
  34. Debnath & Shah 2015, p.  40.
  35. Carothers 2000, p.  176.
  36. Stewart 2002, p. 193.
  37. 1 2 Stewart 2002, p. 893.
  38. Stewart 2002, p.  947; Christopher 2013, p. 682.
  39. Stewart 2002, p.  949.
  40. Silverman 1989, p.  216; Bhardwaj 2005, See p. 6.4.
  41. Mathai & Haubold 2017, p.  52.
  42. Gbur 2011, pp. 36–37.
  43. Varberg, Purcell & Rigdon 2007, p. 642.
  44. Guzman 2003, p.  35.
  45. 1 2 3 Davvaz 2023, p.  266.
  46. Lee 2013, p. 72.
  47. Davvaz 2023, p.  267.
  48. Roussos 2014, p. 303.
  49. Gbur 2011, pp. 261–264.
  50. Gray, Abbena & Salamon 2006, p.  826.
  51. Azegami 2020. See p. 209 for the Gateaux derivative, and p. 211 for the Fréchet derivative.
  52. Funaro 1992, p.  8485.
  53. Kolchin 1973, p.  58, 126.
  54. Georgiev 2018, p.  8.
  55. Barbeau 1961.

Related Research Articles

In calculus, the chain rule is a formula that expresses the derivative of the composition of two differentiable functions f and g in terms of the derivatives of f and g. More precisely, if is the function such that for every x, then the chain rule is, in Lagrange's notation, or, equivalently,

<span class="mw-page-title-main">Curl (mathematics)</span> Circulation density in a vector field

In vector calculus, the curl, also known as rotor, is a vector operator that describes the infinitesimal circulation of a vector field in three-dimensional Euclidean space. The curl at a point in the field is represented by a vector whose length and direction denote the magnitude and axis of the maximum circulation. The curl of a field is formally defined as the circulation density at each point of the field.

<span class="mw-page-title-main">Gradient</span> Multivariate derivative (mathematics)

In vector calculus, the gradient of a scalar-valued differentiable function of several variables is the vector field whose value at a point gives the direction and the rate of fastest increase. The gradient transforms like a vector under change of basis of the space of variables of . If the gradient of a function is non-zero at a point , the direction of the gradient is the direction in which the function increases most quickly from , and the magnitude of the gradient is the rate of increase in that direction, the greatest absolute directional derivative. Further, a point where the gradient is the zero vector is known as a stationary point. The gradient thus plays a fundamental role in optimization theory, where it is used to minimize a function by gradient descent. In coordinate-free terms, the gradient of a function may be defined by:

In mathematics, a partial derivative of a function of several variables is its derivative with respect to one of those variables, with the others held constant. Partial derivatives are used in vector calculus and differential geometry.

In vector calculus, the Jacobian matrix of a vector-valued function of several variables is the matrix of all its first-order partial derivatives. When this matrix is square, that is, when the function takes the same number of variables as input as the number of vector components of its output, its determinant is referred to as the Jacobian determinant. Both the matrix and the determinant are often referred to simply as the Jacobian in literature. They are named after Carl Gustav Jacob Jacobi.

In vector calculus, Green's theorem relates a line integral around a simple closed curve C to a double integral over the plane region D bounded by C. It is the two-dimensional special case of Stokes' theorem. In one dimension, it is equivalent to the fundamental theorem of calculus. In three dimensions, it is equivalent to the divergence theorem.

<span class="mw-page-title-main">Product rule</span> Formula for the derivative of a product

In calculus, the product rule is a formula used to find the derivatives of products of two or more functions. For two functions, it may be stated in Lagrange's notation as or in Leibniz's notation as

<span class="mw-page-title-main">Legendre transformation</span> Mathematical transformation

In mathematics, the Legendre transformation, first introduced by Adrien-Marie Legendre in 1787 when studying the minimal surface problem, is an involutive transformation on real-valued functions that are convex on a real variable. Specifically, if a real-valued multivariable function is convex on one of its independent real variables, then the Legendre transform with respect to this variable is applicable to the function.

In mathematics, especially vector calculus and differential topology, a closed form is a differential form α whose exterior derivative is zero, and an exact form is a differential form, α, that is the exterior derivative of another differential form β. Thus, an exact form is in the image of d, and a closed form is in the kernel of d.

In multivariable calculus, the implicit function theorem is a tool that allows relations to be converted to functions of several real variables. It does so by representing the relation as the graph of a function. There may not be a single function whose graph can represent the entire relation, but there may be such a function on a restriction of the domain of the relation. The implicit function theorem gives a sufficient condition to ensure that there is such a function.

In mathematics, differential refers to several related notions derived from the early days of calculus, put on a rigorous footing, such as infinitesimal differences and the derivatives of functions.

In mathematics, the total derivative of a function f at a point is the best linear approximation near this point of the function with respect to its arguments. Unlike partial derivatives, the total derivative approximates the function with respect to all of its arguments, not just a single one. In many situations, this is the same as considering all partial derivatives simultaneously. The term "total derivative" is primarily used when f is a function of several variables, because when f is a function of a single variable, the total derivative is the same as the ordinary derivative of the function.

In mathematical analysis, and applications in geometry, applied mathematics, engineering, and natural sciences, a function of a real variable is a function whose domain is the real numbers , or a subset of that contains an interval of positive length. Most real functions that are considered and studied are differentiable in some interval. The most widely considered such functions are the real functions, which are the real-valued functions of a real variable, that is, the functions of a real variable whose codomain is the set of real numbers.

In mathematics, matrix calculus is a specialized notation for doing multivariable calculus, especially over spaces of matrices. It collects the various partial derivatives of a single function with respect to many variables, and/or of a multivariate function with respect to a single variable, into vectors and matrices that can be treated as single entities. This greatly simplifies operations such as finding the maximum or minimum of a multivariate function and solving systems of differential equations. The notation used here is commonly used in statistics and engineering, while the tensor index notation is preferred in physics.

In calculus, the Leibniz integral rule for differentiation under the integral sign, named after Gottfried Wilhelm Leibniz, states that for an integral of the form where and the integrands are functions dependent on the derivative of this integral is expressible as where the partial derivative indicates that inside the integral, only the variation of with is considered in taking the derivative.

A vector-valued function, also referred to as a vector function, is a mathematical function of one or more variables whose range is a set of multidimensional vectors or infinite-dimensional vectors. The input of a vector-valued function could be a scalar or a vector ; the dimension of the function's domain has no relation to the dimension of its range.

In differential calculus, there is no single uniform notation for differentiation. Instead, various notations for the derivative of a function or variable have been proposed by various mathematicians. The usefulness of each notation varies with the context, and it is sometimes advantageous to use more than one notation in a given context. The most common notations for differentiation are listed below.

This is a summary of differentiation rules, that is, rules for computing the derivative of a function in calculus.

In calculus, the differential represents the principal part of the change in a function with respect to changes in the independent variable. The differential is defined by where is the derivative of f with respect to , and is an additional real variable. The notation is such that the equation

In mathematical analysis and its applications, a function of several real variables or real multivariate function is a function with more than one argument, with all arguments being real variables. This concept extends the idea of a function of a real variable to several variables. The "input" variables take real values, while the "output", also called the "value of the function", may be real or complex. However, the study of the complex-valued functions may be easily reduced to the study of the real-valued functions, by considering the real and imaginary parts of the complex function; therefore, unless explicitly specified, only real-valued functions will be considered in this article.

References