Directional derivative

Last updated

A directional derivative is a concept in multivariable calculus that measures the rate at which a function changes in a particular direction at a given point.[ citation needed ]

Contents

The directional derivative of a multivariable differentiable (scalar) function along a given vector v at a given point x intuitively represents the instantaneous rate of change of the function, moving through x with a velocity specified by v.

The directional derivative of a scalar function f with respect to a vector v at a point (e.g., position) x may be denoted by any of the following:

It therefore generalizes the notion of a partial derivative, in which the rate of change is taken along one of the curvilinear coordinate curves, all other coordinates being constant. The directional derivative is a special case of the Gateaux derivative.

Definition

A contour plot of
f
(
x
,
y
)
=
x
2
+
y
2
{\displaystyle f(x,y)=x^{2}+y^{2}}
, showing the gradient vector in black, and the unit vector
u
{\displaystyle \mathbf {u} }
scaled by the directional derivative in the direction of
u
{\displaystyle \mathbf {u} }
in orange. The gradient vector is longer because the gradient points in the direction of greatest rate of increase of a function. Directional derivative contour plot.svg
A contour plot of , showing the gradient vector in black, and the unit vector scaled by the directional derivative in the direction of in orange. The gradient vector is longer because the gradient points in the direction of greatest rate of increase of a function.

The directional derivative of a scalar function

along a vector

is the function defined by the limit [1]

This definition is valid in a broad range of contexts, for example where the norm of a vector (and hence a unit vector) is undefined. [2]

For differentiable functions

If the function f is differentiable at x, then the directional derivative exists along any unit vector v at x, and one has

where the on the right denotes the gradient , is the dot product and v is a unit vector. [3] This follows from defining a path and using the definition of the derivative as a limit which can be calculated along this path to get:

Intuitively, the directional derivative of f at a point x represents the rate of change of f, in the direction of v with respect to time, when moving past x.

Using only direction of vector

The angle a between the tangent A and the horizontal will be maximum if the cutting plane contains the direction of the gradient A. Geometrical interpretation of a directional derivative.svg
The angle α between the tangent A and the horizontal will be maximum if the cutting plane contains the direction of the gradient A.

In a Euclidean space, some authors [4] define the directional derivative to be with respect to an arbitrary nonzero vector v after normalization, thus being independent of its magnitude and depending only on its direction. [5]

This definition gives the rate of increase of f per unit of distance moved in the direction given by v. In this case, one has

or in case f is differentiable at x,

Restriction to a unit vector

In the context of a function on a Euclidean space, some texts restrict the vector v to being a unit vector. With this restriction, both the above definitions are equivalent. [6]

Properties

Many of the familiar properties of the ordinary derivative hold for the directional derivative. These include, for any functions f and g defined in a neighborhood of, and differentiable at, p:

  1. sum rule :
  2. constant factor rule : For any constant c,
  3. product rule (or Leibniz's rule):
  4. chain rule : If g is differentiable at p and h is differentiable at g(p), then

In differential geometry

Let M be a differentiable manifold and p a point of M. Suppose that f is a function defined in a neighborhood of p, and differentiable at p. If v is a tangent vector to M at p, then the directional derivative of f along v, denoted variously as df(v) (see Exterior derivative), (see Covariant derivative), (see Lie derivative), or (see Tangent space § Definition via derivations), can be defined as follows. Let γ : [−1, 1] → M be a differentiable curve with γ(0) = p and γ′(0) = v. Then the directional derivative is defined by

This definition can be proven independent of the choice of γ, provided γ is selected in the prescribed manner so that γ(0) = p and γ′(0) = v.

The Lie derivative

The Lie derivative of a vector field along a vector field is given by the difference of two directional derivatives (with vanishing torsion):

In particular, for a scalar field , the Lie derivative reduces to the standard directional derivative:

The Riemann tensor

Directional derivatives are often used in introductory derivations of the Riemann curvature tensor. Consider a curved rectangle with an infinitesimal vector along one edge and along the other. We translate a covector along then and then subtract the translation along and then . Instead of building the directional derivative using partial derivatives, we use the covariant derivative. The translation operator for is thus

and for ,

The difference between the two paths is then

It can be argued [7] that the noncommutativity of the covariant derivatives measures the curvature of the manifold:

where is the Riemann curvature tensor and the sign depends on the sign convention of the author.

In group theory

Translations

In the Poincaré algebra, we can define an infinitesimal translation operator P as

(the i ensures that P is a self-adjoint operator) For a finite displacement λ, the unitary Hilbert space representation for translations is [8]

By using the above definition of the infinitesimal translation operator, we see that the finite translation operator is an exponentiated directional derivative:

This is a translation operator in the sense that it acts on multivariable functions f(x) as

Proof of the last equation

In standard single-variable calculus, the derivative of a smooth function f(x) is defined by (for small ε)

This can be rearranged to find f(x+ε):

It follows that is a translation operator. This is instantly generalized [9] to multivariable functions f(x)

Here is the directional derivative along the infinitesimal displacement ε. We have found the infinitesimal version of the translation operator:

It is evident that the group multiplication law [10] U(g)U(f)=U(gf) takes the form

So suppose that we take the finite displacement λ and divide it into N parts (N→∞ is implied everywhere), so that λ/N=ε. In other words,

Then by applying U(ε) N times, we can construct U(λ):

We can now plug in our above expression for U(ε):

Using the identity [11]

we have

And since U(ε)f(x) = f(x+ε) we have

Q.E.D.

As a technical note, this procedure is only possible because the translation group forms an Abelian subgroup (Cartan subalgebra) in the Poincaré algebra. In particular, the group multiplication law U(a)U(b) = U(a+b) should not be taken for granted. We also note that Poincaré is a connected Lie group. It is a group of transformations T(ξ) that are described by a continuous set of real parameters . The group multiplication law takes the form

Taking as the coordinates of the identity, we must have

The actual operators on the Hilbert space are represented by unitary operators U(T(ξ)). In the above notation we suppressed the T; we now write U(λ) as U(P(λ)). For a small neighborhood around the identity, the power series representation

is quite good. Suppose that U(T(ξ)) form a non-projective representation, i.e.,

The expansion of f to second power is

After expanding the representation multiplication equation and equating coefficients, we have the nontrivial condition

Since is by definition symmetric in its indices, we have the standard Lie algebra commutator:

with C the structure constant. The generators for translations are partial derivative operators, which commute:

This implies that the structure constants vanish and thus the quadratic coefficients in the f expansion vanish as well. This means that f is simply additive:

and thus for abelian groups,

Q.E.D.

Rotations

The rotation operator also contains a directional derivative. The rotation operator for an angle θ, i.e. by an amount θ = |θ| about an axis parallel to is

Here L is the vector operator that generates SO(3):

It may be shown geometrically that an infinitesimal right-handed rotation changes the position vector x by

So we would expect under infinitesimal rotation:

It follows that

Following the same exponentiation procedure as above, we arrive at the rotation operator in the position basis, which is an exponentiated directional derivative: [12]

Normal derivative

A normal derivative is a directional derivative taken in the direction normal (that is, orthogonal) to some surface in space, or more generally along a normal vector field orthogonal to some hypersurface. See for example Neumann boundary condition. If the normal direction is denoted by , then the normal derivative of a function f is sometimes denoted as . In other notations,

In the continuum mechanics of solids

Several important results in continuum mechanics require the derivatives of vectors with respect to vectors and of tensors with respect to vectors and tensors. [13] The directional directive provides a systematic way of finding these derivatives.

The definitions of directional derivatives for various situations are given below. It is assumed that the functions are sufficiently smooth that derivatives can be taken.

Derivatives of scalar valued functions of vectors

Let f(v) be a real valued function of the vector v. Then the derivative of f(v) with respect to v (or at v) is the vector defined through its dot product with any vector u being

for all vectors u. The above dot product yields a scalar, and if u is a unit vector gives the directional derivative of f at v, in the u direction.

Properties:

  1. If then
  2. If then
  3. If then

Derivatives of vector valued functions of vectors

Let f(v) be a vector valued function of the vector v. Then the derivative of f(v) with respect to v (or at v) is the second order tensor defined through its dot product with any vector u being

for all vectors u. The above dot product yields a vector, and if u is a unit vector gives the direction derivative of f at v, in the directional u.

Properties:

  1. If then
  2. If then
  3. If then

Derivatives of scalar valued functions of second-order tensors

Let be a real valued function of the second order tensor . Then the derivative of with respect to (or at ) in the direction is the second order tensor defined as

for all second order tensors .

Properties:

  1. If then
  2. If then
  3. If then

Derivatives of tensor valued functions of second-order tensors

Let be a second order tensor valued function of the second order tensor . Then the derivative of with respect to (or at ) in the direction is the fourth order tensor defined as

for all second order tensors .

Properties:

  1. If then
  2. If then
  3. If then
  4. If then

See also


Notes

  1. R. Wrede; M.R. Spiegel (2010). Advanced Calculus (3rd ed.). Schaum's Outline Series. ISBN   978-0-07-162366-7.
  2. The applicability extends to functions over spaces without a metric and to differentiable manifolds, such as in general relativity.
  3. If the dot product is undefined, the gradient is also undefined; however, for differentiable f, the directional derivative is still defined, and a similar relation exists with the exterior derivative.
  4. Thomas, George B. Jr.; and Finney, Ross L. (1979) Calculus and Analytic Geometry, Addison-Wesley Publ. Co., fifth edition, p. 593.
  5. This typically assumes a Euclidean space – for example, a function of several variables typically has no definition of the magnitude of a vector, and hence of a unit vector.
  6. Hughes Hallett, Deborah; McCallum, William G.; Gleason, Andrew M. (2012-01-01). Calculus : Single and multivariable. John wiley. p. 780. ISBN   9780470888612. OCLC   828768012.
  7. Zee, A. (2013). Einstein gravity in a nutshell. Princeton: Princeton University Press. p. 341. ISBN   9780691145587.
  8. Weinberg, Steven (1999). The quantum theory of fields (Reprinted (with corr.). ed.). Cambridge [u.a.]: Cambridge Univ. Press. ISBN   9780521550017.
  9. Zee, A. (2013). Einstein gravity in a nutshell. Princeton: Princeton University Press. ISBN   9780691145587.
  10. Cahill, Kevin Cahill (2013). Physical mathematics (Repr. ed.). Cambridge: Cambridge University Press. ISBN   978-1107005211.
  11. Larson, Ron; Edwards, Bruce H. (2010). Calculus of a single variable (9th ed.). Belmont: Brooks/Cole. ISBN   9780547209982.
  12. Shankar, R. (1994). Principles of quantum mechanics (2nd ed.). New York: Kluwer Academic / Plenum. p. 318. ISBN   9780306447907.
  13. J. E. Marsden and T. J. R. Hughes, 2000, Mathematical Foundations of Elasticity, Dover.

Related Research Articles

<span class="mw-page-title-main">Navier–Stokes equations</span> Equations describing the motion of viscous fluid substances

The Navier–Stokes equations are partial differential equations which describe the motion of viscous fluid substances. They were named after French engineer and physicist Claude-Louis Navier and the Irish physicist and mathematician George Gabriel Stokes. They were developed over several decades of progressively building the theories, from 1822 (Navier) to 1842–1850 (Stokes).

<span class="mw-page-title-main">Fokker–Planck equation</span> Partial differential equation

In statistical mechanics and information theory, the Fokker–Planck equation is a partial differential equation that describes the time evolution of the probability density function of the velocity of a particle under the influence of drag forces and random forces, as in Brownian motion. The equation can be generalized to other observables as well. The Fokker-Planck equation has multiple applications in information theory, graph theory, data science, finance, economics etc.

In mathematics, the Laplace operator or Laplacian is a differential operator given by the divergence of the gradient of a scalar function on Euclidean space. It is usually denoted by the symbols , (where is the nabla operator), or . In a Cartesian coordinate system, the Laplacian is given by the sum of second partial derivatives of the function with respect to each independent variable. In other coordinate systems, such as cylindrical and spherical coordinates, the Laplacian also has a useful form. Informally, the Laplacian Δf (p) of a function f at a point p measures by how much the average value of f over small spheres or balls centered at p deviates from f (p).

In continuum mechanics, the infinitesimal strain theory is a mathematical approach to the description of the deformation of a solid body in which the displacements of the material particles are assumed to be much smaller than any relevant dimension of the body; so that its geometry and the constitutive properties of the material at each point of space can be assumed to be unchanged by the deformation.

<span class="mw-page-title-main">Hooke's law</span> Physical law: force needed to deform a spring scales linearly with distance

In physics, Hooke's law is an empirical law which states that the force needed to extend or compress a spring by some distance scales linearly with respect to that distance—that is, Fs = kx, where k is a constant factor characteristic of the spring, and x is small compared to the total possible deformation of the spring. The law is named after 17th-century British physicist Robert Hooke. He first stated the law in 1676 as a Latin anagram. He published the solution of his anagram in 1678 as: ut tensio, sic vis. Hooke states in the 1678 work that he was aware of the law since 1660.

In vector calculus, Green's theorem relates a line integral around a simple closed curve C to a double integral over the plane region D bounded by C. It is the two-dimensional special case of Stokes' theorem.

In the calculus of variations, a field of mathematical analysis, the functional derivative relates a change in a functional to a change in a function on which the functional depends.

<span class="mw-page-title-main">Bloch's theorem</span> Fundamental theorem in condensed matter physics

In condensed matter physics, Bloch's theorem states that solutions to the Schrödinger equation in a periodic potential can be expressed as plane waves modulated by periodic functions. The theorem is named after the physicist Felix Bloch, who discovered the theorem in 1929. Mathematically, they are written

In Hamiltonian mechanics, a canonical transformation is a change of canonical coordinates (q, p) → that preserves the form of Hamilton's equations. This is sometimes known as form invariance. Although Hamilton's equations are preserved, it need not preserve the explicit form of the Hamiltonian itself. Canonical transformations are useful in their own right, and also form the basis for the Hamilton–Jacobi equations and Liouville's theorem.

In mathematics, Green's identities are a set of three identities in vector calculus relating the bulk with the boundary of a region on which differential operators act. They are named after the mathematician George Green, who discovered Green's theorem.

In physics, the Hamilton–Jacobi equation, named after William Rowan Hamilton and Carl Gustav Jacob Jacobi, is an alternative formulation of classical mechanics, equivalent to other formulations such as Newton's laws of motion, Lagrangian mechanics and Hamiltonian mechanics.

<span class="mw-page-title-main">Cartesian tensor</span>

In geometry and linear algebra, a Cartesian tensor uses an orthonormal basis to represent a tensor in a Euclidean space in the form of components. Converting a tensor's components from one such basis to another is done through an orthogonal transformation.

In continuum mechanics, the finite strain theory—also called large strain theory, or large deformation theory—deals with deformations in which strains and/or rotations are large enough to invalidate assumptions inherent in infinitesimal strain theory. In this case, the undeformed and deformed configurations of the continuum are significantly different, requiring a clear distinction between them. This is commonly the case with elastomers, plastically-deforming materials and other fluids and biological soft tissue.

<span class="mw-page-title-main">Maxwell stress tensor</span> Mathematical description in electromagnetism

The Maxwell stress tensor is a symmetric second-order tensor used in classical electromagnetism to represent the interaction between electromagnetic forces and mechanical momentum. In simple situations, such as a point charge moving freely in a homogeneous magnetic field, it is easy to calculate the forces on the charge from the Lorentz force law. When the situation becomes more complicated, this ordinary procedure can become impractically difficult, with equations spanning multiple lines. It is therefore convenient to collect many of these terms in the Maxwell stress tensor, and to use tensor arithmetic to find the answer to the problem at hand.

<span class="mw-page-title-main">Mathematical descriptions of the electromagnetic field</span> Formulations of electromagnetism

There are various mathematical descriptions of the electromagnetic field that are used in the study of electromagnetism, one of the four fundamental interactions of nature. In this article, several approaches are discussed, although the equations are in terms of electric and magnetic fields, potentials, and charges with currents, generally speaking.

<span class="mw-page-title-main">Liénard–Wiechert potential</span> Electromagnetic effect of point charges

The Liénard–Wiechert potentials describe the classical electromagnetic effect of a moving electric point charge in terms of a vector potential and a scalar potential in the Lorenz gauge. Stemming directly from Maxwell's equations, these describe the complete, relativistically correct, time-varying electromagnetic field for a point charge in arbitrary motion, but are not corrected for quantum mechanical effects. Electromagnetic radiation in the form of waves can be obtained from these potentials. These expressions were developed in part by Alfred-Marie Liénard in 1898 and independently by Emil Wiechert in 1900.

The derivatives of scalars, vectors, and second-order tensors with respect to second-order tensors are of considerable use in continuum mechanics. These derivatives are used in the theories of nonlinear elasticity and plasticity, particularly in the design of algorithms for numerical simulations.

In continuum mechanics, a compatible deformation tensor field in a body is that unique tensor field that is obtained when the body is subjected to a continuous, single-valued, displacement field. Compatibility is the study of the conditions under which such a displacement field can be guaranteed. Compatibility conditions are particular cases of integrability conditions and were first derived for linear elasticity by Barré de Saint-Venant in 1864 and proved rigorously by Beltrami in 1886.

The optical metric was defined by German theoretical physicist Walter Gordon in 1923 to study the geometrical optics in curved space-time filled with moving dielectric materials.

<span class="mw-page-title-main">Rock mass plasticity</span>

Plasticity theory for rocks is concerned with the response of rocks to loads beyond the elastic limit. Historically, conventional wisdom has it that rock is brittle and fails by fracture while plasticity is identified with ductile materials. In field scale rock masses, structural discontinuities exist in the rock indicating that failure has taken place. Since the rock has not fallen apart, contrary to expectation of brittle behavior, clearly elasticity theory is not the last word.

References

Commons-logo.svg Media related to Directional derivative at Wikimedia Commons