For a function in three-dimensional Cartesian coordinate variables, the gradient is the vector field:
where i, j, k are the standardunit vectors for the x, y, z-axes. More generally, for a function of n variables , also called a scalar field, the gradient is the vector field: where are mutually orthogonal unit vectors.
As the name implies, the gradient is proportional to, and points in the direction of, the function's most rapid (positive) change.
As the name implies, the divergence is a (local) measure of the degree to which vectors in the field diverge.
The divergence of a tensor field of non-zero order k is written as , a contraction of a tensor field of order k − 1. Specifically, the divergence of a vector is a scalar. The divergence of a higher-order tensor field may be found by decomposing the tensor field into a sum of outer products and using the identity, where is the directional derivative in the direction of multiplied by its magnitude. Specifically, for the outer product of two vectors,[2]
For a tensor field of order k > 1, the tensor field of order k − 1 is defined by the recursive relation where is an arbitrary constant vector.
For a tensor field of order k > 1, the tensor field of order k is defined by the recursive relation where is an arbitrary constant vector.
A tensor field of order greater than one may be decomposed into a sum of outer products, and then the following identity may be used: Specifically, for the outer product of two vectors,[3]
The Laplacian is a measure of how much a function is changing over a small sphere centered at the point.
When the Laplacian is equal to 0, the function is called a harmonic function. That is,
For a tensor field, , the Laplacian is generally written as: and is a tensor field of the same order.
For a tensor field of order k > 0, the tensor field of order k is defined by the recursive relation where is an arbitrary constant vector.
Special notations
In Feynman subscript notation, where the notation ∇B means the subscripted gradient operates on only the factor B.[4][5][6]
More general but similar is the Hestenesoverdot notation in geometric algebra.[7][8] The above identity is then expressed as: where overdots define the scope of the vector derivative. The dotted vector, in this case B, is differentiated, while the (undotted) A is held constant.
The utility of the Feynman subscript notation lies in its use in the derivation of vector and tensor derivative identities, as in the following example which uses the algebraic identity C⋅(A×B) = (C×A)⋅B:
An alternative method is to use the Cartesian components of the del operator as follows (with implicit summation over the indexi):
Another method of deriving vector and tensor derivative identities is to replace all occurrences of a vector in an algebraic identity by the del operator, provided that no variable occurs both inside and outside the scope of an operator or both inside the scope of one operator in a term and outside the scope of another operator in the same term (i.e., the operators must be nested). The validity of this rule follows from the validity of the Feynman method, for one may always substitute a subscripted del and then immediately drop the subscript under the condition of the rule. For example, from the identity A⋅(B×C) = (A×B)⋅C we may derive A⋅(∇×C) = (A×∇)⋅C but not ∇⋅(B×C) = (∇×B)⋅C, nor from A⋅(B×A) = 0 may we derive A⋅(∇×A) = 0. On the other hand, a subscripted del operates on all occurrences of the subscript in the term, so that A⋅(∇A×A) = ∇A⋅(A×A) = ∇⋅(A×A) = 0. Also, from A×(A×C) = A(A⋅C) − (A⋅A)C we may derive ∇×(∇×C) = ∇(∇⋅C) − ∇2C, but from (Aψ)⋅(Aφ) = (A⋅A)(ψφ) we may not derive (∇ψ)⋅(∇φ) = ∇2(ψφ).
A subscript c on a quantity indicates that it is temporarily considered to be a constant. Since a constant is not a variable, when the substitution rule (see the preceding paragraph) is used it, unlike a variable, may be moved into or out of the scope of a del operator, as in the following example:[9]
Another way to indicate that a quantity is a constant is to affix it as a subscript to the scope of a del operator, as follows:[10]
For the remainder of this article, Feynman subscript notation will be used where appropriate.
First derivative identities
For scalar fields , and vector fields , , we have the following derivative identities.
Distributive properties
First derivative associative properties
Product rule for multiplication by a scalar
We have the following generalizations of the product rule in single-variable calculus.
Quotient rule for division by a scalar
Chain rule
Let be a one-variable function from scalars to scalars, a parametrized curve, a function from vectors to scalars, and a vector field. We have the following special cases of the multi-variable chain rule.
For a vector transformation we have:
Here we take the trace of the dot product of two second-order tensors, which corresponds to the product of their matrices.
Here ∇2 is the vector Laplacian operating on the vector field A.
Curl of divergence is not defined
The divergence of a vector field A is a scalar, and the curl of a scalar quantity is undefined. Therefore,
Second derivative associative properties
DCG chart: Some rules for second derivatives.
A mnemonic
The figure to the right is a mnemonic for some of these identities. The abbreviations used are:
D: divergence,
C: curl,
G: gradient,
L: Laplacian,
CC: curl of curl.
Each arrow is labeled with the result of an identity, specifically, the result of applying the operator at the arrow's tail to the operator at its head. The blue circle in the middle means curl of curl exists, whereas the other two red circles (dashed) mean that DD and GG do not exist.
In the following surface–volume integral theorems, V denotes a three-dimensional volume with a corresponding two-dimensional boundaryS = ∂V (a closed surface):
Integration around a closed curve in the clockwise sense is the negative of the same line integral in the counterclockwise sense (analogous to interchanging the limits in a definite integral):
Endpoint-curve integrals
In the following endpoint–curve integral theorems, P denotes a 1d open path with signed 0d boundary points and integration along P is from to :
A tensor form of a vector integral theorem may be obtained by replacing the vector (or one of them) by a tensor, provided that the vector is first made to appear only as the right-most vector of each integrand. For example, Stokes' theorem becomes[18]
.
A scalar field may also be treated as a vector and replaced by a vector or tensor. For example, Green's first identity becomes
.
Similar rules apply to algebraic and differentiation formulas. For algebraic formulas one may alternatively use the left-most vector position.
↑ Wangsness, Roald K.; Cloud, Michael J. (1986). Electromagnetic Fields (2nded.). Wiley. ISBN978-0-471-81186-2.
↑ Page, Leigh; Adams, Norman Ilsley, Jr. (1940). Electrodynamics. New York: D. Van Nostrand Company, Inc. pp. 44–45, Eq. (18-3).{{cite book}}: CS1 maint: multiple names: authors list (link)
↑ Pérez-Garrido, Antonio (2024). "Recovering seldom-used theorems of vector calculus and their application to problems of electromagnetism". American Journal of Physics. 92 (5): 354–359. arXiv:2312.17268. doi:10.1119/5.0182191.
This page is based on this Wikipedia article Text is available under the CC BY-SA 4.0 license; additional terms may apply. Images, videos and audio are available under their respective licenses.