WikiMili The Free Encyclopedia

In mathematics, the **Dirac delta function** (**δ function**) is a generalized function or distribution introduced by the physicist Paul Dirac. It is used to model the density of an idealized point mass or point charge as a function equal to zero everywhere except for zero and whose integral over the entire real line is equal to one.^{ [1] }^{ [2] }^{ [3] } As there is no function that has these properties, the computations made by the theoretical physicists appeared to mathematicians as nonsense until the introduction of distributions by Laurent Schwartz to formalize and validate the computations. As a distribution, the Dirac delta function is a linear functional that maps every function to its value at zero.^{ [4] }^{ [5] } The Kronecker delta function, which is usually defined on a discrete domain and takes values 0 and 1, is a discrete analog of the Dirac delta function.

**Mathematics** includes the study of such topics as quantity, structure, space, and change.

In mathematics, **generalized functions**, or **distributions**, are objects extending the notion of functions. There is more than one recognized theory. Generalized functions are especially useful in making discontinuous functions more like smooth functions, and describing discrete physical phenomena such as point charges. They are applied extensively, especially in physics and engineering.

**Distributions** are objects that generalize the classical notion of functions in mathematical analysis. Distributions make it possible to differentiate functions whose derivatives do not exist in the classical sense. In particular, any locally integrable function has a distributional derivative. Distributions are widely used in the theory of partial differential equations, where it may be easier to establish the existence of distributional solutions than classical solutions, or appropriate classical solutions may not exist. Distributions are also important in physics and engineering where many problems naturally lead to differential equations whose solutions or initial conditions are distributions, such as the Dirac delta function.

- Motivation and overview
- History
- Definitions
- As a measure
- As a distribution
- Generalizations
- Properties
- Scaling and symmetry
- Algebraic properties
- Translation
- Composition with a function
- Properties in n dimensions
- Fourier transform
- Distributional derivatives
- Higher dimensions
- Representations of the delta function
- Approximations to the identity
- Probabilistic considerations
- Semigroups
- Oscillatory integrals
- Plane wave decomposition
- Fourier kernels
- Hilbert space theory
- Infinitesimal delta functions
- Dirac comb
- Sokhotski–Plemelj theorem
- Relationship to the Kronecker delta
- Applications
- Probability theory
- Quantum mechanics
- Structural mechanics
- See also
- Notes
- References
- External links

In engineering and signal processing, the delta function, also known as the **unit impulse symbol**,^{ [6] } may be regarded through its Laplace transform, as coming from the boundary values of a complex analytic function of a complex variable. The formal rules obeyed by this function are part of the operational calculus, a standard tool kit of physics and engineering. In many applications, the Dirac delta is regarded as a kind of limit (a weak limit) of a sequence of functions having a tall spike at the origin (in theory of distributions, this is a true limit). The approximating functions of the sequence are thus "approximate" or "nascent" delta functions.

**Signal processing** is a subfield of mathematics, information and electrical engineering that concerns the analysis, synthesis, and modification of signals, which are broadly defined as functions conveying "information about the behavior or attributes of some phenomenon", such as sound, images, and biological measurements. For example, signal processing techniques are used to improve signal transmission fidelity, storage efficiency, and subjective quality, and to emphasize or detect components of interest in a measured signal.

In mathematics, the **Laplace transform** is an integral transform named after its inventor Pierre-Simon Laplace. It takes a function of a real variable *t* to a function of a complex variable s.

**Operational calculus**, also known as **operational analysis**, is a technique by which problems in analysis, in particular differential equations, are transformed into algebraic problems, usually the problem of solving a polynomial equation.

The graph of the delta function is usually thought of as following the whole *x*-axis and the positive *y*-axis. The Dirac delta is used to model a tall narrow spike function (an *impulse*), and other similar abstractions such as a point charge, point mass or electron point. For example, to calculate the dynamics of a billiard ball being struck, one can approximate the force of the impact by a delta function. In doing so, one not only simplifies the equations, but one also is able to calculate the motion of the ball by only considering the total impulse of the collision without a detailed model of all of the elastic energy transfer at subatomic levels (for instance).

In mathematics, the **graph** of a function *f* is, formally, the set of all ordered pairs (*x*, *f* ), such that *x* is in the domain of the function *f*, and, in practice, the graphical representation of this set. If the function input *x*, and the values *f*(*x*), are real numbers, the graph is a two-dimensional graph, and, for a continuous function, is a curve. If the function input *x* is an ordered pair (*x*_{1}, *x*_{2}) of real numbers, the graph is the collection of all pairs (, *f* ). These pairs can be identified with the ordered triples (*x*_{1}, *x*_{2}, *f* ). For a continuous function the graph of such a function is a surface.

**Abstraction** in its main sense is a conceptual process where general rules and concepts are derived from the usage and classification of specific examples, literal signifiers, first principles, or other methods.

The **electron** is a subatomic particle, symbol ^{}e^{−}_{} or ^{}β^{−}_{}, whose electric charge is negative one elementary charge. Electrons belong to the first generation of the lepton particle family, and are generally thought to be elementary particles because they have no known components or substructure. The electron has a mass that is approximately 1/1836 that of the proton. Quantum mechanical properties of the electron include an intrinsic angular momentum (spin) of a half-integer value, expressed in units of the reduced Planck constant, *ħ*. As it is a fermion, no two electrons can occupy the same quantum state, in accordance with the Pauli exclusion principle. Like all elementary particles, electrons exhibit properties of both particles and waves: they can collide with other particles and can be diffracted like light. The wave properties of electrons are easier to observe with experiments than those of other particles like neutrons and protons because electrons have a lower mass and hence a longer de Broglie wavelength for a given energy.

To be specific, suppose that a billiard ball is at rest. At time it is struck by another ball, imparting it with a momentum *P*, in . The exchange of momentum is not actually instantaneous, being mediated by elastic processes at the molecular and subatomic level, but for practical purposes it is convenient to consider that energy transfer as effectively instantaneous. The force therefore is . (The units of are .)

In Newtonian mechanics, **linear momentum**, **translational momentum**, or simply **momentum** is the product of the mass and velocity of an object. It is a vector quantity, possessing a magnitude and a direction in three-dimensional space. If *m* is an object's mass and **v** is the velocity, then the momentum is

In physics, a **force** is any interaction that, when unopposed, will change the motion of an object. A force can cause an object with mass to change its velocity, i.e., to accelerate. Force can also be described intuitively as a push or a pull. A force has both magnitude and direction, making it a vector quantity. It is measured in the SI unit of newtons and represented by the symbol **F**.

To model this situation more rigorously, suppose that the force instead is uniformly distributed over a small time interval . That is,

Then the momentum at any time *t* is found by integration:

Now, the model situation of an instantaneous transfer of momentum requires taking the limit as , giving

Here the functions are thought of as useful approximations to the idea of instantaneous transfer of momentum.

The delta function allows us to construct an idealized limit of these approximations. Unfortunately, the actual limit of the functions (in the sense of ordinary calculus) is zero everywhere but a single point, where it is infinite. To make proper sense of the delta function, we should instead insist that the property

which holds for all , should continue to hold in the limit. So, in the equation , it is understood that the limit is always taken *outside the integral*.

In applied mathematics, as we have done here, the delta function is often manipulated as a kind of limit (a weak limit) of a sequence of functions, each member of which has a tall spike at the origin: for example, a sequence of Gaussian distributions centered at the origin with variance tending to zero.

In mathematics, a **sequence** is an enumerated collection of objects in which repetitions are allowed. Like a set, it contains members. The number of elements is called the *length* of the sequence. Unlike a set, the same elements can appear multiple times at different positions in a sequence, and order matters. Formally, a sequence can be defined as a function whose domain is either the set of the natural numbers or the set of the first *n* natural numbers. The position of an element in a sequence is its *rank* or *index*; it is the natural number from which the element is the image. It depends on the context or a specific convention, if the first element has index 0 or 1. When a symbol has been chosen for denoting a sequence, the *n*th element of the sequence is denoted by this symbol with *n* as subscript; for example, the *n*th element of the Fibonacci sequence is generally denoted *F*_{n}.

In probability theory and statistics, **variance** is the expectation of the squared deviation of a random variable from its mean. Informally, it measures how far a set of (random) numbers are spread out from their average value. Variance has a central role in statistics, where some ideas that use it include descriptive statistics, statistical inference, hypothesis testing, goodness of fit, and Monte Carlo sampling. Variance is an important tool in the sciences, where statistical analysis of data is common. The variance is the square of the standard deviation, the second central moment of a distribution, and the covariance of the random variable with itself, and it is often represented by , , or .

Despite its name, the delta function is not truly a function, at least not a usual one with range in real numbers. For example, the objects *f*(*x*) = *δ*(*x*) and *g*(*x*) = 0 are equal everywhere except at *x* = 0 yet have integrals that are different. According to Lebesgue integration theory, if *f* and *g* are functions such that *f* = *g* almost everywhere, then *f* is integrable if and only if *g* is integrable and the integrals of *f* and *g* are identical. A rigorous approach to regarding the Dirac delta function as a mathematical object in its own right requires measure theory or the theory of distributions.

In mathematics, a **real number** is a value of a continuous quantity that can represent a distance along a line. The adjective *real* in this context was introduced in the 17th century by René Descartes, who distinguished between real and imaginary roots of polynomials. The real numbers include all the rational numbers, such as the integer −5 and the fraction 4/3, and all the irrational numbers, such as √2. Included within the irrationals are the transcendental numbers, such as π (3.14159265...). In addition to measuring distance, real numbers can be used to measure quantities such as time, mass, energy, velocity, and many more.

In measure theory, a property holds **almost everywhere** if, in a technical sense, the set for which the property holds takes up nearly all possibilities. The notion of almost everywhere is a companion notion to the concept of measure zero. In the subject of probability, which is largely based in measure theory, the notion is referred to as *almost surely*.

In logic and related fields such as mathematics and philosophy, **if and only if** is a biconditional logical connective between statements.

Joseph Fourier presented what is now called the Fourier integral theorem in his treatise *Théorie analytique de la chaleur* in the form:^{ [7] }

which is tantamount to the introduction of the *δ*-function in the form:^{ [8] }

Later, Augustin Cauchy expressed the theorem using exponentials:^{ [9] }^{ [10] }

Cauchy pointed out that in some circumstances the *order* of integration in this result is significant (contrast Fubini's theorem).^{ [11] }^{ [12] }

As justified using the theory of distributions, the Cauchy equation can be rearranged to resemble Fourier's original formulation and expose the *δ*-function as:

where the *δ*-function is expressed as:

A rigorous interpretation of the exponential form and the various limitations upon the function *f* necessary for its application extended over several centuries. The problems with a classical interpretation are explained as follows:^{ [13] }

- The greatest drawback of the classical Fourier transformation is a rather narrow class of functions (originals) for which it can be effectively computed. Namely, it is necessary that these functions decrease sufficiently rapidly to zero (in the neighborhood of infinity) in order to ensure the existence of the Fourier integral. For example, the Fourier transform of such simple functions as polynomials does not exist in the classical sense. The extension of the classical Fourier transformation to distributions considerably enlarged the class of functions that could be transformed and this removed many obstacles.

Further developments included generalization of the Fourier integral, "beginning with Plancherel's pathbreaking *L*^{2}-theory (1910), continuing with Wiener's and Bochner's works (around 1930) and culminating with the amalgamation into L. Schwartz's theory of distributions (1945) ...",^{ [14] } and leading to the formal development of the Dirac delta function.

An infinitesimal formula for an infinitely tall, unit impulse delta function (infinitesimal version of Cauchy distribution) explicitly appears in an 1827 text of Augustin Louis Cauchy.^{ [15] } Siméon Denis Poisson considered the issue in connection with the study of wave propagation as did Gustav Kirchhoff somewhat later. Kirchhoff and Hermann von Helmholtz also introduced the unit impulse as a limit of Gaussians, which also corresponded to Lord Kelvin's notion of a point heat source. At the end of the 19th century, Oliver Heaviside used formal Fourier series to manipulate the unit impulse.^{ [16] } The Dirac delta function as such was introduced as a "convenient notation" by Paul Dirac in his influential 1930 book * The Principles of Quantum Mechanics *.^{ [17] } He called it the "delta function" since he used it as a continuous analogue of the discrete Kronecker delta.

The Dirac delta can be loosely thought of as a function on the real line which is zero everywhere except at the origin, where it is infinite,

and which is also constrained to satisfy the identity

^{ [18] }

This is merely a heuristic characterization. The Dirac delta is not a function in the traditional sense as no function defined on the real numbers has these properties.^{ [17] } The Dirac delta function can be rigorously defined either as a distribution or as a measure.

One way to rigorously capture the notion of the Dirac delta function is to define a measure, which accepts a subset *A* of the real line **R** as an argument, and returns *δ*(*A*) = 1 if 0 ∈ *A*, and *δ*(*A*) = 0 otherwise.^{ [19] } If the delta function is conceptualized as modeling an idealized point mass at 0, then *δ*(*A*) represents the mass contained in the set *A*. One may then define the integral against *δ* as the integral of a function against this mass distribution. Formally, the Lebesgue integral provides the necessary analytic device. The Lebesgue integral with respect to the measure *δ* satisfies

for all continuous compactly supported functions *f*. The measure *δ* is not absolutely continuous with respect to the Lebesgue measure — in fact, it is a singular measure. Consequently, the delta measure has no Radon–Nikodym derivative — no true function for which the property

holds.^{ [20] } As a result, the latter notation is a convenient abuse of notation, and not a standard (Riemann or Lebesgue) integral.

As a probability measure on **R**, the delta measure is characterized by its cumulative distribution function, which is the unit step function ^{ [21] }

This means that *H*(*x*) is the integral of the cumulative indicator function **1**_{(−∞, x]} with respect to the measure *δ*; to wit,

Thus in particular the integral of the delta function against a continuous function can be properly understood as a Riemann–Stieltjes integral:^{ [22] }

All higher moments of *δ* are zero. In particular, characteristic function and moment generating function are both equal to one.

In the theory of distributions, a generalized function is considered not a function in itself but only in relation to how it affects other functions when "integrated" against them. In keeping with this philosophy, to define the delta function properly, it is enough to say what the "integral" of the delta function is against a sufficiently "good" **test function** *φ*. Test functions are also known as bump functions. If the delta function is already understood as a measure, then the Lebesgue integral of a test function against that measure supplies the necessary integral.

A typical space of test functions consists of all smooth functions on **R** with compact support that have as many derivatives as required. As a distribution, the Dirac delta is a linear functional on the space of test functions and is defined by^{ [23] }

**(1)**

for every test function *.*

For *δ* to be properly a distribution, it must be continuous in a suitable topology on the space of test functions. In general, for a linear functional *S* on the space of test functions to define a distribution, it is necessary and sufficient that, for every positive integer *N* there is an integer *M*_{N} and a constant *C*_{N} such that for every test function *φ*, one has the inequality^{ [24] }

With the *δ* distribution, one has such an inequality (with *C*_{N} = 1) with *M*_{N} = 0 for all *N*. Thus *δ* is a distribution of order zero. It is, furthermore, a distribution with compact support (the support being {0}).

The delta distribution can also be defined in a number of equivalent ways. For instance, it is the distributional derivative of the Heaviside step function. This means that, for every test function *φ*, one has

Intuitively, if integration by parts were permitted, then the latter integral should simplify to

and indeed, a form of integration by parts is permitted for the Stieltjes integral, and in that case one does have

In the context of measure theory, the Dirac measure gives rise to a distribution by integration. Conversely, equation (** 1 **) defines a Daniell integral on the space of all compactly supported continuous functions *φ* which, by the Riesz representation theorem, can be represented as the Lebesgue integral of *φ* with respect to some Radon measure.

Generally, when the term "*Dirac delta function*" is used, it is in the sense of distributions rather than measures, the Dirac measure being among several terms for the corresponding notion in measure theory. Some sources may also use the term *Dirac delta distribution*.

The delta function can be defined in *n*-dimensional Euclidean space **R**^{n} as the measure such that

for every compactly supported continuous function *f*. As a measure, the *n*-dimensional delta function is the product measure of the 1-dimensional delta functions in each variable separately. Thus, formally, with **x** = (*x*_{1}, *x*_{2}, ..., *x*_{n}), one has^{ [6] }

**(2)**

The delta function can also be defined in the sense of distributions exactly as above in the one-dimensional case.^{ [25] } However, despite widespread use in engineering contexts, (** 2 **) should be manipulated with care, since the product of distributions can only be defined under quite narrow circumstances.^{ [26] }

The notion of a ** Dirac measure ** makes sense on any set.^{ [19] } Thus if *X* is a set, *x*_{0} ∈ *X* is a marked point, and Σ is any sigma algebra of subsets of *X*, then the measure defined on sets *A* ∈ Σ by

is the delta measure or unit mass concentrated at *x*_{0}.

Another common generalization of the delta function is to a differentiable manifold where most of its properties as a distribution can also be exploited because of the differentiable structure. The delta function on a manifold *M* centered at the point *x*_{0} ∈ *M* is defined as the following distribution:

**(3)**

for all compactly supported smooth real-valued functions *φ* on *M*.^{ [27] } A common special case of this construction is when *M* is an open set in the Euclidean space **R**^{n}.

On a locally compact Hausdorff space *X*, the Dirac delta measure concentrated at a point *x* is the Radon measure associated with the Daniell integral (** 3 **) on compactly supported continuous functions *φ*.^{ [28] } At this level of generality, calculus as such is no longer possible, however a variety of techniques from abstract analysis are available. For instance, the mapping is a continuous embedding of *X* into the space of finite Radon measures on *X*, equipped with its vague topology. Moreover, the convex hull of the image of *X* under this embedding is dense in the space of probability measures on *X*.^{ [29] }

The delta function satisfies the following scaling property for a non-zero scalar α:^{ [30] }

and so

**(4)**

In particular, the delta function is an even distribution, in the sense that

which is homogeneous of degree −1.

The distributional product of *δ* with *x* is equal to zero:

Conversely, if *xf*(*x*) = *xg*(*x*), where *f* and *g* are distributions, then

for some constant *c*.^{ [31] }

The integral of the time-delayed Dirac delta is given by:

This is sometimes referred to as the *sifting property*^{ [32] } or the *sampling property*. The delta function is said to "sift out" the value at *t* = *T*.

It follows that the effect of convolving a function *f*(*t*) with the time-delayed Dirac delta is to time-delay *f*(*t*) by the same amount:

(using ( **4**): )

This holds under the precise condition that *f* be a tempered distribution (see the discussion of the Fourier transform below). As a special case, for instance, we have the identity (understood in the distribution sense)

More generally, the delta distribution may be composed with a smooth function *g*(*x*) in such a way that the familiar change of variables formula holds, that

provided that *g* is a continuously differentiable function with *g*′ nowhere zero.^{ [33] } That is, there is a unique way to assign meaning to the distribution so that this identity holds for all compactly supported test functions *f*. Therefore, the domain must be broken up to exclude the *g*′ = 0 point. This distribution satisfies *δ*(*g*(*x*)) = 0 if *g* is nowhere zero, and otherwise if *g* has a real root at *x*_{0}, then

It is natural therefore to *define* the composition *δ*(*g*(*x*)) for continuously differentiable functions *g* by

where the sum extends over all roots of *g*(*x*), which are assumed to be simple.^{ [33] } Thus, for example

In the integral form the generalized scaling property may be written as

The delta distribution in an *n*-dimensional space satisfies the following scaling property instead,

so that *δ* is a homogeneous distribution of degree −*n*.

Under any reflection or rotation ρ, the delta function is invariant,

As in the one-variable case, it is possible to define the composition of *δ* with a bi-Lipschitz function ^{ [34] }*g*: **R**^{n} → **R**^{n} uniquely so that the identity

for all compactly supported functions *f*.

Using the coarea formula from geometric measure theory, one can also define the composition of the delta function with a submersion from one Euclidean space to another one of different dimension; the result is a type of current. In the special case of a continuously differentiable function *g*: **R**^{n} → **R** such that the gradient of *g* is nowhere zero, the following identity holds^{ [35] }

where the integral on the right is over *g*^{−1}(0), the (*n* − 1)-dimensional surface defined by *g*(**x**) = 0 with respect to the Minkowski content measure. This is known as a *simple layer* integral.

More generally, if *S* is a smooth hypersurface of **R**^{n}, then we can associate to *S* the distribution that integrates any compactly supported smooth function *g* over *S*:

where σ is the hypersurface measure associated to *S*. This generalization is associated with the potential theory of simple layer potentials on *S*. If *D* is a domain in **R**^{n} with smooth boundary *S*, then *δ*_{S} is equal to the normal derivative of the indicator function of *D* in the distribution sense,

where *n* is the outward normal.^{ [36] }^{ [37] } For a proof, see e.g. the article on the surface delta function.

The delta function is a tempered distribution, and therefore it has a well-defined Fourier transform. Formally, one finds^{ [38] }

Properly speaking, the Fourier transform of a distribution is defined by imposing self-adjointness of the Fourier transform under the duality pairing of tempered distributions with Schwartz functions. Thus is defined as the unique tempered distribution satisfying

for all Schwartz functions *φ*. And indeed it follows from this that

As a result of this identity, the convolution of the delta function with any other tempered distribution *S* is simply *S*:

That is to say that *δ* is an identity element for the convolution on tempered distributions, and in fact the space of compactly supported distributions under convolution is an associative algebra with identity the delta function. This property is fundamental in signal processing, as convolution with a tempered distribution is a linear time-invariant system, and applying the linear time-invariant system measures its impulse response. The impulse response can be computed to any desired degree of accuracy by choosing a suitable approximation for *δ*, and once it is known, it characterizes the system completely. See LTI system theory § Impulse response and convolution.

The inverse Fourier transform of the tempered distribution *f*(*ξ*) = 1 is the delta function. Formally, this is expressed

and more rigorously, it follows since

for all Schwartz functions *f*.

In these terms, the delta function provides a suggestive statement of the orthogonality property of the Fourier kernel on **R**. Formally, one has

This is, of course, shorthand for the assertion that the Fourier transform of the tempered distribution

is

which again follows by imposing self-adjointness of the Fourier transform.

By analytic continuation of the Fourier transform, the Laplace transform of the delta function is found to be^{ [39] }

The distributional derivative of the Dirac delta distribution is the distribution *δ*′ defined on compactly supported smooth test functions *φ* by^{ [40] }

The first equality here is a kind of integration by parts, for if *δ* were a true function then

The *k*-th derivative of *δ* is defined similarly as the distribution given on test functions by

In particular, *δ* is an infinitely differentiable distribution.

The first derivative of the delta function is the distributional limit of the difference quotients:^{ [41] }

More properly, one has

where τ_{h} is the translation operator, defined on functions by *τ*_{h}*φ*(*x*) = *φ*(*x* + *h*), and on a distribution *S* by

In the theory of electromagnetism, the first derivative of the delta function represents a point magnetic dipole situated at the origin. Accordingly, it is referred to as a dipole or the doublet function.^{ [42] }

The derivative of the delta function satisfies a number of basic properties, including:

^{ [43] }

Furthermore, the convolution of *δ*′ with a compactly supported smooth function *f* is

which follows from the properties of the distributional derivative of a convolution.

More generally, on an open set *U* in the *n*-dimensional Euclidean space **R**^{n}, the Dirac delta distribution centered at a point *a* ∈ *U* is defined by^{ [44] }

for all *φ* ∈ *S*(*U*), the space of all smooth compactly supported functions on *U*. If *α* = (*α*_{1}, ..., *α*_{n}) is any multi-index and ∂^{α} denotes the associated mixed partial derivative operator, then the *α*th derivative ∂^{α}*δ*_{a} of *δ*_{a} is given by^{ [44] }

That is, the *α*th derivative of *δ*_{a} is the distribution whose value on any test function *φ* is the *α*th derivative of *φ* at *a* (with the appropriate positive or negative sign).

The first partial derivatives of the delta function are thought of as double layers along the coordinate planes. More generally, the normal derivative of a simple layer supported on a surface is a double layer supported on that surface, and represents a laminar magnetic monopole. Higher derivatives of the delta function are known in physics as multipoles.

Higher derivatives enter into mathematics naturally as the building blocks for the complete structure of distributions with point support. If *S* is any distribution on *U* supported on the set {*a*} consisting of a single point, then there is an integer *m* and coefficients *c*_{α} such that^{ [45] }

The delta function can be viewed as the limit of a sequence of functions

where *η _{ε}*(

**(5)**

for all continuous functions *f* having compact support, or that this limit holds for all smooth functions *f* with compact support. The difference between these two slightly different modes of weak convergence is often subtle: the former is convergence in the vague topology of measures, and the latter is convergence in the sense of distributions.

Typically a nascent delta function *η _{ε}* can be constructed in the following manner. Let

In *n* dimensions, one uses instead the scaling

Then a simple change of variables shows that *η _{ε}* also has integral 1. One may show that (

The *η _{ε}* constructed in this way are known as an

This limit holds in the sense of mean convergence (convergence in *L*^{1}). Further conditions on the *η _{ε}*, for instance that it be a mollifier associated to a compactly supported function,

If the initial *η* = *η*_{1} is itself smooth and compactly supported then the sequence is called a mollifier. The standard mollifier is obtained by choosing *η* to be a suitably normalized bump function, for instance

In some situations such as numerical analysis, a piecewise linear approximation to the identity is desirable. This can be obtained by taking *η*_{1} to be a hat function. With this choice of *η*_{1}, one has

which are all continuous and compactly supported, although not smooth and so not a mollifier.

In the context of probability theory, it is natural to impose the additional condition that the initial *η*_{1} in an approximation to the identity should be positive, as such a function then represents a probability distribution. Convolution with a probability distribution is sometimes favorable because it does not result in overshoot or undershoot, as the output is a convex combination of the input values, and thus falls between the maximum and minimum of the input function. Taking *η*_{1} to be any probability distribution at all, and letting *η _{ε}*(

Another example is with the Wigner semicircle distribution

This is continuous and compactly supported, but not a mollifier because it is not smooth.

Nascent delta functions often arise as convolution semigroups. This amounts to the further constraint that the convolution of *η _{ε}* with

for all *ε*, *δ* > 0. Convolution semigroups in *L*^{1} that form a nascent delta function are always an approximation to the identity in the above sense, however the semigroup condition is quite a strong restriction.

In practice, semigroups approximating the delta function arise as fundamental solutions or Green's functions to physically motivated elliptic or parabolic partial differential equations. In the context of applied mathematics, semigroups arise as the output of a linear time-invariant system. Abstractly, if *A* is a linear operator acting on functions of *x*, then a convolution semigroup arises by solving the initial value problem

in which the limit is as usual understood in the weak sense. Setting *η _{ε}*(

Some examples of physically important convolution semigroups arising from such a fundamental solution include the following.

- The heat kernel

The heat kernel, defined by

represents the temperature in an infinite wire at time *t* > 0, if a unit of heat energy is stored at the origin of the wire at time *t* = 0. This semigroup evolves according to the one-dimensional heat equation:

In probability theory, *η _{ε}*(

In higher-dimensional Euclidean space **R**^{n}, the heat kernel is

and has the same physical interpretation, * mutatis mutandis *. It also represents a nascent delta function in the sense that *η _{ε}* →

- The Poisson kernel

The Poisson kernel

is the fundamental solution of the Laplace equation in the upper half-plane.^{ [50] } It represents the electrostatic potential in a semi-infinite plate whose potential along the edge is held at fixed at the delta function. The Poisson kernel is also closely related to the Cauchy distribution. This semigroup evolves according to the equation

where the operator is rigorously defined as the Fourier multiplier

In areas of physics such as wave propagation and wave mechanics, the equations involved are hyperbolic and so may have more singular solutions. As a result, the nascent delta functions that arise as fundamental solutions of the associated Cauchy problems are generally oscillatory integrals. An example, which comes from a solution of the Euler–Tricomi equation of transonic gas dynamics,^{ [51] } is the rescaled Airy function

Although using the Fourier transform, it is easy to see that this generates a semigroup in some sense—it is not absolutely integrable and so cannot define a semigroup in the above strong sense. Many nascent delta functions constructed as oscillatory integrals only converge in the sense of distributions (an example is the Dirichlet kernel below), rather than in the sense of measures.

Another example is the Cauchy problem for the wave equation in **R**^{1+1}:^{ [52] }

The solution *u* represents the displacement from equilibrium of an infinite elastic string, with an initial disturbance at the origin.

Other approximations to the identity of this kind include the sinc function (used widely in electronics and telecommunications)

and the Bessel function

One approach to the study of a linear partial differential equation

where *L* is a differential operator on **R**^{n}, is to seek first a fundamental solution, which is a solution of the equation

When *L* is particularly simple, this problem can often be resolved using the Fourier transform directly (as in the case of the Poisson kernel and heat kernel already mentioned). For more complicated operators, it is sometimes easier first to consider an equation of the form

where *h* is a plane wave function, meaning that it has the form

for some vector ξ. Such an equation can be resolved (if the coefficients of *L* are analytic functions) by the Cauchy–Kovalevskaya theorem or (if the coefficients of *L* are constant) by quadrature. So, if the delta function can be decomposed into plane waves, then one can in principle solve linear partial differential equations.

Such a decomposition of the delta function into plane waves was part of a general technique first introduced essentially by Johann Radon, and then developed in this form by Fritz John (1955).^{ [53] } Choose *k* so that *n* + *k* is an even integer, and for a real number *s*, put

Then *δ* is obtained by applying a power of the Laplacian to the integral with respect to the unit sphere measure dω of *g*(*x* · *ξ*) for *ξ* in the unit sphere *S*^{n−1}:

The Laplacian here is interpreted as a weak derivative, so that this equation is taken to mean that, for any test function *φ*,

The result follows from the formula for the Newtonian potential (the fundamental solution of Poisson's equation). This is essentially a form of the inversion formula for the Radon transform, because it recovers the value of *φ*(*x*) from its integrals over hyperplanes. For instance, if *n* is odd and *k* = 1, then the integral on the right hand side is

where *Rφ*(*ξ*, *p*) is the Radon transform of *φ*:

An alternative equivalent expression of the plane wave decomposition, from Gel'fand & Shilov (1966–1968 , I, §3.10), is

for *n* even, and

for *n* odd.

In the study of Fourier series, a major question consists of determining whether and in what sense the Fourier series associated with a periodic function converges to the function. The *n*th partial sum of the Fourier series of a function *f* of period 2π is defined by convolution (on the interval [−π,π]) with the Dirichlet kernel:

Thus,

where

A fundamental result of elementary Fourier series states that the Dirichlet kernel tends to the a multiple of the delta function as *N* → ∞. This is interpreted in the distribution sense, that

for every compactly supported *smooth* function *f*. Thus, formally one has

on the interval [−π,π].

In spite of this, the result does not hold for all compactly supported *continuous* functions: that is *D _{N}* does not converge weakly in the sense of measures. The lack of convergence of the Fourier series has led to the introduction of a variety of summability methods in order to produce convergence. The method of Cesàro summation leads to the Fejér kernel

The Fejér kernels tend to the delta function in a stronger sense that^{ [55] }

for every compactly supported *continuous* function *f*. The implication is that the Fourier series of any continuous function is Cesàro summable to the value of the function at every point.

The Dirac delta distribution is a densely defined unbounded linear functional on the Hilbert space L^{2} of square-integrable functions. Indeed, smooth compactly support functions are dense in *L*^{2}, and the action of the delta distribution on such functions is well-defined. In many applications, it is possible to identify subspaces of *L*^{2} and to give a stronger topology on which the delta function defines a bounded linear functional.

- Sobolev spaces

The Sobolev embedding theorem for Sobolev spaces on the real line **R** implies that any square-integrable function *f* such that

is automatically continuous, and satisfies in particular

Thus *δ* is a bounded linear functional on the Sobolev space *H*^{1}. Equivalently *δ* is an element of the continuous dual space *H*^{−1} of *H*^{1}. More generally, in *n* dimensions, one has *δ* ∈ *H*^{−s}(**R**^{n}) provided *s* > *n* / 2.

In complex analysis, the delta function enters via Cauchy's integral formula, which asserts that if *D* is a domain in the complex plane with smooth boundary, then

for all holomorphic functions *f* in *D* that are continuous on the closure of *D*. As a result, the delta function *δ*_{z} is represented in this class of holomorphic functions by the Cauchy integral:

Moreover, let *H*^{2}(∂*D*) be the Hardy space consisting of the closure in *L*^{2}(∂*D*) of all holomorphic functions in *D* continuous up to the boundary of *D*. Then functions in *H*^{2}(∂*D*) uniquely extend to holomorphic functions in *D*, and the Cauchy integral formula continues to hold. In particular for *z* ∈ *D*, the delta function *δ*_{z} is a continuous linear functional on *H*^{2}(∂*D*). This is a special case of the situation in several complex variables in which, for smooth domains *D*, the Szegő kernel plays the role of the Cauchy integral.

Given a complete orthonormal basis set of functions {*φ*_{n}} in a separable Hilbert space, for example, the normalized eigenvectors of a compact self-adjoint operator, any vector *f* can be expressed as:

The coefficients {α_{n}} are found as:

which may be represented by the notation:

a form of the bra–ket notation of Dirac.^{ [56] } Adopting this notation, the expansion of *f* takes the dyadic form:^{ [57] }

Letting *I* denote the identity operator on the Hilbert space, the expression

is called a resolution of the identity. When the Hilbert space is the space *L*^{2}(*D*) of square-integrable functions on a domain *D*, the quantity:

is an integral operator, and the expression for *f* can be rewritten as:

The right-hand side converges to *f* in the *L*^{2} sense. It need not hold in a pointwise sense, even when *f* is a continuous function. Nevertheless, it is common to abuse notation and write

resulting in the representation of the delta function:^{ [58] }

With a suitable rigged Hilbert space (Φ, *L*^{2}(*D*), Φ*) where Φ ⊂ *L*^{2}(*D*) contains all compactly supported smooth functions, this summation may converge in Φ*, depending on the properties of the basis *φ*_{n}. In most cases of practical interest, the orthonormal basis comes from an integral or differential operator, in which case the series converges in the distribution sense.^{ [59] }

Cauchy used an infinitesimal α to write down a unit impulse, infinitely tall and narrow Dirac-type delta function *δ _{α}* satisfying in a number of articles in 1827.

Non-standard analysis allows one to rigorously treat infinitesimals. The article by Yamashita (2007) contains a bibliography on modern Dirac delta functions in the context of an infinitesimal-enriched continuum provided by the hyperreals. Here the Dirac delta can be given by an actual function, having the property that for every real function *F* one has as anticipated by Fourier and Cauchy.

A so-called uniform "pulse train" of Dirac delta measures, which is known as a Dirac comb, or as the Shah distribution, creates a sampling function, often used in digital signal processing (DSP) and discrete time signal analysis. The Dirac comb is given as the infinite sum, whose limit is understood in the distribution sense,

which is a sequence of point masses at each of the integers.

Up to an overall normalizing constant, the Dirac comb is equal to its own Fourier transform. This is significant because if *f* is any Schwartz function, then the periodization of *f* is given by the convolution

In particular,

is precisely the Poisson summation formula.^{ [61] }

The Sokhotski–Plemelj theorem, important in quantum mechanics, relates the delta function to the distribution p.v. 1/*x*, the Cauchy principal value of the function 1/*x*, defined by

Sokhotsky's formula states that^{ [62] }

Here the limit is understood in the distribution sense, that for all compactly supported smooth functions *f*,

The Kronecker delta *δ _{ij}* is the quantity defined by

for all integers *i*, *j*. This function then satisfies the following analog of the sifting property: if is any doubly infinite sequence, then

Similarly, for any real or complex valued continuous function *f* on **R**, the Dirac delta satisfies the sifting property

This exhibits the Kronecker delta function as a discrete analog of the Dirac delta function.^{ [63] }

In probability theory and statistics, the Dirac delta function is often used to represent a discrete distribution, or a partially discrete, partially continuous distribution, using a probability density function (which is normally used to represent fully continuous distributions). For example, the probability density function *f*(*x*) of a discrete distribution consisting of points **x** = {*x*_{1}, ..., *x _{n}*}, with corresponding probabilities

As another example, consider a distribution which 6/10 of the time returns a standard normal distribution, and 4/10 of the time returns exactly the value 3.5 (i.e. a partly continuous, partly discrete mixture distribution). The density function of this distribution can be written as

The delta function is also used in a completely different way to represent the local time of a diffusion process (like Brownian motion). The local time of a stochastic process *B*(*t*) is given by

and represents the amount of time that the process spends at the point *x* in the range of the process. More precisely, in one dimension this integral can be written

where **1**_{[x−ε, x+ε]} is the indicator function of the interval [*x*−*ε*, *x*+*ε*].

The delta function is expedient in quantum mechanics. The wave function of a particle gives the probability amplitude of finding a particle within a given region of space. Wave functions are assumed to be elements of the Hilbert space *L*^{2} of square-integrable functions, and the total probability of finding a particle within a given interval is the integral of the magnitude of the wave function squared over the interval. A set {*φ*_{n}} of wave functions is orthonormal if they are normalized by

where *δ* here refers to the Kronecker delta. A set of orthonormal wave functions is complete in the space of square-integrable functions if any wave function *ψ* can be expressed as a combination of the *φ*_{n}:

with . Complete orthonormal systems of wave functions appear naturally as the eigenfunctions of the Hamiltonian (of a bound system) in quantum mechanics that measures the energy levels, which are called the eigenvalues. The set of eigenvalues, in this case, is known as the spectrum of the Hamiltonian. In bra–ket notation, as above, this equality implies the resolution of the identity:

Here the eigenvalues are assumed to be discrete, but the set of eigenvalues of an observable may be continuous rather than discrete. An example is the position observable, *Qψ*(*x*) = *x*ψ(*x*). The spectrum of the position (in one dimension) is the entire real line, and is called a continuous spectrum. However, unlike the Hamiltonian, the position operator lacks proper eigenfunctions. The conventional way to overcome this shortcoming is to widen the class of available functions by allowing distributions as well: that is, to replace the Hilbert space of quantum mechanics by an appropriate rigged Hilbert space.^{ [64] } In this context, the position operator has a complete set of eigen-distributions, labeled by the points *y* of the real line, given by

The eigenfunctions of position are denoted by in Dirac notation, and are known as position eigenstates.

Similar considerations apply to the eigenstates of the momentum operator, or indeed any other self-adjoint unbounded operator *P* on the Hilbert space, provided the spectrum of *P* is continuous and there are no degenerate eigenvalues. In that case, there is a set Ω of real numbers (the spectrum), and a collection *φ*_{y} of distributions indexed by the elements of Ω, such that

That is, *φ*_{y} are the eigenvectors of *P*. If the eigenvectors are normalized so that

in the distribution sense, then for any test function ψ,

where

That is, as in the discrete case, there is a resolution of the identity

where the operator-valued integral is again understood in the weak sense. If the spectrum of *P* has both continuous and discrete parts, then the resolution of the identity involves a summation over the discrete spectrum *and* an integral over the continuous spectrum.

The delta function also has many more specialized applications in quantum mechanics, such as the delta potential models for a single and double potential well.

The delta function can be used in structural mechanics to describe transient loads or point loads acting on structures. The governing equation of a simple mass–spring system excited by a sudden force impulse *I* at time *t* = 0 can be written

where *m* is the mass, ξ the deflection and *k* the spring constant.

As another example, the equation governing the static deflection of a slender beam is, according to Euler–Bernoulli theory,

where *EI* is the bending stiffness of the beam, *w* the deflection, *x* the spatial coordinate and *q*(*x*) the load distribution. If a beam is loaded by a point force *F* at *x* = *x*_{0}, the load distribution is written

As integration of the delta function results in the Heaviside step function, it follows that the static deflection of a slender beam subject to multiple point loads is described by a set of piecewise polynomials.

Also a point moment acting on a beam can be described by delta functions. Consider two opposing point forces *F* at a distance *d* apart. They then produce a moment *M* = *Fd* acting on the beam. Now, let the distance *d* approach the limit zero, while *M* is kept constant. The load distribution, assuming a clockwise moment acting at *x* = 0, is written

Point moments can thus be represented by the derivative of the delta function. Integration of the beam equation again results in piecewise polynomial deflection.

- ↑ Arfken & Weber 2000 , p. 84
- ↑ Dirac 1958 , §15 The
*δ*function, p. 58 - ↑ Gel'fand & Shilov 1968 , Volume I, §1.1
- ↑ Gel'fand & Shilov 1968 , Volume I, §1.3
- ↑ Schwartz 1950 , p. 3
- 1 2 Bracewell 1986 , Chapter 5
- ↑ JB Fourier (1822).
*The Analytical Theory of Heat*(English translation by Alexander Freeman, 1878 ed.). The University Press. p. 408., cf p 449 and pp 546–551. The original French text can be found**here**. - ↑ Hikosaburo Komatsu (2002). "Fourier's hyperfunctions and Heaviside's pseudodifferential operators". In Takahiro Kawai; Keiko Fujita.
*Microlocal Analysis and Complex Fourier Analysis*. World Scientific. p. 200. ISBN 978-981-238-161-3. - ↑ Tyn Myint-U.; Lokenath Debnath (2007).
*Linear Partial Differential Equations for Scientists And Engineers*(4th ed.). Springer. p. 4. ISBN 978-0-8176-4393-5. - ↑ Lokenath Debnath; Dambaru Bhatta (2007).
*Integral Transforms And Their Applications*(2nd ed.). CRC Press. p. 2. ISBN 978-1-58488-575-7. - ↑ Ivor Grattan-Guinness (2009).
*Convolutions in French Mathematics, 1800–1840: From the Calculus and Mechanics to Mathematical Analysis and Mathematical Physics, Volume 2*. Birkhäuser. p. 653. ISBN 978-3-7643-2238-0. - ↑ See, for example,
*Des intégrales doubles qui se présentent sous une forme indéterminèe* - ↑ Dragiša Mitrović; Darko Žubrinić (1998).
*Fundamentals of Applied Functional Analysis: Distributions, Sobolev Spaces*. CRC Press. p. 62. ISBN 978-0-582-24694-2. - ↑ Manfred Kracht; Erwin Kreyszig (1989). "On singular integral operators and generalizations". In Themistocles M. Rassias.
*Topics in Mathematical Analysis: A Volume Dedicated to the Memory of A.L. Cauchy*. World Scientific. p. 553. ISBN 978-9971-5-0666-7. - ↑ Laugwitz 1989 , p. 230
- ↑ A more complete historical account can be found in van der Pol & Bremmer 1987 , §V.4.
- 1 2 Dirac 1958 , §15
- ↑ Gel'fand & Shilov 1968 , Volume I, §1.1, p. 1
- 1 2 Rudin 1966 , §1.20
- ↑ Hewitt & Stromberg 1963 , §19.61
- ↑ Driggers 2003 , p. 2321. See also Bracewell 1986 , Chapter 5 for a different interpretation. Other conventions for the assigning the value of the Heaviside function at zero exist, and some of these are not consistent with what follows.
- ↑ Hewitt & Stromberg 1965 , §9.19
- ↑ Strichartz 1994 , §2.2
- ↑ Hörmander 1983 , Theorem 2.1.5
- ↑ Hörmander 1983 , §3.1
- ↑ Strichartz 1994 , §2.3; Hörmander 1983 , §8.2
- ↑ Dieudonné 1972 , §17.3.3
- ↑ Krantz, S. G., & Parks, H. R.,
*Geometric Integration Theory*(Boston: Birkhäuser, 2008), pp. 67–69. - ↑ Federer 1969 , §2.5.19
- ↑ Strichartz 1994 , Problem 2.6.2
- ↑ Vladimirov 1971 , Chapter 2, Example 3(d)
- ↑ Weisstein, Eric W. "Sifting Property".
*MathWorld*. - 1 2 Gel'fand & Shilov 1966–1968 , Vol. 1, §II.2.5
- ↑ Further refinement is possible, namely to submersions, although these require a more involved change of variables formula.
- ↑ Hörmander 1983 , §6.1
- ↑ Lange 2012 , pp.29–30
- ↑ Gelfand & Shilov , p. 212
- ↑ In some conventions for the Fourier transform.
- ↑ Bracewell 1986
- ↑ Gel'fand & Shilov 1966 , p. 26
- ↑ Gel'fand & Shilov 1966 , §2.1
- ↑ Weisstein, Eric W. "Doublet Function".
*MathWorld*. - ↑ The property follows by applying a test function and integration by parts.
- 1 2 Hörmander 1983 , p. 56
- ↑ Hörmander 1983 , p. 56; Rudin 1991 , Theorem 6.25
- ↑ Stein & Weiss , Theorem 1.18
- ↑ Rudin 1991 , §II.6.31
- ↑ More generally, one only needs
*η*=*η*_{1}to have an integrable radially symmetric decreasing rearrangement. - ↑ Saichev & Woyczyński 1997 , §1.1 The "delta function" as viewed by a physicist and an engineer, p. 3
- ↑ Stein & Weiss 1971 , §I.1
- ↑ Vallée & Soares 2004 , §7.2
- ↑ Hörmander 1983 , §7.8
- ↑ See also Courant & Hilbert 1962 , §14.
- ↑ Lang 1997 , p. 312
- ↑ In the terminology of Lang (1997), the Fejér kernel is a Dirac sequence, whereas the Dirichlet kernel is not.
- ↑ The development of this section in bra–ket notation is found in ( Levin 2002 , Coordinate-space wave functions and completeness, pp.=109
*ff*) - ↑ Davis & Thomson 2000 , Perfect operators, p.344
- ↑ Davis & Thomson 2000 , Equation 8.9.11, p. 344
- ↑ de la Madrid, Bohm & Gadella 2002
- ↑ See Laugwitz (1989).
- ↑ Córdoba 1988; Hörmander 1983 , §7.2
- ↑ Vladimirov 1971 , §5.7
- ↑ Hartmann 1997 , pp. 154–155
- ↑ Isham 1995 , §6.2

The **Heaviside step function**, or the **unit step function**, usually denoted by H or θ, is a discontinuous function, named after Oliver Heaviside (1850–1925), whose value is zero for negative arguments and one for positive arguments. It is an example of the general class of step functions, all of which can be represented as linear combinations of translations of this one.

**Noether's** (**first**) **theorem** states that every differentiable symmetry of the action of a physical system has a corresponding conservation law. The theorem was proven by mathematician Emmy Noether in 1915 and published in 1918, after a special case was proven by E. Cosserat & F. Cosserat in 1909. The action of a physical system is the integral over time of a Lagrangian function, from which the system's behavior can be determined by the principle of least action. This theorem only applies to continuous and smooth symmetries over physical space.

The **heat equation** is a parabolic partial differential equation that describes the distribution of heat in a given region over time.

**Synchrotron radiation** is the electromagnetic radiation emitted when charged particles are accelerated radially, i.e., when they are subject to an acceleration perpendicular to their velocity. It is produced, for example, in synchrotrons using bending magnets, undulators and/or wigglers. If the particle is non-relativistic, then the emission is called cyclotron emission. If, on the other hand, the particles are relativistic, sometimes referred to as ultrarelativistic, the emission is called synchrotron emission. Synchrotron radiation may be achieved artificially in synchrotrons or storage rings, or naturally by fast electrons moving through magnetic fields. The radiation produced in this way has a characteristic polarization and the frequencies generated can range over the entire electromagnetic spectrum which is also called continuum radiation.

In mathematics, the **Laplace operator** or **Laplacian** is a differential operator given by the divergence of the gradient of a function on Euclidean space. It is usually denoted by the symbols ∇·∇, ∇^{2}, or Δ. The Laplacian Δ*f*(*p*) of a function *f* at a point *p*, is the rate at which the average value of *f* over spheres centered at *p* deviates from *f*(*p*) as the radius of the sphere grows. In a Cartesian coordinate system, the Laplacian is given by the sum of second partial derivatives of the function with respect to each independent variable. In other coordinate systems such as cylindrical and spherical coordinates, the Laplacian also has a useful form.

In mathematics, the **Kronecker delta** is a function of two variables, usually just non-negative integers. The function is 1 if the variables are equal, and 0 otherwise:

In the calculus of variations, a field of mathematical analysis, the **functional derivative** relates a change in a functional to a change in a function on which the functional depends.

In mathematics, a **Colombeau algebra** is an algebra of a certain kind containing the space of Schwartz distributions. While in classical distribution theory a general multiplication of distributions is not possible, Colombeau algebras provide a rigorous framework for this.

The **path integral formulation** of quantum mechanics is a description of quantum theory that generalizes the action principle of classical mechanics. It replaces the classical notion of a single, unique classical trajectory for a system with a sum, or functional integral, over an infinity of quantum-mechanically possible trajectories to compute a quantum amplitude.

In the theory of stochastic processes, the **Karhunen–Loève theorem**, also known as the **Kosambi–Karhunen–Loève theorem** is a representation of a stochastic process as an infinite linear combination of orthogonal functions, analogous to a Fourier series representation of a function on a bounded interval. The transformation is also known as Hotelling transform and eigenvector transform, and is closely related to principal component analysis (PCA) technique widely used in image processing and in data analysis in many fields.

The **Havriliak–Negami relaxation** is an empirical modification of the Debye relaxation model in electromagnetism. Unlike the Debye model, the Havriliak–Negami relaxation accounts for the asymmetry and broadness of the dielectric dispersion curve. The model was first used to describe the dielectric relaxation of some polymers, by adding two exponential parameters to the Debye equation:

In calculus, **Leibniz's rule** for differentiation under the integral sign, named after Gottfried Leibniz, states that for an integral of the form

In mathematical analysis an **oscillatory integral** is a type of distribution. Oscillatory integrals make rigorous many arguments that, on a naive level, appear to use divergent integrals. It is possible to represent approximate solution operators for many differential equations as oscillatory integrals.

In physics, **Hamilton's principle** is William Rowan Hamilton's formulation of the principle of stationary action. It states that the dynamics of a physical system is determined by a variational problem for a functional based on a single function, the Lagrangian, which contains all physical information concerning the system and the forces acting on it. The variational problem is equivalent to and allows for the derivation of the *differential* equations of motion of the physical system. Although formulated originally for classical mechanics, Hamilton's principle also applies to classical fields such as the electromagnetic and gravitational fields, and plays an important role in quantum mechanics, quantum field theory and criticality theories.

In physics, the **Green's function for Laplace's equation in three variables** is used to describe the response of a particular type of physical system to a point source. In particular, this Green's function arises in systems that can be described by Poisson's equation, a partial differential equation (PDE) of the form

In many-body theory, the term **Green's function** is sometimes used interchangeably with correlation function, but refers specifically to correlators of field operators or creation and annihilation operators.

The **Sokhotski–Plemelj theorem** is a theorem in complex analysis, which helps in evaluating certain integrals. The real-line version of it is often used in physics, although rarely referred to by name. The theorem is named after Julian Sochocki, who proved it in 1868, and Josip Plemelj, who rediscovered it as a main ingredient of his solution of the Riemann-Hilbert problem in 1908.

In mathematics, a **homogeneous distribution** is a distribution *S* on Euclidean space **R**^{n} or **R**^{n} \ {0} that is homogeneous in the sense that, roughly speaking,

**Lagrangian field theory** is a formalism in classical field theory. It is the field-theoretic analogue of Lagrangian mechanics. Lagrangian mechanics is used for discrete particles each with a finite number of degrees of freedom. Lagrangian field theory applies to continua and fields, which have an infinite number of degrees of freedom.

- Aratyn, Henrik; Rasinariu, Constantin (2006),
*A short course in mathematical methods with Maple*, World Scientific, ISBN 978-981-256-461-0 . - Arfken, G. B.; Weber, H. J. (2000),
*Mathematical Methods for Physicists*(5th ed.), Boston, Massachusetts: Academic Press, ISBN 978-0-12-059825-0 . - Bracewell, R. (1986),
*The Fourier Transform and Its Applications*(2nd ed.), McGraw-Hill. - Córdoba, A., "La formule sommatoire de Poisson",
*Comptes Rendus de l'Académie des Sciences, Série I*,**306**: 373–376. - Courant, Richard; Hilbert, David (1962),
*Methods of Mathematical Physics, Volume II*, Wiley-Interscience. - Davis, Howard Ted; Thomson, Kendall T (2000),
*Linear algebra and linear operators in engineering with applications in Mathematica*, Academic Press, ISBN 978-0-12-206349-7 - Dieudonné, Jean (1976),
*Treatise on analysis. Vol. II*, New York: Academic Press [Harcourt Brace Jovanovich Publishers], ISBN 978-0-12-215502-4, MR 0530406 . - Dieudonné, Jean (1972),
*Treatise on analysis. Vol. III*, Boston, Massachusetts: Academic Press, MR 0350769 - Dirac, Paul (1958),
*The Principles of Quantum Mechanics*(4th ed.), Oxford at the Clarendon Press, ISBN 978-0-19-852011-5 . - Driggers, Ronald G. (2003),
*Encyclopedia of Optical Engineering*, CRC Press, ISBN 978-0-8247-0940-2 . - Duistermaat, Hans; Kolk (2010),
*Distributions: Theory and applications*, Springer. - Federer, Herbert (1969),
*Geometric measure theory*, Die Grundlehren der mathematischen Wissenschaften,**153**, New York: Springer-Verlag, pp. xiv+676, ISBN 978-3-540-60656-7, MR 0257325 . - Gannon, Terry (2008), "Vertex operator algebras",
*Princeton Companion to Mathematics*, Princeton University Press. - Gel'fand, I. M.; Shilov, G. E. (1966–1968),
*Generalized functions*,**1–5**, Academic Press. - Hartman, William M. (1997),
*Signals, sound, and sensation*, Springer, ISBN 978-1-56396-283-7 . - Hewitt, E; Stromberg, K (1963),
*Real and abstract analysis*, Springer-Verlag. - Hörmander, L. (1983),
*The analysis of linear partial differential operators I*, Grundl. Math. Wissenschaft.,**256**, Springer, doi:10.1007/978-3-642-96750-4, ISBN 978-3-540-12104-6, MR 0717035 . - Isham, C. J. (1995),
*Lectures on quantum theory: mathematical and structural foundations*, Imperial College Press, ISBN 978-81-7764-190-5 . - John, Fritz (1955),
*Plane waves and spherical means applied to partial differential equations*, Interscience Publishers, New York-London, MR 0075429 . - Lang, Serge (1997),
*Undergraduate analysis*, Undergraduate Texts in Mathematics (2nd ed.), Berlin, New York: Springer-Verlag, doi:10.1007/978-1-4757-2698-5, ISBN 978-0-387-94841-6, MR 1476913 . - Lange, Rutger-Jan (2012), "Potential theory, path integrals and the Laplacian of the indicator",
*Journal of High Energy Physics*,**2012**(11): 29–30, arXiv: 1302.0864 , Bibcode:2012JHEP...11..032L, doi:10.1007/JHEP11(2012)032 . - Laugwitz, D. (1989), "Definite values of infinite sums: aspects of the foundations of infinitesimal analysis around 1820",
*Arch. Hist. Exact Sci.*,**39**(3): 195–245, doi:10.1007/BF00329867 . - Levin, Frank S. (2002), "Coordinate-space wave functions and completeness",
*An introduction to quantum theory*, Cambridge University Press, pp. 109*ff*, ISBN 978-0-521-59841-5 - Li, Y. T.; Wong, R. (2008), "Integral and series representations of the Dirac delta function",
*Commun. Pure Appl. Anal.*,**7**(2): 229–247, arXiv: 1303.1943 , doi:10.3934/cpaa.2008.7.229, MR 2373214 . - de la Madrid, R.; Bohm, A.; Gadella, M. (2002), "Rigged Hilbert Space Treatment of Continuous Spectrum",
*Fortschr. Phys.*,**50**(2): 185–216, arXiv: quant-ph/0109154 , Bibcode:2002ForPh..50..185D, doi:10.1002/1521-3978(200203)50:2<185::AID-PROP185>3.0.CO;2-S . - McMahon, D. (2005-11-22), "An Introduction to State Space",
*Quantum Mechanics Demystified, A Self-Teaching Guide*, Demystified Series, New York: McGraw-Hill, p. 108, doi:10.1036/0071455469, ISBN 978-0-07-145546-6 , retrieved 2008-03-17. - van der Pol, Balth.; Bremmer, H. (1987),
*Operational calculus*(3rd ed.), New York: Chelsea Publishing Co., ISBN 978-0-8284-0327-6, MR 0904873 . - Rudin, W. (1991),
*Functional Analysis*(2nd ed.), McGraw-Hill, ISBN 978-0-07-054236-5 . - Soares, Manuel; Vallée, Olivier (2004),
*Airy functions and applications to physics*, London: Imperial College Press. - Saichev, A I; Woyczyński, Wojbor Andrzej (1997), "Chapter1: Basic definitions and operations",
*Distributions in the Physical and Engineering Sciences: Distributional and fractal calculus, integral transforms, and wavelets*, Birkhäuser, ISBN 978-0-8176-3924-2 - Schwartz, L. (1950),
*Théorie des distributions*,**1**, Hermann. - Schwartz, L. (1951),
*Théorie des distributions*,**2**, Hermann. - Stein, Elias; Weiss, Guido (1971),
*Introduction to Fourier Analysis on Euclidean Spaces*, Princeton University Press, ISBN 978-0-691-08078-9 . - Strichartz, R. (1994),
*A Guide to Distribution Theory and Fourier Transforms*, CRC Press, ISBN 978-0-8493-8273-4 . - Vladimirov, V. S. (1971),
*Equations of mathematical physics*, Marcel Dekker, ISBN 978-0-8247-1713-1 . - Weisstein, Eric W. "Delta Function".
*MathWorld*. - Yamashita, H. (2006), "Pointwise analysis of scalar fields: A nonstandard approach",
*Journal of Mathematical Physics*,**47**(9): 092301, Bibcode:2006JMP....47i2301Y, doi:10.1063/1.2339017 - Yamashita, H. (2007), "Comment on "Pointwise analysis of scalar fields: A nonstandard approach" [J. Math. Phys. 47, 092301 (2006)]",
*Journal of Mathematical Physics*,**48**(8): 084101, Bibcode:2007JMP....48h4101Y, doi:10.1063/1.2771422

- Hazewinkel, Michiel, ed. (2001) [1994], "Delta-function",
*Encyclopedia of Mathematics*, Springer Science+Business Media B.V. / Kluwer Academic Publishers, ISBN 978-1-55608-010-4 - KhanAcademy.org video lesson
- The Dirac Delta function, a tutorial on the Dirac delta function.
- Video Lectures – Lecture 23, a lecture by Arthur Mattuck.
- The Dirac delta measure is a hyperfunction
- We show the existence of a unique solution and analyze a finite element approximation when the source term is a Dirac delta measure
- Non-Lebesgue measures on R. Lebesgue-Stieltjes measure, Dirac delta measure.

This page is based on this Wikipedia article

Text is available under the CC BY-SA 4.0 license; additional terms may apply.

Images, videos and audio are available under their respective licenses.

Text is available under the CC BY-SA 4.0 license; additional terms may apply.

Images, videos and audio are available under their respective licenses.