Discrete calculus

Last updated

Discrete calculus or the calculus of discrete functions, is the mathematical study of incremental change, in the same way that geometry is the study of shape and algebra is the study of generalizations of arithmetic operations. The word calculus is a Latin word, meaning originally "small pebble"; as such pebbles were used for calculation, the meaning of the word has evolved and today usually means a method of computation. Meanwhile, calculus , originally called infinitesimal calculus or "the calculus of infinitesimals", is the study of continuous change.

Contents

Discrete calculus has two entry points, differential calculus and integral calculus. Differential calculus concerns incremental rates of change and the slopes of piece-wise linear curves. Integral calculus concerns accumulation of quantities and the areas under piece-wise constant curves. These two points of view are related to each other by the fundamental theorem of discrete calculus.

The study of the concepts of change starts with their discrete form. The development is dependent on a parameter, the increment of the independent variable. If we so choose, we can make the increment smaller and smaller and find the continuous counterparts of these concepts as limits. Informally, the limit of discrete calculus as is infinitesimal calculus. Even though it serves as a discrete underpinning of calculus, the main value of discrete calculus is in applications.

Two initial constructions

Discrete differential calculus is the study of the definition, properties, and applications of the difference quotient of a function. The process of finding the difference quotient is called differentiation. Given a function defined at several points of the real line, the difference quotient at that point is a way of encoding the small-scale (i.e., from the point to the next) behavior of the function. By finding the difference quotient of a function at every pair of consecutive points in its domain, it is possible to produce a new function, called the difference quotient function or just the difference quotient of the original function. In formal terms, the difference quotient is a linear operator which takes a function as its input and produces a second function as its output. This is more abstract than many of the processes studied in elementary algebra, where functions usually input a number and output another number. For example, if the doubling function is given the input three, then it outputs six, and if the squaring function is given the input three, then it outputs nine. The derivative, however, can take the squaring function as an input. This means that the derivative takes all the information of the squaring function—such as that two is sent to four, three is sent to nine, four is sent to sixteen, and so on—and uses this information to produce another function. The function produced by differentiating the squaring function turns out to be something close to the doubling function.

Suppose the functions are defined at points separated by an increment :

The "doubling function" may be denoted by and the "squaring function" by . The "difference quotient" is the rate of change of the function over one of the intervals defined by the formula:

It takes the function as an input, that is all the information—such as that two is sent to four, three is sent to nine, four is sent to sixteen, and so on—and uses this information to output another function, the function , as will turn out. As a matter of convenience, the new function may defined at the middle points of the above intervals:

As the rate of change is that for the whole interval , any point within it can be used as such a reference or, even better, the whole interval which makes the difference quotient a -cochain.

The most common notation for the difference quotient is:

If the input of the function represents time, then the difference quotient represents change with respect to time. For example, if is a function that takes a time as input and gives the position of a ball at that time as output, then the difference quotient of is how the position is changing in time, that is, it is the velocity of the ball.

If a function is linear (that is, if the points of the graph of the function lie on a straight line), then the function can be written as , where is the independent variable, is the dependent variable, is the -intercept, and:

Slope:
m
=
D
y
D
x
=
tan
[?]
(
th
)
{\displaystyle m={\frac {\Delta y}{\Delta x}}=\tan(\theta )} Wiki slope in 2d.svg
Slope:

This gives an exact value for the slope of a straight line.

If the function is not linear, however, then the change in divided by the change in varies. The difference quotient give an exact meaning to the notion of change in output with respect to change in input. To be concrete, let be a function, and fix a point in the domain of . is a point on the graph of the function. If is the increment of , then is the next value of . Therefore, is the increment of . The slope of the line between these two points is

So is the slope of the line between and .

Here is a particular example, the difference quotient of the squaring function. Let be the squaring function. Then:

The difference quotient of the difference quotient is called the second difference quotient and it is defined at

and so on.

Discrete integral calculus is the study of the definitions, properties, and applications of the Riemann sums. The process of finding the value of a sum is called integration. In technical language, integral calculus studies a certain linear operator.

The Riemann sum inputs a function and outputs a function, which gives the algebraic sum of areas between the part of the graph of the input and the x-axis.

A motivating example is the distances traveled in a given time.

If the speed is constant, only multiplication is needed, but if the speed changes, we evaluate the distance traveled by breaking up the time into many short intervals of time, then multiplying the time elapsed in each interval by one of the speeds in that interval, and then taking the sum (a Riemann sum) of the distance traveled in each interval.

Constant velocity Constant velocity.png
Constant velocity
The Riemann sum is measuring the total area of the bars, defined by
f
{\displaystyle f}
, between two points (here
a
{\displaystyle a}
and
b
{\displaystyle b}
). Riemann sum convergence.png
The Riemann sum is measuring the total area of the bars, defined by , between two points (here and ).

When velocity is constant, the total distance traveled over the given time interval can be computed by multiplying velocity and time. For example, travelling a steady 50 mph for 3 hours results in a total distance of 150 miles. In the diagram on the left, when constant velocity and time are graphed, these two values form a rectangle with height equal to the velocity and width equal to the time elapsed. Therefore, the product of velocity and time also calculates the rectangular area under the (constant) velocity curve. This connection between the area under a curve and distance traveled can be extended to any irregularly shaped region exhibiting an incrementally varying velocity over a given time period. If the bars in the diagram on the right represents speed as it varies from an interval to the next, the distance traveled (between the times represented by and ) is the area of the shaded region .

So, the interval between and is divided into a number of equal segments, the length of each segment represented by the symbol . For each small segment, we have one value of the function . Call that value . Then the area of the rectangle with base and height gives the distance (time multiplied by speed ) traveled in that segment. Associated with each segment is the value of the function above it, . The sum of all such rectangles gives the area between the axis and the piece-wise constant curve, which is the total distance traveled.

Suppose a function is defined at the mid-points of the intervals of equal length :

Then the Riemann sum from to in sigma notation is:

As this computation is carried out for each , the new function is defined at the points:

The fundamental theorem of calculus states that differentiation and integration are inverse operations. More precisely, it relates the difference quotients to the Riemann sums. It can also be interpreted as a precise statement of the fact that differentiation is the inverse of integration.

The fundamental theorem of calculus: If a function is defined on a partition of the interval , , and if is a function whose difference quotient is , then we have:

Furthermore, for every , we have:

This is also a prototype solution of a difference equation. Difference equations relate an unknown function to its difference or difference quotient, and are ubiquitous in the sciences.

History

The early history of discrete calculus is the history of calculus. Such basic ideas as the difference quotients and the Riemann sums appear implicitly or explicitly in definitions and proofs. After the limit is taken, however, they are never to be seen again. However, the Kirchhoff's voltage law (1847) can be expressed in terms of the one-dimensional discrete exterior derivative.

During the 20th century discrete calculus remains interlinked with infinitesimal calculus especially differential forms but also starts to draw from algebraic topology as both develop. The main contributions come from the following individuals: [1]

The recent development of discrete calculus, starting with Whitney, has been driven by the needs of applied modeling. [2] [3] [4]

Applications

Discrete calculus is used for modeling either directly or indirectly as a discretization of infinitesimal calculus in every branch of the physical sciences, actuarial science, computer science, statistics, engineering, economics, business, medicine, demography, and in other fields wherever a problem can be mathematically modeled. It allows one to go from (non-constant) rates of change to the total change or vice versa, and many times in studying a problem we know one and are trying to find the other.

Physics makes particular use of calculus; all discrete concepts in classical mechanics and electromagnetism are related through discrete calculus. The mass of an object of known density that varies incrementally, the moment of inertia of such objects, as well as the total energy of an object within a discrete conservative field can be found by the use of discrete calculus. An example of the use of discrete calculus in mechanics is Newton's second law of motion: historically stated it expressly uses the term "change of motion" which implies the difference quotient saying The change of momentum of a body is equal to the resultant force acting on the body and is in the same direction. Commonly expressed today as Force = Mass × Acceleration, it invokes discrete calculus when the change is incremental because acceleration is the difference quotient of velocity with respect to time or second difference quotient of the spatial position. Starting from knowing how an object is accelerating, we use the Riemann sums to derive its path.

Maxwell's theory of electromagnetism and Einstein's theory of general relativity have been expressed in the language of discrete calculus.

Chemistry uses calculus in determining reaction rates and radioactive decay (exponential decay).

In biology, population dynamics starts with reproduction and death rates to model population changes (population modeling).

In engineering, difference equations are used to plot a course of a spacecraft within zero gravity environments, to model heat transfer, diffusion, and wave propagation.

The discrete analogue of Green's theorem is applied in an instrument known as a planimeter, which is used to calculate the area of a flat surface on a drawing. For example, it can be used to calculate the amount of area taken up by an irregularly shaped flower bed or swimming pool when designing the layout of a piece of property. It can be used to efficiently calculate sums of rectangular domains in images, to rapidly extract features and detect object; another algorithm that could be used is the summed area table.

In the realm of medicine, calculus can be used to find the optimal branching angle of a blood vessel so as to maximize flow. From the decay laws for a particular drug's elimination from the body, it is used to derive dosing laws. In nuclear medicine, it is used to build models of radiation transport in targeted tumor therapies.

In economics, calculus allows for the determination of maximal profit by calculating both marginal cost and marginal revenue, as well as modeling of markets. [5]

In signal processing and machine learning, discrete calculus allows for appropriate definitions of operators (e.g., convolution), level set optimization and other key functions for neural network analysis on graph structures. [3]

Discrete calculus can be used in conjunction with other mathematical disciplines. For example, it can be used in probability theory to determine the probability of a discrete random variable from an assumed density function.

Calculus of differences and sums

Suppose a function (a -cochain) is defined at points separated by an increment :

The difference (or the exterior derivative, or the coboundary operator) of the function is given by:

It is defined at each of the above intervals; it is a -cochain.

Suppose a -cochain is defined at each of the above intervals. Then its sum is a function (a -cochain) defined at each of the points by:

These are their properties:

The definitions are applied to graphs as follows. If a function (a -cochain) is defined at the nodes of a graph:

then its exterior derivative (or the differential) is the difference, i.e., the following function defined on the edges of the graph (-cochain):

If is a -cochain, then its integral over a sequence of edges of the graph is the sum of its values over all edges of ("path integral"):

These are the properties:

where a -chain consists of for some fixed , then

See references. [6] [7] [8] [9] [3] [10]

Chains of simplices and cubes

A simplicial complex. Simplicial complex example.svg
A simplicial complex.

A simplicial complex is a set of simplices that satisfies the following conditions:

1. Every face of a simplex from is also in .
2. The non-empty intersection of any two simplices is a face of both and .
The boundary of a boundary of a 2-simplex (left) and the boundary of a 1-chain (right) are taken. Both are 0, being sums in which both the positive and negative of a 0-simplex occur once. The boundary of a boundary is always 0. A nontrivial cycle is something that closes up like the boundary of a simplex, in that its boundary sums to 0, but which isn't actually the boundary of a simplex or chain. Simplicial homology - exactness of boundary maps.svg
The boundary of a boundary of a 2-simplex (left) and the boundary of a 1-chain (right) are taken. Both are 0, being sums in which both the positive and negative of a 0-simplex occur once. The boundary of a boundary is always 0. A nontrivial cycle is something that closes up like the boundary of a simplex, in that its boundary sums to 0, but which isn't actually the boundary of a simplex or chain.

By definition, an orientation of a k-simplex is given by an ordering of the vertices, written as , with the rule that two orderings define the same orientation if and only if they differ by an even permutation. Thus every simplex has exactly two orientations, and switching the order of two vertices changes an orientation to the opposite orientation. For example, choosing an orientation of a 1-simplex amounts to choosing one of the two possible directions, and choosing an orientation of a 2-simplex amounts to choosing what "counterclockwise" should mean.

Let be a simplicial complex. A simplicial k-chain is a finite formal sum

where each ci is an integer and σi is an oriented k-simplex. In this definition, we declare that each oriented simplex is equal to the negative of the simplex with the opposite orientation. For example,

The vector space of k-chains on is written . It has a basis in one-to-one correspondence with the set of k-simplices in . To define a basis explicitly, one has to choose an orientation of each simplex. One standard way to do this is to choose an ordering of all the vertices and give each simplex the orientation corresponding to the induced ordering of its vertices.

Let be an oriented k-simplex, viewed as a basis element of . The boundary operator

is the linear operator defined by:

where the oriented simplex

is the th face of , obtained by deleting its th vertex.

In , elements of the subgroup

are referred to as cycles, and the subgroup

is said to consist of boundaries.

A direct computation shows that . In geometric terms, this says that the boundary of anything has no boundary. Equivalently, the vector spaces form a chain complex. Another equivalent statement is that is contained in .

A cubical complex is a set composed of points, line segments, squares, cubes, and their n-dimensional counterparts. They are used analogously to simplices to form complexes. An elementary interval is a subset of the form

for some . An elementary cube is the finite product of elementary intervals, i.e.

where are elementary intervals. Equivalently, an elementary cube is any translate of a unit cube embedded in Euclidean space (for some with ). A set is a cubicalcomplex if it can be written as a union of elementary cubes (or possibly, is homeomorphic to such a set) and it contains all of the faces of all of its cubes. The boundary operator and the chain complex are defined similarly to those for simplicial complexes.

More general are cell complexes.

A chain complex is a sequence of vector spaces connected by linear operators (called boundary operators) , such that the composition of any two consecutive maps is the zero map. Explicitly, the boundary operators satisfy , or with indices suppressed, . The complex may be written out as follows.

A simplicial map is a map between simplicial complexes with the property that the images of the vertices of a simplex always span a simplex (therefore, vertices have vertices for images). A simplicial map from a simplicial complex to another is a function from the vertex set of to the vertex set of such that the image of each simplex in (viewed as a set of vertices) is a simplex in . It generates a linear map, called a chain map, from the chain complex of to the chain complex of . Explicitly, it is given on -chains by

if are all distinct, and otherwise it is set equal to .

A chain map between two chain complexes and is a sequence of homomorphisms for each that commutes with the boundary operators on the two chain complexes, so . This is written out in the following commutative diagram:

Chain map.svg

A chain map sends cycles to cycles and boundaries to boundaries.

See references. [11] [10] [12]

Discrete differential forms: cochains

For each vector space Ci in the chain complex we consider its dual space and is its dual linear operator

This has the effect of "reversing all the arrows" of the original complex, leaving a cochain complex

The cochain complex is the dual notion to a chain complex. It consists of a sequence of vector spaces connected by linear operators satisfying . The cochain complex may be written out in a similar fashion to the chain complex.

The index in either or is referred to as the degree (or dimension). The difference between chain and cochain complexes is that, in chain complexes, the differentials decrease dimension, whereas in cochain complexes they increase dimension.

The elements of the individual vector spaces of a (co)chain complex are called cochains. The elements in the kernel of are called cocycles (or closed elements), and the elements in the image of are called coboundaries (or exact elements). Right from the definition of the differential, all boundaries are cycles.

The Poincaré lemma states that if is an open ball in , any closed -form defined on is exact, for any integer with .

When we refer to cochains as discrete (differential) forms, we refer to as the exterior derivative. We also use the calculus notation for the values of the forms:

Stokes' theorem is a statement about the discrete differential forms on manifolds, which generalizes the fundamental theorem of discrete calculus for a partition of an interval:

Stokes' theorem says that the sum of a form over the boundary of some orientable manifold is equal to the sum of its exterior derivative over the whole of , i.e.,

Stokes patch.svg

It is worthwhile to examine the underlying principle by considering an example for dimensions. The essential idea can be understood by the diagram on the left, which shows that, in an oriented tiling of a manifold, the interior paths are traversed in opposite directions; their contributions to the path integral thus cancel each other pairwise. As a consequence, only the contribution from the boundary remains.

See references. [11] [10]

The wedge product of forms

In discrete calculus, this is a construction that creates from forms higher order forms: adjoining two cochains of degree and to form a composite cochain of degree .

For cubical complexes, the wedge product is defined on every cube seen as a vector space of the same dimension.

For simplicial complexes, the wedge product is implemented as the cup product: if is a -cochain and is a -cochain, then

where is a -simplex and , is the simplex spanned by into the -simplex whose vertices are indexed by . So, is the -th front face and is the -th back face of , respectively.

The coboundary of the cup product of cochains and is given by

The cup product of two cocycles is again a cocycle, and the product of a coboundary with a cocycle (in either order) is a coboundary.

The cup product operation satisfies the identity

In other words, the corresponding multiplication is graded-commutative.

See references. [11]

Laplace operator

The Laplace operator of a function at a vertex , is (up to a factor) the rate at which the average value of over a cellular neighborhood of deviates from . The Laplace operator represents the flux density of the gradient flow of a function. For instance, the net rate at which a chemical dissolved in a fluid moves toward or away from some point is proportional to the Laplace operator of the chemical concentration at that point; expressed symbolically, the resulting equation is the diffusion equation. For these reasons, it is extensively used in the sciences for modelling various physical phenomena.

The codifferential

is an operator defined on -forms by:

where is the exterior derivative or differential and is the Hodge star operator.

The codifferential is the adjoint of the exterior derivative according to Stokes' theorem:

Since the differential satisfies , the codifferential has the corresponding property

The Laplace operator is defined by:

See references. [10]

See also

Related Research Articles

In number theory, an arithmetic, arithmetical, or number-theoretic function is generally any function f(n) whose domain is the positive integers and whose range is a subset of the complex numbers. Hardy & Wright include in their definition the requirement that an arithmetical function "expresses some arithmetical property of n". There is a larger class of number-theoretic functions that do not fit this definition, for example, the prime-counting functions. This article provides links to functions of both classes.

In a chemical reaction, chemical equilibrium is the state in which both the reactants and products are present in concentrations which have no further tendency to change with time, so that there is no observable change in the properties of the system. This state results when the forward reaction proceeds at the same rate as the reverse reaction. The reaction rates of the forward and backward reactions are generally not zero, but they are equal. Thus, there are no net changes in the concentrations of the reactants and products. Such a state is known as dynamic equilibrium.

A finite difference is a mathematical expression of the form f (x + b) − f (x + a). If a finite difference is divided by ba, one gets a difference quotient. The approximation of derivatives by finite differences plays a central role in finite difference methods for the numerical solution of differential equations, especially boundary value problems.

In mathematics, the Kronecker delta is a function of two variables, usually just non-negative integers. The function is 1 if the variables are equal, and 0 otherwise:

In single-variable calculus, the difference quotient is usually the name for the expression

In mathematics, a Gaussian function, often simply referred to as a Gaussian, is a function of the base form

<span class="mw-page-title-main">Product rule</span> Formula for the derivative of a product

In calculus, the product rule is a formula used to find the derivatives of products of two or more functions. For two functions, it may be stated in Lagrange's notation as or in Leibniz's notation as

In algebraic topology, singular homology refers to the study of a certain set of algebraic invariants of a topological space X, the so-called homology groups Intuitively, singular homology counts, for each dimension n, the n-dimensional holes of a space. Singular homology is a particular example of a homology theory, which has now grown to be a rather broad collection of theories. Of the various theories, it is perhaps one of the simpler ones to understand, being built on fairly concrete constructions.

In mathematics, specifically in algebraic topology, the cup product is a method of adjoining two cocycles of degree p and q to form a composite cocycle of degree p + q. This defines an associative (and distributive) graded commutative product operation in cohomology, turning the cohomology of a space X into a graded ring, H(X), called the cohomology ring. The cup product was introduced in work of J. W. Alexander, Eduard Čech and Hassler Whitney from 1935–1938, and, in full generality, by Samuel Eilenberg in 1944.

<span class="mw-page-title-main">Čech cohomology</span>

In mathematics, specifically algebraic topology, Čech cohomology is a cohomology theory based on the intersection properties of open covers of a topological space. It is named for the mathematician Eduard Čech.

The equilibrium constant of a chemical reaction is the value of its reaction quotient at chemical equilibrium, a state approached by a dynamic chemical system after sufficient time has elapsed at which its composition has no measurable tendency towards further change. For a given set of reaction conditions, the equilibrium constant is independent of the initial analytical concentrations of the reactant and product species in the mixture. Thus, given the initial composition of a system, known equilibrium constant values can be used to determine the composition of the system at equilibrium. However, reaction parameters like temperature, solvent, and ionic strength may all influence the value of the equilibrium constant.

In mechanics, virtual work arises in the application of the principle of least action to the study of forces and movement of a mechanical system. The work of a force acting on a particle as it moves along a displacement is different for different displacements. Among all the possible displacements that a particle may follow, called virtual displacements, one will minimize the action. This displacement is therefore the displacement followed by the particle according to the principle of least action.

The work of a force on a particle along a virtual displacement is known as the virtual work.

In general relativity, the Gibbons–Hawking–York boundary term is a term that needs to be added to the Einstein–Hilbert action when the underlying spacetime manifold has a boundary.

Differential entropy is a concept in information theory that began as an attempt by Claude Shannon to extend the idea of (Shannon) entropy, a measure of average (surprisal) of a random variable, to continuous probability distributions. Unfortunately, Shannon did not derive this formula, and rather just assumed it was the correct continuous analogue of discrete entropy, but it is not. The actual continuous version of discrete entropy is the limiting density of discrete points (LDDP). Differential entropy is commonly encountered in the literature, but it is a limiting case of the LDDP, and one that loses its fundamental association with discrete entropy.

<span class="mw-page-title-main">Lattice Boltzmann methods</span> Class of computational fluid dynamics methods

The lattice Boltzmann methods (LBM), originated from the lattice gas automata (LGA) method (Hardy-Pomeau-Pazzis and Frisch-Hasslacher-Pomeau models), is a class of computational fluid dynamics (CFD) methods for fluid simulation. Instead of solving the Navier–Stokes equations directly, a fluid density on a lattice is simulated with streaming and collision (relaxation) processes. The method is versatile as the model fluid can straightforwardly be made to mimic common fluid behaviour like vapour/liquid coexistence, and so fluid systems such as liquid droplets can be simulated. Also, fluids in complex environments such as porous media can be straightforwardly simulated, whereas with complex boundaries other CFD methods can be hard to work with.

In algebraic topology the cap product is a method of adjoining a chain of degree p with a cochain of degree q, such that qp, to form a composite chain of degree pq. It was introduced by Eduard Čech in 1936, and independently by Hassler Whitney in 1938.

In mathematics, specifically in algebraic topology, the Eilenberg–Zilber theorem is an important result in establishing the link between the homology groups of a product space and those of the spaces and . The theorem first appeared in a 1953 paper in the American Journal of Mathematics by Samuel Eilenberg and Joseph A. Zilber. One possible route to a proof is the acyclic model theorem.

In mathematics and theoretical physics, Noether's second theorem relates symmetries of an action functional with a system of differential equations. The action S of a physical system is an integral of a so-called Lagrangian function L, from which the system's behavior can be determined by the principle of least action.

In number theory, an average order of an arithmetic function is some simpler or better-understood function which takes the same values "on average".

In mathematics, Maass forms or Maass wave forms are studied in the theory of automorphic forms. Maass forms are complex-valued smooth functions of the upper half plane, which transform in a similar way under the operation of a discrete subgroup of as modular forms. They are eigenforms of the hyperbolic Laplace operator defined on and satisfy certain growth conditions at the cusps of a fundamental domain of . In contrast to modular forms, Maass forms need not be holomorphic. They were studied first by Hans Maass in 1949.

References

  1. Dieudonné, Jean (1988). A History of Algebraic and Differential Topology 1900–1960. Birkhäuser Boston. ISBN   9780817649074.
  2. Auclair-Fortier, Marie-Flavie; Ziou, Djemel; Allili, Madjid (2004). "Global computational algebraic topology approach for diffusion". In Bouman, Charles A; Miller, Eric L (eds.). Computational Imaging II. Vol. 5299. SPIE. p. 357. doi:10.1117/12.525975. S2CID   2211593.
  3. 1 2 3 Grady, Leo; Polimeni, Jonathan (2011). Discrete Calculus: Applied Analysis on Graphs for Computational Science (PDF). Springer.
  4. Desbrun, Mathieu; Kanso, Eva; Tong, Yiying (2008). "Discrete Differential Forms for Computational Modeling". In Bobenko, A.I.; Sullivan, J.M.; Schröder, P.; Ziegler, G.M. (eds.). Discrete Differential Geometry. Oberwolfach Seminars. Vol. 38. Basel: Birkhäuser.
  5. Wilmott, Paul; Howison, Sam; Dewynne, Jeff (1995). The Mathematics of Financial Derivatives: A Student Introduction. Cambridge University Press. p.  137. ISBN   978-0-521-49789-3.
  6. Chaudhry, M. Hanif (2007). Open-Channel Flow. Springer. p. 369. ISBN   978-0-387-68648-6.
  7. Levy, H.; Lessman, F. (1992). Finite Difference Equations. Dover Publications. ISBN   0-486-67260-3.
  8. Ames, W.F. (1977). "Section 1.6". Numerical Methods for Partial Differential Equations. Academic Press. ISBN   0-12-056760-1.
  9. Hildebrand, F.B. (1968). "Section 2.2". Finite-Difference Equations and Simulations. Prentice-Hall. OCLC   780785195.
  10. 1 2 3 4 Saveliev, Peter (2016). Topology Illustrated. Peter Saveliev. ISBN   978-1495188756.
  11. 1 2 3 Bredon, Glen E. (1997). Topology and Geometry. Graduate Texts in Mathematics. Springer. ISBN   0387979263.
  12. Kaczynski, Tomasz; Mischaikow, Konstantin; Mrozek, Marian (2004). Computational Topology. ISBN   0-387-40853-3.