Non-standard calculus

Last updated

In mathematics, non-standard calculus is the modern application of infinitesimals, in the sense of non-standard analysis, to infinitesimal calculus. It provides a rigorous justification that were previously considered merely heuristic.

Contents

Nonrigourous calculations with infinitesimals were widely used before Karl Weierstrass sought to replace them with the (ε, δ)-definition of limit starting in the 1870s. (See history of calculus.) For almost one hundred years thereafter, mathematicians like Richard Courant viewed infinitesimals as being naive and vague or meaningless. [1]

Contrary to such views, Abraham Robinson showed in 1960 that infinitesimals are precise, clear, and meaningful, building upon work by Edwin Hewitt and Jerzy Łoś. According to Howard Keisler, "Robinson solved a three hundred year old problem by giving a precise treatment of infinitesimals. Robinson's achievement will probably rank as one of the major mathematical advances of the twentieth century." [2]

History

The history of non-standard calculus began with the use of infinitely small quantities, called infinitesimals in calculus. The use of infinitesimals can be found in the foundations of calculus independently developed by Gottfried Leibniz and Isaac Newton starting in the 1660s. John Wallis refined earlier techniques of indivisibles of Cavalieri and others by exploiting an infinitesimal quantity he denoted in area calculations, preparing the ground for integral calculus. [3] They drew on the work of such mathematicians as Pierre de Fermat, Isaac Barrow and René Descartes.

In early calculus the use of infinitesimal quantities was criticized by a number of authors, most notably Michel Rolle and Bishop Berkeley in his book The Analyst .

Several mathematicians, including Maclaurin and d'Alembert, advocated the use of limits. Augustin Louis Cauchy developed a versatile spectrum of foundational approaches, including a definition of continuity in terms of infinitesimals and a (somewhat imprecise) prototype of an ε, δ argument in working with differentiation. Karl Weierstrass formalized the concept of limit in the context of a (real) number system without infinitesimals. Following the work of Weierstrass, it eventually became common to base calculus on ε, δ arguments instead of infinitesimals.

This approach formalized by Weierstrass came to be known as the standard calculus. After many years of the infinitesimal approach to calculus having fallen into disuse other than as an introductory pedagogical tool, use of infinitesimal quantities was finally given a rigorous foundation by Abraham Robinson in the 1960s. Robinson's approach is called non-standard analysis to distinguish it from the standard use of limits. This approach used technical machinery from mathematical logic to create a theory of hyperreal numbers that interpret infinitesimals in a manner that allows a Leibniz-like development of the usual rules of calculus. An alternative approach, developed by Edward Nelson, finds infinitesimals on the ordinary real line itself, and involves a modification of the foundational setting by extending ZFC through the introduction of a new unary predicate "standard".

Motivation

To calculate the derivative of the function at x, both approaches agree on the algebraic manipulations:

This becomes a computation of the derivatives using the hyperreals if is interpreted as an infinitesimal and the symbol "" is the relation "is infinitely close to".

In order to make f ' a real-valued function, the final term is dispensed with. In the standard approach using only real numbers, that is done by taking the limit as tends to zero. In the hyperreal approach, the quantity is taken to be an infinitesimal, a nonzero number that is closer to 0 than to any nonzero real. The manipulations displayed above then show that is infinitely close to 2x, so the derivative of f at x is then 2x.

Discarding the "error term" is accomplished by an application of the standard part function. Dispensing with infinitesimal error terms was historically considered paradoxical by some writers, most notably George Berkeley.

Once the hyperreal number system (an infinitesimal-enriched continuum) is in place, one has successfully incorporated a large part of the technical difficulties at the foundational level. Thus, the epsilon, delta techniques that some believe to be the essence of analysis can be implemented once and for all at the foundational level, and the students needn't be "dressed to perform multiple-quantifier logical stunts on pretense of being taught infinitesimal calculus", to quote a recent study. [4] More specifically, the basic concepts of calculus such as continuity, derivative, and integral can be defined using infinitesimals without reference to epsilon, delta (see next section).

Keisler's textbook

Keisler's Elementary Calculus: An Infinitesimal Approach defines continuity on page 125 in terms of infinitesimals, to the exclusion of epsilon, delta methods. The derivative is defined on page 45 using infinitesimals rather than an epsilon-delta approach. The integral is defined on page 183 in terms of infinitesimals. Epsilon, delta definitions are introduced on page 282.

Definition of derivative

The hyperreals can be constructed in the framework of Zermelo–Fraenkel set theory, the standard axiomatisation of set theory used elsewhere in mathematics. To give an intuitive idea for the hyperreal approach, note that, naively speaking, non-standard analysis postulates the existence of positive numbers ε which are infinitely small, meaning that ε is smaller than any standard positive real, yet greater than zero. Every real number x is surrounded by an infinitesimal "cloud" of hyperreal numbers infinitely close to it. To define the derivative of f at a standard real number x in this approach, one no longer needs an infinite limiting process as in standard calculus. Instead, one sets

where st is the standard part function, yielding the real number infinitely close to the hyperreal argument of st, and is the natural extension of to the hyperreals.

Continuity

A real function f is continuous at a standard real number x if for every hyperreal x' infinitely close to x, the value f(x' ) is also infinitely close to f(x). This captures Cauchy's definition of continuity as presented in his 1821 textbook Cours d'Analyse, p. 34.

Here to be precise, f would have to be replaced by its natural hyperreal extension usually denoted f* (see discussion of Transfer principle in main article at non-standard analysis).

Using the notation for the relation of being infinitely close as above, the definition can be extended to arbitrary (standard or non-standard) points as follows:

A function f is microcontinuous at x if whenever , one has

Here the point x' is assumed to be in the domain of (the natural extension of) f.

The above requires fewer quantifiers than the (ε, δ)-definition familiar from standard elementary calculus:

f is continuous at x if for every ε > 0, there exists a δ > 0 such that for every x' , whenever |x  x'| < δ, one has |f(x)  f(x')| < ε.

Uniform continuity

A function f on an interval I is uniformly continuous if its natural extension f* in I* has the following property (see Keisler, Foundations of Infinitesimal Calculus ('07), p. 45):

for every pair of hyperreals x and y in I*, if then .

In terms of microcontinuity defined in the previous section, this can be stated as follows: a real function is uniformly continuous if its natural extension f* is microcontinuous at every point of the domain of f*.

This definition has a reduced quantifier complexity when compared with the standard (ε, δ)-definition. Namely, the epsilon-delta definition of uniform continuity requires four quantifiers, while the infinitesimal definition requires only two quantifiers. It has the same quantifier complexity as the definition of uniform continuity in terms of sequences in standard calculus, which however is not expressible in the first-order language of the real numbers.

The hyperreal definition can be illustrated by the following three examples.

Example 1: a function f is uniformly continuous on the semi-open interval (0,1], if and only if its natural extension f* is microcontinuous (in the sense of the formula above) at every positive infinitesimal, in addition to continuity at the standard points of the interval.

Example 2: a function f is uniformly continuous on the semi-open interval [0,∞) if and only if it is continuous at the standard points of the interval, and in addition, the natural extension f* is microcontinuous at every positive infinite hyperreal point.

Example 3: similarly, the failure of uniform continuity for the squaring function

is due to the absence of microcontinuity at a single infinite hyperreal point, see below.

Concerning quantifier complexity, the following remarks were made by Kevin Houston: [5]

The number of quantifiers in a mathematical statement gives a rough measure of the statement’s complexity. Statements involving three or more quantifiers can be difficult to understand. This is the main reason why it is hard to understand the rigorous definitions of limit, convergence, continuity and differentiability in analysis as they have many quantifiers. In fact, it is the alternation of the and that causes the complexity.

Andreas Blass wrote as follows:

Often ... the nonstandard definition of a concept is simpler than the standard definition (both intuitively simpler and simpler in a technical sense, such as quantifiers over lower types or fewer alternations of quantifiers). [6]

Compactness

A set A is compact if and only if its natural extension A* has the following property: every point in A* is infinitely close to a point of A. Thus, the open interval (0,1) is not compact because its natural extension contains positive infinitesimals which are not infinitely close to any positive real number.

Heine–Cantor theorem

The fact that a continuous function on a compact interval I is necessarily uniformly continuous (the Heine–Cantor theorem) admits a succinct hyperreal proof. Let x, y be hyperreals in the natural extension I* of I. Since I is compact, both st(x) and st(y) belong to I. If x and y were infinitely close, then by the triangle inequality, they would have the same standard part

Since the function is assumed continuous at c,

and therefore f(x) and f(y) are infinitely close, proving uniform continuity of f.

Why is the squaring function not uniformly continuous?

Let f(x) = x2 defined on . Let be an infinite hyperreal. The hyperreal number is infinitely close to N. Meanwhile, the difference

is not infinitesimal. Therefore, f* fails to be microcontinuous at the hyperreal point N. Thus, the squaring function is not uniformly continuous, according to the definition in uniform continuity above.

A similar proof may be given in the standard setting ( Fitzpatrick 2006 , Example 3.15).

Example: Dirichlet function

Consider the Dirichlet function

It is well known that, under the standard definition of continuity, the function is discontinuous at every point. Let us check this in terms of the hyperreal definition of continuity above, for instance let us show that the Dirichlet function is not continuous at π. Consider the continued fraction approximation an of π. Now let the index n be an infinite hypernatural number. By the transfer principle, the natural extension of the Dirichlet function takes the value 1 at an. Note that the hyperrational point an is infinitely close to π. Thus the natural extension of the Dirichlet function takes different values (0 and 1) at these two infinitely close points, and therefore the Dirichlet function is not continuous at π.

Limit

While the thrust of Robinson's approach is that one can dispense with the approach using multiple quantifiers, the notion of limit can be easily recaptured in terms of the standard part function st, namely

if and only if whenever the difference x  a is infinitesimal, the difference f(x)  L is infinitesimal, as well, or in formulas:

if st(x) = a  then st(f(x)) = L,

cf. (ε, δ)-definition of limit.

Limit of sequence

Given a sequence of real numbers , if L is the limit of the sequence and

if for every infinite hypernatural n, st(xn)=L (here the extension principle is used to define xn for every hyperinteger n).

This definition has no quantifier alternations. The standard (ε, δ)-style definition on the other hand does have quantifier alternations:

Extreme value theorem

To show that a real continuous function f on [0,1] has a maximum, let N be an infinite hyperinteger. The interval [0, 1] has a natural hyperreal extension. The function f is also naturally extended to hyperreals between 0 and 1. Consider the partition of the hyperreal interval [0,1] into N subintervals of equal infinitesimal length 1/N, with partition points xi = i /N as i "runs" from 0 to N. In the standard setting (when N is finite), a point with the maximal value of f can always be chosen among the N+1 points xi, by induction. Hence, by the transfer principle, there is a hyperinteger i0 such that 0 i0 N and for all i = 0, …, N (an alternative explanation is that every hyperfinite set admits a maximum). Consider the real point

where st is the standard part function. An arbitrary real point x lies in a suitable sub-interval of the partition, namely , so that st(xi) = x. Applying st to the inequality , . By continuity of f,

.

Hence f(c) f(x), for all x, proving c to be a maximum of the real function f. See Keisler (1986 , p. 164).

Intermediate value theorem

As another illustration of the power of Robinson's approach, a short proof of the intermediate value theorem (Bolzano's theorem) using infinitesimals is done by the following.

Let f be a continuous function on [a,b] such that f(a)<0 while f(b)>0. Then there exists a point c in [a,b] such that f(c)=0.

The proof proceeds as follows. Let N be an infinite hyperinteger. Consider a partition of [a,b] into N intervals of equal length, with partition points xi as i runs from 0 to N. Consider the collection I of indices such that f(xi)>0. Let i0 be the least element in I (such an element exists by the transfer principle, as I is a hyperfinite set). Then the real number

is the desired zero of f. Such a proof reduces the quantifier complexity of a standard proof of the IVT.

Basic theorems

If f is a real valued function defined on an interval [a, b], then the transfer operator applied to f, denoted by *f, is an internal, hyperreal-valued function defined on the hyperreal interval [*a, *b].

Theorem. Let f be a real-valued function defined on an interval [a, b]. Then f is differentiable at a < x < b if and only if for every non-zero infinitesimal h, the value

is independent of h. In that case, the common value is the derivative of f at x.

This fact follows from the transfer principle of non-standard analysis and overspill.

Note that a similar result holds for differentiability at the endpoints a, b provided the sign of the infinitesimal h is suitably restricted.

For the second theorem, the Riemann integral is defined as the limit, if it exists, of a directed family of Riemann sums; these are sums of the form

where

Such a sequence of values is called a partition or mesh and

the width of the mesh. In the definition of the Riemann integral, the limit of the Riemann sums is taken as the width of the mesh goes to 0.

Theorem. Let f be a real-valued function defined on an interval [a, b]. Then f is Riemann-integrable on [a, b] if and only if for every internal mesh of infinitesimal width, the quantity

is independent of the mesh. In this case, the common value is the Riemann integral of f over [a, b].

Applications

One immediate application is an extension of the standard definitions of differentiation and integration to internal functions on intervals of hyperreal numbers.

An internal hyperreal-valued function f on [a, b] is S-differentiable at x, provided

exists and is independent of the infinitesimal h. The value is the S derivative at x.

Theorem. Suppose f is S-differentiable at every point of [a, b] where ba is a bounded hyperreal. Suppose furthermore that

Then for some infinitesimal ε

To prove this, let N be a non-standard natural number. Divide the interval [a, b] into N subintervals by placing N 1 equally spaced intermediate points:

Then

Now the maximum of any internal set of infinitesimals is infinitesimal. Thus all the εk's are dominated by an infinitesimal ε. Therefore,

from which the result follows.

See also

Notes

  1. Courant described infinitesimals on page 81 of Differential and Integral Calculus, Vol I, as "devoid of any clear meaning" and "naive befogging". Similarly on page 101, Courant described them as "incompatible with the clarity of ideas demanded in mathematics", "entirely meaningless", "fog which hung round the foundations", and a "hazy idea".
  2. Elementary Calculus: An Infinitesimal Approach
  3. Scott, J.F. 1981. "The Mathematical Work of John Wallis, D.D., F.R.S. (16161703)". Chelsea Publishing Co. New York, NY. p. 18.
  4. Katz, Mikhail; Tall, David (2011), Tension between Intuitive Infinitesimals and Formal Mathematical Analysis, Bharath Sriraman, Editor. Crossroads in the History of Mathematics and Mathematics Education. The Montana Mathematics Enthusiast Monographs in Mathematics Education 12, Information Age Publishing, Inc., Charlotte, NC, arXiv: 1110.5747 , Bibcode:2011arXiv1110.5747K
  5. Kevin Houston, How to Think Like a Mathematician, ISBN   978-0-521-71978-0
  6. Blass, Andreas (1978), "Review: Martin Davis, Applied nonstandard analysis, and K. D. Stroyan and W. A. J. Luxemburg, Introduction to the theory of infinitesimals, and H. Jerome Keisler, Foundations of infinitesimal calculus", Bull. Amer. Math. Soc., 84 (1): 34–41, doi:10.1090/S0002-9904-1978-14401-2 , p. 37.

Related Research Articles

In mathematics, a continuous function is a function that does not have any abrupt changes in value, known as discontinuities. More precisely, sufficiently small changes in the input of a continuous function result in arbitrarily small changes in its output. Otherwise, a function is said to be a discontinuous function. A continuous function with a continuous inverse function is called a homeomorphism.

Intermediate value theorem theorem

In mathematical analysis, the intermediate value theorem states that if f is a continuous function whose domain contains the interval [a, b], then it takes on any given value between f(a) and f(b) at some point within the interval.

Non-standard analysis calculus using a logically rigorous notion of infinitesimal numbers

The history of calculus is fraught with philosophical debates about the meaning and logical validity of fluxions or infinitesimal numbers. The standard way to resolve these debates is to define the operations of calculus using epsilon–delta procedures rather than infinitesimals. Non-standard analysis instead reformulates the calculus using a logically rigorous notion of infinitesimal numbers.

Real analysis Mathematics of real numbers and real functions

In mathematics, real analysis is the branch of mathematical analysis that studies the behavior of real numbers, sequences and series of real numbers, and real functions. Some particular properties of real-valued sequences and functions that real analysis studies include convergence, limits, continuity, smoothness, differentiability and integrability.

Uniform continuity Function limiting the "growth" of distances of outputs uniformly across its domain

In mathematics, a function f is uniformly continuous if, roughly speaking, it is possible to guarantee that f(x) and f(y) be as close to each other as we please by requiring only that x and y are sufficiently close to each other; unlike ordinary continuity, where the maximum distance between f(x) and f(y) may depend on x and y themselves.

Hyperreal number element of a nonstandard model of the reals, which can be infinite or infinitesimal

In mathematics, the system of hyperreal numbers is a way of treating infinite and infinitesimal quantities. The hyperreals, or nonstandard reals, *R, are an extension of the real numbers R that contains numbers greater than anything of the form

Infinitesimal extremely small quantity in calculus; thing so small that there is no way to measure them

In mathematics, infinitesimals are things so small that there is no way to measure them. The insight with exploiting infinitesimals was that entities could still retain certain specific properties, such as angle or slope, even though these entities were quantitatively small. The word infinitesimal comes from a 17th-century Modern Latin coinage infinitesimus, which originally referred to the "infinity-th" item in a sequence. Infinitesimals are a basic ingredient in the procedures of infinitesimal calculus as developed by Leibniz, including the law of continuity and the transcendental law of homogeneity. In common speech, an infinitesimal object is an object that is smaller than any feasible measurement, but not zero in size—or, so small that it cannot be distinguished from zero by any available means. Hence, when used as an adjective, "infinitesimal" means "extremely small". To give it a meaning, it usually must be compared to another infinitesimal object in the same context. Infinitely many infinitesimals are summed to produce an integral.

In calculus, absolute continuity is a smoothness property of functions that is stronger than continuity and uniform continuity. The notion of absolute continuity allows one to obtain generalizations of the relationship between the two central operations of calculus—differentiation and integration. This relationship is commonly characterized in the framework of Riemann integration, but with absolute continuity it may be formulated in terms of Lebesgue integration. For real-valued functions on the real line, two interrelated notions appear: absolute continuity of functions and absolute continuity of measures. These two notions are generalized in different directions. The usual derivative of a function is related to the Radon–Nikodym derivative, or density, of a measure.

Extreme value theorem theorem that states that the image of real function having real closed interval as domain, has maximum and minimum

In calculus, the extreme value theorem states that if a real-valued function is continuous on the closed interval , then must attain a maximum and a minimum, each at least once. That is, there exist numbers and in such that:

In mathematics, the limit of a function is a fundamental concept in calculus and analysis concerning the behavior of that function near a particular input.

Limit of a sequence null sequence; value that the terms of a sequence "tend to"

In mathematics, the limit of a sequence is the value that the terms of a sequence "tend to". If such a limit exists, the sequence is called convergent. A sequence that does not converge is said to be divergent. The limit of a sequence is said to be the fundamental notion on which the whole of analysis ultimately rests.

Squeeze theorem theorem of calculus

In calculus, the squeeze theorem, also known as the pinching theorem, the sandwich theorem, the sandwich rule, and sometimes the squeeze lemma, is a theorem regarding the limit of a function. In Italy, the theorem is also known as theorem of Carabinieri.

In non-standard analysis, a branch of mathematics, overspill is a widely used proof technique. It is based on the fact that the set of standard natural numbers N is not an internal subset of the internal set *N of hypernatural numbers.

In model theory, a transfer principle states that all statements of some language that are true for some structure are true for another structure. One of the first examples was the Lefschetz principle, which states that any sentence in the first-order language of fields that is true for the complex numbers is also true for any algebraically closed field of characteristic 0.

In non-standard analysis, a field of mathematics, the increment theorem states the following: Suppose a function y = f(x) is differentiable at x and that Δx is infinitesimal. Then

In non-standard analysis, the standard part function is a function from the limited (finite) hyperreal numbers to the real numbers. Briefly, the standard part function "rounds off" a finite hyperreal to the nearest real. It associates to every such hyperreal , the unique real infinitely close to it, i.e. is infinitesimal. As such, it is a mathematical implementation of the historical concept of adequality introduced by Pierre de Fermat, as well as Leibniz's Transcendental law of homogeneity.

(<i>ε</i>, <i>δ</i>)-definition of limit Mathematical definition of a limit

In calculus, the (εδ)-definition of limit is a formalization of the notion of limit. The concept is due to Augustin-Louis Cauchy, who never gave an definition of limit in his Cours d'Analyse, but occasionally used arguments in proofs. It was first given as a formal definition by Bernard Bolzano in 1817, and the definitive modern statement was ultimately provided by Karl Weierstrass. It makes rigorous the following informal notion: the dependent expression f(x) approaches the value L as the variable x approaches the value c if f(x) can be made as close as desired to L by taking x sufficiently close to c.

Elementary Calculus: An Infinitesimal approach is a textbook by H. Jerome Keisler. The subtitle alludes to the infinitesimal numbers of the hyperreal number system of Abraham Robinson and is sometimes given as An approach using infinitesimals. The book is available freely online and is currently published by Dover.

In calculus, the differential represents the principal part of the change in a function y = f(x) with respect to changes in the independent variable. The differential dy is defined by

In mathematics, a limit is the value that a function "approaches" as the input "approaches" some value. Limits are essential to calculus and are used to define continuity, derivatives, and integrals.

References