Calculus

Last updated

Calculus, originally called infinitesimal calculus or "the calculus of infinitesimals", is the mathematical study of continuous change, in the same way that geometry is the study of shape, and algebra is the study of generalizations of arithmetic operations.

Contents

It has two major branches, differential calculus and integral calculus; differential calculus concerns instantaneous rates of change, and the slopes of curves, while integral calculus concerns accumulation of quantities, and areas under or between curves. These two branches are related to each other by the fundamental theorem of calculus, and they make use of the fundamental notions of convergence of infinite sequences and infinite series to a well-defined limit. [1]

Infinitesimal calculus was developed independently in the late 17th century by Isaac Newton and Gottfried Wilhelm Leibniz. [2] [3] Later work, including codifying the idea of limits, put these developments on a more solid conceptual footing. Today, calculus has widespread uses in science, engineering, and social science. [4]

In mathematics education, calculus denotes courses of elementary mathematical analysis, which are mainly devoted to the study of functions and limits. The word calculus is Latin for "small pebble" (the diminutive of calx, meaning "stone"), a meaning which still persists in medicine. Because such pebbles were used for counting out distances, [5] tallying votes, and doing abacus arithmetic, the word came to mean a method of computation. In this sense, it was used in English at least as early as 1672, several years prior to the publications of Leibniz and Newton. [6]

In addition to the differential calculus and integral calculus, the term is also used for naming specific methods of calculation and related theories which seek to model a particular concept in terms of mathematics. Examples of this convention include propositional calculus, Ricci calculus, calculus of variations, lambda calculus, and process calculus. Furthermore, the term "calculus" has variously been applied in ethics and philosophy, for such systems as Bentham's felicific calculus, and the ethical calculus.

History

Modern calculus was developed in 17th-century Europe by Isaac Newton and Gottfried Wilhelm Leibniz (independently of each other, first publishing around the same time) but elements of it appeared in ancient Greece, then in China and the Middle East, and still later again in medieval Europe and in India.

Ancient precursors

Egypt

Calculations of volume and area, one goal of integral calculus, can be found in the Egyptian Moscow papyrus (c.1820 BC), but the formulae are simple instructions, with no indication as to how they were obtained. [7] [8]

Greece

Laying the foundations for integral calculus and foreshadowing the concept of the limit, ancient Greek mathematician Eudoxus of Cnidus (c. 390 – 337 BCE) developed the method of exhaustion to prove the formulas for cone and pyramid volumes.

During the Hellenistic period, this method was further developed by Archimedes, who combined it with a concept of the indivisibles—a precursor to infinitesimals—allowing him to solve several problems now treated by integral calculus. These problems include, for example, calculating the center of gravity of a solid hemisphere, the center of gravity of a frustum of a circular paraboloid, and the area of a region bounded by a parabola and one of its secant lines. [9]

China

The method of exhaustion was later discovered independently in China by Liu Hui in the 3rd century AD in order to find the area of a circle. [10] [11] In the 5th century AD, Zu Gengzhi, son of Zu Chongzhi, established a method [12] [13] that would later be called Cavalieri's principle to find the volume of a sphere.

Medieval

Middle East

In the Middle East, Hasan Ibn al-Haytham, Latinized as Alhazen (c.965 – c.1040 CE) derived a formula for the sum of fourth powers. He used the results to carry out what would now be called an integration of this function, where the formulae for the sums of integral squares and fourth powers allowed him to calculate the volume of a paraboloid. [14]

India

In the 14th century, Indian mathematicians gave a non-rigorous method, resembling differentiation, applicable to some trigonometric functions. Madhava of Sangamagrama and the Kerala School of Astronomy and Mathematics thereby stated components of calculus. A complete theory encompassing these components is now well known in the Western world as the Taylor series or infinite series approximations. [15] [16] However, they were not able to "combine many differing ideas under the two unifying themes of the derivative and the integral, show the connection between the two, and turn calculus into the great problem-solving tool we have today". [14]

Modern

The calculus was the first achievement of modern mathematics and it is difficult to overestimate its importance. I think it defines more unequivocally than anything else the inception of modern mathematics, and the system of mathematical analysis, which is its logical development, still constitutes the greatest technical advance in exact thinking.

John von Neumann [17]

Johannes Kepler's work Stereometrica Doliorum formed the basis of integral calculus. [18] Kepler developed a method to calculate the area of an ellipse by adding up the lengths of many radii drawn from a focus of the ellipse. [19]

A significant work was a treatise, the origin being Kepler's methods, [19] written by Bonaventura Cavalieri, who argued that volumes and areas should be computed as the sums of the volumes and areas of infinitesimally thin cross-sections. The ideas were similar to Archimedes' in The Method , but this treatise is believed to have been lost in the 13th century, and was only rediscovered in the early 20th century, and so would have been unknown to Cavalieri. Cavalieri's work was not well respected since his methods could lead to erroneous results, and the infinitesimal quantities he introduced were disreputable at first.

The formal study of calculus brought together Cavalieri's infinitesimals with the calculus of finite differences developed in Europe at around the same time. Pierre de Fermat, claiming that he borrowed from Diophantus, introduced the concept of adequality, which represented equality up to an infinitesimal error term. [20] The combination was achieved by John Wallis, Isaac Barrow, and James Gregory, the latter two proving predecessors to the second fundamental theorem of calculus around 1670. [21] [22]

The product rule and chain rule, [23] the notions of higher derivatives and Taylor series, [24] and of analytic functions [25] were used by Isaac Newton in an idiosyncratic notation which he applied to solve problems of mathematical physics. In his works, Newton rephrased his ideas to suit the mathematical idiom of the time, replacing calculations with infinitesimals by equivalent geometrical arguments which were considered beyond reproach. He used the methods of calculus to solve the problem of planetary motion, the shape of the surface of a rotating fluid, the oblateness of the earth, the motion of a weight sliding on a cycloid, and many other problems discussed in his Principia Mathematica (1687). In other work, he developed series expansions for functions, including fractional and irrational powers, and it was clear that he understood the principles of the Taylor series. He did not publish all these discoveries, and at this time infinitesimal methods were still considered disreputable. [26]

These ideas were arranged into a true calculus of infinitesimals by Gottfried Wilhelm Leibniz, who was originally accused of plagiarism by Newton. [27] He is now regarded as an independent inventor of and contributor to calculus. His contribution was to provide a clear set of rules for working with infinitesimal quantities, allowing the computation of second and higher derivatives, and providing the product rule and chain rule, in their differential and integral forms. Unlike Newton, Leibniz put painstaking effort into his choices of notation. [28]

Today, Leibniz and Newton are usually both given credit for independently inventing and developing calculus. Newton was the first to apply calculus to general physics and Leibniz developed much of the notation used in calculus today. The basic insights that both Newton and Leibniz provided were the laws of differentiation and integration, second and higher derivatives, and the notion of an approximating polynomial series.

When Newton and Leibniz first published their results, there was great controversy over which mathematician (and therefore which country) deserved credit. Newton derived his results first (later to be published in his Method of Fluxions ), but Leibniz published his "Nova Methodus pro Maximis et Minimis" first. Newton claimed Leibniz stole ideas from his unpublished notes, which Newton had shared with a few members of the Royal Society. This controversy divided English-speaking mathematicians from continental European mathematicians for many years, to the detriment of English mathematics. [29] A careful examination of the papers of Leibniz and Newton shows that they arrived at their results independently, with Leibniz starting first with integration and Newton with differentiation. It is Leibniz, however, who gave the new discipline its name. Newton called his calculus "the science of fluxions", a term that endured in English schools into the 19th century. [30] :100 The first complete treatise on calculus to be written in English and use the Leibniz notation was not published until 1815. [31]

Since the time of Leibniz and Newton, many mathematicians have contributed to the continuing development of calculus. One of the first and most complete works on both infinitesimal and integral calculus was written in 1748 by Maria Gaetana Agnesi. [32] [33]

Foundations

In calculus, foundations refers to the rigorous development of the subject from axioms and definitions. In early calculus the use of infinitesimal quantities was thought unrigorous, and was fiercely criticized by a number of authors, most notably Michel Rolle and Bishop Berkeley. Berkeley famously described infinitesimals as the ghosts of departed quantities in his book The Analyst in 1734. Working out a rigorous foundation for calculus occupied mathematicians for much of the century following Newton and Leibniz, and is still to some extent an active area of research today. [34]

Several mathematicians, including Maclaurin, tried to prove the soundness of using infinitesimals, but it would not be until 150 years later when, due to the work of Cauchy and Weierstrass, a way was finally found to avoid mere "notions" of infinitely small quantities. [35] The foundations of differential and integral calculus had been laid. In Cauchy's Cours d'Analyse , we find a broad range of foundational approaches, including a definition of continuity in terms of infinitesimals, and a (somewhat imprecise) prototype of an (ε, δ)-definition of limit in the definition of differentiation. [36] In his work Weierstrass formalized the concept of limit and eliminated infinitesimals (although his definition can actually validate nilsquare infinitesimals). Following the work of Weierstrass, it eventually became common to base calculus on limits instead of infinitesimal quantities, though the subject is still occasionally called "infinitesimal calculus". Bernhard Riemann used these ideas to give a precise definition of the integral. [37] It was also during this period that the ideas of calculus were generalized to the complex plane with the development of complex analysis. [38]

In modern mathematics, the foundations of calculus are included in the field of real analysis, which contains full definitions and proofs of the theorems of calculus. The reach of calculus has also been greatly extended. Henri Lebesgue invented measure theory, based on earlier developments by Émile Borel, and used it to define integrals of all but the most pathological functions. [39] Laurent Schwartz introduced distributions, which can be used to take the derivative of any function whatsoever. [40]

Limits are not the only rigorous approach to the foundation of calculus. Another way is to use Abraham Robinson's non-standard analysis. Robinson's approach, developed in the 1960s, uses technical machinery from mathematical logic to augment the real number system with infinitesimal and infinite numbers, as in the original Newton-Leibniz conception. The resulting numbers are called hyperreal numbers, and they can be used to give a Leibniz-like development of the usual rules of calculus. [41] There is also smooth infinitesimal analysis, which differs from non-standard analysis in that it mandates neglecting higher-power infinitesimals during derivations. [34]

Significance

While many of the ideas of calculus had been developed earlier in Greece, China, India, Iraq, Persia, and Japan, the use of calculus began in Europe, during the 17th century, when Isaac Newton and Gottfried Wilhelm Leibniz built on the work of earlier mathematicians to introduce its basic principles. [11] [26] [42] The development of calculus was built on earlier concepts of instantaneous motion and area underneath curves.

Applications of differential calculus include computations involving velocity and acceleration, the slope of a curve, and optimization. Applications of integral calculus include computations involving area, volume, arc length, center of mass, work, and pressure. More advanced applications include power series and Fourier series.

Calculus is also used to gain a more precise understanding of the nature of space, time, and motion. For centuries, mathematicians and philosophers wrestled with paradoxes involving division by zero or sums of infinitely many numbers. These questions arise in the study of motion and area. The ancient Greek philosopher Zeno of Elea gave several famous examples of such paradoxes. Calculus provides tools, especially the limit and the infinite series, that resolve the paradoxes. [43]

Principles

Limits and infinitesimals

Calculus is usually developed by working with very small quantities. Historically, the first method of doing so was by infinitesimals. These are objects which can be treated like real numbers but which are, in some sense, "infinitely small". For example, an infinitesimal number could be greater than 0, but less than any number in the sequence 1, 1/2, 1/3, ... and thus less than any positive real number. From this point of view, calculus is a collection of techniques for manipulating infinitesimals. The symbols ${\displaystyle dx}$ and ${\displaystyle dy}$ were taken to be infinitesimal, and the derivative ${\displaystyle dy/dx}$ was simply their ratio. [34]

The infinitesimal approach fell out of favor in the 19th century because it was difficult to make the notion of an infinitesimal precise. In the late 19th century, infinitesimals were replaced within academia by the epsilon, delta approach to limits. Limits describe the behavior of a function at a certain input in terms of its values at nearby inputs. They capture small-scale behavior using the intrinsic structure of the real number system (as a metric space with the least-upper-bound property). In this treatment, calculus is a collection of techniques for manipulating certain limits. Infinitesimals get replaced by sequences of smaller and smaller numbers, and the infinitely small behavior of a function is found by taking the limiting behavior for these sequences. Limits were thought to provide a more rigorous foundation for calculus, and for this reason they became the standard approach during the 20th century. However, the infinitesimal concept was revived in the 20th century with the introduction of non-standard analysis and smooth infinitesimal analysis, which provided solid foundations for the manipulation of infinitesimals. [34]

Differential calculus

Differential calculus is the study of the definition, properties, and applications of the derivative of a function. The process of finding the derivative is called differentiation. Given a function and a point in the domain, the derivative at that point is a way of encoding the small-scale behavior of the function near that point. By finding the derivative of a function at every point in its domain, it is possible to produce a new function, called the derivative function or just the derivative of the original function. In formal terms, the derivative is a linear operator which takes a function as its input and produces a second function as its output. This is more abstract than many of the processes studied in elementary algebra, where functions usually input a number and output another number. For example, if the doubling function is given the input three, then it outputs six, and if the squaring function is given the input three, then it outputs nine. The derivative, however, can take the squaring function as an input. This means that the derivative takes all the information of the squaring function—such as that two is sent to four, three is sent to nine, four is sent to sixteen, and so on—and uses this information to produce another function. The function produced by differentiating the squaring function turns out to be the doubling function. [44] :32

In more explicit terms the "doubling function" may be denoted by g(x) = 2x and the "squaring function" by f(x) = x2. The "derivative" now takes the function f(x), defined by the expression "x2", as an input, that is all the information—such as that two is sent to four, three is sent to nine, four is sent to sixteen, and so on—and uses this information to output another function, the function g(x) = 2x, as will turn out.

In Lagrange's notation, the symbol for a derivative is an apostrophe-like mark called a prime. Thus, the derivative of a function called f is denoted by f′, pronounced "f prime". For instance, if f(x) = x2 is the squaring function, then f′(x) = 2x is its derivative (the doubling function g from above).

If the input of the function represents time, then the derivative represents change with respect to time. For example, if f is a function that takes a time as input and gives the position of a ball at that time as output, then the derivative of f is how the position is changing in time, that is, it is the velocity of the ball. [44] :18–20

If a function is linear (that is, if the graph of the function is a straight line), then the function can be written as y = mx + b, where x is the independent variable, y is the dependent variable, b is the y-intercept, and:

${\displaystyle m={\frac {\text{rise}}{\text{run}}}={\frac {{\text{change in }}y}{{\text{change in }}x}}={\frac {\Delta y}{\Delta x}}.}$

This gives an exact value for the slope of a straight line. If the graph of the function is not a straight line, however, then the change in y divided by the change in x varies. Derivatives give an exact meaning to the notion of change in output with respect to change in input. To be concrete, let f be a function, and fix a point a in the domain of f. (a, f(a)) is a point on the graph of the function. If h is a number close to zero, then a + h is a number close to a. Therefore, (a + h, f(a + h)) is close to (a, f(a)). The slope between these two points is

${\displaystyle m={\frac {f(a+h)-f(a)}{(a+h)-a}}={\frac {f(a+h)-f(a)}{h}}.}$

This expression is called a difference quotient . A line through two points on a curve is called a secant line, so m is the slope of the secant line between (a, f(a)) and (a + h, f(a + h)). The secant line is only an approximation to the behavior of the function at the point a because it does not account for what happens between a and a + h. It is not possible to discover the behavior at a by setting h to zero because this would require dividing by zero, which is undefined. The derivative is defined by taking the limit as h tends to zero, meaning that it considers the behavior of f for all small values of h and extracts a consistent value for the case when h equals zero:

${\displaystyle \lim _{h\to 0}{f(a+h)-f(a) \over {h}}.}$

Geometrically, the derivative is the slope of the tangent line to the graph of f at a. The tangent line is a limit of secant lines just as the derivative is a limit of difference quotients. For this reason, the derivative is sometimes called the slope of the function f.

Here is a particular example, the derivative of the squaring function at the input 3. Let f(x) = x2 be the squaring function.

{\displaystyle {\begin{aligned}f'(3)&=\lim _{h\to 0}{(3+h)^{2}-3^{2} \over {h}}\\&=\lim _{h\to 0}{9+6h+h^{2}-9 \over {h}}\\&=\lim _{h\to 0}{6h+h^{2} \over {h}}\\&=\lim _{h\to 0}(6+h)\\&=6\end{aligned}}}

The slope of the tangent line to the squaring function at the point (3, 9) is 6, that is to say, it is going up six times as fast as it is going to the right. The limit process just described can be performed for any point in the domain of the squaring function. This defines the derivative function of the squaring function or just the derivative of the squaring function for short. A computation similar to the one above shows that the derivative of the squaring function is the doubling function.

Leibniz notation

A common notation, introduced by Leibniz, for the derivative in the example above is

{\displaystyle {\begin{aligned}y&=x^{2}\\{\frac {dy}{dx}}&=2x.\end{aligned}}}

In an approach based on limits, the symbol dy/dx is to be interpreted not as the quotient of two numbers but as a shorthand for the limit computed above. Leibniz, however, did intend it to represent the quotient of two infinitesimally small numbers, dy being the infinitesimally small change in y caused by an infinitesimally small change dx applied to x. We can also think of d/dx as a differentiation operator, which takes a function as an input and gives another function, the derivative, as the output. For example:

${\displaystyle {\frac {d}{dx}}(x^{2})=2x.}$

In this usage, the dx in the denominator is read as "with respect to x". Another example of correct notation could be:

{\displaystyle {\begin{aligned}g(t)&=t^{2}+2t+4\\{d \over dt}g(t)&=2t+2\end{aligned}}}

Even when calculus is developed using limits rather than infinitesimals, it is common to manipulate symbols like dx and dy as if they were real numbers; although it is possible to avoid such manipulations, they are sometimes notationally convenient in expressing operations such as the total derivative.

Integral calculus

Integral calculus is the study of the definitions, properties, and applications of two related concepts, the indefinite integral and the definite integral. The process of finding the value of an integral is called integration. In technical language, integral calculus studies two related linear operators.

The indefinite integral, also known as the antiderivative , is the inverse operation to the derivative. F is an indefinite integral of f when f is a derivative of F. (This use of lower- and upper-case letters for a function and its indefinite integral is common in calculus.)

The definite integral inputs a function and outputs a number, which gives the algebraic sum of areas between the graph of the input and the x-axis. The technical definition of the definite integral involves the limit of a sum of areas of rectangles, called a Riemann sum. [45] :282

A motivating example is the distance traveled in a given time. If the speed is constant, only multiplication is needed:

${\displaystyle \mathrm {Distance} =\mathrm {Speed} \cdot \mathrm {Time} }$

But if the speed changes, a more powerful method of finding the distance is necessary. One such method is to approximate the distance traveled by breaking up the time into many short intervals of time, then multiplying the time elapsed in each interval by one of the speeds in that interval, and then taking the sum (a Riemann sum) of the approximate distance traveled in each interval. The basic idea is that if only a short time elapses, then the speed will stay more or less the same. However, a Riemann sum only gives an approximation of the distance traveled. We must take the limit of all such Riemann sums to find the exact distance traveled.

When velocity is constant, the total distance traveled over the given time interval can be computed by multiplying velocity and time. For example, travelling a steady 50 mph for 3 hours results in a total distance of 150 miles. In the diagram on the left, when constant velocity and time are graphed, these two values form a rectangle with height equal to the velocity and width equal to the time elapsed. Therefore, the product of velocity and time also calculates the rectangular area under the (constant) velocity curve. This connection between the area under a curve and distance traveled can be extended to any irregularly shaped region exhibiting a fluctuating velocity over a given time period. If f(x) in the diagram on the right represents speed as it varies over time, the distance traveled (between the times represented by a and b) is the area of the shaded region s.

To approximate that area, an intuitive method would be to divide up the distance between a and b into a number of equal segments, the length of each segment represented by the symbol Δx. For each small segment, we can choose one value of the function f(x). Call that value h. Then the area of the rectangle with base Δx and height h gives the distance (time Δx multiplied by speed h) traveled in that segment. Associated with each segment is the average value of the function above it, f(x) = h. The sum of all such rectangles gives an approximation of the area between the axis and the curve, which is an approximation of the total distance traveled. A smaller value for Δx will give more rectangles and in most cases a better approximation, but for an exact answer we need to take a limit as Δx approaches zero.

The symbol of integration is ${\displaystyle \int }$, an elongated S (the S stands for "sum"). The definite integral is written as:

${\displaystyle \int _{a}^{b}f(x)\,dx.}$

and is read "the integral from a to b of f-of-x with respect to x." The Leibniz notation dx is intended to suggest dividing the area under the curve into an infinite number of rectangles, so that their width Δx becomes the infinitesimally small dx. [44] :44 In a formulation of the calculus based on limits, the notation

${\displaystyle \int _{a}^{b}\cdots \,dx}$

is to be understood as an operator that takes a function as an input and gives a number, the area, as an output. The terminating differential, dx, is not a number, and is not being multiplied by f(x), although, serving as a reminder of the Δx limit definition, it can be treated as such in symbolic manipulations of the integral. Formally, the differential indicates the variable over which the function is integrated and serves as a closing bracket for the integration operator.

The indefinite integral, or antiderivative, is written:

${\displaystyle \int f(x)\,dx.}$

Functions differing by only a constant have the same derivative, and it can be shown that the antiderivative of a given function is actually a family of functions differing only by a constant. [45] :326 Since the derivative of the function y = x2 + C, where C is any constant, is y′ = 2x, the antiderivative of the latter is given by:

${\displaystyle \int 2x\,dx=x^{2}+C.}$

The unspecified constant C present in the indefinite integral or antiderivative is known as the constant of integration.

Fundamental theorem

The fundamental theorem of calculus states that differentiation and integration are inverse operations. [45] :290 More precisely, it relates the values of antiderivatives to definite integrals. Because it is usually easier to compute an antiderivative than to apply the definition of a definite integral, the fundamental theorem of calculus provides a practical way of computing definite integrals. It can also be interpreted as a precise statement of the fact that differentiation is the inverse of integration.

The fundamental theorem of calculus states: If a function f is continuous on the interval [a, b] and if F is a function whose derivative is f on the interval (a, b), then

${\displaystyle \int _{a}^{b}f(x)\,dx=F(b)-F(a).}$

Furthermore, for every x in the interval (a, b),

${\displaystyle {\frac {d}{dx}}\int _{a}^{x}f(t)\,dt=f(x).}$

This realization, made by both Newton and Leibniz, was key to the proliferation of analytic results after their work became known. (The extent to which Newton and Leibniz were influenced by immediate predecessors, and particularly what Leibniz may have learned from the work of Isaac Barrow, is difficult to determine thanks to the priority dispute between them. [46] ) The fundamental theorem provides an algebraic method of computing many definite integrals—without performing limit processes—by finding formulae for antiderivatives. It is also a prototype solution of a differential equation. Differential equations relate an unknown function to its derivatives, and are ubiquitous in the sciences.

Applications

Calculus is used in every branch of the physical sciences, [47] :1 actuarial science, computer science, statistics, engineering, economics, business, medicine, demography, and in other fields wherever a problem can be mathematically modeled and an optimal solution is desired. It allows one to go from (non-constant) rates of change to the total change or vice versa, and many times in studying a problem we know one and are trying to find the other. [48] Calculus can be used in conjunction with other mathematical disciplines. For example, it can be used with linear algebra to find the "best fit" linear approximation for a set of points in a domain. Or, it can be used in probability theory to determine the expectation value of a continuous random variable given a probability density function. [49] :37 In analytic geometry, the study of graphs of functions, calculus is used to find high points and low points (maxima and minima), slope, concavity and inflection points. Calculus is also used to find approximate solutions to equations; in practice it is the standard way to solve differential equations and do root finding in most applications. Examples are methods such as Newton's method, fixed point iteration, and linear approximation. For instance, spacecraft use a variation of the Euler method to approximate curved courses within zero gravity environments.

Physics makes particular use of calculus; all concepts in classical mechanics and electromagnetism are related through calculus. The mass of an object of known density, the moment of inertia of objects, and the potential energies due to gravitational and electromagnetic forces can all be found by the use of calculus. An example of the use of calculus in mechanics is Newton's second law of motion, which states that the derivative of an object's momentum with respect to time equals the net force upon it. Alternatively, Newton's second law can be expressed by saying that the net force is equal to the object's mass times its acceleration, which is the time derivative of velocity and thus the second time derivative of spatial position. Starting from knowing how an object is accelerating, we use calculus to derive its path. [50]

Maxwell's theory of electromagnetism and Einstein's theory of general relativity are also expressed in the language of differential calculus. [51] [52] :52–55 Chemistry also uses calculus in determining reaction rates [53] :599 and in studying radioactive decay. [53] :814 In biology, population dynamics starts with reproduction and death rates to model population changes. [54] [55] :631

Green's theorem, which gives the relationship between a line integral around a simple closed curve C and a double integral over the plane region D bounded by C, is applied in an instrument known as a planimeter, which is used to calculate the area of a flat surface on a drawing. [56] For example, it can be used to calculate the amount of area taken up by an irregularly shaped flower bed or swimming pool when designing the layout of a piece of property.

In the realm of medicine, calculus can be used to find the optimal branching angle of a blood vessel so as to maximize flow. [57] Calculus can be applied to understand how quickly a drug is eliminated from a body or how quickly a cancerous tumour grows. [58]

In economics, calculus allows for the determination of maximal profit by providing a way to easily calculate both marginal cost and marginal revenue. [59] :387

Varieties

Over the years, many reformulations of calculus have been investigated for different purposes.

Non-standard calculus

Imprecise calculations with infinitesimals were widely replaced with the rigorous (ε, δ)-definition of limit starting in the 1870s. Meanwhile, calculations with infinitesimals persisted and often led to correct results. This led Abraham Robinson to investigate if it were possible to develop a number system with infinitesimal quantities over which the theorems of calculus were still valid. In 1960, building upon the work of Edwin Hewitt and Jerzy Łoś, he succeeded in developing non-standard analysis. The theory of non-standard analysis is rich enough to be applied in many branches of mathematics. As such, books and articles dedicated solely to the traditional theorems of calculus often go by the title non-standard calculus.

Smooth infinitesimal analysis

This is another reformulation of the calculus in terms of infinitesimals. Based on the ideas of F. W. Lawvere and employing the methods of category theory, it views all functions as being continuous and incapable of being expressed in terms of discrete entities. One aspect of this formulation is that the law of excluded middle does not hold in this formulation. [34]

Constructive analysis

Constructive mathematics is a branch of mathematics that insists that proofs of the existence of a number, function, or other mathematical object should give a construction of the object. As such constructive mathematics also rejects the law of excluded middle. Reformulations of calculus in a constructive framework are generally part of the subject of constructive analysis. [34]

Related Research Articles

In mathematics, the derivative of a function of a real variable measures the sensitivity to change of the function value with respect to a change in its argument. Derivatives are a fundamental tool of calculus. For example, the derivative of the position of a moving object with respect to time is the object's velocity: this measures how quickly the position of the object changes when time advances.

In mathematics, an integral assigns numbers to functions in a way that describes displacement, area, volume, and other concepts that arise by combining infinitesimal data. The process of finding integrals is called integration. Along with differentiation, integration is a fundamental, essential operation of calculus, and serves as a tool to solve problems in mathematics and physics involving the area of an arbitrary shape, the length of a curve, and the volume of a solid, among others.

In mathematics, differential calculus is a subfield of calculus that studies the rates at which quantities change. It is one of the two traditional divisions of calculus, the other being integral calculus—the study of the area beneath a curve.

In mathematics, the system of hyperreal numbers is a way of treating infinite and infinitesimal quantities. The hyperreals, or nonstandard reals, *R, are an extension of the real numbers R that contains numbers greater than anything of the form

In mathematics, an infinitesimal or infinitesimal number is a quantity that is closer to zero than any standard real number, but that is not zero. The word infinitesimal comes from a 17th-century Modern Latin coinage infinitesimus, which originally referred to the "infinity-th" item in a sequence.

In analysis, numerical integration comprises a broad family of algorithms for calculating the numerical value of a definite integral, and by extension, the term is also sometimes used to describe the numerical solution of differential equations. This article focuses on calculation of definite integrals.

In calculus, integration by substitution, also known as u-substitution, reverse chain rule or change of variables, is a method for evaluating integrals and antiderivatives. It is the counterpart to the chain rule for differentiation, and can loosely be thought of as using the chain rule "backwards".

In calculus, Leibniz's notation, named in honor of the 17th-century German philosopher and mathematician Gottfried Wilhelm Leibniz, uses the symbols dx and dy to represent infinitely small increments of x and y, respectively, just as Δx and Δy represent finite increments of x and y, respectively.

In mathematics, nonstandard calculus is the modern application of infinitesimals, in the sense of nonstandard analysis, to infinitesimal calculus. It provides a rigorous justification for some arguments in calculus that were previously considered merely heuristic.

In mathematics, differential refers to several related notions derived from the early days of calculus, put on a rigorous footing, such as infinitesimal differences and the derivatives of functions.

Calculus, known in its early history as infinitesimal calculus, is a mathematical discipline focused on limits, continuity, derivatives, integrals, and infinite series. Isaac Newton and Gottfried Wilhelm Leibniz independently developed the theory of infinitesimal calculus in the later 17th century. By the end of the 17th century, both Leibniz and Newton claimed that the other had stolen his work, and the Leibniz–Newton calculus controversy continued until the death of Leibniz in 1716.

In calculus, symbolic integration is the problem of finding a formula for the antiderivative, or indefinite integral, of a given function f(x), i.e. to find a differentiable function F(x) such that

In calculus, the second derivative, or the second order derivative, of a function f is the derivative of the derivative of f. Roughly speaking, the second derivative measures how the rate of change of a quantity is itself changing; for example, the second derivative of the position of an object with respect to time is the instantaneous acceleration of the object, or the rate at which the velocity of the object is changing with respect to time. In Leibniz notation:

In differential calculus, there is no single uniform notation for differentiation. Instead, various notations for the derivative of a function or variable have been proposed by various mathematicians. The usefulness of each notation varies with the context, and it is sometimes advantageous to use more than one notation in a given context. The most common notations for differentiation are listed below.

The fundamental theorem of calculus is a theorem that links the concept of differentiating a function with the concept of integrating a function. The two operations are inverses of each other apart from a constant value which is dependent on where one starts to compute area.

In calculus, the differential represents the principal part of the change in a function y = f(x) with respect to changes in the independent variable. The differential dy is defined by

In calculus, Cavalieri's quadrature formula, named for 17th-century Italian mathematician Bonaventura Cavalieri, is the integral

Most of the terms listed in Wikipedia glossaries are already defined and explained within Wikipedia itself. However, glossaries like this one are useful for looking up, comparing and reviewing large numbers of terms together. You can help enhance this page by adding new terms or writing definitions for existing ones.

References

1. DeBaggis, Henry F.; Miller, Kenneth S. (1966). Foundations of the Calculus. Philadelphia: Saunders. OCLC   527896.
2. Boyer, Carl B. (1959). . New York: Dover. OCLC   643872.
3. Bardi, Jason Socrates (2006). The Calculus Wars : Newton, Leibniz, and the Greatest Mathematical Clash of All Time. New York: Thunder's Mouth Press. ISBN   1-56025-706-7.
4. Hoffmann, Laurence D.; Bradley, Gerald L. (2004). Calculus for Business, Economics, and the Social and Life Sciences (8th ed.). Boston: McGraw Hill. ISBN   0-07-242432-X.
5. See, for example:
6. . Oxford English Dictionary (Online ed.). Oxford University Press. (Subscription or participating institution membership required.)
7. Kline, Morris (16 August 1990). Mathematical Thought from Ancient to Modern Times: Volume 1. Oxford University Press. pp. 15–21. ISBN   978-0-19-506135-2.
8. Imhausen, Annette (2016). Mathematics in Ancient Egypt: A Contextual History. Princeton University Press. p. 112. ISBN   978-1-4008-7430-9. OCLC   934433864.
9. See, for example:
10. Dun, Liu; Fan, Dainian; Cohen, Robert Sonné (1966). A comparison of Archimdes' and Liu Hui's studies of circles. Chinese studies in the history and philosophy of science and technology. Vol. 130. Springer. p. 279. ISBN   978-0-7923-3463-7.
11. Dainian Fan; R. S. Cohen (1996). Chinese studies in the history and philosophy of science and technology. Dordrecht: Kluwer Academic Publishers. ISBN   0-7923-3463-9. OCLC   32272485.
12. Katz, Victor J. (2008). A history of mathematics (3rd ed.). Boston, MA: Addison-Wesley. p. 203. ISBN   978-0-321-38700-4.
13. Zill, Dennis G.; Wright, Scott; Wright, Warren S. (2009). Calculus: Early Transcendentals (3 ed.). Jones & Bartlett Learning. p. xxvii. ISBN   978-0-7637-5995-7.
14. Katz, Victor J. (June 1995). "Ideas of Calculus in Islam and India". Mathematics Magazine . 68 (3): 163–174. doi:10.1080/0025570X.1995.11996307. ISSN   0025-570X. JSTOR   2691411.
15. Plofker, Kim (2009). Mathematics in India. Princeton: Princeton University Press. ISBN   978-1-4008-3407-5. OCLC   650305544.
16. von Neumann, J. (1947). "The Mathematician". In Heywood, R. B. (ed.). The Works of the Mind. University of Chicago Press. pp. 180–196. Reprinted in Bródy, F.; Vámos, T., eds. (1995). The Neumann Compendium. World Scientific Publishing Co. Pte. Ltd. pp. 618–626. ISBN   981-02-2201-7.
17. "Johannes Kepler: His Life, His Laws and Times". NASA. 24 September 2016. Retrieved 10 June 2021.
18. Chisholm, Hugh, ed. (1911). . Encyclopædia Britannica . Vol. 14 (11th ed.). Cambridge University Press. p. 537.
19. Weil, André (1984). Number theory: An approach through History from Hammurapi to Legendre. Boston: Birkhauser Boston. p. 28. ISBN   0-8176-4565-9.
20. Hollingdale, Stuart (1991). "Review of Before Newton: The Life and Times of Isaac Barrow". Notes and Records of the Royal Society of London . 45 (2): 277–279. doi:10.1098/rsnr.1991.0027. ISSN   0035-9149. JSTOR   531707. S2CID   165043307. The most interesting to us are Lectures X-XII, in which Barrow comes close to providing a geometrical demonstration of the fundamental theorem of the calculus... He did not realize, however, the full significance of his results, and his rejection of algebra means that his work must remain a piece of mid-17th century geometrical analysis of mainly historic interest.
21. Blank, Brian E.; Krantz, Steven George (2006). Calculus: Single Variable, Volume 1 (Illustrated ed.). Springer Science & Business Media. p. 248. ISBN   978-1-931914-59-8.
22. Ferraro, Giovanni (2007). The Rise and Development of the Theory of Series up to the Early 1820s (Illustrated ed.). Springer Science & Business Media. p. 87. ISBN   978-0-387-73468-2.
23. Guicciardini, Niccolò (2005). "Isaac Newton, Philosophiae naturalis principia mathematica, first edition (1687)". Landmark Writings in Western Mathematics 1640-1940. Elsevier. pp. 59–87. doi:10.1016/b978-044450871-3/50086-3. ISBN   978-0-444-50871-3. [Newton] immediately realised that quadrature problems (the inverse problems) could be tackled via infinite series: as we would say nowadays, by expanding the integrand in power series and integrating term-wise.
24. Grattan-Guinness, I., ed. (2005). Landmark writings in Western mathematics 1640-1940. Amsterdam: Elsevier. ISBN   0-444-50871-6. OCLC   60416766.
25. Leibniz, Gottfried Wilhelm (2008). The Early Mathematical Manuscripts of Leibniz. Cosimo, Inc. p. 228. ISBN   978-1-605-20533-5.
26. Mazur, Joseph (2014). Enlightening Symbols / A Short History of Mathematical Notation and Its Hidden Powers. Princeton University Press. p. 166. ISBN   978-0-691-17337-5. Leibniz understood symbols, their conceptual powers as well as their limitations. He would spend years experimenting with some—adjusting, rejecting, and corresponding with everyone he knew, consulting with as many of the leading mathematicians of the time who were sympathetic to his fastidiousness.
27. Schrader, Dorothy V. (1962). "The Newton-Leibniz controversy concerning the discovery of the calculus". The Mathematics Teacher. 55 (5): 385–396. doi:10.5951/MT.55.5.0385. ISSN   0025-5769. JSTOR   27956626.
28. Stedall, Jacqueline (2012). The History of Mathematics: A Very Short Introduction. Oxford University Press. ISBN   978-0-191-63396-6.
29. Stenhouse, Brigitte (May 2020). "Mary Somerville's early contributions to the circulation of differential calculus". Historia Mathematica . 51: 1–25. doi:10.1016/j.hm.2019.12.001. S2CID   214472568.
30. Allaire, Patricia R. (2007). Foreword. A Biography of Maria Gaetana Agnesi, an Eighteenth-century Woman Mathematician. By Cupillari, Antonella. Edwin Mellen Press. p. iii. ISBN   978-0-7734-5226-8.
31. Unlu, Elif (April 1995). "Maria Gaetana Agnesi". Agnes Scott College.
32. Bell, John L. (6 September 2013). "Continuity and Infinitesimals". Stanford Encyclopedia of Philosophy . Retrieved 20 February 2022.
33. Russell, Bertrand (1946). History of Western Philosophy. London: George Allen & Unwin Ltd. p.  857. The great mathematicians of the seventeenth century were optimistic and anxious for quick results; consequently they left the foundations of analytical geometry and the infinitesimal calculus insecure. Leibniz believed in actual infinitesimals, but although this belief suited his metaphysics it had no sound basis in mathematics. Weierstrass, soon after the middle of the nineteenth century, showed how to establish the calculus without infinitesimals, and thus at last made it logically secure. Next came Georg Cantor, who developed the theory of continuity and infinite number. "Continuity" had been, until he defined it, a vague word, convenient for philosophers like Hegel, who wished to introduce metaphysical muddles into mathematics. Cantor gave a precise significance to the word, and showed that continuity, as he defined it, was the concept needed by mathematicians and physicists. By this means a great deal of mysticism, such as that of Bergson, was rendered antiquated.
34. Grabiner, Judith V. (1981). . Cambridge: MIT Press. ISBN   978-0-387-90527-3.
35. Archibald, Tom (2008). "The Development of Rigor in Mathematical Analysis". In Gowers, Timothy; Barrow-Green, June; Leader, Imre (eds.). The Princeton Companion to Mathematics. Princeton University Press. pp. 117–129. ISBN   978-0-691-11880-2. OCLC   682200048.
36. Rice, Adrian (2008). "A Chronology of Mathematical Events". In Gowers, Timothy; Barrow-Green, June; Leader, Imre (eds.). The Princeton Companion to Mathematics. Princeton University Press. pp. 1010–1014. ISBN   978-0-691-11880-2. OCLC   682200048.
37. Siegmund-Schultze, Reinhard (2008). "Henri Lebesgue". In Gowers, Timothy; Barrow-Green, June; Leader, Imre (eds.). The Princeton Companion to Mathematics. Princeton University Press. pp. 796–797. ISBN   978-0-691-11880-2. OCLC   682200048.
38. Barany, Michael J.; Paumier, Anne-Sandrine; Lützen, Jesper (November 2017). "From Nancy to Copenhagen to the World: The internationalization of Laurent Schwartz and his theory of distributions". Historia Mathematica . 44 (4): 367–394. doi:10.1016/j.hm.2017.04.002.
39. Daubin, Joseph W. (2008). "Abraham Robinson". In Gowers, Timothy; Barrow-Green, June; Leader, Imre (eds.). The Princeton Companion to Mathematics. Princeton University Press. pp. 822–823. ISBN   978-0-691-11880-2. OCLC   682200048.
40. Kline, Morris (1990). Mathematical thought from ancient to modern times. v. 3. New York: Oxford University Press. ISBN   978-0-19-977048-9. OCLC   726764443.
41. Cheng, Eugenia (2017). Beyond Infinity: An Expedition to the Outer Limits of Mathematics. Basic Books. pp. 206–210. ISBN   978-1-541-64413-7. OCLC   1003309980.
42. Frautschi, Steven C.; Olenick, Richard P.; Apostol, Tom M.; Goodstein, David L. (2007). The Mechanical Universe: Mechanics and Heat (Advanced ed.). Cambridge [Cambridgeshire]: Cambridge University Press. ISBN   978-0-521-71590-4. OCLC   227002144.
43. Hughes-Hallett, Deborah; McCallum, William G.; Gleason, Andrew M.; et al. (2013). Calculus: Single and Multivariable (6th ed.). Hoboken, NJ: Wiley. ISBN   978-0-470-88861-2. OCLC   794034942.
44. See, for example:
45. Baron, Margaret E. (1969). The origins of the infinitesimal calculus (1st ed.). Oxford. ISBN   978-1-483-28092-9. OCLC   892067655.
46. Hu, Zhiying (14 April 2021). "The Application and Value of Calculus in Daily Life". 2021 2nd Asia-Pacific Conference on Image Processing, Electronics and Computers. Ipec2021. Dalian China: ACM: 562–564. doi:10.1145/3452446.3452583. ISBN   978-1-4503-8981-5. S2CID   233384462.
47. Garber, Elizabeth (2001). The language of physics : the calculus and the development of theoretical physics in Europe, 1750-1914. Springer Science+Business Media. ISBN   978-1-4612-7272-4. OCLC   921230825.
48. Hall, Graham (2008). "Maxwell's Electromagnetic Theory and Special Relativity". Philosophical Transactions: Mathematical, Physical and Engineering Sciences. 366 (1871): 1849–1860. Bibcode:2008RSPTA.366.1849H. doi:10.1098/rsta.2007.2192. ISSN   1364-503X. JSTOR   25190792. PMID   18218598. S2CID   502776.
49. Gbur, Greg (2011). Mathematical Methods for Optical Physics and Engineering. Cambridge, U.K.: Cambridge University Press. ISBN   978-0-511-91510-9. OCLC   704518582.
50. Atkins, Peter W.; Jones, Loretta (2010). Chemical principles : the quest for insight (5th ed.). New York: W.H. Freeman. ISBN   978-1-4292-1955-6. OCLC   501943698.
51. Murray, J. D. (2002). Mathematical biology. I, Introduction (3rd ed.). New York: Springer. ISBN   0-387-22437-8. OCLC   53165394.
52. Neuhauser, Claudia (2011). Calculus for biology and medicine (3rd ed.). Boston: Prentice Hall. ISBN   978-0-321-64468-8. OCLC   426065941.
53. Gatterdam, R. W. (1981). "The planimeter as an example of Green's theorem". The American Mathematical Monthly . 88 (9): 701–704. doi:10.2307/2320679. JSTOR   2320679.
54. Adam, John A. (June 2011). "Blood Vessel Branching: Beyond the Standard Calculus Problem". Mathematics Magazine . 84 (3): 196–207. doi:10.4169/math.mag.84.3.196. ISSN   0025-570X. S2CID   8259705.
55. Mackenzie, Dana (2004). "Mathematical Modeling and Cancer" (PDF). SIAM News . 37 (1).
56. Perloff, Jeffrey M. (2018). Microeconomics : Theory and Applications with Calculus (4th global ed.). Harlow, United Kingdom. ISBN   978-1-292-15446-6. OCLC   1064041906.