Large numbers

Last updated

Large numbers are numbers significantly larger than those typically used in everyday life (for instance in simple counting or in monetary transactions), appearing frequently in fields such as mathematics, cosmology, cryptography, and statistical mechanics. They are typically large positive integers, or more generally, large positive real numbers, but may also be other numbers in other contexts. Googology is the study of nomenclature and properties of large numbers. [1] [2] [ better source needed ]

Contents

In the everyday world

Scientific notation was created to handle the wide range of values that occur in scientific study. 1.0 × 109, for example, means one billion, or a 1 followed by nine zeros: 1 000 000 000. The reciprocal, 1.0 × 10−9, means one billionth, or 0.000 000 001. Writing 109 instead of nine zeros saves readers the effort and hazard of counting a long series of zeros to see how large the number is. In addition to scientific (powers of 10) notation, the following examples include (short scale) systematic nomenclature of large numbers.

Examples of large numbers describing everyday real-world objects include:

Astronomical

Other large numbers, as regards length and time, are found in astronomy and cosmology. For example, the current Big Bang model suggests that the universe is 13.8 billion years (4.355 × 1017 seconds) old, and that the observable universe is 93 billion light years across (8.8 × 1026 metres), and contains about 5 × 1022 stars, organized into around 125 billion (1.25 × 1011) galaxies, according to Hubble Space Telescope observations. There are about 1080 atoms in the observable universe, by rough estimation. [7]

According to Don Page, physicist at the University of Alberta, Canada, the longest finite time that has so far been explicitly calculated by any physicist is

which corresponds to the scale of an estimated Poincaré recurrence time for the quantum state of a hypothetical box containing a black hole with the estimated mass of the entire universe, observable or not, assuming a certain inflationary model with an inflaton whose mass is 10−6 Planck masses. [8] [9] This time assumes a statistical model subject to Poincaré recurrence. A much simplified way of thinking about this time is in a model where the universe's history repeats itself arbitrarily many times due to properties of statistical mechanics; this is the time scale when it will first be somewhat similar (for a reasonable choice of "similar") to its current state again.

Combinatorial processes rapidly generate even larger numbers. The factorial function, which defines the number of permutations on a set of fixed objects, grows very rapidly with the number of objects. Stirling's formula gives a precise asymptotic expression for this rate of growth.

Combinatorial processes generate very large numbers in statistical mechanics. These numbers are so large that they are typically only referred to using their logarithms.

Gödel numbers, and similar numbers used to represent bit-strings in algorithmic information theory, are very large, even for mathematical statements of reasonable length. However, some pathological numbers are even larger than the Gödel numbers of typical mathematical propositions.

Logician Harvey Friedman has done work related to very large numbers, such as with Kruskal's tree theorem and the Robertson–Seymour theorem.

"Billions and billions"

To help viewers of Cosmos distinguish between "millions" and "billions", astronomer Carl Sagan stressed the "b". Sagan never did, however, say "billions and billions". The public's association of the phrase and Sagan came from a Tonight Show skit. Parodying Sagan's effect, Johnny Carson quipped "billions and billions". [10] The phrase has, however, now become a humorous fictitious number—the Sagan. Cf., Sagan Unit.

Examples

Standardized system of writing

A standardized way of writing very large numbers allows them to be easily sorted in increasing order, and one can get a good idea of how much larger a number is than another one.

To compare numbers in scientific notation, say 5×104 and 2×105, compare the exponents first, in this case 5 > 4, so 2×105 > 5×104. If the exponents are equal, the mantissa (or coefficient) should be compared, thus 5×104 > 2×104 because 5 > 2.

Tetration with base 10 gives the sequence , the power towers of numbers 10, where denotes a functional power of the function (the function also expressed by the suffix "-plex" as in googolplex, see the googol family).

These are very round numbers, each representing an order of magnitude in a generalized sense. A crude way of specifying how large a number is, is specifying between which two numbers in this sequence it is.

More precisely, numbers in between can be expressed in the form , i.e., with a power tower of 10s and a number at the top, possibly in scientific notation, e.g. , a number between and (note that if ). (See also extension of tetration to real heights.)

Thus googolplex is

Another example:

(between and )

Thus the "order of magnitude" of a number (on a larger scale than usually meant), can be characterized by the number of times (n) one has to take the to get a number between 1 and 10. Thus, the number is between and . As explained, a more precise description of a number also specifies the value of this number between 1 and 10, or the previous number (taking the logarithm one time less) between 10 and 1010, or the next, between 0 and 1.

Note that

I.e., if a number x is too large for a representation the power tower can be made one higher, replacing x by log10x, or find x from the lower-tower representation of the log10 of the whole number. If the power tower would contain one or more numbers different from 10, the two approaches would lead to different results, corresponding to the fact that extending the power tower with a 10 at the bottom is then not the same as extending it with a 10 at the top (but, of course, similar remarks apply if the whole power tower consists of copies of the same number, different from 10).

If the height of the tower is large, the various representations for large numbers can be applied to the height itself. If the height is given only approximately, giving a value at the top does not make sense, so the double-arrow notation (e.g. ) can be used. If the value after the double arrow is a very large number itself, the above can recursively be applied to that value.

Examples:

(between and )
(between and )

Similarly to the above, if the exponent of is not exactly given then giving a value at the right does not make sense, and instead of using the power notation of , it is possible to add to the exponent of , to obtain e.g. .

If the exponent of is large, the various representations for large numbers can be applied to this exponent itself. If this exponent is not exactly given then, again, giving a value at the right does not make sense, and instead of using the power notation of it is possible use the triple arrow operator, e.g. .

If the right-hand argument of the triple arrow operator is large the above applies to it, obtaining e.g. (between and ). This can be done recursively, so it is possible to have a power of the triple arrow operator.

Then it is possible to proceed with operators with higher numbers of arrows, written .

Compare this notation with the hyper operator and the Conway chained arrow notation:

= ( abn ) = hyper(a, n + 2, b)

An advantage of the first is that when considered as function of b, there is a natural notation for powers of this function (just like when writing out the n arrows): . For example:

= ( 10 → ( 10 → ( 10 → b → 2 ) → 2 ) → 2 )

and only in special cases the long nested chain notation is reduced; for obtains:

= ( 10 → 3 → 3 )

Since the b can also be very large, in general it can be written instead a number with a sequence of powers with decreasing values of n (with exactly given integer exponents ) with at the end a number in ordinary scientific notation. Whenever a is too large to be given exactly, the value of is increased by 1 and everything to the right of is rewritten.

For describing numbers approximately, deviations from the decreasing order of values of n are not needed. For example, , and . Thus is obtained the somewhat counterintuitive result that a number x can be so large that, in a way, x and 10x are "almost equal" (for arithmetic of large numbers see also below).

If the superscript of the upward arrow is large, the various representations for large numbers can be applied to this superscript itself. If this superscript is not exactly given then there is no point in raising the operator to a particular power or to adjust the value on which it act, instead it is possible to simply use a standard value at the right, say 10, and the expression reduces to with an approximate n. For such numbers the advantage of using the upward arrow notation no longer applies, so the chain notation can be used instead.

The above can be applied recursively for this n, so the notation is obtained in the superscript of the first arrow, etc., or a nested chain notation, e.g.:

(10 → 10 → (10 → 10 → ) ) =

If the number of levels gets too large to be convenient, a notation is used where this number of levels is written down as a number (like using the superscript of the arrow instead of writing many arrows). Introducing a function = (10 → 10 → n), these levels become functional powers of f, allowing us to write a number in the form where m is given exactly and n is an integer which may or may not be given exactly (for example: ). If n is large, any of the above can be used for expressing it. The "roundest" of these numbers are those of the form fm(1) = (10→10→m→2). For example,

Compare the definition of Graham's number: it uses numbers 3 instead of 10 and has 64 arrow levels and the number 4 at the top; thus , but also .

If m in is too large to give exactly, it is possible to use a fixed n, e.g. n = 1, and apply the above recursively to m, i.e., the number of levels of upward arrows is itself represented in the superscripted upward-arrow notation, etc. Using the functional power notation of f this gives multiple levels of f. Introducing a function these levels become functional powers of g, allowing us to write a number in the form where m is given exactly and n is an integer which may or may not be given exactly. For example, if (10→10→m→3) = gm(1). If n is large any of the above can be used for expressing it. Similarly a function h, etc. can be introduced. If many such functions are required, they can be numbered instead of using a new letter every time, e.g. as a subscript, such that there are numbers of the form where k and m are given exactly and n is an integer which may or may not be given exactly. Using k=1 for the f above, k=2 for g, etc., obtains (10→10→nk) = . If n is large any of the above can be used to express it. Thus is obtained a nesting of forms where going inward the k decreases, and with as inner argument a sequence of powers with decreasing values of n (where all these numbers are exactly given integers) with at the end a number in ordinary scientific notation.

When k is too large to be given exactly, the number concerned can be expressed as =(10→10→10→n) with an approximate n. Note that the process of going from the sequence =(10→n) to the sequence =(10→10→n) is very similar to going from the latter to the sequence =(10→10→10→n): it is the general process of adding an element 10 to the chain in the chain notation; this process can be repeated again (see also the previous section). Numbering the subsequent versions of this function a number can be described using functions , nested in lexicographical order with q the most significant number, but with decreasing order for q and for k; as inner argument yields a sequence of powers with decreasing values of n (where all these numbers are exactly given integers) with at the end a number in ordinary scientific notation.

For a number too large to write down in the Conway chained arrow notation it size can be described by the length of that chain, for example only using elements 10 in the chain; in other words, one could specify its position in the sequence 10, 10→10, 10→10→10, .. If even the position in the sequence is a large number same techniques can be applied again.

Examples

Numbers expressible in decimal notation:

Numbers expressible in scientific notation:

Numbers expressible in (10 ↑)nk notation:

Bigger numbers:

Other notations

Some notations for extremely large numbers:

These notations are essentially functions of integer variables, which increase very rapidly with those integers. Ever-faster-increasing functions can easily be constructed recursively by applying these functions with large integers as argument.

A function with a vertical asymptote is not helpful in defining a very large number, although the function increases very rapidly: one has to define an argument very close to the asymptote, i.e. use a very small number, and constructing that is equivalent to constructing a very large number, e.g. the reciprocal.

Comparison of base values

The following illustrates the effect of a base different from 10, base 100. It also illustrates representations of numbers and the arithmetic.

, with base 10 the exponent is doubled.

, ditto.

, the highest exponent is very little more than doubled (increased by log102).

Accuracy

For a number , one unit change in n changes the result by a factor 10. In a number like , with the 6.2 the result of proper rounding using significant figures, the true value of the exponent may be 50 less or 50 more. Hence the result may be a factor too large or too small. This seems like extremely poor accuracy, but for such a large number it may be considered fair (a large error in a large number may be "relatively small" and therefore acceptable).

For very large numbers

In the case of an approximation of an extremely large number, the relative error may be large, yet there may still be a sense in which one wants to consider the numbers as "close in magnitude". For example, consider

and

The relative error is

a large relative error. However, one can also consider the relative error in the logarithms; in this case, the logarithms (to base 10) are 10 and 9, so the relative error in the logarithms is only 10%.

The point is that exponential functions magnify relative errors greatly – if a and b have a small relative error,

and

the relative error is larger, and

and

will have an even larger relative error. The question then becomes: on which level of iterated logarithms do to compare two numbers? There is a sense in which one may want to consider

and

to be "close in magnitude". The relative error between these two numbers is large, and the relative error between their logarithms is still large; however, the relative error in their second-iterated logarithms is small:

and

Such comparisons of iterated logarithms are common, e.g., in analytic number theory.

Classes

One solution to the problem of comparing large numbers is to define classes of numbers, such as the system devised by Robert Munafo, [13] which is based on different "levels" of perception of an average person. Class 0 – numbers between zero and six – is defined to contain numbers that are easily subitized, that is, numbers that show up very frequently in daily life and are almost instantly comparable. Class 1 – numbers between six and 1,000,000=106 – is defined to contain numbers whose decimal expressions are easily subitized, that is, numbers who are easily comparable not by cardinality, but "at a glance" given the decimal expansion.

Each class after these are defined in terms of iterating this base-10 exponentiation, to simulate the effect of another "iteration" of human indistinguishibility. For example, class 5 is defined to include numbers between 101010106 and 10101010106, which are numbers where X becomes humanly indistinguishable from X2 [14] (taking iterated logarithms of such X yields indistinguishibility firstly between log(X) and 2log(X), secondly between log(log(X)) and 1+log(log(X)), and finally an extremely long decimal expansion whose length can't be subitized).

Approximate arithmetic

There are some general rules relating to the usual arithmetic operations performed on very large numbers:

Hence:

Systematically creating ever-faster-increasing sequences

Given a strictly increasing integer sequence/function (n≥1), it is possible to produce a faster-growing sequence (where the superscript n denotes the nth functional power). This can be repeated any number of times by letting , each sequence growing much faster than the one before it. Thus it is possible to define , which grows much faster than any for finite k (here ω is the first infinite ordinal number, representing the limit of all finite numbers k). This is the basis for the fast-growing hierarchy of functions, in which the indexing subscript is extended to ever-larger ordinals.

For example, starting with f0(n) = n + 1:

In some noncomputable sequences

The busy beaver function Σ is an example of a function which grows faster than any computable function. Its value for even relatively small input is huge. The values of Σ(n) for n = 1, 2, 3, 4 are 1, 4, 6, 13 (sequence A028444 in the OEIS ). Σ(5) is not known but is definitely ≥ 4098. Σ(6) is at least 10↑↑15.

Infinite numbers

Although all the numbers discussed above are very large, they are all still decidedly finite. Certain fields of mathematics define infinite and transfinite numbers. For example, aleph-null is the cardinality of the infinite set of natural numbers, and aleph-one is the next greatest cardinal number. is the cardinality of the reals. The proposition that is known as the continuum hypothesis.

See also

Related Research Articles

<span class="mw-page-title-main">Exponential function</span> Mathematical function, denoted exp(x) or e^x

The exponential function is a mathematical function denoted by or . Unless otherwise specified, the term generally refers to the positive-valued function of a real variable, although it can be extended to the complex numbers or generalized to other mathematical objects like matrices or Lie algebras. The exponential function originated from the operation of taking powers of a number, but various modern definitions allow it to be rigorously extended to all real arguments , including irrational numbers. Its ubiquitous occurrence in pure and applied mathematics led mathematician Walter Rudin to consider the exponential function to be "the most important function in mathematics".

In mathematics, the factorial of a non-negative integer , denoted by , is the product of all positive integers less than or equal to . The factorial of also equals the product of with the next smaller factorial:

<span class="mw-page-title-main">Logarithm</span> Mathematical function, inverse of an exponential function

In mathematics, the logarithm is the inverse function to exponentiation. That means that the logarithm of a number x to the base b is the exponent to which b must be raised to produce x. For example, since 1000 = 103, the logarithm base  of 1000 is 3, or log10 (1000) = 3. The logarithm of x to base b is denoted as logb (x), or without parentheses, logbx. When the base is clear from the context or is irrelevant, such as in big O notation, it is sometimes written log x.

<span class="mw-page-title-main">Natural logarithm</span> Logarithm to the base of the mathematical constant e

The natural logarithm of a number is its logarithm to the base of the mathematical constant e, which is an irrational and transcendental number approximately equal to 2.718281828459. The natural logarithm of x is generally written as ln x, logex, or sometimes, if the base e is implicit, simply log x. Parentheses are sometimes added for clarity, giving ln(x), loge(x), or log(x). This is done particularly when the argument to the logarithm is not a single symbol, so as to prevent ambiguity.

An order of magnitude is an approximation of the logarithm of a value relative to some contextually understood reference value, usually 10, interpreted as the base of the logarithm and the representative of values of magnitude one. Logarithmic distributions are common in nature and considering the order of magnitude of values sampled from such a distribution can be more intuitive. When the reference value is 10, the order of magnitude can be understood as the number of digits in the base-10 representation of the value. Similarly, if the reference value is one of some powers of 2 since computers store data in a binary format, the magnitude can be understood in terms of the amount of computer memory needed to store that value.

Big <i>O</i> notation Describes limiting behavior of a function

Big O notation is a mathematical notation that describes the limiting behavior of a function when the argument tends towards a particular value or infinity. Big O is a member of a family of notations invented by German mathematicians Paul Bachmann, Edmund Landau, and others, collectively called Bachmann–Landau notation or asymptotic notation. The letter O was chosen by Bachmann to stand for Ordnung, meaning the order of approximation.

2 (two) is a number, numeral and digit. It is the natural number following 1 and preceding 3. It is the smallest and only even prime number. Because it forms the basis of a duality, it has religious and spiritual significance in many cultures.

<span class="mw-page-title-main">Exponentiation</span> Arithmetic operation

In mathematics, exponentiation is an operation involving two numbers: the base and the exponent or power. Exponentiation is written as bn, where b is the base and n is the power; this is pronounced as "b (raised) to the n". When n is a positive integer, exponentiation corresponds to repeated multiplication of the base: that is, bn is the product of multiplying n bases:

<span class="mw-page-title-main">Common logarithm</span> Mathematical function

In mathematics, the common logarithm is the logarithm with base 10. It is also known as the decadic logarithm and as the decimal logarithm, named after its base, or Briggsian logarithm, after Henry Briggs, an English mathematician who pioneered its use, as well as standard logarithm. Historically, it was known as logarithmus decimalis or logarithmus decadis. It is indicated by log(x), log10(x), or sometimes Log(x) with a capital L; on calculators, it is printed as "log", but mathematicians usually mean natural logarithm (logarithm with base e ≈ 2.71828) rather than common logarithm when writing "log". To mitigate this ambiguity, the ISO 80000 specification recommends that log10(x) should be written lg(x), and loge(x) should be ln(x).

Graham's number is an immense number that arose as an upper bound on the answer of a problem in the mathematical field of Ramsey theory. It is much larger than many other large numbers such as Skewes's number and Moser's number, both of which are in turn much larger than a googolplex. As with these, it is so large that the observable universe is far too small to contain an ordinary digital representation of Graham's number, assuming that each digit occupies one Planck volume, possibly the smallest measurable space. But even the number of digits in this digital representation of Graham's number would itself be a number so large that its digital representation cannot be represented in the observable universe. Nor even can the number of digits of that number—and so forth, for a number of times far exceeding the total number of Planck volumes in the observable universe. Thus Graham's number cannot be expressed even by physical universe-scale power towers of the form .

<span class="mw-page-title-main">Binary logarithm</span> Exponent of a power of two

In mathematics, the binary logarithm is the power to which the number 2 must be raised to obtain the value n. That is, for any real number x,

In mathematics, Knuth's up-arrow notation is a method of notation for very large integers, introduced by Donald Knuth in 1976.

Conway chained arrow notation, created by mathematician John Horton Conway, is a means of expressing certain extremely large numbers. It is simply a finite sequence of positive integers separated by rightward arrows, e.g. .

In mathematics, Steinhaus–Moser notation is a notation for expressing certain large numbers. It is an extension of Hugo Steinhaus's polygon notation.

<span class="mw-page-title-main">Tetration</span> Arithmetic operation

In mathematics, tetration is an operation based on iterated, or repeated, exponentiation. There is no standard notation for tetration, though Knuth's up arrow notation and the left-exponent xb are common.

In information theory, the information content, self-information, surprisal, or Shannon information is a basic quantity derived from the probability of a particular event occurring from a random variable. It can be thought of as an alternative way of expressing probability, much like odds or log-odds, but which has particular mathematical advantages in the setting of information theory.

<span class="mw-page-title-main">Pentation</span> Arithmetic operation

In mathematics, pentation is the fifth hyperoperation. Pentation is defined to be repeated tetration, similarly to how tetration is repeated exponentiation, exponentiation is repeated multiplication, and multiplication is repeated addition. The concept of "pentation" was named by English mathematician Reuben Goodstein in 1947, when he came up with the naming scheme for hyperoperations.

Methods of computing square roots are algorithms for approximating the non-negative square root of a positive real number . Since all square roots of natural numbers, other than of perfect squares, are irrational, square roots can usually only be computed to some finite precision: these methods typically construct a series of increasingly accurate approximations.

<span class="mw-page-title-main">Complex logarithm</span> Logarithm of a complex number

In mathematics, a complex logarithm is a generalization of the natural logarithm to nonzero complex numbers. The term refers to one of the following, which are strongly related:

In mathematics, the super-logarithm is one of the two inverse functions of tetration. Just as exponentiation has two inverse functions, roots and logarithms, tetration has two inverse functions, super-roots and super-logarithms. There are several ways of interpreting super-logarithms:

References

  1. One Million Things: A Visual Encyclopedia [ unreliable source? ]
  2. «The study of large numbers is called googology» [ unreliable source? ]
  3. Bianconi, Eva; Piovesan, Allison; Facchin, Federica; Beraudi, Alina; Casadei, Raffaella; Frabetti, Flavia; Vitale, Lorenza; Pelleri, Maria Chiara; Tassani, Simone (Nov–Dec 2013). "An estimation of the number of cells in the human body". Annals of Human Biology. 40 (6): 463–471. doi: 10.3109/03014460.2013.807878 . hdl:11585/152451. ISSN   1464-5033. PMID   23829164. S2CID   16247166.
  4. Landenmark HK, Forgan DH, Cockell CS (June 2015). "An Estimate of the Total DNA in the Biosphere". PLOS Biology. 13 (6): e1002168. doi: 10.1371/journal.pbio.1002168 . PMC   4466264 . PMID   26066900.
  5. Nuwer R (18 July 2015). "Counting All the DNA on Earth". The New York Times. New York. ISSN   0362-4331 . Retrieved 2015-07-18.
  6. Shannon, Claude (March 1950). "XXII. Programming a Computer for Playing Chess" (PDF). Philosophical Magazine. Series 7. 41 (314). Archived from the original (PDF) on 2010-07-06. Retrieved 2019-01-25.
  7. Atoms in the Universe. Universe Today. 30-07-2009. Retrieved 02-03-13.
  8. Information Loss in Black Holes and/or Conscious Beings?, Don N. Page, Heat Kernel Techniques and Quantum Gravity (1995), S. A. Fulling (ed), p. 461. Discourses in Mathematics and its Applications, No. 4, Texas A&M University Department of Mathematics. arXiv : hep-th/9411193. ISBN   0-9630728-3-8.
  9. How to Get A Googolplex
  10. Carl Sagan takes questions more from his 'Wonder and Skepticism' CSICOP 1994 keynote, Skeptical Inquirer Archived December 21, 2016, at the Wayback Machine
  11. "GIMPS Discovers Largest Known Prime Number". Great Internet Mersenne Prime Search. 2018-12-21.
  12. Regarding the comparison with the previous value: , so starting the 64 steps with 1 instead of 4 more than compensates for replacing the numbers 3 by 10
  13. "Large Numbers at MROB". www.mrob.com. Retrieved 2021-05-13.
  14. "Large Numbers (page 2) at MROB". www.mrob.com. Retrieved 2021-05-13.