Associative property

Last updated
Associative property
Associativity of binary operations (without question marks).svg
A visual graph representing associative operations;
Type Law, rule of replacement
Field
Symbolic statement
  1. Elementary algebra
  2. Propositional calculus

In mathematics, the associative property [1] is a property of some binary operations, which means that rearranging the parentheses in an expression will not change the result. In propositional logic, associativity is a valid rule of replacement for expressions in logical proofs.

Contents

Within an expression containing two or more occurrences in a row of the same associative operator, the order in which the operations are performed does not matter as long as the sequence of the operands is not changed. That is (after rewriting the expression with parentheses and in infix notation if necessary), rearranging the parentheses in such an expression will not change its value. Consider the following equations:

Even though the parentheses were rearranged on each line, the values of the expressions were not altered. Since this holds true when performing addition and multiplication on any real numbers, it can be said that "addition and multiplication of real numbers are associative operations".

Associativity is not the same as commutativity, which addresses whether the order of two operands affects the result. For example, the order does not matter in the multiplication of real numbers, that is, a × b = b × a, so we say that the multiplication of real numbers is a commutative operation. However, operations such as function composition and matrix multiplication are associative, but not (generally) commutative.

Associative operations are abundant in mathematics; in fact, many algebraic structures (such as semigroups and categories) explicitly require their binary operations to be associative.

However, many important and interesting operations are non-associative; some examples include subtraction, exponentiation, and the vector cross product. In contrast to the theoretical properties of real numbers, the addition of floating point numbers in computer science is not associative, and the choice of how to associate an expression can have a significant effect on rounding error.

Definition

A binary operation * on the set S is associative when this diagram commutes. That is, when the two paths from SxSxS to S compose to the same function from SxSxS to S. Semigroup associative.svg
A binary operation ∗ on the set S is associative when this diagram commutes. That is, when the two paths from S×S×S to S compose to the same function from S×S×S to S.

Formally, a binary operation on a set S is called associative if it satisfies the associative law:

(xy) ∗ z = x ∗ (yz) for all x, y, z in S.

Here, ∗ is used to replace the symbol of the operation, which may be any symbol, and even the absence of symbol (juxtaposition) as for multiplication.

(xy)z = x(yz) = xyz for all x, y, z in S.

The associative law can also be expressed in functional notation thus: f(f(x, y), z) = f(x, f(y, z)).

Generalized associative law

In the absence of the associative property, five factors a, b,c, d, e result in a Tamari lattice of order four, possibly different products. Tamari lattice.svg
In the absence of the associative property, five factors a, b,c, d, e result in a Tamari lattice of order four, possibly different products.

If a binary operation is associative, repeated application of the operation produces the same result regardless of how valid pairs of parentheses are inserted in the expression. [2] This is called the generalized associative law.

The number of possible bracketings is just the Catalan number, Cn, for n operations on n+1 values. For instance, a product of 3 operations on 4 elements may be written (ignoring permutations of the arguments), in C3 = 5 possible ways:

If the product operation is associative, the generalized associative law says that all these expressions will yield the same result. So unless the expression with omitted parentheses already has a different meaning (see below), the parentheses can be considered unnecessary and "the" product can be written unambiguously as

abcd.

As the number of elements increases, the number of possible ways to insert parentheses grows quickly, but they remain unnecessary for disambiguation.

An example where this does not work is the logical biconditional . It is associative; thus, A ↔ (BC) is equivalent to (AB) ↔ C, but ABC most commonly means (AB) and (BC), which is not equivalent.

Examples

The addition of real numbers is associative. Associativity of real number addition.svg
The addition of real numbers is associative.

Some examples of associative operations include the following.

Propositional logic

Rule of replacement

In standard truth-functional propositional logic, association, [4] [5] or associativity [6] are two valid rules of replacement. The rules allow one to move parentheses in logical expressions in logical proofs. The rules (using logical connectives notation) are:

and

where "" is a metalogical symbol representing "can be replaced in a proof with".

Truth functional connectives

Associativity is a property of some logical connectives of truth-functional propositional logic. The following logical equivalences demonstrate that associativity is a property of particular connectives. The following (and their converses, since is commutative) are truth-functional tautologies.[ citation needed ]

Associativity of disjunction
Associativity of conjunction
Associativity of equivalence

Joint denial is an example of a truth functional connective that is not associative.

Non-associative operation

A binary operation on a set S that does not satisfy the associative law is called non-associative. Symbolically,

For such an operation the order of evaluation does matter. For example:

Subtraction
Division
Exponentiation
Vector cross product

Also although addition is associative for finite sums, it is not associative inside infinite sums (series). For example,

whereas

Some non-associative operations are fundamental in mathematics. They appear often as the multiplication in structures called non-associative algebras, which have also an addition and a scalar multiplication. Examples are the octonions and Lie algebras. In Lie algebras, the multiplication satisfies Jacobi identity instead of the associative law; this allows abstracting the algebraic nature of infinitesimal transformations.

Other examples are quasigroup, quasifield, non-associative ring, and commutative non-associative magmas.

Nonassociativity of floating point calculation

In mathematics, addition and multiplication of real numbers is associative. By contrast, in computer science, the addition and multiplication of floating point numbers is not associative, as rounding errors are introduced when dissimilar-sized values are joined together. [7]

To illustrate this, consider a floating point representation with a 4-bit mantissa:

(1.0002×20 + 1.0002×20) + 1.0002×24 = 1.0002×21 + 1.0002×24 = 1.0012×24
1.0002×20 + (1.0002×20 + 1.0002×24) = 1.0002×20 + 1.0002×24 = 1.0002×24

Even though most computers compute with 24 or 53 bits of mantissa, [8] this is an important source of rounding error, and approaches such as the Kahan summation algorithm are ways to minimise the errors. It can be especially problematic in parallel computing. [9] [10]

Notation for non-associative operations

In general, parentheses must be used to indicate the order of evaluation if a non-associative operation appears more than once in an expression (unless the notation specifies the order in another way, like ). However, mathematicians agree on a particular order of evaluation for several common non-associative operations. This is simply a notational convention to avoid parentheses.

A left-associative operation is a non-associative operation that is conventionally evaluated from left to right, i.e.,

while a right-associative operation is conventionally evaluated from right to left:

Both left-associative and right-associative operations occur. Left-associative operations include the following:

Subtraction and division of real numbers [11] [12] [13] [14] [15]
Function application

This notation can be motivated by the currying isomorphism, which enables partial application.

Right-associative operations include the following:

Exponentiation of real numbers in superscript notation

Exponentiation is commonly used with brackets or right-associatively because a repeated left-associative exponentiation operation is of little use. Repeated powers would mostly be rewritten with multiplication:

Formatted correctly, the superscript inherently behaves as a set of parentheses; e.g. in the expression the addition is performed before the exponentiation despite there being no explicit parentheses wrapped around it. Thus given an expression such as , the full exponent of the base is evaluated first. However, in some contexts, especially in handwriting, the difference between , and can be hard to see. In such a case, right-associativity is usually implied.

Function definition

Using right-associative notation for these operations can be motivated by the Curry–Howard correspondence and by the currying isomorphism.

Non-associative operations for which no conventional evaluation order is defined include the following.

Exponentiation of real numbers in infix notation [16]
Knuth's up-arrow operators
Taking the cross product of three vectors
Taking the pairwise average of real numbers
Taking the relative complement of sets
.

(Compare material nonimplication in logic.)

History

William Rowan Hamilton seems to have coined the term "associative property" [17] around 1844, a time when he was contemplating the non-associative algebra of the octonions he had learned about from John T. Graves. [18]

See also

Related Research Articles

In mathematics, an associative algebraA over a commutative ring K is a ring A together with a ring homomorphism from K into the center of A. This is thus an algebraic structure with an addition, a multiplication, and a scalar multiplication. The addition and multiplication operations together give A the structure of a ring; the addition and scalar multiplication operations together give A the structure of a module or vector space over K. In this article we will also use the term K-algebra to mean an associative algebra over K. A standard first example of a K-algebra is a ring of square matrices over a commutative ring K, with the usual matrix multiplication.

<span class="mw-page-title-main">Binary operation</span> Mathematical operation with two operands

In mathematics, a binary operation or dyadic operation is a rule for combining two elements to produce another element. More formally, a binary operation is an operation of arity two.

In mathematics, the greatest common divisor (GCD) of two or more integers, which are not all zero, is the largest positive integer that divides each of the integers. For two integers x, y, the greatest common divisor of x and y is denoted . For example, the GCD of 8 and 12 is 4, that is, gcd(8, 12) = 4.

<span class="mw-page-title-main">Least common multiple</span> Smallest positive number divisible by two integers

In arithmetic and number theory, the least common multiple, lowest common multiple, or smallest common multiple of two integers a and b, usually denoted by lcm(ab), is the smallest positive integer that is divisible by both a and b. Since division of integers by zero is undefined, this definition has meaning only if a and b are both different from zero. However, some authors define lcm(a, 0) as 0 for all a, since 0 is the only common multiple of a and 0.

<span class="mw-page-title-main">Logical connective</span> Symbol connecting sentential formulas in logic

In logic, a logical connective is a logical constant. Connectives can be used to connect logical formulas. For instance in the syntax of propositional logic, the binary connective can be used to join the two atomic formulas and , rendering the complex formula .

<span class="mw-page-title-main">Exponentiation</span> Arithmetic operation

In mathematics, exponentiation is an operation involving two numbers: the base and the exponent or power. Exponentiation is written as bn, where b is the base and n is the power; this is pronounced as "b (raised) to the n". When n is a positive integer, exponentiation corresponds to repeated multiplication of the base: that is, bn is the product of multiplying n bases:

In mathematics, the distributive property of binary operations is a generalization of the distributive law, which asserts that the equality

<span class="mw-page-title-main">Matrix multiplication</span> Mathematical operation in linear algebra

In mathematics, particularly in linear algebra, matrix multiplication is a binary operation that produces a matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the number of rows in the second matrix. The resulting matrix, known as the matrix product, has the number of rows of the first and the number of columns of the second matrix. The product of matrices A and B is denoted as AB.

<span class="mw-page-title-main">Unitary group</span> Group of unitary matrices

In mathematics, the unitary group of degree n, denoted U(n), is the group of n × n unitary matrices, with the group operation of matrix multiplication. The unitary group is a subgroup of the general linear group GL(n, C). Hyperorthogonal group is an archaic name for the unitary group, especially over finite fields. For the group of unitary matrices with determinant 1, see Special unitary group.

<span class="mw-page-title-main">Commutative property</span> Property of some mathematical operations

In mathematics, a binary operation is commutative if changing the order of the operands does not change the result. It is a fundamental property of many binary operations, and many mathematical proofs depend on it. Perhaps most familiar as a property of arithmetic, e.g. "3 + 4 = 4 + 3" or "2 × 5 = 5 × 2", the property can also be used in more advanced settings. The name is needed because there are operations, such as division and subtraction, that do not have it ; such operations are not commutative, and so are referred to as noncommutative operations. The idea that simple operations, such as the multiplication and addition of numbers, are commutative was for many years implicitly assumed. Thus, this property was not named until the 19th century, when mathematics started to become formalized. A similar property exists for binary relations; a binary relation is said to be symmetric if the relation applies regardless of the order of its operands; for example, equality is symmetric as two equal mathematical objects are equal regardless of their order.

In mathematics, coalgebras or cogebras are structures that are dual to unital associative algebras. The axioms of unital associative algebras can be formulated in terms of commutative diagrams. Turning all arrows around, one obtains the axioms of coalgebras. Every coalgebra, by duality, gives rise to an algebra, but not in general the other way. In finite dimensions, this duality goes in both directions.

<span class="mw-page-title-main">Wheel theory</span> Algebra where division is always defined

A wheel is a type of algebra where division is always defined. In particular, division by zero is meaningful. The real numbers can be extended to a wheel, as can any commutative ring.

In linear algebra, a rotation matrix is a transformation matrix that is used to perform a rotation in Euclidean space. For example, using the convention below, the matrix

In mathematics, the Cayley transform, named after Arthur Cayley, is any of a cluster of related things. As originally described by Cayley (1846), the Cayley transform is a mapping between skew-symmetric matrices and special orthogonal matrices. The transform is a homography used in real analysis, complex analysis, and quaternionic analysis. In the theory of Hilbert spaces, the Cayley transform is a mapping between linear operators.

In mathematics, a GCD domain is an integral domain R with the property that any two elements have a greatest common divisor (GCD); i.e., there is a unique minimal principal ideal containing the ideal generated by two given elements. Equivalently, any two elements of R have a least common multiple (LCM).

In mathematics, the tensor product of modules is a construction that allows arguments about bilinear maps to be carried out in terms of linear maps. The module construction is analogous to the construction of the tensor product of vector spaces, but can be carried out for a pair of modules over a commutative ring resulting in a third module, and also for a pair of a right-module and a left-module over any ring, with result an abelian group. Tensor products are important in areas of abstract algebra, homological algebra, algebraic topology, algebraic geometry, operator algebras and noncommutative geometry. The universal property of the tensor product of vector spaces extends to more general situations in abstract algebra. The tensor product of an algebra and a module can be used for extension of scalars. For a commutative ring, the tensor product of modules can be iterated to form the tensor algebra of a module, allowing one to define multiplication in the module in a universal way.

In mathematics, specifically abstract algebra, the opposite of a ring is another ring with the same elements and addition operation, but with the multiplication performed in the reverse order. More explicitly, the opposite of a ring (R, +, ) is the ring (R, +, ∗) whose multiplication ∗ is defined by ab = ba for all a, b in R. The opposite ring can be used to define multimodules, a generalization of bimodules. They also help clarify the relationship between left and right modules (see § Properties).

In mathematical logic, Skolem arithmetic is the first-order theory of the natural numbers with multiplication, named in honor of Thoralf Skolem. The signature of Skolem arithmetic contains only the multiplication operation and equality, omitting the addition operation entirely.

<span class="mw-page-title-main">Classical group</span>

In mathematics, the classical groups are defined as the special linear groups over the reals R, the complex numbers C and the quaternions H together with special automorphism groups of symmetric or skew-symmetric bilinear forms and Hermitian or skew-Hermitian sesquilinear forms defined on real, complex and quaternionic finite-dimensional vector spaces. Of these, the complex classical Lie groups are four infinite families of Lie groups that together with the exceptional groups exhaust the classification of simple Lie groups. The compact classical groups are compact real forms of the complex classical groups. The finite analogues of the classical groups are the classical groups of Lie type. The term "classical group" was coined by Hermann Weyl, it being the title of his 1939 monograph The Classical Groups.

<span class="mw-page-title-main">Hadamard product (matrices)</span> Matrix operation

In mathematics, the Hadamard product is a binary operation that takes in two matrices of the same dimensions and returns a matrix of the multiplied corresponding elements. This operation can be thought as a "naive matrix multiplication" and is different from the matrix product. It is attributed to, and named after, either French mathematician Jacques Hadamard or German mathematician Issai Schur.

References

  1. Hungerford, Thomas W. (1974). Algebra (1st ed.). Springer. p. 24. ISBN   978-0387905181. Definition 1.1 (i) a(bc) = (ab)c for all a, b, c in G.
  2. Durbin, John R. (1992). Modern Algebra: an Introduction (3rd ed.). New York: Wiley. p. 78. ISBN   978-0-471-51001-7. If are elements of a set with an associative operation, then the product is unambiguous; this is, the same element will be obtained regardless of how parentheses are inserted in the product.
  3. "Matrix product associativity". Khan Academy. Retrieved 5 June 2016.
  4. Moore, Brooke Noel; Parker, Richard (2017). Critical Thinking (12th ed.). New York: McGraw-Hill Education. p. 321. ISBN   9781259690877.
  5. Copi, Irving M.; Cohen, Carl; McMahon, Kenneth (2014). Introduction to Logic (14th ed.). Essex: Pearson Education. p. 387. ISBN   9781292024820.
  6. Hurley, Patrick J.; Watson, Lori (2016). A Concise Introduction to Logic (13th ed.). Boston: Cengage Learning. p. 427. ISBN   9781305958098.
  7. Knuth, Donald, The Art of Computer Programming, Volume 3, section 4.2.2
  8. IEEE Computer Society (29 August 2008). IEEE Standard for Floating-Point Arithmetic. doi:10.1109/IEEESTD.2008.4610935. ISBN   978-0-7381-5753-5. IEEE Std 754-2008.
  9. Villa, Oreste; Chavarría-mir, Daniel; Gurumoorthi, Vidhya; Márquez, Andrés; Krishnamoorthy, Sriram, Effects of Floating-Point non-Associativity on Numerical Computations on Massively Multithreaded Systems (PDF), archived from the original (PDF) on 15 February 2013, retrieved 8 April 2014
  10. Goldberg, David (March 1991). "What Every Computer Scientist Should Know About Floating-Point Arithmetic" (PDF). ACM Computing Surveys . 23 (1): 5–48. doi:10.1145/103162.103163. S2CID   222008826. Archived (PDF) from the original on 2022-05-19. Retrieved 20 January 2016.
  11. George Mark Bergman "Order of arithmetic operations"
  12. "The Order of Operations". Education Place.
  13. "The Order of Operations", timestamp 5m40s. Khan Academy.
  14. "Using Order of Operations and Exploring Properties" Archived 2022-07-16 at the Wayback Machine , section 9. Virginia Department of Education.
  15. Bronstein, de:Taschenbuch der Mathematik , pages 115-120, chapter: 2.4.1.1, ISBN   978-3-8085-5673-3
  16. Exponentiation Associativity and Standard Math Notation Codeplea. 23 August 2016. Retrieved 20 September 2016.
  17. Hamilton, W.R. (1844–1850). "On quaternions or a new system of imaginaries in algebra". David R. Wilkins collection. Philosophical Magazine . Trinity College Dublin.
  18. Baez, John C. (2002). "The Octonions" (PDF). Bulletin of the American Mathematical Society. 39 (2): 145–205. arXiv: math/0105155 . doi:10.1090/S0273-0979-01-00934-X. ISSN   0273-0979. MR   1886087. S2CID   586512.