DeGroot learning refers to a rule-of-thumb type of social learning process. The idea was stated in its general form by the American statistician Morris H. DeGroot; [1] antecedents were articulated by John R. P. French [2] and Frank Harary. [3] The model has been used in physics, computer science and most widely in the theory of social networks. [4] [5]
Take a society of agents where everybody has an opinion on a subject, represented by a vector of probabilities . Agents obtain no new information based on which they can update their opinions but they communicate with other agents. Links between agents (who knows whom) and the weight they put on each other's opinions is represented by a trust matrix where is the weight that agent puts on agent 's opinion. The trust matrix is thus in a one-to-one relationship with a weighted, directed graph where there is an edge between and if and only if . The trust matrix is stochastic, its rows consists of nonnegative real numbers, with each row summing to 1.
Formally, the beliefs are updated in each period as
so the th period opinions are related to the initial opinions by
An important question is whether beliefs converge to a limit and to each other in the long run. As the trust matrix is stochastic, standard results in Markov chain theory can be used to state conditions under which the limit
exists for any initial beliefs . The following cases are treated in Golub and Jackson [6] (2010).
If the social network graph (represented by the trust matrix) is strongly connected, convergence of beliefs is equivalent to each of the following properties:
The equivalence between the last two is a direct consequence from Perron–Frobenius theorem.
It is not necessary to have a strongly connected social network to have convergent beliefs, however, the equality of limiting beliefs does not hold in general.
We say that a group of agents is closed if for any , only if . Beliefs are convergent if and only if every set of nodes (representing individuals) that is strongly connected and closed is also aperiodic.
A group of individuals is said to reach a consensus if for any . This means that, as a result of the learning process, in the limit they have the same belief on the subject.
With a strongly connected and aperiodic network the whole group reaches a consensus. In general, any strongly connected and closed group of individuals reaches a consensus for every initial vector of beliefs if and only if it is aperiodic. If, for example, there are two groups satisfying these assumptions, they reach a consensus inside the groups but there is not necessarily a consensus at the society level.
Take a strongly connected and aperiodic social network. In this case, the common limiting belief is determined by the initial beliefs through
where is the unique unit length left eigenvector of corresponding to the eigenvalue 1. The vector shows the weights that agents put on each other's initial beliefs in the consensus limit. Thus, the higher is , the more influence individual has on the consensus belief.
The eigenvector property implies that
This means that the influence of is a weighted average of those agents' influence who pay attention to , with weights of their level of trust. Hence influential agents are characterized by being trusted by other individuals with high influence.
These examples appear in Jackson [4] (2008).
Consider a three-individual society with the following trust matrix:
Hence the first person weights the beliefs of the other two with equally, while the second listens only to the first, the third only to the second individual. For this social trust structure, the limit exists and equals
so the influence vector is and the consensus belief is . In words, independently of the initial beliefs, individuals reach a consensus where the initial belief of the first and the second person has twice as high influence than the third one's.
If we change the previous example such that the third person also listens exclusively to the first one, we have the following trust matrix:
In this case for any we have
and
so does not exist and beliefs do not converge in the limit. Intuitively, 1 is updating based on 2 and 3's beliefs while 2 and 3 update solely based on 1's belief so they interchange their beliefs in each period.
It is possible to examine the outcome of the DeGroot learning process in large societies, that is, in the limit.
Let the subject on which people have opinions be a "true state" . Assume that individuals have independent noisy signals of (now superscript refers to time, the argument to the size of the society). Assume that for all the trust matrix is such that the limiting beliefs exists independently from the initial beliefs. Then the sequence of societies is called wise if
where denotes convergence in probability. This means that if the society grows without bound, over time they will have a common and accurate belief on the uncertain subject.
A necessary and sufficient condition for wisdom can be given with the help of influence vectors. A sequence of societies is wise if and only if
that is, the society is wise precisely when even the most influential individual's influence vanishes in the large society limit. For further characterization and examples see Golub and Jackson [6] (2010).
In mathematics, a sequence is an enumerated collection of objects in which repetitions are allowed and order matters. Like a set, it contains members. The number of elements is called the length of the sequence. Unlike a set, the same elements can appear multiple times at different positions in a sequence, and unlike a set, the order does matter. Formally, a sequence can be defined as a function from natural numbers to the elements at each position. The notion of a sequence can be generalized to an indexed family, defined as a function from an arbitrary index set.
In probability theory, the central limit theorem (CLT) establishes that, in many situations, for independent and identically distributed random variables, the sampling distribution of the standardized sample mean tends towards the standard normal distribution even if the original variables themselves are not normally distributed.
A Markov chain or Markov process is a stochastic model describing a sequence of possible events in which the probability of each event depends only on the state attained in the previous event. Informally, this may be thought of as, "What happens next depends only on the state of affairs now." A countably infinite sequence, in which the chain moves state at discrete time steps, gives a discrete-time Markov chain (DTMC). A continuous-time process is called a continuous-time Markov chain (CTMC). It is named after the Russian mathematician Andrey Markov.
In linear algebra, the Cholesky decomposition or Cholesky factorization is a decomposition of a Hermitian, positive-definite matrix into the product of a lower triangular matrix and its conjugate transpose, which is useful for efficient numerical solutions, e.g., Monte Carlo simulations. It was discovered by André-Louis Cholesky for real matrices, and posthumously published in 1924. When it is applicable, the Cholesky decomposition is roughly twice as efficient as the LU decomposition for solving systems of linear equations.
In linear algebra, an n-by-n square matrix A is called invertible, if there exists an n-by-n square matrix B such that
In mathematics, and in particular linear algebra, the Moore–Penrose inverse of a matrix is the most widely known generalization of the inverse matrix. It was independently described by E. H. Moore in 1920, Arne Bjerhammar in 1951, and Roger Penrose in 1955. Earlier, Erik Ivar Fredholm had introduced the concept of a pseudoinverse of integral operators in 1903. When referring to a matrix, the term pseudoinverse, without further specification, is often used to indicate the Moore–Penrose inverse. The term generalized inverse is sometimes used as a synonym for pseudoinverse.
In mathematics, the spectral radius of a square matrix is the maximum of the absolute values of its eigenvalues. More generally, the spectral radius of a bounded linear operator is the supremum of the absolute values of the elements of its spectrum. The spectral radius is often denoted by ρ(·).
In physics, the S-matrix or scattering matrix relates the initial state and the final state of a physical system undergoing a scattering process. It is used in quantum mechanics, scattering theory and quantum field theory (QFT).
In mathematical analysis, Cesàro summation assigns values to some infinite sums that are not necessarily convergent in the usual sense. The Cesàro sum is defined as the limit, as n tends to infinity, of the sequence of arithmetic means of the first n partial sums of the series.
In epidemiology, the basic reproduction number, or basic reproductive number, denoted , of an infection is the expected number of cases directly generated by one case in a population where all individuals are susceptible to infection. The definition assumes that no other individuals are infected or immunized. Some definitions, such as that of the Australian Department of Health, add the absence of "any deliberate intervention in disease transmission". The basic reproduction number is not necessarily the same as the effective reproduction number , which is the number of cases generated in the current state of a population, which does not have to be the uninfected state. is a dimensionless number and not a time rate, which would have units of time−1, or units of time like doubling time.
A continuous-time Markov chain (CTMC) is a continuous stochastic process in which, for each state, the process will change state according to an exponential random variable and then move to a different state as specified by the probabilities of a stochastic matrix. An equivalent formulation describes the process as changing state according to the least value of a set of exponential random variables, one for each possible state it can move to, with the parameters determined by the current state.
In matrix theory, the Perron–Frobenius theorem, proved by Oskar Perron (1907) and Georg Frobenius (1912), asserts that a real square matrix with positive entries has a unique eigenvalue of largest magnitude and that eigenvalue is real. The corresponding eigenvector can be chosen to have strictly positive components, and also asserts a similar statement for certain classes of nonnegative matrices. This theorem has important applications to probability theory ; to the theory of dynamical systems ; to economics ; to demography ; to social networks ; to Internet search engines (PageRank); and even to ranking of football teams. The first to discuss the ordering of players within tournaments using Perron–Frobenius eigenvectors is Edmund Landau.
In mathematics, the Silverman–Toeplitz theorem, first proved by Otto Toeplitz, is a result in summability theory characterizing matrix summability methods that are regular. A regular matrix summability method is a matrix transformation of a convergent sequence which preserves the limit.
A Neumann series is a mathematical series of the form
In mathematics, the square root of a matrix extends the notion of square root from numbers to matrices. A matrix B is said to be a square root of A if the matrix product BB is equal to A.
In mathematics, a logarithm of a matrix is another matrix such that the matrix exponential of the latter matrix equals the original matrix. It is thus a generalization of the scalar logarithm and in some sense an inverse function of the matrix exponential. Not all matrices have a logarithm and those matrices that do have a logarithm may have more than one logarithm. The study of logarithms of matrices leads to Lie theory since when a matrix has a logarithm then it is in an element of a Lie group and the logarithm is the corresponding element of the vector space of the Lie algebra.
In probability, a discrete-time Markov chain (DTMC) is a sequence of random variables, known as a stochastic process, in which the value of the next variable depends only on the value of the current variable, and not any variables in the past. For instance, a machine may have two states, A and E. When it is in state A, there is a 40% chance of it moving to state E and a 60% chance of it remaining in state A. When it is in state E, there is a 70% chance of it moving to A and a 30% chance of it staying in E. The sequence of states of the machine is a Markov chain. If we denote the chain by then is the state which the machine starts in and is the random variable describing its state after 10 transitions. The process continues forever, indexed by the natural numbers.
In numerical linear algebra, the Jacobi eigenvalue algorithm is an iterative method for the calculation of the eigenvalues and eigenvectors of a real symmetric matrix. It is named after Carl Gustav Jacob Jacobi, who first proposed the method in 1846, but only became widely used in the 1950s with the advent of computers.
In the mathematical theory of probability, the entropy rate or source information rate of a stochastic process is, informally, the time density of the average information in a stochastic process. For stochastic processes with a countable index, the entropy rate is the limit of the joint entropy of members of the process divided by , as tends to infinity:
In linear algebra, a convergent matrix is a matrix that converges to the zero matrix under matrix exponentiation.