Ergodicity economics is a research programme that applies the concept of ergodicity to problems in economics and decision-making under uncertainty. [1] The programme's main goal is to understand how traditional economic theory, framed in terms of the expectation values , changes when replacing expectation value with time averages. In particular, the programme is interested in understanding how behaviour is shaped by non-ergodic economic processes, that is processes where the expectation value of an observable does not equal its time average.
Mean values and expected values are used extensively in economic theory, most commonly as a summary statistic, e.g. used for modelling agents’ decisions under uncertainty. Early economic theory was developed at a time when the expected value had been invented but its relation to the time average had not been studied. No clear distinction was made between the two mathematical objects, which can be interpreted as an implicit assumption of ergodicity. Ergodicity economics explores what aspects of economics can be informed by avoiding this implicit assumption.
While one common critique of modelling decisions based on expected values is the sensitivity of the mean to outliers, ergodicity economics focuses on a different critique. It emphasizes the physical meaning of expected values as averages across a statistical ensemble of parallel systems. It insists on a physical justification when expected values are used. In essence, at least one of two conditions must hold:
In ergodicity economics, expected values are replaced, where necessary, by averages that account for the ergodicity or non-ergodicity of the observables involved. Non-ergodicity is closely related to the problems of irreversibility and path dependence [2] that are common themes in economics.
In mathematics and physics, the concept of ergodicity is used to characterise dynamical systems and stochastic processes. A system is said to be ergodic, if a point of a moving system will eventually visit all parts of the space that the system moves in, in a uniform and random sense. Ergodicity implies that the average behaviour along a single trajectory through time (time average) is equivalent to the average behaviour of a large ensemble at one point in time (ensemble average). For an infinitely large ensemble, the ensemble average of an observable is equivalent to the expected value.
Ergodicity economics inherits from these ideas the probing of the ergodic properties of stochastic processes used as economic models.
Ergodicity economics questions whether expected value is a useful indicator of an economic observable's behaviour over time. In doing so it builds on existing critiques of the use of expected value in the modeling of economic decisions. Such critiques started soon after the introduction of expected value in 1654. For instance, expected-utility theory was proposed in 1738 by Daniel Bernoulli [3] as a way of modeling behavior which is inconsistent with expected-value maximization. In 1956, John Kelly devised the Kelly criterion by optimizing the use of available information, and Leo Breiman later noted that this is equivalent to optimizing time-average performance, as opposed to expected value. [4]
The ergodicity economics research programme originates in two papers by Ole Peters in 2011, a theoretical physicist and current external professor at the Santa Fe Institute. [5] The first studied the problem of optimal leverage in finance and how this may be achieved by considering the non-ergodic properties of geometric brownian motion. [6] The second paper applied principles of non-ergodicity to propose a possible solution for the St. Petersburg paradox. [7] More recent work has suggested possible solutions for the equity premium puzzle, the insurance puzzle, gamble-selection, probability weighting, and has provided insights into the dynamics of income inequality. [8]
Ergodicity economics emphasizes what happens to an agent's wealth over time . From this follows a possible decision theory where agents maximize the time-average growth rate of wealth. [9] [10] The functional form of the growth rate, , depends on the wealth process . In general, a growth rate takes the form , where the function , linearizes , such that growth rates evaluated at different times can be meaningfully compared.
Growth processes generally violate ergodicity, but their growth rates may nonetheless be ergodic. In this case, the time-average growth rate, can be computed as the rate of change of the expected value of , i.e.
In this context, is called the ergodicity transformation.
An influential class of models for economic decision-making is known as expected utility theory. The following specific model can be mapped to the growth-rate optimization highlighted by ergodicity economics. Here, agents evaluate monetary wealth according to a utility function , and it is postulated that decisions maximize the expected value of the change in utility,
This model was proposed as an improvement of expected-value maximization, where agents maximize . A non-linear utility function allows the encoding of behavioral patterns not represented in expected-value maximization. Specifically, expected-utility maximizing agents can have idiosyncratic risk preferences. An agent specified by a convex utility function is more risk-seeking than an expected wealth maximizer, and a concave utility function implies greater risk aversion.
Comparing (2) to (1), we can identify the utility function with the linearization , and make the two expressions identical by dividing (2) by . Division by simply implements a preference for faster utility growth in the expected-utility-theory decision protocol.
This mapping shows that the two models will yield identical predictions if the utility function applied under expected-utility theory is the same as the ergodicity transformation, needed to compute an ergodic growth rate.
Ergodicity economics thus emphasizes the dynamic circumstances under which a decision is made, whereas expected-utility theory emphasizes idiosyncratic preferences to explain behavior. Different ergodicity transformations indicate different types of wealth dynamics, whereas different utility functions indicate different personal preferences. The mapping highlights the relationship between the two approaches, showing that differences in personal preferences can arise purely as a result of different dynamic contexts of decision makers.
A simple example for an agent's wealth process, , is geometric Brownian motion (GBM), commonly used in mathematical finance and other fields. is said to follow GBM if it satisfies the stochastic differential equation
where is the increment in a Wiener process, and ('drift') and ('volatility') are constants. Solving (3) gives
In this case the ergodicity transformation is , as is easily verified: grows linearly in time.
Following the recipe laid out above, this leads to the time-average growth rate
It follows that for geometric Brownian motion, maximizing the rate of change in the logarithmic utility function, , is equivalent to maximizing the time-average growth rate of wealth, i.e. what happens to the agent's wealth over time.
Stochastic processes other than (3) possess different ergodicity transformations, where growth-optimal agents maximize the expected value of utility functions other than the logarithm. Trivially, replacing (3) with additive dynamics implies a linear ergodicity transformation, and many similar pairs of dynamics and transformations can be derived.
A popular illustration of non-ergodicity in economic processes is a repeated multiplicative coin toss, an instance of the binomial multiplicative process. [11] It demonstrates how an expected-value analysis can indicate that a gamble is favorable although the gambler is guaranteed to lose over time.
In this thought experiment, discussed in, [8] a person participates in a simple game where they toss a fair coin. If the coin lands heads, the person gains 50% on their current wealth; if it lands tails, the person loses 40%.
The game shows the difference between the expected value of an investment, or bet, and the time-average or real-world outcome of repeatedly engaging in that bet over time.
Denoting current wealth by , and the time when the payout is received by , we find that wealth after one round is given by the random variable , which takes the values (for heads) and (for tails), each with probability . The expected value of the gambler's wealth after one round is therefore
By induction, after rounds expected wealth is , increasing exponentially at 5% per round in the game.
This calculation shows that the game is favorable in expectation—its expected value increases with each round played.
The time-average performance indicates what happens to the wealth of a single gambler who plays repeatedly, reinvesting their entire wealth every round. Due to compounding, after rounds the wealth will be
where we have written to denote the realized random factor by which wealth is multiplied in the round of the game (either for heads; or , for tails). Averaged over time, wealth has grown per round by a factor
Introducing the notation for the number of heads in a sequence of coin tosses we re-write this as
For any finite , the time-average per-round growth factor, , is a random variable. The long-time limit, found by letting the number of rounds diverge , provides a characteristic scalar which can be compared with the per-round growth factor of the expected value. The proportion of heads tossed then converges to the probability of heads (namely 1/2), and the time-average growth factor is
The comparison between expected value and time-average performance illustrates an effect of broken ergodicity: over time, with probability one, wealth decreases by about 5% per round, in contrast to the increase by 5% per round of the expected value.
To explain the danger of betting in a non-stationary system, a simple game is used. We have two people sitting opposite each other separated by a black cloth, so that they cannot see each other. They are playing the following game: the person we will call A tosses a coin and the person we will call B tries to guess the state in which the coin is on the table. This game lasts an arbitrary interval of time and person A is free to choose how many tosses to make during the chosen interval of time, person B does not see the toss of the coin but can at any time, within the interval of time, make a bet. When he makes a bet, if he guesses the state in which the coin is at that moment, he wins. The game begins, A tosses only once (result: heads), while B bets twice on heads, winning both times.
Question: What is the overall probability of the outcome?
• B calculates a probability of 25% (0.5 × 0.5), considering bets as independent.
• A calculates a probability of 50% since there was only one toss and not two separate events.
The difference arises from the estimate of the conditional probability:
• B estimates the conditional probability in this way P(E2 | E1) = P(E2) treating the events (bets) as completely independent.
• A estimates the conditional probability in this other way P(E2 | E1) = 1 treating the events as completely dependent.
E1=first bet
E2=second bet
The correct answer depends on information: only the person tossing the coin (A) knows the number of tosses and can correctly estimate the probability.
Application to financial markets
The game highlights how traders (player B) often treat their trades as independent, ignoring the non-ergodic structure of financial markets (player A). Markets are not ergodic because sequences of events cannot be simply represented by long-term statistical averages. In other words, returns do not follow independent and identically distributed (i.i.d.) processes, and historical conditions profoundly influence future outcomes [12] . This error leads to an overestimation of predictive capabilities and excessive risk-taking [13] .
![]() | This section gives self-sourcing popular culture examples .(November 2023) |
In December 2020, Bloomberg news published an article titled "Everything We’ve Learned About Modern Economic Theory Is Wrong" [14] discussing the implications of ergodicity in economics following the publication of a review of the subject in Nature Physics. [8] Morningstar covered the story to discuss the investment case for stock diversification. [15]
In the book Skin in the Game , Nassim Nicholas Taleb suggests that the ergodicity problem requires a rethinking of how economists use probabilities. [16] A summary of the arguments was published by Taleb in a Medium article in August 2017. [17]
In the book The End of Theory , Richard Bookstaber lists non-ergodicity as one of four characteristics of our economy that are part of financial crises, that conventional economics fails to adequately account for, and that any model of such crises needs to take adequate account of. [18] The other three are: computational irreducibility, emergent phenomena, and radical uncertainty.[ citation needed ]
In the book The Ergodic Investor and Entrepreneur , Boyd and Reardon tackle the practical implications of non-ergodic capital growth for investors and entrepreneurs, especially for those with a sustainability, circular economy, net positive, or regenerative focus. [19]
James White and Victor Haghani discuss the field of ergodicity economics in their book The Missing Billionaires . [20]
It has been claimed that expected utility theory implicitly assumes ergodicity in the sense that it optimizes an expected value which is only relevant to the long-term benefit of the decision-maker if the relevant observable is ergodic. [8] Doctor, Wakker, and Tang argue that this is wrong because such assumptions are “outside the scope of expected utility theory as a static theory”. [21] They further argue that ergodicity economics overemphasizes the importance of long-term growth as “the primary factor that explains economic phenomena,” and downplays the importance of individual preferences. They also caution against optimizing long-term growth inappropriately.
Doctor, Wakker, and Tang gives the example of a short-term decision between A) a great loss incurred with certainty and B) a gain enjoyed with almost-certainty paired with an even greater loss at negligible probability. In the example the long-term growth rate favors the certain loss and seems an inappropriate criterion for the short-term decision horizon.
Finally, an experiment by Meder and colleagues claims to find that individual risk preferences change with dynamical conditions in ways predicted by ergodicity economics. [22] Doctor, Wakker, and Tang criticize the experiment for being confounded by differences in ambiguity and the complexity of probability calculations. Further, they criticize the analysis for applying static expected utility theory models to a context where dynamic versions are more appropriate. In support of this, Goldstein claims to show that multi-period EUT predicts a similar change in risk preferences as observed in the experiment. [23]
In information theory, the entropy of a random variable quantifies the average level of uncertainty or information associated with the variable's potential states or possible outcomes. This measures the expected amount of information needed to describe the state of the variable, considering the distribution of probabilities across all potential states. Given a discrete random variable , which takes values in the set and is distributed according to , the entropy is where denotes the sum over the variable's possible values. The choice of base for , the logarithm, varies for different applications. Base 2 gives the unit of bits, while base e gives "natural units" nat, and base 10 gives units of "dits", "bans", or "hartleys". An equivalent definition of entropy is the expected value of the self-information of a variable.
In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of possible outcomes for an experiment. It is a mathematical description of a random phenomenon in terms of its sample space and the probabilities of events.
In game theory, the Nash equilibrium is the most commonly-used solution concept for non-cooperative games. A Nash equilibrium is a situation where no player could gain by changing their own strategy. The idea of Nash equilibrium dates back to the time of Cournot, who in 1838 applied it to his model of competition in an oligopoly.
In economics, utility is a measure of a certain person's satisfaction from a certain state of the world. Over time, the term has been used with at least two meanings.
In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data. This is achieved by maximizing a likelihood function so that, under the assumed statistical model, the observed data is most probable. The point in the parameter space that maximizes the likelihood function is called the maximum likelihood estimate. The logic of maximum likelihood is both intuitive and flexible, and as such the method has become a dominant means of statistical inference.
In physics, a partition function describes the statistical properties of a system in thermodynamic equilibrium. Partition functions are functions of the thermodynamic state variables, such as the temperature and volume. Most of the aggregate thermodynamic variables of the system, such as the total energy, free energy, entropy, and pressure, can be expressed in terms of the partition function or its derivatives. The partition function is dimensionless.
In mathematical optimization and decision theory, a loss function or cost function is a function that maps an event or values of one or more variables onto a real number intuitively representing some "cost" associated with the event. An optimization problem seeks to minimize a loss function. An objective function is either a loss function or its opposite, in which case it is to be maximized. The loss function could include terms from several levels of the hierarchy.
The St. Petersburg paradox or St. Petersburg lottery is a paradox involving the game of flipping a coin where the expected payoff of the lottery game is infinite but nevertheless seems to be worth only a very small amount to the participants. The St. Petersburg paradox is a situation where a naïve decision criterion that takes only the expected value into account predicts a course of action that presumably no actual person would be willing to take. Several resolutions to the paradox have been proposed, including the impossible amount of money a casino would need to continue the game indefinitely.
The expected utility hypothesis is a foundational assumption in mathematical economics concerning decision making under uncertainty. It postulates that rational agents maximize utility, meaning the subjective desirability of their actions. Rational choice theory, a cornerstone of microeconomics, builds this postulate to model aggregate social behaviour.
In mathematical statistics, the Kullback–Leibler (KL) divergence, denoted , is a type of statistical distance: a measure of how much a model probability distribution Q is different from a true probability distribution P. Mathematically, it is defined as
A Bellman equation, named after Richard E. Bellman, is a necessary condition for optimality associated with the mathematical optimization method known as dynamic programming. It writes the "value" of a decision problem at a certain point in time in terms of the payoff from some initial choices and the "value" of the remaining decision problem that results from those initial choices. This breaks a dynamic optimization problem into a sequence of simpler subproblems, as Bellman's “principle of optimality" prescribes. The equation applies to algebraic structures with a total ordering; for algebraic structures with a partial ordering, the generic Bellman's equation can be used.
Mode choice analysis is the third step in the conventional four-step transportation forecasting model of transportation planning, following trip distribution and preceding route assignment. From origin-destination table inputs provided by trip distribution, mode choice analysis allows the modeler to determine probabilities that travelers will use a certain mode of transport. These probabilities are called the modal share, and can be used to produce an estimate of the amount of trips taken using each feasible mode.
The birth–death process is a special case of continuous-time Markov process where the state transitions are of only two types: "births", which increase the state variable by one and "deaths", which decrease the state by one. It was introduced by William Feller. The model's name comes from a common application, the use of such models to represent the current size of a population where the transitions are literal births and deaths. Birth–death processes have many applications in demography, queueing theory, performance engineering, epidemiology, biology and other areas. They may be used, for example, to study the evolution of bacteria, the number of people with a disease within a population, or the number of customers in line at the supermarket.
Flory–Huggins solution theory is a lattice model of the thermodynamics of polymer solutions which takes account of the great dissimilarity in molecular sizes in adapting the usual expression for the entropy of mixing. The result is an equation for the Gibbs free energy change for mixing a polymer with a solvent. Although it makes simplifying assumptions, it generates useful results for interpreting experiments.
In probability theory, the Kelly criterion is a formula for sizing a sequence of bets by maximizing the long-term expected value of the logarithm of wealth, which is equivalent to maximizing the long-term expected geometric growth rate. John Larry Kelly Jr., a researcher at Bell Labs, described the criterion in 1956.
In mathematical economics, an isoelastic function, sometimes constant elasticity function, is a function that exhibits a constant elasticity, i.e. has a constant elasticity coefficient. The elasticity is the ratio of the percentage change in the dependent variable to the percentage causative change in the independent variable, in the limit as the changes approach zero in magnitude.
In the mathematical theory of probability, the drift-plus-penalty method is used for optimization of queueing networks and other stochastic systems.
Intertemporal portfolio choice is the process of allocating one's investable wealth to various assets, especially financial assets, repeatedly over time, in such a way as to optimize some criterion. The set of asset proportions at any time defines a portfolio. Since the returns on almost all assets are not fully predictable, the criterion has to take financial risk into account. Typically the criterion is the expected value of some concave function of the value of the portfolio after a certain number of time periods—that is, the expected utility of final wealth. Alternatively, it may be a function of the various levels of goods and services consumption that are attained by withdrawing some funds from the portfolio after each time period.
A Prior-independent mechanism (PIM) is a mechanism in which the designer knows that the agents' valuations are drawn from some probability distribution, but does not know the distribution.
Adding controlled noise from predetermined distributions is a way of designing differentially private mechanisms. This technique is useful for designing private mechanisms for real-valued functions on sensitive data. Some commonly used distributions for adding noise include Laplace and Gaussian distributions.
{{cite journal}}
: CS1 maint: multiple names: authors list (link) CS1 maint: numeric names: authors list (link){{cite book}}
: CS1 maint: location missing publisher (link){{cite book}}
: CS1 maint: location missing publisher (link){{cite book}}
: CS1 maint: location missing publisher (link)