Ergodicity economics is a research programme aimed at reworking the theoretical foundations of economics around the concept of ergodicity. [1] The programme's main goal is to understand how traditional economic theory, framed in terms of the expectation values , changes when replacing expectation value with time averages. In particular, the programme is interested in understanding the effect of non-ergodic processes in economics, that is processes where the expectation value of an observable does not equal its time average.
Ergodicity economics questions whether expected value is a useful indicator of an economic observable's behaviour over time. In doing so it builds on existing critiques of the use of expected value in the modeling of economic decisions. Such critiques started soon after the introduction of expected value in 1654. For instance, expected-utility theory was proposed in 1738 by Daniel Bernoulli [2] as a way of modeling behavior which is inconsistent with expected-value maximization. In 1956, John Kelly devised the Kelly criterion by optimizing the use of available information, and Leo Breiman later noted that this is equivalent to optimizing time-average performance, as opposed to expected value. [3]
The ergodicity economics research programme originates in two papers by Ole Peters in 2011, a theoretical physicist and current external professor at the Santa Fe Institute. [4] The first studied the problem of optimal leverage in finance and how this may be achieved by considering the non-ergodic properties of geometric brownian motion. [5] The second paper applied principles of non-ergodicity to propose a possible solution for the St. Petersburg paradox. [6] More recent work has suggested possible solutions for the equity premium puzzle, the insurance puzzle, gamble-selection, probability weighting, and has provided insights into the dynamics of income inequality. [7]
In mathematics and physics, the concept of ergodicity is used to characterise dynamical systems and stochastic processes. A system is said to be ergodic, if a point of a moving system will eventually visit all parts of the space that the system moves in, in a uniform and random sense. Ergodicity implies that the average behaviour along a single trajectory through time (time average) is equivalent to the average behaviour of a large ensemble at one point in time (ensemble average). For an infinitely large ensemble, the ensemble average of an observable is equivalent to the expected value.
Ergodicity economics inherits from these ideas the probing of the ergodic properties of stochastic processes used as economic models. Early economic theory was developed at a time when the expected value had been invented but its relation to the time average was unclear. No clear distinction was made between the two mathematical objects, which amounts to an implicit assumption of ergodicity.
Ergodicity economics explores what aspects of economics can be informed by avoiding this implicit assumption.
Mean values and expected values are used extensively in economic theory, most commonly as a summary statistic. One common critique of this practice is the sensitivity of mean values to outliers. Ergodicity economics focuses on a different critique and emphasizes the physical meaning of expected values as averages across a statistical ensemble of parallel systems. It insists on a physical justification when expected values are used. In essence, at least one of two conditions must hold:
In ergodicity economics, expected values are replaced, where necessary, by averages that account for the ergodicity or non-ergodicity of the observables involved.
Ergodicity economics emphasizes what happens to an agent's wealth over time . From this follows a possible decision theory where agents maximize the time-average growth rate of wealth. [8] [9] The functional form of the growth rate, , depends on the wealth process . In general, a growth rate takes the form , where the function , linearizes , such that growth rates evaluated at different times can be meaningfully compared.
Growth processes generally violate ergodicity, but their growth rates may nonetheless be ergodic. In this case, the time-average growth rate, can be computed as the rate of change of the expected value of , i.e.
In this context, is called the ergodicity transformation.
An influential class of models for economic decision-making is known as expected utility theory. The following specific model can be mapped to the growth-rate optimization highlighted by ergodicity economics. Here, agents evaluate monetary wealth according to a utility function , and it is postulated that decisions maximize the expected value of the change in utility,
This model was proposed as an improvement of expected-value maximization, where agents maximize . A non-linear utility function allows the encoding of behavioral patterns not represented in expected-value maximization. Specifically, expected-utility maximizing agents can have idiosyncratic risk preferences. An agent specified by a convex utility function is more risk-seeking than an expected wealth maximizer, and a concave utility function implies greater risk aversion.
Comparing (2) to (1), we can identify the utility function with the linearization , and make the two expressions identical by dividing (2) by . Division by simply implements a preference for faster utility growth in the expected-utility-theory decision protocol.
This mapping shows that the two models will yield identical predictions if the utility function applied under expected-utility theory is the same as the ergodicity transformation, needed to compute an ergodic growth rate.
Ergodicity economics thus emphasizes the dynamic circumstances under which a decision is made, whereas expected-utility theory emphasizes idiosyncratic preferences to explain behavior. Different ergodicity transformations indicate different types of wealth dynamics, whereas different utility functions indicate different personal preferences. The mapping highlights the relationship between the two approaches, showing that differences in personal preferences can arise purely as a result of different dynamic contexts of decision makers.
A simple example for an agent's wealth process, , is geometric Brownian motion (GBM), commonly used in mathematical finance and other fields. is said to follow GBM if it satisfies the stochastic differential equation
where is the increment in a Wiener process, and ('drift') and ('volatility') are constants. Solving (3) gives
In this case the ergodicity transformation is , as is easily verified: grows linearly in time.
Following the recipe laid out above, this leads to the time-average growth rate
It follows that for geometric Brownian motion, maximizing the rate of change in the logarithmic utility function, , is equivalent to maximizing the time-average growth rate of wealth, i.e. what happens to the agent's wealth over time.
Stochastic processes other than (3) possess different ergodicity transformations, where growth-optimal agents maximize the expected value of utility functions other than the logarithm. Trivially, replacing (3) with additive dynamics implies a linear ergodicity transformation, and many similar pairs of dynamics and transformations can be derived.
A popular illustration of non-ergodicity in economic processes is a repeated multiplicative coin toss, an instance of the binomial multiplicative process. [10] It demonstrates how an expected-value analysis can indicate that a gamble is favorable although the gambler is guaranteed to lose over time.
In this thought experiment, discussed in, [7] a person participates in a simple game where they toss a fair coin. If the coin lands heads, the person gains 50% on their current wealth; if it lands tails, the person loses 40%.
The game shows the difference between the expected value of an investment, or bet, and the time-average or real-world outcome of repeatedly engaging in that bet over time.
Denoting current wealth by , and the time when the payout is received by , we find that wealth after one round is given by the random variable , which takes the values (for heads) and (for tails), each with probability . The expected value of the gambler's wealth after one round is therefore
By induction, after rounds expected wealth is , increasing exponentially at 5% per round in the game.
This calculation shows that the game is favorable in expectation—its expected value increases with each round played.
The time-average performance indicates what happens to the wealth of a single gambler who plays repeatedly, reinvesting their entire wealth every round. Due to compounding, after rounds the wealth will be
where we have written to denote the realized random factor by which wealth is multiplied in the round of the game (either for heads; or , for tails). Averaged over time, wealth has grown per round by a factor
Introducing the notation for the number of heads in a sequence of coin tosses we re-write this as
For any finite , the time-average per-round growth factor, , is a random variable. The long-time limit, found by letting the number of rounds diverge , provides a characteristic scalar which can be compared with the per-round growth factor of the expected value. The proportion of heads tossed then converges to the probability of heads (namely 1/2), and the time-average growth factor is
The comparison between expected value and time-average performance illustrates an effect of broken ergodicity: over time, with probability one, wealth decreases by about 5% per round, in contrast to the increase by 5% per round of the expected value.
This section gives self-sourcing popular culture examples .(November 2023) |
In December 2020, Bloomberg news published an article titled "Everything We’ve Learned About Modern Economic Theory Is Wrong" [11] discussing the implications of ergodicity in economics following the publication of a review of the subject in Nature Physics. [7] Morningstar covered the story to discuss the investment case for stock diversification. [12]
In the book Skin in the Game , Nassim Nicholas Taleb suggests that the ergodicity problem requires a rethinking of how economists use probabilities. [13] A summary of the arguments was published by Taleb in a Medium article in August 2017. [14]
In the book The End of Theory , Richard Bookstaber lists non-ergodicity as one of four characteristics of our economy that are part of financial crises, that conventional economics fails to adequately account for, and that any model of such crises needs to take adequate account of. [15] The other three are: computational irreducibility, emergent phenomena, and radical uncertainty.[ citation needed ]
In the book The Ergodic Investor and Entrepreneur , Boyd and Reardon tackle the practical implications of non-ergodic capital growth for investors and entrepreneurs, especially for those with a sustainability, circular economy, net positive, or regenerative focus. [16]
James White and Victor Haghani discuss the field of ergodicity economics in their book The Missing Billionaires . [17]
It has been claimed that expected utility theory implicitly assumes ergodicity in the sense that it optimizes an expected value which is only relevant to the long-term benefit of the decision-maker if the relevant observable is ergodic. [7] Doctor, Wakker, and Tang argue that this is wrong because such assumptions are “outside the scope of expected utility theory as a static theory”. [18] They further argue that ergodicity economics overemphasizes the importance of long-term growth as “the primary factor that explains economic phenomena,” and downplays the importance of individual preferences. They also caution against optimizing long-term growth inappropriately. An example is given of a short-term decision between A) a great loss incurred with certainty and B) a gain enjoyed with almost-certainty paired with an even greater loss at negligible probability. In the example the long-term growth rate favors the certain loss and seems an inappropriate criterion for the short-term decision horizon. Finally, an experiment by Meder and colleagues claims to find that individual risk preferences change with dynamical conditions in ways predicted by ergodicity economics. [19] Doctor, Wakker, and Tang criticize the experiment for being confounded by differences in ambiguity and the complexity of probability calculations. Further, they criticize the analysis for applying static expected utility theory models to a context where dynamic versions are more appropriate. In support of this, Goldstein claims to show that multi-period EUT predicts a similar change in risk preferences as observed in the experiment. [20]
In information theory, the entropy of a random variable quantifies the average level of uncertainty or information associated with the variable's potential states or possible outcomes. This measures the expected amount of information needed to describe the state of the variable, considering the distribution of probabilities across all potential states. Given a discrete random variable , which takes values in the set and is distributed according to , the entropy is where denotes the sum over the variable's possible values. The choice of base for , the logarithm, varies for different applications. Base 2 gives the unit of bits, while base e gives "natural units" nat, and base 10 gives units of "dits", "bans", or "hartleys". An equivalent definition of entropy is the expected value of the self-information of a variable.
In game theory, the Nash equilibrium is the most commonly-used solution concept for non-cooperative games. A Nash equilibrium is a situation where no player could gain by changing their own strategy. The idea of Nash equilibrium dates back to the time of Cournot, who in 1838 applied it to his model of competition in an oligopoly.
In physical chemistry, the Arrhenius equation is a formula for the temperature dependence of reaction rates. The equation was proposed by Svante Arrhenius in 1889, based on the work of Dutch chemist Jacobus Henricus van 't Hoff who had noted in 1884 that the van 't Hoff equation for the temperature dependence of equilibrium constants suggests such a formula for the rates of both forward and reverse reactions. This equation has a vast and important application in determining the rate of chemical reactions and for calculation of energy of activation. Arrhenius provided a physical justification and interpretation for the formula. Currently, it is best seen as an empirical relationship. It can be used to model the temperature variation of diffusion coefficients, population of crystal vacancies, creep rates, and many other thermally induced processes and reactions. The Eyring equation, developed in 1935, also expresses the relationship between rate and energy.
The Pareto distribution, named after the Italian civil engineer, economist, and sociologist Vilfredo Pareto, is a power-law probability distribution that is used in description of social, quality control, scientific, geophysical, actuarial, and many other types of observable phenomena; the principle originally applied to describing the distribution of wealth in a society, fitting the trend that a large portion of wealth is held by a small fraction of the population. The Pareto principle or "80-20 rule" stating that 80% of outcomes are due to 20% of causes was named in honour of Pareto, but the concepts are distinct, and only Pareto distributions with shape value of log45 ≈ 1.16 precisely reflect it. Empirical observation has shown that this 80-20 distribution fits a wide range of cases, including natural phenomena and human activities.
In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data. This is achieved by maximizing a likelihood function so that, under the assumed statistical model, the observed data is most probable. The point in the parameter space that maximizes the likelihood function is called the maximum likelihood estimate. The logic of maximum likelihood is both intuitive and flexible, and as such the method has become a dominant means of statistical inference.
In mathematical finance, the Greeks are the quantities representing the sensitivity of the price of a derivative instrument such as an option to changes in one or more underlying parameters on which the value of an instrument or portfolio of financial instruments is dependent. The name is used because the most common of these sensitivities are denoted by Greek letters. Collectively these have also been called the risk sensitivities, risk measures or hedge parameters.
The Lotka–Volterra equations, also known as the Lotka–Volterra predator–prey model, are a pair of first-order nonlinear differential equations, frequently used to describe the dynamics of biological systems in which two species interact, one as a predator and the other as prey. The populations change through time according to the pair of equations:
In microeconomics, a consumer's Marshallian demand function is the quantity they demand of a particular good as a function of its price, their income, and the prices of other goods, a more technical exposition of the standard demand function. It is a solution to the utility maximization problem of how the consumer can maximize their utility for given income and prices. A synonymous term is uncompensated demand function, because when the price rises the consumer is not compensated with higher nominal income for the fall in their real income, unlike in the Hicksian demand function. Thus the change in quantity demanded is a combination of a substitution effect and a wealth effect. Although Marshallian demand is in the context of partial equilibrium theory, it is sometimes called Walrasian demand as used in general equilibrium theory.
In mathematical statistics, the Kullback–Leibler (KL) divergence, denoted , is a type of statistical distance: a measure of how one reference probability distribution P is different from a second probability distribution Q. Mathematically, it is defined as
The equilibrium constant of a chemical reaction is the value of its reaction quotient at chemical equilibrium, a state approached by a dynamic chemical system after sufficient time has elapsed at which its composition has no measurable tendency towards further change. For a given set of reaction conditions, the equilibrium constant is independent of the initial analytical concentrations of the reactant and product species in the mixture. Thus, given the initial composition of a system, known equilibrium constant values can be used to determine the composition of the system at equilibrium. However, reaction parameters like temperature, solvent, and ionic strength may all influence the value of the equilibrium constant.
A satellite galaxy is a smaller companion galaxy that travels on bound orbits within the gravitational potential of a more massive and luminous host galaxy. Satellite galaxies and their constituents are bound to their host galaxy, in the same way that planets within the Solar System are gravitationally bound to the Sun. While most satellite galaxies are dwarf galaxies, satellite galaxies of large galaxy clusters can be much more massive. The Milky Way is orbited by about fifty satellite galaxies, the largest of which is the Large Magellanic Cloud.
In probability theory, the Kelly criterion is a formula for sizing a sequence of bets by maximizing the long-term expected value of the logarithm of wealth, which is equivalent to maximizing the long-term expected geometric growth rate. John Larry Kelly Jr., a researcher at Bell Labs, described the criterion in 1956.
The Ramsey–Cass–Koopmans model, or Ramsey growth model, is a neoclassical model of economic growth based primarily on the work of Frank P. Ramsey, with significant extensions by David Cass and Tjalling Koopmans. The Ramsey–Cass–Koopmans model differs from the Solow–Swan model in that the choice of consumption is explicitly microfounded at a point in time and so endogenizes the savings rate. As a result, unlike in the Solow–Swan model, the saving rate may not be constant along the transition to the long run steady state. Another implication of the model is that the outcome is Pareto optimal or Pareto efficient.
In mathematical finance, the Cox–Ingersoll–Ross (CIR) model describes the evolution of interest rates. It is a type of "one factor model" as it describes interest rate movements as driven by only one source of market risk. The model can be used in the valuation of interest rate derivatives. It was introduced in 1985 by John C. Cox, Jonathan E. Ingersoll and Stephen A. Ross as an extension of the Vasicek model, itself an Ornstein–Uhlenbeck_process.
Expected shortfall (ES) is a risk measure—a concept used in the field of financial risk measurement to evaluate the market risk or credit risk of a portfolio. The "expected shortfall at q% level" is the expected return on the portfolio in the worst of cases. ES is an alternative to value at risk that is more sensitive to the shape of the tail of the loss distribution.
In economics and finance, exponential utility is a specific form of the utility function, used in some contexts because of its convenience when risk is present, in which case expected utility is maximized. Formally, exponential utility is given by:
The Larson–Miller relation, also widely known as the Larson–Miller parameter and often abbreviated LMP, is a parametric relation used to extrapolate experimental data on creep and rupture life of engineering materials.
In mathematical finance, the Black–Scholes equation, also called the Black–Scholes–Merton equation, is a partial differential equation (PDE) governing the price evolution of derivatives under the Black–Scholes model. Broadly speaking, the term may refer to a similar PDE that can be derived for a variety of options, or more generally, derivatives.
The population dynamics of pest insects is a subject of interest to farmers, agricultural economists, ecologists, and those concerned with animal welfare.
Adding controlled noise from predetermined distributions is a way of designing differentially private mechanisms. This technique is useful for designing private mechanisms for real-valued functions on sensitive data. Some commonly used distributions for adding noise include Laplace and Gaussian distributions.
{{cite journal}}
: CS1 maint: multiple names: authors list (link) CS1 maint: numeric names: authors list (link){{cite book}}
: CS1 maint: location missing publisher (link){{cite book}}
: CS1 maint: location missing publisher (link){{cite book}}
: CS1 maint: location missing publisher (link)