Stochastic ordering

Last updated

In probability theory and statistics, a stochastic order quantifies the concept of one random variable being "bigger" than another. These are usually partial orders, so that one random variable may be neither stochastically greater than, less than, nor equal to another random variable . Many different orders exist, which have different applications.

Contents

Usual stochastic order

A real random variable is less than a random variable in the "usual stochastic order" if

where denotes the probability of an event. This is sometimes denoted or .

If additionally for some , then is stochastically strictly less than , sometimes denoted . In decision theory, under this circumstance, B is said to be first-order stochastically dominant over A.

Characterizations

The following rules describe situations when one random variable is stochastically less than or equal to another. Strict version of some of these rules also exist.

  1. if and only if for all non-decreasing functions , .
  2. If is non-decreasing and then
  3. If is increasing in each variable and and are independent sets of random variables with for each , then and in particular Moreover, the th order statistics satisfy .
  4. If two sequences of random variables and , with for all each converge in distribution, then their limits satisfy .
  5. If , and are random variables such that and for all and such that , then .

Other properties

If and then (the random variables are equal in distribution).

Stochastic dominance

Stochastic dominance relations are a family of stochastic orderings used in decision theory: [1]

There also exist higher-order notions of stochastic dominance. With the definitions above, we have .

Multivariate stochastic order

An -valued random variable is less than an -valued random variable in the "usual stochastic order" if

Other types of multivariate stochastic orders exist. For instance the upper and lower orthant order which are similar to the usual one-dimensional stochastic order. is said to be smaller than in upper orthant order if

and is smaller than in lower orthant order if [2]

All three order types also have integral representations, that is for a particular order is smaller than if and only if for all in a class of functions . [3] is then called generator of the respective order.

Other dominance orders

The following stochastic orders are useful in the theory of random social choice. They are used to compare the outcomes of random social choice functions, in order to check them for efficiency or other desirable criteria. [4] The dominance orders below are ordered from the most conservative to the least conservative. They are exemplified on random variables over the finite support {30,20,10}.

Deterministic dominance, denoted , means that every possible outcome of is at least as good as every possible outcome of : for all x < y, . In other words: . For example, .

Bilinear dominance, denoted , means that, for every possible outcome, the probability that yields the better one and yields the worse one is at least as large as the probability the other way around: for all x<y, For example, .

Stochastic dominance (already mentioned above), denoted , means that, for every possible outcome x, the probability that yields at least x is at least as large as the probability that yields at least x: for all x, . For example, .

Pairwise-comparison dominance, denoted , means that the probability that that yields a better outcome than is larger than the other way around: . For example, .

Downward-lexicographic dominance, denoted , means that has a larger probability than of returning the best outcome, or both and have the same probability to return the best outcome but has a larger probability than of returning the second-best best outcome, etc. Upward-lexicographic dominance is defined analogously based on the probability to return the worst outcomes. See lexicographic dominance.

Other stochastic orders

Hazard rate order

The hazard rate of a non-negative random variable with absolutely continuous distribution function and density function is defined as

Given two non-negative variables and with absolutely continuous distribution and , and with hazard rate functions and , respectively, is said to be smaller than in the hazard rate order (denoted as ) if

for all ,

or equivalently if

is decreasing in .

Likelihood ratio order

Let and two continuous (or discrete) random variables with densities (or discrete densities) and , respectively, so that increases in over the union of the supports of and ; in this case, is smaller than in the likelihood ratio order ().

Variability orders

If two variables have the same mean, they can still be compared by how "spread out" their distributions are. This is captured to a limited extent by the variance, but more fully by a range of stochastic orders.[ citation needed ]

Convex order

Convex order is a special kind of variability order. Under the convex ordering, is less than if and only if for all convex , .

Laplace transform order

Laplace transform order compares both size and variability of two random variables. Similar to convex order, Laplace transform order is established by comparing the expectation of a function of the random variable where the function is from a special class: . This makes the Laplace transform order an integral stochastic order with the generator set given by the function set defined above with a positive real number.

Realizable monotonicity

Considering a family of probability distributions on partially ordered space indexed with (where is another partially ordered space, the concept of complete or realizable monotonicity may be defined. It means, there exists a family of random variables on the same probability space, such that the distribution of is and almost surely whenever . It means the existence of a monotone coupling. [5]

See also

Related Research Articles

<span class="mw-page-title-main">Cumulative distribution function</span> Probability that random variable X is less than or equal to x

In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable , or just distribution function of , evaluated at , is the probability that will take a value less than or equal to .

<span class="mw-page-title-main">Expected value</span> Average value of a random variable

In probability theory, the expected value is a generalization of the weighted average. Informally, the expected value is the mean of the possible values a random variable can take, weighted by the probability of those outcomes. Since it is obtained through arithmetic, the expected value sometimes may not even be included in the sample data set; it is not the value you would "expect" to get in reality.

<span class="mw-page-title-main">Random variable</span> Variable representing a random phenomenon

A random variable is a mathematical formalization of a quantity or object which depends on random events. The term 'random variable' in its mathematical definition refers to neither randomness nor variability but instead is a mathematical function in which

<span class="mw-page-title-main">Inverse transform sampling</span> Basic method for pseudo-random number sampling

Inverse transform sampling is a basic method for pseudo-random number sampling, i.e., for generating sample numbers at random from any probability distribution given its cumulative distribution function.

<span class="mw-page-title-main">Geometric distribution</span> Probability distribution

In probability theory and statistics, the geometric distribution is either one of two discrete probability distributions:

<span class="mw-page-title-main">Wiener process</span> Stochastic process generalizing Brownian motion

In mathematics, the Wiener process is a real-valued continuous-time stochastic process named in honor of American mathematician Norbert Wiener for his investigations on the mathematical properties of the one-dimensional Brownian motion. It is often also called Brownian motion due to its historical connection with the physical process of the same name originally observed by Scottish botanist Robert Brown. It is one of the best known Lévy processes and occurs frequently in pure and applied mathematics, economics, quantitative finance, evolutionary biology, and physics.

In probability theory, Chebyshev's inequality provides an upper bound on the probability of deviation of a random variable from its mean. More specifically, the probability that a random variable deviates from its mean by more than is at most , where is any positive constant and is the standard deviation.

In probability theory, Markov's inequality gives an upper bound on the probability that a non-negative random variable is greater than or equal to some positive constant. Markov's inequality is tight in the sense that for each chosen positive constant, there exists a random variable such that the inequality is in fact an equality.

In probability theory, the conditional expectation, conditional expected value, or conditional mean of a random variable is its expected value evaluated with respect to the conditional probability distribution. If the random variable can take on only a finite number of values, the "conditions" are that the variable can only take on a subset of those values. More formally, in the case when the random variable is defined over a discrete probability space, the "conditions" are a partition of this probability space.

In probability theory, the Vysochanskij–Petunin inequality gives a lower bound for the probability that a random variable with finite variance lies within a certain number of standard deviations of the variable's mean, or equivalently an upper bound for the probability that it lies further away. The sole restrictions on the distribution are that it be unimodal and have finite variance; here unimodal implies that it is a continuous probability distribution except at the mode, which may have a non-zero probability.

In numerical analysis and computational statistics, rejection sampling is a basic technique used to generate observations from a distribution. It is also commonly called the acceptance-rejection method or "accept-reject algorithm" and is a type of exact simulation method. The method works for any distribution in with a density.

A stochastic differential equation (SDE) is a differential equation in which one or more of the terms is a stochastic process, resulting in a solution which is also a stochastic process. SDEs have many applications throughout pure mathematics and are used to model various behaviours of stochastic models such as stock prices, random growth models or physical systems that are subjected to thermal fluctuations.

In statistics and information theory, a maximum entropy probability distribution has entropy that is at least as great as that of all other members of a specified class of probability distributions. According to the principle of maximum entropy, if nothing is known about a distribution except that it belongs to a certain class, then the distribution with the largest entropy should be chosen as the least-informative default. The motivation is twofold: first, maximizing entropy minimizes the amount of prior information built into the distribution; second, many physical systems tend to move towards maximal entropy configurations over time.

Probability theory and statistics have some commonly used conventions, in addition to standard mathematical notation and mathematical symbols.

Stochastic dominance is a partial order between random variables. It is a form of stochastic ordering. The concept arises in decision theory and decision analysis in situations where one gamble can be ranked as superior to another gamble for a broad class of decision-makers. It is based on shared preferences regarding sets of possible outcomes and their associated probabilities. Only limited knowledge of preferences is required for determining dominance. Risk aversion is a factor only in second order stochastic dominance.

In mathematics – specifically, in the theory of stochastic processes – Doob's martingale convergence theorems are a collection of results on the limits of supermartingales, named after the American mathematician Joseph L. Doob. Informally, the martingale convergence theorem typically refers to the result that any supermartingale satisfying a certain boundedness condition must converge. One may think of supermartingales as the random variable analogues of non-increasing sequences; from this perspective, the martingale convergence theorem is a random variable analogue of the monotone convergence theorem, which states that any bounded monotone sequence converges. There are symmetric results for submartingales, which are analogous to non-decreasing sequences.

Financial models with long-tailed distributions and volatility clustering have been introduced to overcome problems with the realism of classical financial models. These classical models of financial time series typically assume homoskedasticity and normality and as such cannot explain stylized phenomena such as skewness, heavy tails, and volatility clustering of the empirical asset returns in finance. In 1963, Benoit Mandelbrot first used the stable distribution to model the empirical distributions which have the skewness and heavy-tail property. Since -stable distributions have infinite -th moments for all , the tempered stable processes have been proposed for overcoming this limitation of the stable distribution.

<span class="mw-page-title-main">Poisson distribution</span> Discrete probability distribution

In probability theory and statistics, the Poisson distribution is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time if these events occur with a known constant mean rate and independently of the time since the last event. It can also be used for the number of events in other types of intervals than time, and in dimension greater than 1.

In probability theory, concentration inequalities provide mathematical bounds on the probability of a random variable deviating from some value. The deviation or other function of the random variable can be thought of as a secondary random variable. The simplest example of the concentration of such a secondary random variable is the CDF of the first random variable which concentrates the probability to unity. If an analytic form of the CDF is available this provides a concentration equality that provides the exact probability of concentration. It is precisely when the CDF is difficult to calculate or even the exact form of the first random variable is unknown that the applicable concentration inequalities provide useful insight.

For certain applications in linear algebra, it is useful to know properties of the probability distribution of the largest eigenvalue of a finite sum of random matrices. Suppose is a finite sequence of random matrices. Analogous to the well-known Chernoff bound for sums of scalars, a bound on the following is sought for a given parameter t:

References

  1. Perrakis, Stylianos (2019). Stochastic Dominance Option Pricing. Palgrave Macmillan, Cham. doi:10.1007/978-3-030-11590-6_1. ISBN   978-3-030-11589-0.
  2. Definition 2.3 in Thibaut Lux, Antonin Papapantoleon: "Improved Fréchet-Hoeffding bounds for d-copulas and applications in model-free finance." Annals of Applied Probability 27, 3633-3671, 2017
  3. Alfred Müller, Dietrich Stoyan: Comparison methods for stochastic models and risks. Wiley, Chichester 2002, ISBN   0-471-49446-1, S. 2.
  4. Felix Brandt (2017-10-26). "Roling the Dice: Recent Results in Probabilistic Social Choice". In Endriss, Ulle (ed.). Trends in Computational Social Choice. Lulu.com. ISBN   978-1-326-91209-3.
  5. Fill, James Allen; Machida, Motoya (2001). "Stochastic Monotonicity and Realizable Monotonicity". The Annals of Probability. 29 (2): 938–978. ISSN   0091-1798.

Bibliography