Prophet inequality

Last updated

In the theory of online algorithms and optimal stopping, a prophet inequality is a bound on the expected value of a decision-making process that handles a sequence of random inputs from known probability distributions, relative to the expected value that could be achieved by a "prophet" who knows all the inputs (and not just their distributions) ahead of time. [1] [2] These inequalities have applications in the theory of algorithmic mechanism design and mathematical finance. [3]

Contents

Single item

The classical single-item prophet inequality was published by Krengel & Sucheston (1978), crediting its tight form to D. J. H. (Ben) Garling. It concerns a process in which a sequence of random variables arrive from known distributions . When each arrives, the decision-making process must decide whether to accept it and stop the process, or whether to reject it and go on to the next variable in the sequence. The value of the process is the single accepted variable, if there is one, or zero otherwise. It may be assumed that all variables are non-negative; otherwise, replacing negative values by zero does not change the outcome. This can model, for instance, financial situations in which the variables are offers to buy some indivisible good at a certain price, and the seller must decide which (if any) offer to accept. A prophet, knowing the whole sequence of variables, can obviously select the largest of them, achieving value for any specific instance of this process, and expected value . The prophet inequality states the existence of an online algorithm for this process whose expected value is at least half that of the prophet: . No algorithm can achieve a greater expected value for all distributions of inputs. [3] [4]

One method for proving the single-item prophet inequality is to use a "threshold algorithm" that sets a parameter and then accepts the first random variable that is at least as large as . If the probability that this process accepts an item is , then its expected value is plus the expected excess over that the selected variable (if there is one) has. Each variable will be considered by the threshold algorithm with probability at least , and if it is considered will contribute to the excess, so by linearity of expectation the expected excess is at least

Setting to the median of the distribution of , so that , and adding to this bound on expected excess, causes the and terms to cancel each other, showing that for this setting of the threshold algorithm achieves an expected value of at least . [3] [5] A different threshold, , also achieves at least this same expected value. [3] [6]

Generalizations

Various generalizations of the single-item prophet inequality to other online scenarios are known, and are also called prophet inequalities. [3]

Comparison to competitive analysis

Prophet inequalities are related to the competitive analysis of online algorithms, but differ in two ways. First, much of competitive analysis assumes worst case inputs, chosen to maximize the ratio between the computed value and the optimal value that could have been achieved with knowledge of the future, whereas for prophet inequalities some knowledge of the input, its distribution, is assumed to be known. And second, in order to achieve a certain competitive ratio, an online algorithm must perform within that ratio of the optimal performance on all inputs. Instead, a prophet inequality only bounds the performance in expectation, allowing some input sequences to produce worse performance as long as the average is good. [3]

Related Research Articles

<span class="mw-page-title-main">Convolution</span> Integral expressing the amount of overlap of one function as it is shifted over another

In mathematics, convolution is a mathematical operation on two functions that produces a third function that expresses how the shape of one is modified by the other. The term convolution refers to both the result function and to the process of computing it. It is defined as the integral of the product of the two functions after one is reflected about the y-axis and shifted. The choice of which function is reflected and shifted before the integral does not change the integral result. The integral is evaluated for all values of shift, producing the convolution function.

<span class="mw-page-title-main">Expected value</span> Average value of a random variable

In probability theory, the expected value is a generalization of the weighted average. Informally, the expected value is the arithmetic mean of a large number of independently selected outcomes of a random variable.

<span class="mw-page-title-main">Normal distribution</span> Probability distribution

In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is

<span class="mw-page-title-main">Heaviside step function</span> Indicator function of positive numbers

The Heaviside step function, or the unit step function, usually denoted by H or θ, is a step function, named after Oliver Heaviside (1850–1925), the value of which is zero for negative arguments and one for positive arguments. It is an example of the general class of step functions, all of which can be represented as linear combinations of translations of this one.

Students <i>t</i>-distribution Probability distribution

In probability and statistics, Student's t-distribution is a continuous probability distribution that generalizes the standard normal distribution. Like the standard normal distribution it is symmetric around zero and bell-shaped.

<span class="mw-page-title-main">Martingale (probability theory)</span> Model in probability theory

In probability theory, a martingale is a sequence of random variables for which, at a particular time, the conditional expectation of the next value in the sequence is equal to the present value, regardless of all prior values.

<span class="mw-page-title-main">Expectation–maximization algorithm</span> Iterative method for finding maximum likelihood estimates in statistical models

In statistics, an expectation–maximization (EM) algorithm is an iterative method to find (local) maximum likelihood or maximum a posteriori (MAP) estimates of parameters in statistical models, where the model depends on unobserved latent variables. The EM iteration alternates between performing an expectation (E) step, which creates a function for the expectation of the log-likelihood evaluated using the current estimate for the parameters, and a maximization (M) step, which computes parameters maximizing the expected log-likelihood found on the E step. These parameter-estimates are then used to determine the distribution of the latent variables in the next E step.

<span class="mw-page-title-main">Stopping time</span> Time at which a random variable stops exhibiting a behavior of interest

In probability theory, in particular in the study of stochastic processes, a stopping time is a specific type of “random time”: a random variable whose value is interpreted as the time at which a given stochastic process exhibits a certain behavior of interest. A stopping time is often defined by a stopping rule, a mechanism for deciding whether to continue or stop a process on the basis of the present position and past events, and which will almost always lead to a decision to stop at some finite time.

The drawdown is the measure of the decline from a historical peak in some variable.

<span class="mw-page-title-main">Softmax function</span> Smooth approximation of one-hot arg max

The softmax function, also known as softargmax or normalized exponential function, converts a vector of K real numbers into a probability distribution of K possible outcomes. It is a generalization of the logistic function to multiple dimensions, and used in multinomial logistic regression. The softmax function is often used as the last activation function of a neural network to normalize the output of a network to a probability distribution over predicted output classes, based on Luce's choice axiom.

In the study of stochastic processes in mathematics, a hitting time is the first time at which a given process "hits" a given subset of the state space. Exit times and return times are also examples of hitting times.

In mathematics, the theory of optimal stopping or early stopping is concerned with the problem of choosing a time to take a particular action, in order to maximise an expected reward or minimise an expected cost. Optimal stopping problems can be found in areas of statistics, economics, and mathematical finance. A key example of an optimal stopping problem is the secretary problem. Optimal stopping problems can often be written in the form of a Bellman equation, and are therefore often solved using dynamic programming.

<span class="mw-page-title-main">All-pay auction</span>

In economics and game theory, an all-pay auction is an auction in which every bidder must pay regardless of whether they win the prize, which is awarded to the highest bidder as in a conventional auction. As shown by Riley and Samuelson (1981), equilibrium bidding in an all pay auction with private information is revenue equivalent to bidding in a sealed high bid or open ascending price auction.

<span class="mw-page-title-main">Quantile regression</span> Statistics concept

Quantile regression is a type of regression analysis used in statistics and econometrics. Whereas the method of least squares estimates the conditional mean of the response variable across values of the predictor variables, quantile regression estimates the conditional median of the response variable. Quantile regression is an extension of linear regression used when the conditions of linear regression are not met.

In actuarial science and applied probability, ruin theory uses mathematical models to describe an insurer's vulnerability to insolvency/ruin. In such models key quantities of interest are the probability of ruin, distribution of surplus immediately prior to ruin and deficit at time of ruin.

In probability theory, the optional stopping theorem says that, under certain conditions, the expected value of a martingale at a stopping time is equal to its initial expected value. Since martingales can be used to model the wealth of a gambler participating in a fair game, the optional stopping theorem says that, on average, nothing can be gained by stopping play based on the information obtainable so far. Certain conditions are necessary for this result to hold true. In particular, the theorem applies to doubling strategies.

<span class="mw-page-title-main">Reflection principle (Wiener process)</span>

In the theory of probability for stochastic processes, the reflection principle for a Wiener process states that if the path of a Wiener process f(t) reaches a value f(s) = a at time t = s, then the subsequent path after time s has the same distribution as the reflection of the subsequent path about the value a. More formally, the reflection principle refers to a lemma concerning the distribution of the supremum of the Wiener process, or Brownian motion. The result relates the distribution of the supremum of Brownian motion up to time t to the distribution of the process at time t. It is a corollary of the strong Markov property of Brownian motion.

In the mathematical theory of probability, the drift-plus-penalty method is used for optimization of queueing networks and other stochastic systems.

Exponential Tilting (ET), Exponential Twisting, or Exponential Change of Measure (ECM) is a distribution shifting technique used in many parts of mathematics. The different exponential tiltings of a random variable is known as the natural exponential family of .

<span class="mw-page-title-main">Tuza's conjecture</span> Problem on triangles in graph theory

Tuza's conjecture is an unsolved problem in graph theory, a branch of mathematics, concerning triangles in undirected graphs.

References

  1. Correa, Jose; Foncea, Patricio; Hoeksma, Ruben; Oosterwijk, Tim; Vredeveld, Tjark (2018), "Recent Developments in Prophet Inequalities" (PDF), ACM SIGecom Exchanges, 17 (1): 61–70
  2. Hill, Theodore P.; Kertz, Robert P. (1992), "A survey of prophet inequalities in optimal stopping theory", in Bruss, F. Thomas; Ferguson, Thomas S.; Samuels, Stephen M. (eds.), Strategies for Sequential Search and Selection in Real Time: Proceedings of the AMS–IMS–SIAM Joint Summer Research Conference held at the University of Massachusetts, Amherst, Massachusetts, June 21–27, 1990, Contemporary Mathematics, vol. 125, Providence, Rhode Island: American Mathematical Society, pp. 191–207, doi:10.1090/conm/125/1160620, MR   1160620
  3. 1 2 3 4 5 6 Feldman, Michal; Kesselheim, Thomas; Singla, Sahil (2021), "Tutorial on Prophet Inequalities", EC'21
  4. Krengel, Ulrich; Sucheston, Louis (1978), "On semiamarts, amarts, and processes with finite value", in Kuelbs, James (ed.), Probability on Banach Spaces, Advances in Probability and Related Topics, vol. 4, Dekker, New York, pp. 197–266, MR   0515432
  5. Samuel-Cahn, Ester (1984), "Comparison of threshold stop rules and maximum for independent nonnegative random variables", Annals of Probability , 12 (4): 1213–1216, JSTOR   2243359, MR   0757778
  6. Kleinberg, Robert; Weinberg, S. Matthew (2019), "Matroid prophet inequalities and applications to multi-dimensional mechanism design", Games and Economic Behavior , 113: 97–115, doi:10.1016/j.geb.2014.11.002, MR   3926869