In statistical hypothesis testing, e-values quantify the evidence in the data against a null hypothesis (e.g., "the coin is fair", or, in a medical context, "this new treatment has no effect"). They serve as a more robust alternative to p-values, addressing some shortcomings of the latter.
In contrast to p-values, e-values can deal with optional continuation: e-values of subsequent experiments (e.g. clinical trials concerning the same treatment) may simply be multiplied to provide a new, "product" e-value that represents the evidence in the joint experiment. This works even if, as often happens in practice, the decision to perform later experiments may depend in vague, unknown ways on the data observed in earlier experiments, and it is not known beforehand how many trials will be conducted: the product e-value remains a meaningful quantity, leading to tests with Type-I error control. For this reason, e-values and their sequential extension, the e-process, are the fundamental building blocks for anytime-valid statistical methods (e.g. confidence sequences). Another advantage over p-values is that any weighted average of e-values remains an e-value, even if the individual e-values are arbitrarily dependent. This is one of the reasons why e-values have also turned out to be useful tools in multiple testing. [1]
E-values can be interpreted in a number of different ways: first, an e-value can be interpreted as rescaling of a test that is presented on a more appropriate scale that facilitates merging them. [2] Second, the reciprocal of an e-value is a p-value, but not just any p-value: a special p-value for which a rejection `at level p' retains a generalized Type-I error guarantee. [3] Third, they are broad generalizations of likelihood ratios and are also related to, yet distinct from, Bayes factors. Fourth, they have an interpretation as bets. Fifth, in a sequential context, they can also be interpreted as increments of nonnegative supermartingales. Interest in e-values has exploded since 2019, when the term 'e-value' was coined and a number of breakthrough results were achieved by several research groups. The first overview article appeared in 2023. [4]
Let the null hypothesis be given as a set of distributions for data . Usually with each a single outcome and a fixed sample size or some stopping time. We shall refer to such , which represent the full sequence of outcomes of a statistical experiment, as a sample or batch of outcomes. But in some cases may also be an unordered bag of outcomes or a single outcome.
An e-variable or e-statistic is a nonnegative random variable such that under all , its expected value is bounded by 1:
.
The value taken by e-variable is called the e-value. In practice, the term e-value (a number) is often used when one is really referring to the underlying e-variable (a random variable, that is, a measurable function of the data).
A test for a null hypothesis is traditionally modeled as a function from the data to . A test is said to be valid for level if
This is classically conveniently summarized as a function from the data to that satisfies
.
Moreover, this is sometimes generalized to permit external randomization by letting the test take value in . Here, its value is interpreted as a probability with which one should subsequently reject the hypothesis.
An issue with modelling a test in this manner, is that the traditional decision space or does not encode the level at which the test rejects. This is odd at best, because a rejection at level 1% is a much stronger claim than a rejection at level 10%. A more suitable decision space seems to be .
The e-value can be interpreted as resolving this problem. Indeed, we can rescale from to and to by rescaling the test by its level:
,
where we denote a test on this evidence scale by to avoid confusion. Such a test is then valid if
.
That is: it is valid if it is an e-value.
In fact, this reveals that e-values bounded to are rescaled randomized tests, that are continuously interpreted as evidence against the hypothesis. The standard e-value that takes value in appears as a generalization of a level 0 test. [2]
This interpretation shows that e-values are indeed fundamental to testing: they are equivalent to tests, thinly veiled by a rescaling. From this perspective, it may be surprising that typical e-values look very different from traditional tests: maximizing the objective
for an alternative hypothesis would yield traditional Neyman-Pearson style tests. Indeed, this maximizes the probability under that .
But if we continuously interpret the value of the test as evidence against the hypothesis, then we may also be interested in maximizing different targets such as
.
This yields tests that are remarkably different from traditional Neyman-Pearson tests, and more suitable when merged through multiplication as they are positive with probability 1 under . From this angle, the main innovation of the e-value compared to traditional testing is to maximize a different power target. [2]
For any e-variable and any and all , it holds that
.
This means is a valid p-value. Moreover, the e-value based test with significance level , which rejects if , has a Type-I error bounded by . But, whereas with standard p-values the inequality (*) above is usually an equality (with continuous-valued data) or near-equality (with discrete data), this is not the case with e-variables. This makes e-value-based tests more conservative (less power) than those based on standard p-values.
In exchange for this conservativeness, the p-value comes with a stronger guarantee. In particular, for every possibly data-dependent significance level , we have
if and only if . This means that a p-value satisfies this guarantee if and only if it is the reciprocal of an e-variable . [3]
The interpretation of this guarantee is that, on average, the relative Type-I error distortion caused by using a data-dependent level is controlled for every choice of the data-dependent significance level. Traditional p-values only satisfy this guarantee for data-independent or pre-specified levels.
This stronger guarantee is also called the post-hoc Type-I error, as it allows one to choose the significance level after observing the data: post-hoc. A p-value that satisfies this guarantee is also called a post-hoc p-value. As is a post-hoc p-value if and only if for some e-value , it is possible to view this as an alternative definition of an e-value.
Under this post-hoc Type-I error, the problem of choosing the significance level vanishes: we can simply choose the smallest data-dependent level at which we reject the hypothesis by setting it equal to the post-hoc p-value: . Indeed, at this data-dependent level we have
since is an e-variable. As a consequence, we can truly reject at level and still retain the post-hoc Type-I error guarantee. For a traditional p-value , rejecting at level p comes with no such guarantee.
Moreover, a post-hoc p-value inherits optional continuation and merging properties of e-values. But instead of an arithmetic weighted average, a weighted harmonic average of post-hoc p-values is still a post-hoc p-value.
Let be a simple null hypothesis. Let be any other distribution on , and let
be their likelihood ratio. Then is an e-variable. Conversely, any e-variable relative to a simple null can be written as a likelihood ratio with respect to some distribution . Thus, when the null is simple, e-variables coincide with likelihood ratios. E-variables exist for general composite nulls as well though, and they may then be thought of as generalizations of likelihood ratios. The two main ways of constructing e-variables, UI and RIPr (see below) both lead to expressions that are variations of likelihood ratios as well.
Two other standard generalizations of the likelihood ratio are (a) the generalized likelihood ratio as used in the standard, classical likelihood ratio test and (b) the Bayes factor. Importantly, neither (a) nor (b) are e-variables in general: generalized likelihood ratios in sense (a) are not e-variables unless the alternative is simple (see below under "universal inference"). Bayes factors are e-variables if the null is simple. To see this, note that, if represents a statistical model, and a prior density on , then we can set as above to be the Bayes marginal distribution with density
and then is also a Bayes factor of vs. . If the null is composite, then some special e-variables can be written as Bayes factors with some very special priors, but most Bayes factors one encounters in practice are not e-variables and many e-variables one encounters in practice are not Bayes factors. [4]
Suppose you can buy a ticket for 1 monetary unit, with nonnegative pay-off . The statements " is an e-variable" and "if the null hypothesis is true, you do not expect to gain any money if you engage in this bet" are logically equivalent. This is because being an e-variable means that the expected gain of buying the ticket is the pay-off minus the cost, i.e. , which has expectation . Based on this interpretation, the product e-value for a sequence of tests can be interpreted as the amount of money you have gained by sequentially betting with pay-offs given by the individual e-variables and always re-investing all your gains. [5]
The betting interpretation becomes particularly visible if we rewrite an e-variable as where has expectation under all and is chosen so that a.s. Any e-variable can be written in the form although with parametric nulls, writing it as a likelihood ratio is usually mathematically more convenient. The form on the other hand is often more convenient in nonparametric settings. As a prototypical example, [6] consider the case that with the taking values in the bounded interval . According to , the are i.i.d. according to a distribution with mean ; no other assumptions about are made. Then we may first construct a family of e-variables for single outcomes, , for any (these are the for which is guaranteed to be nonnegative). We may then define a new e-variable for the complete data vector by taking the product
,
where is an estimate for , based only on past data , and designed to make as large as possible in the "e-power" or "GRO" sense (see below). Waudby-Smith and Ramdas use this approach to construct "nonparametric" confidence intervals for the mean that tend to be significantly narrower than those based on more classical methods such as Chernoff, Hoeffding and Bernstein bounds. [6]
E-values are more suitable than p-value when one expects follow-up tests involving the same null hypothesis with different data or experimental set-ups. This includes, for example, combining individual results in a meta-analysis. The advantage of e-values in this setting is that they allow for optional continuation. Indeed, they have been employed in what may be the world's first fully 'online' meta-analysis with explicit Type-I error control. [7]
Informally, optional continuation implies that the product of any number of e-values, , defined on independent samples , is itself an e-value, even if the definition of each e-value is allowed to depend on all previous outcomes, and no matter what rule is used to decide when to stop gathering new samples (e.g. to perform new trials). It follows that, for any significance level , if the null is true, then the probability that a product of e-values will ever become larger than is bounded by . Thus if we decide to combine the samples observed so far and reject the null if the product e-value is larger than , then our Type-I error probability remains bounded by . We say that testing based on e-values remains safe (Type-I valid) under optional continuation.
Mathematically, this is shown by first showing that the product e-variables form a nonnegative discrete-time martingale in the filtration generated by (the individual e-variables are then increments of this martingale). The results then follow as a consequence of Doob's optional stopping theorem and Ville's inequality.
We already implicitly used product e-variables in the example above, where we defined e-variables on individual outcomes and designed a new e-value by taking products. Thus, in the example, the individual outcomes play the role of 'batches' (full samples) above, and we can therefore even engage in optional stopping "within" the original batch : we may stop the data analysis at any individual outcome (not just "batch of outcomes") we like, for whatever reason, and reject if the product so far exceeds . Not all e-variables defined for batches of outcomes can be decomposed as a product of per-outcome e-values in this way though. If this is not possible, we cannot use them for optional stopping (within a sample ) but only for optional continuation (from one sample to the next and so on).
If we set independently of the data we get a trivial e-value: it is an e-variable by definition, but it will never allow us to reject the null hypothesis. This example shows that some e-variables may be better than others, in a sense to be defined below. Intuitively, a good e-variable is one that tends to be large (much larger than 1) if the alternative is true. This is analogous to the situation with p-values: both e-values and p-values can be defined without referring to an alternative, but if an alternative is available, we would like them to be small (p-values) or large (e-values) with high probability. In standard hypothesis tests, the quality of a valid test is formalized by the notion of statistical power but this notion has to be suitably modified in the context of e-values. [4] [8]
The standard notion of quality of an e-variable relative to a given alternative , used by most authors in the field, is a generalization of the Kelly criterion in economics and (since it does exhibit close relations to classical power) is sometimes called e-power; [9] the optimal e-variable in this sense is known as log-optimal or growth-rate optimal (often abbreviated to GRO [8] ). In the case of a simple alternative , the e-power of a given e-variable is simply defined as the expectation ; in case of composite alternatives, there are various versions (e.g. worst-case absolute, worst-case relative) [8] of e-power and GRO.
Let and both be simple. Then the likelihood ratio e-variable has maximal e-power in the sense above, i.e. it is GRO. [4]
Let be simple and be composite, such that all elements of have densities (denoted by lower-case letters) relative to the same underlying measure. Grünwald et al. show that under weak regularity conditions, the GRO e-variable exists, is essentially unique, and is given by
where is the Reverse Information Projection (RIPr) of unto the convex hull of . [8] Under further regularity conditions (and in all practically relevant cases encountered so far), is given by a Bayes marginal density: there exists a specific, unique distribution on such that .
In the same setting as above, [10] show that, under no regularity conditions at all,
is an e-variable (with the second equality holding if the MLE (maximum likelihood estimator) based on data is always well-defined). This way of constructing e-variables has been called the universal inference (UI) method, "universal" referring to the fact that no regularity conditions are required.
Now let be simple and be composite, such that all elements of have densities relative to the same underlying measure. There are now two generic, closely related ways of obtaining e-variables that are close to growth-optimal (appropriately redefined [4] for composite ): Robbins' method of mixtures and the plug-in method, originally due to Wald [11] but, in essence, re-discovered by Philip Dawid as "prequential plug-in" [12] and Jorma Rissanen as "predictive MDL". [13] The method of mixtures essentially amounts to "being Bayesian about the numerator" (the reason it is not called "Bayesian method" is that, when both null and alternative are composite, the numerator may often not be a Bayes marginal): we posit any prior distribution on and set
and use the e-variable .
To explicate the plug-in method, suppose that where constitute a stochastic process and let be an estimator of based on data for . In practice one usually takes a "smoothed" maximum likelihood estimator (such as, for example, the regression coefficients in ridge regression), initially set to some "default value" . One now recursively constructs a density for by setting .
Effectively, both the method of mixtures and the plug-in method can be thought of learning a specific instantiation of the alternative that explains the data well. [4]
In parametric settings, we can simply combine the main methods for the composite alternative (obtaining or ) with the main methods for the composite null (UI or RIPr, using the single distribution or as an alternative). Note in particular that when using the plug-in method together with the UI method, the resulting e-variable will look like
which resembles, but is still fundamentally different from, the generalized likelihood ratio as used in the classical likelihood ratio test.
The advantage of the UI method compared to RIPr is that (a) it can be applied whenever the MLE can be efficiently computed - in many such cases, it is not known whether/how the reverse information projection can be calculated; and (b) that it 'automatically' gives not just an e-variable but a full e-process (see below): if we replace in the formula above by a general stopping time , the resulting ratio is still an e-variable; for the reverse information projection this automatic e-process generation only holds in special cases.
Its main disadvantage compared to RIPr is that it can be substantially sub-optimal in terms of the e-power/GRO criterion, which means that it leads to tests which also have less classical statistical power than RIPr-based methods. Thus, for settings in which the RIPr-method is computationally feasible and leads to e-processes, it is to be preferred. These include the z-test, t-test and corresponding linear regressions, k-sample tests with Bernoulli, Gaussian and Poisson distributions and the logrank test (an R package is available for a subset of these), as well as conditional independence testing under a model-X assumption. [14] However, in many other statistical testing problems, it is currently (2023) unknown whether fast implementations of the reverse information projection exist, and they may very well not exist (e.g. generalized linear models without the model-X assumption).
In nonparametric settings (such as testing a mean as in the example above, or nonparametric 2-sample testing), it is often more natural to consider e-variables of the type. However, while these superficially look very different from likelihood ratios, they can often still be interpreted as such and sometimes can even be re-interpreted as implementing a version of the RIPr-construction. [4]
Finally, in practice, one sometimes resorts to mathematically or computationally convenient combinations of RIPr, UI and other methods. [4] For example, RIPr is applied to get optimal e-variables for small blocks of outcomes and these are then multiplied to obtain e-variables for larger samples - these e-variables work well in practice but cannot be considered optimal anymore.
There exist functions that convert p-values into e-values. [15] [16] [17] Such functions are called p-to-e calibrators. Formally, a calibrator is a nonnegative decreasing function which, when applied to a p-variable (a random variable whose value is a p-value), yields an e-variable. A calibrator is said to dominate another calibrator if , and this domination is strict if the inequality is strict. An admissible calibrator is one that is not strictly dominated by any other calibrator. One can show that for a function to be a calibrator, it must have an integral of at most 1 over the uniform probability measure.
One family of admissible calibrators is given by the set of functions with . Another calibrator is given by integrating out :
Conversely, an e-to-p calibrator transforms e-values back into p-variables. Interestingly, the following calibrator dominates all other e-to-p calibrators:
While of theoretical importance, calibration is not much used in the practical design of e-variables since the resulting e-variables are often far from growth-optimal for any given . [8]
Now consider data arriving sequentially, constituting a discrete-time stochastic process. Let be another discrete-time process where for each can be written as a (measurable) function of the first outcomes. We call an e-process if for any stopping time is an e-variable, i.e. for all .
In basic cases, the stopping time can be defined by any rule that determines, at each sample size , based only on the data observed so far, whether to stop collecting data or not. For example, this could be "stop when you have seen four consecutive outcomes larger than 1", "stop at ", or the level--aggressive rule, "stop as soon as you can reject at level -level, i.e. at the smallest such that ", and so on. With e-processes, we obtain an e-variable with any such rule. Crucially, the data analyst may not know the rule used for stopping. For example, her boss may tell her to stop data collecting and she may not know exactly why - nevertheless, she gets a valid e-variable and Type-I error control. This is in sharp contrast to data analysis based on p-values (which becomes invalid if stopping rules are not determined in advance) or in classical Wald-style sequential analysis (which works with data of varying length but again, with stopping times that need to be determined in advance). In more complex cases, the stopping time has to be defined relative to some slightly reduced filtration, but this is not a big restriction in practice. In particular, the level--aggressive rule is always allowed. Because of this validity under optional stopping, e-processes are the fundamental building block of confidence sequences, also known as anytime-valid confidence intervals. [18] [4]
Technically, e-processes are generalizations of test supermartingales, which are nonnegative supermartingales with starting value 1: any test supermartingale constitutes an e-process but not vice versa.
E-processes can be constructed in a number of ways. Often, one starts with an e-value for whose definition is allowed to depend on previous data, i.e.,
for all
(again, in complex testing problems this definition needs to be modified a bit using reduced filtrations). Then the product process with is a test supermartingale, and hence also an e-process (note that we already used this construction in the example described under "e-values as bets" above: for fixed , the e-values were not dependent on past-data, but by using depending on the past, they became dependent on past data).
Another way to construct an e-process is to use the universal inference construction described above for sample sizes The resulting sequence of e-values will then always be an e-process. [4]
Historically, e-values implicitly appear as building blocks of nonnegative supermartingales in the pioneering work on anytime-valid confidence methods by well-known mathematician Herbert Robbins and some of his students. [18] The first time e-values (or something very much like them) are treated as a quantity of independent interest is by another well-known mathematician, Leonid Levin, in 1976, within the theory of algorithmic randomness. With the exception of contributions by pioneer V. Vovk in various papers with various collaborators (e.g. [16] [15] ), and an independent re-invention of the concept in an entirely different field, [19] the concept did not catch on at all until 2019, when, within just a few months, several pioneering papers by several research groups appeared on arXiv (the corresponding journal publications referenced below sometimes coming years later). In these, the concept was finally given a proper name ("S-Value" [8] and "E-Value"; [17] in later versions of their paper, [8] also adapted "E-Value"); describing their general properties, [17] two generic ways to construct them, [10] and their intimate relation to betting [5] ). Since then, interest by researchers around the world has been surging. In 2023 the first overview paper on "safe, anytime-valid methods", in which e-values play a central role, appeared. [4]
In statistics, the likelihood-ratio test is a hypothesis test that involves comparing the goodness of fit of two competing statistical models, typically one found by maximization over the entire parameter space and another found after imposing some constraint, based on the ratio of their likelihoods. If the more constrained model is supported by the observed data, the two likelihoods should not differ by more than sampling error. Thus the likelihood-ratio test tests whether this ratio is significantly different from one, or equivalently whether its natural logarithm is significantly different from zero.
In probability theory and statistics, the chi-squared distribution with degrees of freedom is the distribution of a sum of the squares of independent standard normal random variables.
In statistics, sufficiency is a property of a statistic computed on a sample dataset in relation to a parametric model of the dataset. A sufficient statistic contains all of the information that the dataset provides about the model parameters. It is closely related to the concepts of an ancillary statistic which contains no information about the model parameters, and of a complete statistic which only contains information about the parameters and no ancillary information.
In probability theory and statistics, the gamma distribution is a versatile two-parameter family of continuous probability distributions. The exponential distribution, Erlang distribution, and chi-squared distribution are special cases of the gamma distribution. There are two equivalent parameterizations in common use:
In statistics, the Neyman–Pearson lemma describes the existence and uniqueness of the likelihood ratio as a uniformly most powerful test in certain contexts. It was introduced by Jerzy Neyman and Egon Pearson in a paper in 1933. The Neyman–Pearson lemma is part of the Neyman–Pearson theory of statistical testing, which introduced concepts like errors of the second kind, power function, and inductive behavior. The previous Fisherian theory of significance testing postulated only one hypothesis. By introducing a competing hypothesis, the Neyman–Pearsonian flavor of statistical testing allows investigating the two types of errors. The trivial cases where one always rejects or accepts the null hypothesis are of little interest but it does prove that one must not relinquish control over one type of error while calibrating the other. Neyman and Pearson accordingly proceeded to restrict their attention to the class of all level tests while subsequently minimizing type II error, traditionally denoted by . Their seminal paper of 1933, including the Neyman–Pearson lemma, comes at the end of this endeavor, not only showing the existence of tests with the most power that retain a prespecified level of type I error, but also providing a way to construct such tests. The Karlin-Rubin theorem extends the Neyman–Pearson lemma to settings involving composite hypotheses with monotone likelihood ratios.
In Bayesian probability theory, if, given a likelihood function , the posterior distribution is in the same probability distribution family as the prior probability distribution , the prior and posterior are then called conjugate distributions with respect to that likelihood function and the prior is called a conjugate prior for the likelihood function .
In mathematical statistics, the Kullback–Leibler (KL) divergence, denoted , is a type of statistical distance: a measure of how much a model probability distribution Q is different from a true probability distribution P. Mathematically, it is defined as
In statistics and information theory, a maximum entropy probability distribution has entropy that is at least as great as that of all other members of a specified class of probability distributions. According to the principle of maximum entropy, if nothing is known about a distribution except that it belongs to a certain class, then the distribution with the largest entropy should be chosen as the least-informative default. The motivation is twofold: first, maximizing entropy minimizes the amount of prior information built into the distribution; second, many physical systems tend to move towards maximal entropy configurations over time.
In statistics, Poisson regression is a generalized linear model form of regression analysis used to model count data and contingency tables. Poisson regression assumes the response variable Y has a Poisson distribution, and assumes the logarithm of its expected value can be modeled by a linear combination of unknown parameters. A Poisson regression model is sometimes known as a log-linear model, especially when used to model contingency tables.
In mathematics, the Weyl character formula in representation theory describes the characters of irreducible representations of compact Lie groups in terms of their highest weights. It was proved by Hermann Weyl. There is a closely related formula for the character of an irreducible representation of a semisimple Lie algebra. In Weyl's approach to the representation theory of connected compact Lie groups, the proof of the character formula is a key step in proving that every dominant integral element actually arises as the highest weight of some irreducible representation. Important consequences of the character formula are the Weyl dimension formula and the Kostant multiplicity formula.
In probability theory, the inverse Gaussian distribution is a two-parameter family of continuous probability distributions with support on (0,∞).
In statistics, the monotone likelihood ratio property is a property of the ratio of two probability density functions (PDFs). Formally, distributions and bear the property if
Covariance matrix adaptation evolution strategy (CMA-ES) is a particular kind of strategy for numerical optimization. Evolution strategies (ES) are stochastic, derivative-free methods for numerical optimization of non-linear or non-convex continuous optimization problems. They belong to the class of evolutionary algorithms and evolutionary computation. An evolutionary algorithm is broadly based on the principle of biological evolution, namely the repeated interplay of variation and selection: in each generation (iteration) new individuals are generated by variation of the current parental individuals, usually in a stochastic way. Then, some individuals are selected to become the parents in the next generation based on their fitness or objective function value . Like this, individuals with better and better -values are generated over the generation sequence.
In mathematics, the Jack function is a generalization of the Jack polynomial, introduced by Henry Jack. The Jack polynomial is a homogeneous, symmetric polynomial which generalizes the Schur and zonal polynomials, and is in turn generalized by the Heckman–Opdam polynomials and Macdonald polynomials.
A ratio distribution is a probability distribution constructed as the distribution of the ratio of random variables having two other known distributions. Given two random variables X and Y, the distribution of the random variable Z that is formed as the ratio Z = X/Y is a ratio distribution.
In classical mechanics, holonomic constraints are relations between the position variables that can be expressed in the following form:
In probability theory and statistics, the Poisson distribution is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time if these events occur with a known constant mean rate and independently of the time since the last event. It can also be used for the number of events in other types of intervals than time, and in dimension greater than 1.
In particle physics, CLs represents a statistical method for setting upper limits on model parameters, a particular form of interval estimation used for parameters that can take only non-negative values. Although CLs are said to refer to Confidence Levels, "The method's name is ... misleading, as the CLs exclusion region is not a confidence interval." It was first introduced by physicists working at the LEP experiment at CERN and has since been used by many high energy physics experiments. It is a frequentist method in the sense that the properties of the limit are defined by means of error probabilities, however it differs from standard confidence intervals in that the stated confidence level of the interval is not equal to its coverage probability. The reason for this deviation is that standard upper limits based on a most powerful test necessarily produce empty intervals with some fixed probability when the parameter value is zero, and this property is considered undesirable by most physicists and statisticians.
For certain applications in linear algebra, it is useful to know properties of the probability distribution of the largest eigenvalue of a finite sum of random matrices. Suppose is a finite sequence of random matrices. Analogous to the well-known Chernoff bound for sums of scalars, a bound on the following is sought for a given parameter t:
In statistical decision theory, a randomised decision rule or mixed decision rule is a decision rule that associates probabilities with deterministic decision rules. In finite decision problems, randomised decision rules define a risk set which is the convex hull of the risk points of the nonrandomised decision rules.