Conditionality principle

Last updated

The conditionality principle is a Fisherian principle of statistical inference that Allan Birnbaum formally defined and studied in an article in the Journal of the American Statistical Association, Birnbaum (1962).

Contents

Informally, the conditionality principle can be taken as the claim that

Experiments which were not actually performed are not relevant to any statistical analysis

and the implicit admonition that unrealized experiments should be ignored: Not included as part of any calculation or discussion of results.

Together with the sufficiency principle, Birnbaum's version of the principle implies the famous likelihood principle. Although the relevance of the proof to data analysis remains controversial among statisticians, many Bayesians and likelihoodists consider the likelihood principle foundational for statistical inference.

Historical background

Some statisticians in the mid 20th century had proposed that a valid statistical analysis must include all of the possible experiments which might have been conducted. Perhaps a series of desired experiments that each require some uncertain opportunity in order to carry out. The uncertain factor could be something such as good weather for a timely astronomical observation (the "experiment" being the search of the telescopic image for traces of some type of object), or availability of more data resources, such as the chance of discovery of some new fossil that would provide more evidence to answer a question covered by another paleontological study. Another resource issue might be the need for special access to private data (patients' medical records, for example) from one of several possible institutions, most of which would be expected to refuse permission; the nature of the data that could possibly be provided and the correct statistical model for its analysis would depend of which institution granted access and how it had collected and curated the private data that might become available for a study (technically, in this case the "experiment" has already been conducted by the medical facility, and some other party is analyzing the collected data to answer their own research question).

All these examples illustrate normal issues of how uncontrolled chance determines the nature of the experiment that can actually be conducted. Some analyses of the statistical significance of the outcomes of particular experiments incorporated the consequences such chance events had on the data that was obtained. Many statisticians were uncomfortable with the idea, and tended to tacitly skip seemingly extraneous random effects in their analyses; many scientists and researchers were baffled by a few statisticians' elaborate efforts to consider circumstantial effects in the statistical analysis of their experiments which the researchers considered irrelevant.[ citation needed ]

A few statisticians in the 1960s and 1970s took the idea even further, and proposed that an experiment could deliberately design-in a random factor, usually by introducing the use of some ancillary statistic, like the roll of a die, or the flip of a coin, and that the contrived random event could later be included in the data analysis, and somehow improve the inferred significance of the observed outcome. Most statisticians were uncomfortable with the idea, and the overwhelming majority of scientists and researchers considered it preposterous, and continue to the present to refute the idea and to reject any analyses based on it.[ citation needed ]

The conditionality principle is a formal rejection of the idea that "the road not taken" can possibly be relevant: In effect, it banishes from statistical analysis any consideration of effects from details of designs for experiments that were not conducted, even if they might have been planed or prepared for. The conditionality principle throws out all speculative considerations about what might have happened, and only allows the statistical analysis of the data obtained to include the procedures, circumstances, and details of the particular experiment actually conducted that produced the data actually collected. Experiments merely contemplated and not conducted, or missed opportunities for plans to obtain data, are all irrelevant and statistical calculations that include them are presumptively wrong.

Formulation

The conditionality principle makes an assertion about a composite experiment, that can be described as a suite or assemblage of several constituent experiments the index is some ancillary statistic , i.e. a statistic whose probability distribution does not depend on any unknown parameter values. This means that obtaining an observation of some specific outcome of the whole experiment requires first observing a value for and then taking an observation from the indicated component experiment

The conditionality principle can be formally stated thus:

Conditionality Principle:
If is any experiment having the form of a mixture of component experiments then for each outcome of the evidential meaning of any outcome of any mixture experiment is the same as that of the corresponding outcome of the corresponding component experiment actually conducted, ignoring the overall structure of the mixed experiment; see Birnbaum (1962).

Examples

An illustration of the conditionality principle, in a bioinformatics context, is given by Barker (2014).

Example scenario

The ancillary statistic could be the roll of die, whose value will be one of This random selection of an experiment is actually a wise precaution to curb the influence of a researchers' biases, if there is reason to suspect that the researcher might consciously or unconsciously select an experiment that seems like it would be likely to produce data that supports a favored hypothesis. The result of the dice roll then determines which of six possible experiments is the one actually conducted to obtain the study's data.

Say that the die rolls a '3'. In that case, the result observed for is actually the outcome of experminent None of the other five experiments is ever conducted, and none of the other possible results is ever seen, that might have been observed if some other number than '3' had come up. The actual observed outcome, is unaffected by any aspect of the other five sub-experiments that were not carried out, and only the procedures and experimental design of the sub-experiment that was conducted to collect the data, had any bearing on the statistical analysis the outcome, regardless of the fact that the experimental designs for the experiments which might have been conducted had been prepared at the time of the actual experiment and might just as likely been performed.

The conditionality principle says that all of the details of must be excluded from the statistical analysis of the actual observation and even the fact that experiment 3 was chosen by the roll of a die: Further, none of the possible randomness brought into the outcome by the statistic (the dice roll) can be included in the analysis either. The only thing that determines the correct statistics to be used for the data analysis is experiment and the only data to consider is not

Citations

    Related Research Articles

    In statistics, the likelihood principle is the proposition that, given a statistical model, all the evidence in a sample relevant to model parameters is contained in the likelihood function.

    <span class="mw-page-title-main">Statistical inference</span> Process of using data analysis

    Statistical inference is the process of using data analysis to infer properties of an underlying distribution of probability. Inferential statistical analysis infers properties of a population, for example by testing hypotheses and deriving estimates. It is assumed that the observed data set is sampled from a larger population.

    The likelihood function is the joint probability mass of observed data viewed as a function of the parameters of a statistical model. Intuitively, the likelihood function is the probability of observing data assuming is the actual parameter.

    Bayesian inference is a method of statistical inference in which Bayes' theorem is used to update the probability for a hypothesis as more evidence or information becomes available. Fundamentally, Bayesian inference uses prior knowledge, in the form of a prior distribution in order to estimate posterior probabilities. Bayesian inference is an important technique in statistics, and especially in mathematical statistics. Bayesian updating is particularly important in the dynamic analysis of a sequence of data. Bayesian inference has found application in a wide range of activities, including science, engineering, philosophy, medicine, sport, and law. In the philosophy of decision theory, Bayesian inference is closely related to subjective probability, often called "Bayesian probability".

    The principle of maximum entropy states that the probability distribution which best represents the current state of knowledge about a system is the one with largest entropy, in the context of precisely stated prior data.

    Minimum Description Length (MDL) is a model selection principle where the shortest description of the data is the best model. MDL methods learn through a data compression perspective and are sometimes described as mathematical applications of Occam's razor. The MDL principle can be extended to other forms of inductive inference and learning, for example to estimation and sequential prediction, without explicitly identifying a single model of the data.

    Bayesian statistics is a theory in the field of statistics based on the Bayesian interpretation of probability, where probability expresses a degree of belief in an event. The degree of belief may be based on prior knowledge about the event, such as the results of previous experiments, or on personal beliefs about the event. This differs from a number of other interpretations of probability, such as the frequentist interpretation, which views probability as the limit of the relative frequency of an event after many trials. More concretely, analysis in Bayesian methods codifies prior knowledge in the form of a prior distribution.

    A prior probability distribution of an uncertain quantity, often simply called the prior, is its assumed probability distribution before some evidence is taken into account. For example, the prior could be the probability distribution representing the relative proportions of voters who will vote for a particular politician in a future election. The unknown quantity may be a parameter of the model or a latent variable rather than an observable variable.

    Fisher's exact test is a statistical significance test used in the analysis of contingency tables. Although in practice it is employed when sample sizes are small, it is valid for all sample sizes. It is named after its inventor, Ronald Fisher, and is one of a class of exact tests, so called because the significance of the deviation from a null hypothesis can be calculated exactly, rather than relying on an approximation that becomes exact in the limit as the sample size grows to infinity, as with many statistical tests.

    <span class="mw-page-title-main">Mathematical statistics</span> Branch of statistics

    Mathematical statistics is the application of probability theory, a branch of mathematics, to statistics, as opposed to techniques for collecting statistical data. Specific mathematical techniques which are used for this include mathematical analysis, linear algebra, stochastic analysis, differential equations, and measure theory.

    This glossary of statistics and probability is a list of definitions of terms and concepts used in the mathematical sciences of statistics and probability, their sub-disciplines, and related fields. For additional related terms, see Glossary of mathematics and Glossary of experimental design.

    <span class="mw-page-title-main">Confounding</span> Variable or factor in causal inference

    In causal inference, a confounder is a variable that influences both the dependent variable and independent variable, causing a spurious association. Confounding is a causal concept, and as such, cannot be described in terms of correlations or associations. The existence of confounders is an important quantitative explanation why correlation does not imply causation. Some notations are explicitly designed to identify the existence, possible existence, or non-existence of confounders in causal relationships between elements of a system.

    <span class="mw-page-title-main">Causal model</span> Conceptual model in philosophy of science

    In the philosophy of science, a causal model is a conceptual model that describes the causal mechanisms of a system. Several types of causal notation may be used in the development of a causal model. Causal models can improve study designs by providing clear rules for deciding which independent variables need to be included/controlled for.

    Bayesian linear regression is a type of conditional modeling in which the mean of one variable is described by a linear combination of other variables, with the goal of obtaining the posterior probability of the regression coefficients and ultimately allowing the out-of-sample prediction of the regressandconditional on observed values of the regressors. The simplest and most widely used version of this model is the normal linear model, in which given is distributed Gaussian. In this model, and under a particular choice of prior probabilities for the parameters—so-called conjugate priors—the posterior can be found analytically. With more arbitrarily chosen priors, the posteriors generally have to be approximated.

    In statistics, missing data, or missing values, occur when no data value is stored for the variable in an observation. Missing data are a common occurrence and can have a significant effect on the conclusions that can be drawn from the data.

    The foundations of statistics consists of the mathematical and philosophical basis for arguments and inferences made using statistics. This includes the justification for methods of statistical inference, estimation and hypothesis testing, the quantification of uncertainty in the conclusions of statistical arguments, and the interpretation of those conclusions in probabilistic terms. A valid foundation can be used to explain statistical paradoxes such as Simpson's paradox, provide a precise description of observed statistical laws, and guide the application of statistical conclusions in social and scientific applications.

    Frequentist inference is a type of statistical inference based in frequentist probability, which treats “probability” in equivalent terms to “frequency” and draws conclusions from sample-data by means of emphasizing the frequency or proportion of findings in the data. Frequentist inference underlies frequentist statistics, in which the well-established methodologies of statistical hypothesis testing and confidence intervals are founded.

    In particle physics, CLs represents a statistical method for setting upper limits on model parameters, a particular form of interval estimation used for parameters that can take only non-negative values. Although CLs are said to refer to Confidence Levels, "The method's name is ... misleading, as the CLs exclusion region is not a confidence interval." It was first introduced by physicists working at the LEP experiment at CERN and has since been used by many high energy physics experiments. It is a frequentist method in the sense that the properties of the limit are defined by means of error probabilities, however it differs from standard confidence intervals in that the stated confidence level of the interval is not equal to its coverage probability. The reason for this deviation is that standard upper limits based on a most powerful test necessarily produce empty intervals with some fixed probability when the parameter value is zero, and this property is considered undesirable by most physicists and statisticians.

    <span class="mw-page-title-main">Bayesian inference in marketing</span>

    In marketing, Bayesian inference allows for decision making and market research evaluation under uncertainty and with limited data.

    <span class="mw-page-title-main">Homoscedasticity and heteroscedasticity</span> Statistical property

    In statistics, a sequence of random variables is homoscedastic if all its random variables have the same finite variance; this is also known as homogeneity of variance. The complementary notion is called heteroscedasticity, also known as heterogeneity of variance. The spellings homoskedasticity and heteroskedasticity are also frequently used. Assuming a variable is homoscedastic when in reality it is heteroscedastic results in unbiased but inefficient point estimates and in biased estimates of standard errors, and may result in overestimating the goodness of fit as measured by the Pearson coefficient.

    References