Deficiency (statistics)

Last updated

In statistics, the deficiency is a measure to compare a statistical model with another statistical model. The term was introduced in the 1960s by the french mathematician Lucien Le Cam, who used it to prove an approximative version of the Blackwell-Sherman-Stein theorem. [1] [2] Closely related is the Le Cam distance, which is a pseudometric that computes the maximum deficiency of two statistical models. If the deficiency of a model in relation to is zero, then one says is better or more informative or stronger than .

Contents

Introduction

Le Cam defined the statistical model more abstract than a probability space with a family of probability measures. He also didn't use the term statistical model and instead used the term experiment. In his publication from 1964 he introduced the statistical experiment to a parameter set as a triple consisting of a set , a vector lattice with unit and a family of normalized positive functionals on . [3] [4] In his book from 1986 he omitted and . [5] We are going to follow his definition from 1986 and use his terminology to emphasize the generalization.

Basic concepts

Let be a parameter space. Given an Abstract L1-space (i.e. a banach lattice such that for elements also holds) consisting of lineare positive functionals . An experiment is a map of the form , such that . is the band induced by and therefore we use the notation . For a denote the . The topological dual of an L-space with the conjugated norm is called an abstract M-space. It's also a lattice with unit defined through for .

Let and be two L-space of two experiments and , then one calls a positive, norm-preserving linear map, i.e. for all , a transition. The adjoint of a transitions is a positive linear map from the dual space of into the dual space of , such that the unit of is the image of the unit of ist. [5]

Deficiency

Let be a parameter space and and be two experiments indexed by . Le and denote the corresponding L-spaces and let be the set of all transitions from to .

The deficiency of in relation to is the number

[6]

where denoted the total variation norm . The factor is just for computational purposes and is sometimes omitted.

Le Cam distance

The Le Cam distance is the following pseudometric

.

This induces an equivalence relation and when , then one says and are equivalent. The equivalent class of is also called the type of .

Often one is interested in families of experiments with and with . If as , then one says and are asymptotically equivalent.

Let be a parameter space and be the set of all types that are induced by , then the Le Cam distance is complete with respect to . The condition induces a partial order on , one says is better or more informative or stronger than . [6]

Bibliographie

Related Research Articles

In statistics, a location parameter of a probability distribution is a scalar- or vector-valued parameter , which determines the "location" or shift of the distribution. In the literature of location parameter estimation, the probability distributions with such parameter are found to be formally defined in one of the following equivalent ways:

In statistics, the likelihood-ratio test assesses the goodness of fit of two competing statistical models, specifically one found by maximization over the entire parameter space and another found after imposing some constraint, based on the ratio of their likelihoods. If the constraint is supported by the observed data, the two likelihoods should not differ by more than sampling error. Thus the likelihood-ratio test tests whether this ratio is significantly different from one, or equivalently whether its natural logarithm is significantly different from zero.

In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data. This is achieved by maximizing a likelihood function so that, under the assumed statistical model, the observed data is most probable. The point in the parameter space that maximizes the likelihood function is called the maximum likelihood estimate. The logic of maximum likelihood is both intuitive and flexible, and as such the method has become a dominant means of statistical inference.

<span class="mw-page-title-main">Cramér–Rao bound</span> Lower bound on variance of an estimator

In estimation theory and statistics, the Cramér–Rao bound (CRB) relates to estimation of a deterministic parameter. The result is named in honor of Harald Cramér and C. R. Rao, but has also been derived independently by Maurice Fréchet, Georges Darmois, and by Alexander Aitken and Harold Silverstone. It is also known as Fréchet-Cramér–Rao or Fréchet-Darmois-Cramér-Rao lower bound. It states that the precision of any unbiased estimator is at most the Fisher information; or (equivalently) the reciprocal of the Fisher information is a lower bound on its variance.

In mathematical statistics, the Fisher information is a way of measuring the amount of information that an observable random variable X carries about an unknown parameter θ of a distribution that models X. Formally, it is the variance of the score, or the expected value of the observed information.

In mathematics, the Riesz–Thorin theorem, often referred to as the Riesz–Thorin interpolation theorem or the Riesz–Thorin convexity theorem, is a result about interpolation of operators. It is named after Marcel Riesz and his student G. Olof Thorin.

<span class="mw-page-title-main">Consistent estimator</span> Statistical estimator converging in probability to a true parameter as sample size increases

In statistics, a consistent estimator or asymptotically consistent estimator is an estimator—a rule for computing estimates of a parameter θ0—having the property that as the number of data points used increases indefinitely, the resulting sequence of estimates converges in probability to θ0. This means that the distributions of the estimates become more and more concentrated near the true value of the parameter being estimated, so that the probability of the estimator being arbitrarily close to θ0 converges to one.

In statistics a minimum-variance unbiased estimator (MVUE) or uniformly minimum-variance unbiased estimator (UMVUE) is an unbiased estimator that has lower variance than any other unbiased estimator for all possible values of the parameter.

In Bayesian probability, the Jeffreys prior, named after Sir Harold Jeffreys, is a non-informative prior distribution for a parameter space; its density function is proportional to the square root of the determinant of the Fisher information matrix:

In statistics, a parametric model or parametric family or finite-dimensional model is a particular class of statistical models. Specifically, a parametric model is a family of probability distributions that has a finite number of parameters.

In estimation theory and decision theory, a Bayes estimator or a Bayes action is an estimator or decision rule that minimizes the posterior expected value of a loss function. Equivalently, it maximizes the posterior expectation of a utility function. An alternative way of formulating an estimator within Bayesian statistics is maximum a posteriori estimation.

In statistics, Basu's theorem states that any boundedly complete minimal sufficient statistic is independent of any ancillary statistic. This is a 1955 result of Debabrata Basu.

In statistical decision theory, where we are faced with the problem of estimating a deterministic parameter (vector) from observations an estimator is called minimax if its maximal risk is minimal among all estimators of . In a sense this means that is an estimator which performs best in the worst possible case allowed in the problem.

In statistics, identifiability is a property which a model must satisfy for precise inference to be possible. A model is identifiable if it is theoretically possible to learn the true values of this model's underlying parameters after obtaining an infinite number of observations from it. Mathematically, this is equivalent to saying that different values of the parameters must generate different probability distributions of the observable variables. Usually the model is identifiable only under certain technical restrictions, in which case the set of these requirements is called the identification conditions.

In statistics, the Hájek–Le Cam convolution theorem states that any regular estimator in a parametric model is asymptotically equivalent to a sum of two independent random variables, one of which is normal with asymptotic variance equal to the inverse of Fisher information, and the other having arbitrary distribution.

In statistics, asymptotic theory, or large sample theory, is a framework for assessing properties of estimators and statistical tests. Within this framework, it is often assumed that the sample size n may grow indefinitely; the properties of estimators and tests are then evaluated under the limit of n → ∞. In practice, a limit evaluation is considered to be approximately valid for large finite sample sizes too.

In statistics, efficiency is a measure of quality of an estimator, of an experimental design, or of a hypothesis testing procedure. Essentially, a more efficient estimator needs fewer input data or observations than a less efficient one to achieve the Cramér–Rao bound. An efficient estimator is characterized by having the smallest possible variance, indicating that there is a small deviance between the estimated value and the "true" value in the L2 norm sense.

In statistics, local asymptotic normality is a property of a sequence of statistical models, which allows this sequence to be asymptotically approximated by a normal location model, after an appropriate rescaling of the parameter. An important example when the local asymptotic normality holds is in the case of i.i.d sampling from a regular parametric model.

In Bayesian inference, the Bernstein–von Mises theorem provides the basis for using Bayesian credible sets for confidence statements in parametric models. It states that under some conditions, a posterior distribution converges in the limit of infinite data to a multivariate normal distribution centered at the maximum likelihood estimator with covariance matrix given by , where is the true population parameter and is the Fisher information matrix at the true population parameter value:

In mathematics, the Poisson boundary is a measure space associated to a random walk. It is an object designed to encode the asymptotic behaviour of the random walk, i.e. how trajectories diverge when the number of steps goes to infinity. Despite being called a boundary it is in general a purely measure-theoretical object and not a boundary in the topological sense. However, in the case where the random walk is on a topological space the Poisson boundary can be related to the Martin boundary, which is an analytic construction yielding a genuine topological boundary. Both boundaries are related to harmonic functions on the space via generalisations of the Poisson formula.

References

  1. Le Cam, Lucien (1964). "Sufficiency and Approximate Sufficiency". Annals of Mathematical Statistics . 35 (4). Institute of Mathematical Statistics: 1429. doi: 10.1214/aoms/1177700372 .
  2. Torgersen, Erik (1991). Comparison of Statistical Experiments. Cambridge University Press, United Kingdom. pp. 222–257. doi:10.1017/CBO9780511666353.
  3. Le Cam, Lucien (1964). "Sufficiency and Approximate Sufficiency". Annals of Mathematical Statistics . 35 (4). Institute of Mathematical Statistics: 1421. doi: 10.1214/aoms/1177700372 .
  4. van der Vaart, Aad (2002). "The Statistical Work of Lucien Le Cam". The Annals of Statistics. 30 (3): 631–82. JSTOR   2699973.
  5. 1 2 Le Cam, Lucien (1986). Asymptotic methods in statistical decision theory. Springer Series in Statistics. Springer, New York. pp. 1–5. doi:10.1007/978-1-4612-4946-7.
  6. 1 2 Le Cam, Lucien (1986). Asymptotic methods in statistical decision theory. Springer Series in Statistics. Springer, New York. pp. 18–19. doi:10.1007/978-1-4612-4946-7.