# Truncation (statistics)

Last updated

In statistics, truncation results in values that are limited above or below, resulting in a truncated sample. [1] A random variable ${\displaystyle y}$ is said to be truncated from below if, for some threshold value ${\displaystyle c}$, the exact value of ${\displaystyle y}$ is known for all cases ${\displaystyle y>c}$, but unknown for all cases ${\displaystyle y\leq c}$. Similarly, truncation from above means the exact value of ${\displaystyle y}$ is known in cases where ${\displaystyle y, but unknown when ${\displaystyle y\geq c}$. [2]

## Contents

Truncation is similar to but distinct from the concept of statistical censoring. A truncated sample can be thought of as being equivalent to an underlying sample with all values outside the bounds entirely omitted, with not even a count of those omitted being kept. With statistical censoring, a note would be recorded documenting which bound (upper or lower) had been exceeded and the value of that bound. With truncated sampling, no note is recorded.

## Applications

Usually the values that insurance adjusters receive are either left-truncated, right-censored, or both. For example, if policyholders are subject to a policy limit u, then any loss amounts that are actually above u are reported to the insurance company as being exactly u because u is the amount the insurance company pays. The insurer knows that the actual loss is greater than u but they don't know what it is. On the other hand, left truncation occurs when policyholders are subject to a deductible. If policyholders are subject to a deductible d, any loss amount that is less than d will not even be reported to the insurance company. If there is a claim on a policy limit of u and a deductible of d, any loss amount that is greater than u will be reported to the insurance company as a loss of ${\displaystyle u-d}$ because that is the amount the insurance company has to pay. Therefore, insurance loss data is left-truncated because the insurance company doesn't know if there are values below the deductible d because policyholders won't make a claim. The insurance loss is also right-censored if the loss is greater than u because u is the most the insurance company will pay. Thus, it only knows that your claim is greater than u, not the exact claim amount.

## Probability distributions

Truncation can be applied to any probability distribution. This will usually lead to a new distribution, not one within the same family. Thus, if a random variable X has F(x) as its distribution function, the new random variable Y defined as having the distribution of X truncated to the semi-open interval (a, b] has the distribution function

${\displaystyle F_{Y}(y)={\frac {F(y)-F(a)}{F(b)-F(a)}}\,}$

for y in the interval (a, b], and 0 or 1 otherwise. If truncation were to the closed interval [a, b], the distribution function would be

${\displaystyle F_{Y}(y)={\frac {F(y)-F(a-)}{F(b)-F(a-)}}\,}$

for y in the interval [a, b], and 0 or 1 otherwise.

## Data analysis

The analysis of data where observations are treated as being from truncated versions of standard distributions can be undertaken using maximum likelihood, where the likelihood would be derived from the distribution or density of the truncated distribution. This involves taking account of the factor ${\displaystyle {F(b)-F(a)}}$ in the modified density function which will depend on the parameters of the original distribution.

In practice, if the fraction truncated is very small the effect of truncation might be ignored when analysing data. For example, it is common to use a normal distribution to model data whose values can only be positive but for which the typical range of values is well away from zero. In such cases, a truncated or censored version of the normal distribution may formally be preferable (although there would be alternatives); there would be very little change in results from the more complicated analysis. However, software is readily available for maximum-likelihood estimation of even moderately complicated models, such as regression models, for truncated data. [3]

In econometrics, truncated dependent variables are variables for which observations cannot be made for certain values in some range. [4] Regression models with such dependent variables require special care that properly recognizes the truncated nature of the variable. Estimation of such truncated regression model can be done in parametric, [5] [6] [7] or semi- and non-parametric frameworks. [8] [9]

## Related Research Articles

The method of least squares is a standard approach in regression analysis to approximate the solution of overdetermined systems by minimizing the sum of the squares of the residuals made in the results of each individual equation.

In statistics, the (binary) logistic model is a statistical model that models the probability of one event taking place by having the log-odds for the event be a linear combination of one or more independent variables ("predictors"). In regression analysis, logistic regression is estimating the parameters of a logistic model. Formally, in binary logistic regression there is a single binary dependent variable, coded by a indicator variable, where the two values are labeled "0" and "1", while the independent variables can each be a binary variable or a continuous variable. The corresponding probability of the value labeled "1" can vary between 0 and 1, hence the labeling; the function that converts log-odds to probability is the logistic function, hence the name. The unit of measurement for the log-odds scale is called a logit, from logistic unit, hence the alternative names. See § Background and § Definition for formal mathematics, and § Example for a worked example.

Survival analysis is a branch of statistics for analyzing the expected duration of time until one event occurs, such as death in biological organisms and failure in mechanical systems. This topic is called reliability theory or reliability analysis in engineering, duration analysis or duration modelling in economics, and event history analysis in sociology. Survival analysis attempts to answer certain questions, such as what is the proportion of a population which will survive past a certain time? Of those that survive, at what rate will they die or fail? Can multiple causes of death or failure be taken into account? How do particular circumstances or characteristics increase or decrease the probability of survival?

The Akaike information criterion (AIC) is an estimator of prediction error and thereby relative quality of statistical models for a given set of data. Given a collection of models for the data, AIC estimates the quality of each model, relative to each of the other models. Thus, AIC provides a means for model selection.

In statistics, a generalized linear model (GLM) is a flexible generalization of ordinary linear regression. The GLM generalizes linear regression by allowing the linear model to be related to the response variable via a link function and by allowing the magnitude of the variance of each measurement to be a function of its predicted value.

In statistical modeling, regression analysis is a set of statistical processes for estimating the relationships between a dependent variable and one or more independent variables. The most common form of regression analysis is linear regression, in which one finds the line that most closely fits the data according to a specific mathematical criterion. For example, the method of ordinary least squares computes the unique line that minimizes the sum of squared differences between the true data and that line. For specific mathematical reasons, this allows the researcher to estimate the conditional expectation of the dependent variable when the independent variables take on a given set of values. Less common forms of regression use slightly different procedures to estimate alternative location parameters or estimate the conditional expectation across a broader collection of non-linear models.

In statistics, the coefficient of determination, denoted R2 or r2 and pronounced "R squared", is the proportion of the variation in the dependent variable that is predictable from the independent variable(s).

In statistics, a probit model is a type of regression where the dependent variable can take only two values, for example married or not married. The word is a portmanteau, coming from probability + unit. The purpose of the model is to estimate the probability that an observation with particular characteristics will fall into a specific one of the categories; moreover, classifying observations based on their predicted probabilities is a type of binary classification model.

In statistics, ordinary least squares (OLS) is a type of linear least squares method for estimating the unknown parameters in a linear regression model. OLS chooses the parameters of a linear function of a set of explanatory variables by the principle of least squares: minimizing the sum of the squares of the differences between the observed dependent variable in the given dataset and those predicted by the linear function of the independent variable.

This glossary of statistics and probability is a list of definitions of terms and concepts used in the mathematical sciences of statistics and probability, their sub-disciplines, and related fields. For additional related terms, see Glossary of mathematics.

In statistics, a tobit model is any of a class of regression models in which the observed range of the dependent variable is censored in some way. The term was coined by Arthur Goldberger in reference to James Tobin, who developed the model in 1958 to mitigate the problem of zero-inflated data for observations of household expenditure on durable goods. Because Tobin's method can be easily extended to handle truncated and other non-randomly selected samples, some authors adopt a broader definition of the tobit model that includes these cases.

Truncated regression models are a class of models in which the sample has been truncated for certain ranges of the dependent variable. That means observations with values in the dependent variable below or above certain thresholds are systematically excluded from the sample. Therefore, whole observations are missing, so that neither the dependent nor the independent variable is known. This is in contrast to censored regression models where only the value of the dependent variable is clustered at a lower threshold, an upper threshold, or both, while the value for independent variables is available.

Bootstrapping is any test or metric that uses random sampling with replacement, and falls under the broader class of resampling methods. Bootstrapping assigns measures of accuracy to sample estimates. This technique allows estimation of the sampling distribution of almost any statistic using random sampling methods.

Omnibus tests are a kind of statistical test. They test whether the explained variance in a set of data is significantly greater than the unexplained variance, overall. One example is the F-test in the analysis of variance. There can be legitimate significant effects within a model even if the omnibus test is not significant. For instance, in a model with two independent variables, if only one variable exerts a significant effect on the dependent variable and the other does not, then the omnibus test may be non-significant. This fact does not affect the conclusions that may be drawn from the one significant variable. In order to test effects within an omnibus test, researchers often use contrasts.

In statistics, censoring is a condition in which the value of a measurement or observation is only partially known.

Quantile regression is a type of regression analysis used in statistics and econometrics. Whereas the method of least squares estimates the conditional mean of the response variable across values of the predictor variables, quantile regression estimates the conditional median of the response variable. Quantile regression is an extension of linear regression used when the conditions of linear regression are not met.

The Heckman correction is a statistical technique to correct bias from non-randomly selected samples or otherwise incidentally truncated dependent variables, a pervasive issue in quantitative social sciences when using observational data. Conceptually, this is achieved by explicitly modelling the individual sampling probability of each observation together with the conditional expectation of the dependent variable. The resulting likelihood function is mathematically similar to the tobit model for censored dependent variables, a connection first drawn by James Heckman in 1974. Heckman also developed a two-step control function approach to estimate this model, which avoids the computational burden of having to estimate both equations jointly, albeit at the cost of inefficiency. Heckman received the Nobel Memorial Prize in Economic Sciences in 2000 for his work in this field.

In probability theory, the Mills ratio of a continuous random variable is the function

In statistics, linear regression is a linear approach for modelling the relationship between a scalar response and one or more explanatory variables. The case of one explanatory variable is called simple linear regression; for more than one, the process is called multiple linear regression. This term is distinct from multivariate linear regression, where multiple correlated dependent variables are predicted, rather than a single scalar variable.

## References

1. Dodge, Y. (2003) The Oxford Dictionary of Statistical Terms. OUP. ISBN   0-19-920613-9
2. Breen, Richard (1996). Regression Models : Censored, Sample Selected, or Truncated Data. Quantitative Applications in the Social Sciences. Vol. 111. Thousand Oaks: Sage. pp. 2–4. ISBN   0-8039-5710-6.
3. Wolynetz, M. S. (1979). "Maximum Likelihood Estimation in a Linear Model from Confined and Censored Normal Data". Journal of the Royal Statistical Society . Series C. 28 (2): 195–206. doi:10.2307/2346749. JSTOR   2346749.
4. "Truncated Dependent Variables". About.com . Retrieved 2008-03-22.
5. Amemiya, T. (1973). "Regression Analysis When the Dependent Variable is Truncated Normal". Econometrica. 41 (6): 997–1016. doi:10.2307/1914031. JSTOR   1914031.
6. Heckman, James (1976). "The Common Structure of Statistical Models of Truncation, Sample Selection and Limited Dependent Variables and a Simple Estimator for Such Models". Annals of Economic and Social Measurement. 5 (4): 475–492.
7. Vancak, V.; Goldberg, Y.; Bar-Lev, S. K.; Boukai, B. (2015). "Continuous statistical models: With or without truncation parameters?". Mathematical Methods of Statistics. 24 (1): 55–73. doi:10.3103/S1066530715010044.{{cite journal}}: CS1 maint: multiple names: authors list (link)
8. Lewbel, A.; Linton, O. (2002). "Nonparametric Censored and Truncated Regression". Econometrica . 70 (2): 765–779. doi:10.1111/1468-0262.00304. JSTOR   2692291.
9. Park, B. U.; Simar, L.; Zelenyuk, V. (2008). "Local Likelihood Estimation of Truncated Regression and its Partial Derivatives: Theory and Application" (PDF). Journal of Econometrics . 146 (1): 185–198. doi:10.1016/j.jeconom.2008.08.007.