Measurement uncertainty

Last updated

In metrology, measurement uncertainty is the expression of the statistical dispersion of the values attributed to a measured quantity. All measurements are subject to uncertainty and a measurement result is complete only when it is accompanied by a statement of the associated uncertainty, such as the standard deviation. By international agreement, this uncertainty has a probabilistic basis and reflects incomplete knowledge of the quantity value. It is a non-negative parameter. [1]

Contents

The measurement uncertainty is often taken as the standard deviation of a state-of-knowledge probability distribution over the possible values that could be attributed to a measured quantity. Relative uncertainty is the measurement uncertainty relative to the magnitude of a particular single choice for the value for the measured quantity, when this choice is nonzero. This particular single choice is usually called the measured value, which may be optimal in some well-defined sense (e.g., a mean, median, or mode). Thus, the relative measurement uncertainty is the measurement uncertainty divided by the absolute value of the measured value, when the measured value is not zero.

Background

The purpose of measurement is to provide information about a quantity of interest – a measurand. For example, the measurand might be the size of a cylindrical feature, the volume of a vessel, the potential difference between the terminals of a battery, or the mass concentration of lead in a flask of water.

No measurement is exact. When a quantity is measured, the outcome depends on the measuring system, the measurement procedure, the skill of the operator, the environment, and other effects. [2] Even if the quantity were to be measured several times, in the same way and in the same circumstances, a different measured value would in general be obtained each time, assuming the measuring system has sufficient resolution to distinguish between the values.

The dispersion of the measured values would relate to how well the measurement is performed. Their average would provide an estimate of the true value of the quantity that generally would be more reliable than an individual measured value. The dispersion and the number of measured values would provide information relating to the average value as an estimate of the true value. However, this information would not generally be adequate.

The measuring system may provide measured values that are not dispersed about the true value, but about some value offset from it. Take a domestic bathroom scale. Suppose it is not set to show zero when there is nobody on the scale, but to show some value offset from zero. Then, no matter how many times the person's mass were re-measured, the effect of this offset would be inherently present in the average of the values.

The "Guide to the Expression of Uncertainty in Measurement" (commonly known as the GUM) is the definitive document on this subject. The GUM has been adopted by all major National Measurement Institutes (NMIs) and by international laboratory accreditation standards such as ISO/IEC 17025 General requirements for the competence of testing and calibration laboratories, which is required for international laboratory accreditation, and is employed in most modern national and international documentary standards on measurement methods and technology. See Joint Committee for Guides in Metrology.

Measurement uncertainty has important economic consequences for calibration and measurement activities. In calibration reports, the magnitude of the uncertainty is often taken as an indication of the quality of the laboratory, and smaller uncertainty values generally are of higher value and of higher cost. The American Society of Mechanical Engineers (ASME) has produced a suite of standards addressing various aspects of measurement uncertainty. For example, ASME standards are used to address the role of measurement uncertainty when accepting or rejecting products based on a measurement result and a product specification, [3] to provide a simplified approach (relative to the GUM) to the evaluation of dimensional measurement uncertainty, [4] to resolve disagreements over the magnitude of the measurement uncertainty statement, [5] and to provide guidance on the risks involved in any product acceptance/rejection decision. [6]

Indirect measurement

The above discussion concerns the direct measurement of a quantity, which incidentally occurs rarely. For example, the bathroom scale may convert a measured extension of a spring into an estimate of the measurand, the mass of the person on the scale. The particular relationship between extension and mass is determined by the calibration of the scale. A measurement model converts a quantity value into the corresponding value of the measurand.

There are many types of measurement in practice and therefore many models. A simple measurement model (for example for a scale, where the mass is proportional to the extension of the spring) might be sufficient for everyday domestic use. Alternatively, a more sophisticated model of a weighing, involving additional effects such as air buoyancy, is capable of delivering better results for industrial or scientific purposes. In general there are often several different quantities, for example temperature, humidity and displacement, that contribute to the definition of the measurand, and that need to be measured.

Correction terms should be included in the measurement model when the conditions of measurement are not exactly as stipulated. These terms correspond to systematic errors. Given an estimate of a correction term, the relevant quantity should be corrected by this estimate. There will be an uncertainty associated with the estimate, even if the estimate is zero, as is often the case. Instances of systematic errors arise in height measurement, when the alignment of the measuring instrument is not perfectly vertical, and the ambient temperature is different from that prescribed. Neither the alignment of the instrument nor the ambient temperature is specified exactly, but information concerning these effects is available, for example the lack of alignment is at most 0.001° and the ambient temperature at the time of measurement differs from that stipulated by at most 2 °C.

As well as raw data representing measured values, there is another form of data that is frequently needed in a measurement model. Some such data relate to quantities representing physical constants, each of which is known imperfectly. Examples are material constants such as modulus of elasticity and specific heat. There are often other relevant data given in reference books, calibration certificates, etc., regarded as estimates of further quantities.

The items required by a measurement model to define a measurand are known as input quantities in a measurement model. The model is often referred to as a functional relationship. The output quantity in a measurement model is the measurand.

Formally, the output quantity, denoted by , about which information is required, is often related to input quantities, denoted by , about which information is available, by a measurement model in the form of

where is known as the measurement function. A general expression for a measurement model is

It is taken that a procedure exists for calculating given , and that is uniquely defined by this equation.

Propagation of distributions

The true values of the input quantities are unknown. In the GUM approach, are characterized by probability distributions and treated mathematically as random variables. These distributions describe the respective probabilities of their true values lying in different intervals, and are assigned based on available knowledge concerning . Sometimes, some or all of are interrelated and the relevant distributions, which are known as joint, apply to these quantities taken together.

Consider estimates , respectively, of the input quantities , obtained from certificates and reports, manufacturers' specifications, the analysis of measurement data, and so on. The probability distributions characterizing are chosen such that the estimates , respectively, are the expectations [7] of . Moreover, for the th input quantity, consider a so-called standard uncertainty, given the symbol , defined as the standard deviation [7] of the input quantity . This standard uncertainty is said to be associated with the (corresponding) estimate .

The use of available knowledge to establish a probability distribution to characterize each quantity of interest applies to the and also to . In the latter case, the characterizing probability distribution for is determined by the measurement model together with the probability distributions for the . The determination of the probability distribution for from this information is known as the propagation of distributions. [7]

The figure below depicts a measurement model in the case where and are each characterized by a (different) rectangular, or uniform, probability distribution. has a symmetric trapezoidal probability distribution in this case.

An additive measurement function with two input quantities
X
1
{\displaystyle X_{1}}
and
X
2
{\displaystyle X_{2}}
characterized by rectangular probability distributions PropDistn.png
An additive measurement function with two input quantities and characterized by rectangular probability distributions

Once the input quantities have been characterized by appropriate probability distributions, and the measurement model has been developed, the probability distribution for the measurand is fully specified in terms of this information. In particular, the expectation of is used as the estimate of , and the standard deviation of as the standard uncertainty associated with this estimate.

Often an interval containing with a specified probability is required. Such an interval, a coverage interval, can be deduced from the probability distribution for . The specified probability is known as the coverage probability. For a given coverage probability, there is more than one coverage interval. The probabilistically symmetric coverage interval is an interval for which the probabilities (summing to one minus the coverage probability) of a value to the left and the right of the interval are equal. The shortest coverage interval is an interval for which the length is least over all coverage intervals having the same coverage probability.

Prior knowledge about the true value of the output quantity can also be considered. For the domestic bathroom scale, the fact that the person's mass is positive, and that it is the mass of a person, rather than that of a motor car, that is being measured, both constitute prior knowledge about the possible values of the measurand in this example. Such additional information can be used to provide a probability distribution for that can give a smaller standard deviation for and hence a smaller standard uncertainty associated with the estimate of . [8] [9] [10]

Type A and Type B evaluation of uncertainty

Knowledge about an input quantity is inferred from repeated measured values ("Type A evaluation of uncertainty"), or scientific judgement or other information concerning the possible values of the quantity ("Type B evaluation of uncertainty").

In Type A evaluations of measurement uncertainty, the assumption is often made that the distribution best describing an input quantity given repeated measured values of it (obtained independently) is a Gaussian distribution. then has expectation equal to the average measured value and standard deviation equal to the standard deviation of the average. When the uncertainty is evaluated from a small number of measured values (regarded as instances of a quantity characterized by a Gaussian distribution), the corresponding distribution can be taken as a t-distribution. [11] Other considerations apply when the measured values are not obtained independently.

For a Type B evaluation of uncertainty, often the only available information is that lies in a specified interval []. In such a case, knowledge of the quantity can be characterized by a rectangular probability distribution [11] with limits and . If different information were available, a probability distribution consistent with that information would be used. [12]

Sensitivity coefficients

Sensitivity coefficients describe how the estimate of would be influenced by small changes in the estimates of the input quantities . For the measurement model , the sensitivity coefficient equals the partial derivative of first order of with respect to evaluated at , , etc. For a linear measurement model

with independent, a change in equal to would give a change in This statement would generally be approximate for measurement models . The relative magnitudes of the terms are useful in assessing the respective contributions from the input quantities to the standard uncertainty associated with . The standard uncertainty associated with the estimate of the output quantity is not given by the sum of the , but these terms combined in quadrature, [1] namely by an expression that is generally approximate for measurement models :

which is known as the law of propagation of uncertainty.

When the input quantities contain dependencies, the above formula is augmented by terms containing covariances, [1] which may increase or decrease .

Uncertainty evaluation

The main stages of uncertainty evaluation constitute formulation and calculation, the latter consisting of propagation and summarizing. The formulation stage constitutes

  1. defining the output quantity (the measurand),
  2. identifying the input quantities on which depends,
  3. developing a measurement model relating to the input quantities, and
  4. on the basis of available knowledge, assigning probability distributions — Gaussian, rectangular, etc. — to the input quantities (or a joint probability distribution to those input quantities that are not independent).

The calculation stage consists of propagating the probability distributions for the input quantities through the measurement model to obtain the probability distribution for the output quantity , and summarizing by using this distribution to obtain

  1. the expectation of , taken as an estimate of ,
  2. the standard deviation of , taken as the standard uncertainty associated with , and
  3. a coverage interval containing with a specified coverage probability.

The propagation stage of uncertainty evaluation is known as the propagation of distributions, various approaches for which are available, including

  1. the GUM uncertainty framework, constituting the application of the law of propagation of uncertainty, and the characterization of the output quantity by a Gaussian or a -distribution,
  2. analytic methods, in which mathematical analysis is used to derive an algebraic form for the probability distribution for , and
  3. a Monte Carlo method, [7] in which an approximation to the distribution function for is established numerically by making random draws from the probability distributions for the input quantities, and evaluating the model at the resulting values.

For any particular uncertainty evaluation problem, approach 1), 2) or 3) (or some other approach) is used, 1) being generally approximate, 2) exact, and 3) providing a solution with a numerical accuracy that can be controlled.

Models with any number of output quantities

When the measurement model is multivariate, that is, it has any number of output quantities, the above concepts can be extended. [13] The output quantities are now described by a joint probability distribution, the coverage interval becomes a coverage region, the law of propagation of uncertainty has a natural generalization, and a calculation procedure that implements a multivariate Monte Carlo method is available.

Uncertainty as an interval

The most common view of measurement uncertainty uses random variables as mathematical models for uncertain quantities and simple probability distributions as sufficient for representing measurement uncertainties. In some situations, however, a mathematical interval might be a better model of uncertainty than a probability distribution. This may include situations involving periodic measurements, binned data values, censoring, detection limits, or plus-minus ranges of measurements where no particular probability distribution seems justified or where one cannot assume that the errors among individual measurements are completely independent.[ citation needed ]

A more robust representation of measurement uncertainty in such cases can be fashioned from intervals. [14] [15] An interval [a, b] is different from a rectangular or uniform probability distribution over the same range in that the latter suggests that the true value lies inside the right half of the range [(a + b)/2, b] with probability one half, and within any subinterval of [a, b] with probability equal to the width of the subinterval divided by b  a. The interval makes no such claims, except simply that the measurement lies somewhere within the interval. Distributions of such measurement intervals can be summarized as probability boxes and Dempster–Shafer structures over the real numbers, which incorporate both aleatoric and epistemic uncertainties.

See also

Related Research Articles

<span class="mw-page-title-main">Entropy (information theory)</span> Expected amount of information needed to specify the output of a stochastic data source

In information theory, the entropy of a random variable is the average level of "information", "surprise", or "uncertainty" inherent to the variable's possible outcomes. Given a discrete random variable , which takes values in the alphabet and is distributed according to :

<span class="mw-page-title-main">Probability distribution</span> Mathematical function for the probability a given outcome occurs in an experiment

In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon in terms of its sample space and the probabilities of events.

A parameter, generally, is any characteristic that can help in defining or classifying a particular system. That is, a parameter is an element of a system that is useful, or critical, when identifying the system, or when evaluating its performance, status, condition, etc.

<span class="mw-page-title-main">Standard deviation</span> In statistics, a measure of variation

In statistics, the standard deviation is a measure of the amount of variation of a random variable expected about its mean. A low standard deviation indicates that the values tend to be close to the mean of the set, while a high standard deviation indicates that the values are spread out over a wider range.

<span class="mw-page-title-main">Uncertainty</span> Situations involving imperfect or unknown information

Uncertainty or Incertitude refers to epistemic situations involving imperfect or unknown information. It applies to predictions of future events, to physical measurements that are already made, or to the unknown. Uncertainty arises in partially observable or stochastic environments, as well as due to ignorance, indolence, or both. It arises in any number of fields, including insurance, philosophy, physics, statistics, economics, finance, medicine, psychology, sociology, engineering, metrology, meteorology, ecology and information science.

<span class="mw-page-title-main">Metrology</span> Science of measurement and its application

Metrology is the scientific study of measurement. It establishes a common understanding of units, crucial in linking human activities. Modern metrology has its roots in the French Revolution's political motivation to standardise units in France when a length standard taken from a natural source was proposed. This led to the creation of the decimal-based metric system in 1795, establishing a set of standards for other types of measurements. Several other countries adopted the metric system between 1795 and 1875; to ensure conformity between the countries, the Bureau International des Poids et Mesures (BIPM) was established by the Metre Convention. This has evolved into the International System of Units (SI) as a result of a resolution at the 11th General Conference on Weights and Measures (CGPM) in 1960.

In statistics, interval estimation is the use of sample data to estimate an interval of possible values of a parameter of interest. This is in contrast to point estimation, which gives a single value.

A Bayesian network is a probabilistic graphical model that represents a set of variables and their conditional dependencies via a directed acyclic graph (DAG). While it is one of several forms of causal notation, causal networks are special cases of Bayesian networks. Bayesian networks are ideal for taking an event that occurred and predicting the likelihood that any one of several possible known causes was the contributing factor. For example, a Bayesian network could represent the probabilistic relationships between diseases and symptoms. Given symptoms, the network can be used to compute the probabilities of the presence of various diseases.

<span class="mw-page-title-main">Confidence interval</span> Range to estimate an unknown parameter

Informally, in frequentist statistics, a confidence interval (CI) is an interval which is expected to typically contain the parameter being estimated. More specifically, given a confidence level , a CI is a random interval which contains the parameter being estimated % of the time. The confidence level, degree of confidence or confidence coefficient represents the long-run proportion of CIs that theoretically contain the true value of the parameter; this is tantamount to the nominal coverage probability. For example, out of all intervals computed at the 95% level, 95% of them should contain the parameter's true value.

In statistical inference, specifically predictive inference, a prediction interval is an estimate of an interval in which a future observation will fall, with a certain probability, given what has already been observed. Prediction intervals are often used in regression analysis.

<span class="mw-page-title-main">Standard error</span> Statistical property

The standard error (SE) of a statistic is the standard deviation of its sampling distribution or an estimate of that standard deviation. If the statistic is the sample mean, it is called the standard error of the mean (SEM). The standard error is a key ingredient in producing confidence intervals.

In statistics, propagation of uncertainty is the effect of variables' uncertainties on the uncertainty of a function based on them. When the variables are the values of experimental measurements they have uncertainties due to measurement limitations which propagate due to the combination of variables in the function.

Significance arithmetic is a set of rules for approximating the propagation of uncertainty in scientific or statistical calculations. These rules can be used to find the appropriate number of significant figures to use to represent the result of a calculation. If a calculation is done without analysis of the uncertainty involved, a result that is written with too many significant figures can be taken to imply a higher precision than is known, and a result that is written with too few significant figures results in an avoidable loss of precision. Understanding these rules requires a good understanding of the concept of significant and insignificant figures.

Weighted least squares (WLS), also known as weighted linear regression, is a generalization of ordinary least squares and linear regression in which knowledge of the unequal variance of observations (heteroscedasticity) is incorporated into the regression. WLS is also a specialization of generalized least squares, when all the off-diagonal entries of the covariance matrix of the errors, are null.

Uncertainty quantification (UQ) is the science of quantitative characterization and estimation of uncertainties in both computational and real world applications. It tries to determine how likely certain outcomes are if some aspects of the system are not exactly known. An example would be to predict the acceleration of a human body in a head-on crash with another car: even if the speed was exactly known, small differences in the manufacturing of individual cars, how tightly every bolt has been tightened, etc., will lead to different results that can only be predicted in a statistical sense.

In statistics, a pivotal quantity or pivot is a function of observations and unobservable parameters such that the function's probability distribution does not depend on the unknown parameters. A pivot quantity need not be a statistic—the function and its value can depend on the parameters of the model, but its distribution must not. If it is a statistic, then it is known as an ancillary statistic.

<span class="mw-page-title-main">Probabilistic design</span> Discipline within engineering design

Probabilistic design is a discipline within engineering design. It deals primarily with the consideration and minimization of the effects of random variability upon the performance of an engineering system during the design phase. Typically, these effects studied and optimized are related to quality and reliability. It differs from the classical approach to design by assuming a small probability of failure instead of using the safety factor. Probabilistic design is used in a variety of different applications to assess the likelihood of failure. Disciplines which extensively use probabilistic design principles include product design, quality control, systems engineering, machine design, civil engineering and manufacturing.

Industrial process data validation and reconciliation, or more briefly, process data reconciliation (PDR), is a technology that uses process information and mathematical methods in order to automatically ensure data validation and reconciliation by correcting measurements in industrial processes. The use of PDR allows for extracting accurate and reliable information about the state of industry processes from raw measurement data and produces a single consistent set of data representing the most likely process operation.

<span class="mw-page-title-main">Probability box</span> Characterization of uncertain numbers consisting of both aleatoric and epistemic uncertainties

A probability box is a characterization of uncertain numbers consisting of both aleatoric and epistemic uncertainties that is often used in risk analysis or quantitative uncertainty modeling where numerical calculations must be performed. Probability bounds analysis is used to make arithmetic and logical calculations with p-boxes.

The Joint Committee for Guides in Metrology (JCGM) is an organization in Sèvres that prepared the Guide to the Expression of Uncertainty in Measurement (GUM) and the International Vocabulary of Metrology (VIM). The JCGM assumed responsibility for these two documents from the ISO Technical Advisory Group 4 (TAG4).

References

  1. 1 2 3 JCGM 100:2008. Evaluation of measurement data – Guide to the expression of uncertainty in measurement, Joint Committee for Guides in Metrology.
  2. Bell, S. Measurement Good Practice Guide No. 11. A Beginner's Guide to Uncertainty of Measurement. Tech. rep., National Physical Laboratory, 1999.
  3. ASME B89.7.3.1, Guidelines for Decision Rules in Determining Conformance to Specifications
  4. ASME B89.7.3.2, Guidelines for the Evaluation of Dimensional Measurement Uncertainty
  5. ASME B89.7.3.3, Guidelines for Assessing the Reliability of Dimensional Measurement Uncertainty Statements
  6. ASME B89.7.4, Measurement Uncertainty and Conformance Testing: Risk Analysis
  7. 1 2 3 4 JCGM 101:2008. Evaluation of measurement data – Supplement 1 to the "Guide to the expression of uncertainty in measurement" – Propagation of distributions using a Monte Carlo method. Joint Committee for Guides in Metrology.
  8. Bernardo, J., and Smith, A. "Bayesian Theory". John Wiley & Sons, New York, USA, 2000. 3.20
  9. Elster, Clemens (2007). "Calculation of uncertainty in the presence of prior knowledge". Metrologia. 44 (2): 111–116. Bibcode:2007Metro..44..111E. doi:10.1088/0026-1394/44/2/002. S2CID   123445853.
  10. EURACHEM/CITAC. "Quantifying uncertainty in analytical measurement". Tech. Rep. Guide CG4, EU-RACHEM/CITEC, EURACHEM/CITAC Guide], 2000. Second edition.
  11. 1 2 JCGM 104:2009. Evaluation of measurement data – An introduction to the "Guide to the expression of uncertainty in measurement" and related documents. Joint Committee for Guides in Metrology.
  12. Weise, K.; Woger, W. (1993). "A Bayesian theory of measurement uncertainty". Measurement Science and Technology. 4 (1): 1–11. Bibcode:1993MeScT...4....1W. doi:10.1088/0957-0233/4/1/001. S2CID   250751314.
  13. Joint Committee for Guides in Metrology (2011). JCGM 102: Evaluation of Measurement Data – Supplement 2 to the "Guide to the Expression of Uncertainty in Measurement" – Extension to Any Number of Output Quantities (PDF) (Technical report). JCGM. Retrieved 13 February 2013.
  14. Manski, C.F. (2003); Partial Identification of Probability Distributions, Springer Series in Statistics, Springer, New York
  15. Ferson, S., V. Kreinovich, J. Hajagos, W. Oberkampf, and L. Ginzburg (2007); Experimental Uncertainty Estimation and Statistics for Data Having Interval Uncertainty, Sandia National Laboratories SAND 2007-0939

Further reading