Reference range

Last updated

In medicine and health-related fields, a reference range or reference interval is the range or the interval of values that is deemed normal for a physiological measurement in healthy persons (for example, the amount of creatinine in the blood, or the partial pressure of oxygen). It is a basis for comparison for a physician or other health professional to interpret a set of test results for a particular patient. Some important reference ranges in medicine are reference ranges for blood tests and reference ranges for urine tests.

Contents

The standard definition of a reference range (usually referred to if not otherwise specified) originates in what is most prevalent in a reference group taken from the general (i.e. total) population. This is the general reference range. However, there are also optimal health ranges (ranges that appear to have the optimal health impact) and ranges for particular conditions or statuses (such as pregnancy reference ranges for hormone levels).

Values within the reference range (WRR) are those within normal limits (WNL). The limits are called the upper reference limit (URL) or upper limit of normal (ULN) and the lower reference limit (LRL) or lower limit of normal (LLN). In health care–related publishing, style sheets sometimes prefer the word reference over the word normal to prevent the nontechnical senses of normal from being conflated with the statistical sense. Values outside a reference range are not necessarily pathologic, and they are not necessarily abnormal in any sense other than statistically. Nonetheless, they are indicators of probable pathosis. Sometimes the underlying cause is obvious; in other cases, challenging differential diagnosis is required to determine what is wrong and thus how to treat it.

A cutoff or threshold is a limit used for binary classification, mainly between normal versus pathological (or probably pathological). Establishment methods for cutoffs include using an upper or a lower limit of a reference range.

Standard definition

The standard definition of a reference range for a particular measurement is defined as the interval between which 95% of values of a reference population fall into, in such a way that 2.5% of the time a value will be less than the lower limit of this interval, and 2.5% of the time it will be larger than the upper limit of this interval, whatever the distribution of these values. [1]

Reference ranges that are given by this definition are sometimes referred as standard ranges.

Since a range is a defined statistical value (Range (statistics)) that describes the interval between the smallest and largest values, many, including the International Federation of Clinical Chemistry prefer to use the expression reference interval rather than reference range. [2]

Regarding the target population, if not otherwise specified, a standard reference range generally denotes the one in healthy individuals, or without any known condition that directly affects the ranges being established. These are likewise established using reference groups from the healthy population, and are sometimes termed normal ranges or normal values (and sometimes "usual" ranges/values). However, using the term normal may not be appropriate as not everyone outside the interval is abnormal, and people who have a particular condition may still fall within this interval.

However, reference ranges may also be established by taking samples from the whole population, with or without diseases and conditions. In some cases, diseased individuals are taken as the population, establishing reference ranges among those having a disease or condition. Preferably, there should be specific reference ranges for each subgroup of the population that has any factor that affects the measurement, such as, for example, specific ranges for each sex, age group, race or any other general determinant.

Establishment methods

Methods for establishing reference ranges can be based on assuming a normal distribution or a log-normal distribution, or directly from percentages of interest, as detailed respectively in following sections. When establishing reference ranges from bilateral organs (e.g., vision or hearing), both results from the same individual can be used, although intra-subject correlation must be taken into account. [3]

Normal distribution

When assuming a normal distribution, the reference range is obtained by measuring the values in a reference group and taking two standard deviations either side of the mean. This encompasses ~95% of the total population. Standard deviation diagram.svg
When assuming a normal distribution, the reference range is obtained by measuring the values in a reference group and taking two standard deviations either side of the mean. This encompasses ~95% of the total population.

The 95% interval, is often estimated by assuming a normal distribution of the measured parameter, in which case it can be defined as the interval limited by 1.96 [4] (often rounded up to 2) population standard deviations from either side of the population mean (also called the expected value). However, in the real world, neither the population mean nor the population standard deviation are known. They both need to be estimated from a sample, whose size can be designated n. The population standard deviation is estimated by the sample standard deviation and the population mean is estimated by the sample mean (also called mean or arithmetic mean). To account for these estimations, the 95% prediction interval (95% PI) is calculated as:

95% PI = mean ± t0.975,n1·(n+1)/n·sd,

where is the 97.5% quantile of a Student's t-distribution with n1 degrees of freedom.

When the sample size is large (n≥30)

This method is often acceptably accurate if the standard deviation, as compared to the mean, is not very large. A more accurate method is to perform the calculations on logarithmized values, as described in separate section later.

The following example of this (not logarithmized) method is based on values of fasting plasma glucose taken from a reference group of 12 subjects: [5]

Fasting plasma glucose
(FPG)
in mmol/L
Deviation from
mean m
Squared deviation
from mean m
Subject 15.50.170.029
Subject 25.2-0.130.017
Subject 35.2-0.130.017
Subject 45.80.470.221
Subject 55.60.270.073
Subject 64.6-0.730.533
Subject 75.60.270.073
Subject 85.90.570.325
Subject 94.7-0.630.397
Subject 105-0.330.109
Subject 115.70.370.137
Subject 125.2-0.130.017
Mean = 5.33 (m)
n=12
Mean = 0.00Sum/(n1) = 1.95/11 =0.18

= standard deviation (s.d.)

As can be given from, for example, a table of selected values of Student's t-distribution, the 97.5% percentile with (12-1) degrees of freedom corresponds to

Subsequently, the lower and upper limits of the standard reference range are calculated as:

Thus, the standard reference range for this example is estimated to be 4.4 to 6.3 mmol/L.

Confidence interval of limit

The 90% confidence interval of a standard reference range limit as estimated assuming a normal distribution can be calculated by: [6]

Lower limit of the confidence interval = percentile limit - 2.81 × SDn
Upper limit of the confidence interval = percentile limit + 2.81 × SDn,

where SD is the standard deviation, and n is the number of samples.

Taking the example from the previous section, the number of samples is 12 and the standard deviation is 0.42 mmol/L, resulting in:

Lower limit of the confidence interval of the lower limit of the standard reference range = 4.4 - 2.81 × 0.4212 ≈ 4.1
Upper limit of the confidence interval of the lower limit of the standard reference range = 4.4 + 2.81 × 0.4212 ≈ 4.7

Thus, the lower limit of the reference range can be written as 4.4 (90% CI 4.1–4.7) mmol/L.

Likewise, with similar calculations, the upper limit of the reference range can be written as 6.3 (90% CI 6.0–6.6) mmol/L.

These confidence intervals reflect random error, but do not compensate for systematic error, which in this case can arise from, for example, the reference group not having fasted long enough before blood sampling.

As a comparison, actual reference ranges used clinically for fasting plasma glucose are estimated to have a lower limit of approximately 3.8 [7] to 4.0, [8] and an upper limit of approximately 6.0 [8] to 6.1. [9]

Log-normal distribution

Some functions of log-normal distribution (here shown with the measurements non-logarithmized), with the same means - m (as calculated after logarithmizing) but different standard deviations - s (after logarithmizing) PDF-log normal distributions.svg
Some functions of log-normal distribution (here shown with the measurements non-logarithmized), with the same means - μ (as calculated after logarithmizing) but different standard deviations - σ (after logarithmizing)

In reality, biological parameters tend to have a log-normal distribution, [10] rather than the normal distribution or Gaussian distribution.

An explanation for this log-normal distribution for biological parameters is: The event where a sample has half the value of the mean or median tends to have almost equal probability to occur as the event where a sample has twice the value of the mean or median. Also, only a log-normal distribution can compensate for the inability of almost all biological parameters to be of negative numbers (at least when measured on absolute scales), with the consequence that there is no definite limit to the size of outliers (extreme values) on the high side, but, on the other hand, they can never be less than zero, resulting in a positive skewness.

As shown in diagram at right, this phenomenon has relatively small effect if the standard deviation (as compared to the mean) is relatively small, as it makes the log-normal distribution appear similar to a normal distribution. Thus, the normal distribution may be more appropriate to use with small standard deviations for convenience, and the log-normal distribution with large standard deviations.

In a log-normal distribution, the geometric standard deviations and geometric mean more accurately estimate the 95% prediction interval than their arithmetic counterparts.

Necessity

Reference ranges for substances that are usually within relatively narrow limits (coefficient of variation less than 0.213, as detailed below) such as electrolytes can be estimated by assuming normal distribution, whereas reference ranges for those that vary significantly (coefficient of variation generally over 0.213) such as most hormones [11] are more accurately established by log-normal distribution.

The necessity to establish a reference range by log-normal distribution rather than normal distribution can be regarded as depending on how much difference it would make to not do so, which can be described as the ratio:

Difference ratio = | Limitlog-normal - Limitnormal|/ Limitlog-normal

where:

  • Limitlog-normal is the (lower or upper) limit as estimated by assuming log-normal distribution
  • Limitnormal is the (lower or upper) limit as estimated by assuming normal distribution.
Coefficient of variation versus deviation in reference ranges established by assuming normal distribution when there is actually a log-normal distribution. Diagram of coefficient of variation versus deviation in reference ranges erroneously not established by log-normal distribution.png
Coefficient of variation versus deviation in reference ranges established by assuming normal distribution when there is actually a log-normal distribution.

This difference can be put solely in relation to the coefficient of variation, as in the diagram at right, where:

Coefficient of variation = s.d./m

where:

  • s.d. is the standard deviation
  • m is the arithmetic mean

In practice, it can be regarded as necessary to use the establishment methods of a log-normal distribution if the difference ratio becomes more than 0.1, meaning that a (lower or upper) limit estimated from an assumed normal distribution would be more than 10% different from the corresponding limit as estimated from a (more accurate) log-normal distribution. As seen in the diagram, a difference ratio of 0.1 is reached for the lower limit at a coefficient of variation of 0.213 (or 21.3%), and for the upper limit at a coefficient of variation at 0.413 (41.3%). The lower limit is more affected by increasing coefficient of variation, and its "critical" coefficient of variation of 0.213 corresponds to a ratio of (upper limit)/(lower limit) of 2.43, so as a rule of thumb, if the upper limit is more than 2.4 times the lower limit when estimated by assuming normal distribution, then it should be considered to do the calculations again by log-normal distribution.

Taking the example from previous section, the standard deviation (s.d.) is estimated at 0.42 and the arithmetic mean (m) is estimated at 5.33. Thus the coefficient of variation is 0.079. This is less than both 0.213 and 0.413, and thus both the lower and upper limit of fasting blood glucose can most likely be estimated by assuming normal distribution. More specifically, the coefficient of variation of 0.079 corresponds to a difference ratio of 0.01 (1%) for the lower limit and 0.007 (0.7%) for the upper limit.

From logarithmized sample values

A method to estimate the reference range for a parameter with log-normal distribution is to logarithmize all the measurements with an arbitrary base (for example e), derive the mean and standard deviation of these logarithms, determine the logarithms located (for a 95% prediction interval) 1.96 standard deviations below and above that mean, and subsequently exponentiate using those two logarithms as exponents and using the same base as was used in logarithmizing, with the two resultant values being the lower and upper limit of the 95% prediction interval.

The following example of this method is based on the same values of fasting plasma glucose as used in the previous section, using e as a base: [5]

Fasting plasma glucose
(FPG)
in mmol/L
log e (FPG)loge(FPG) deviation from
mean μlog
Squared deviation
from mean
Subject 15.51.700.0290.000841
Subject 25.21.650.0210.000441
Subject 35.21.650.0210.000441
Subject 45.81.760.0890.007921
Subject 55.61.720.0490.002401
Subject 64.61.530.1410.019881
Subject 75.61.720.0490.002401
Subject 85.91.770.0990.009801
Subject 94.71.550.1210.014641
Subject 105.01.610.0610.003721
Subject 115.71.740.0690.004761
Subject 125.21.650.0210.000441
Mean: 5.33
(m)
Mean: 1.67
(μlog)
Sum/(n-1) : 0.068/11 = 0.0062

= standard deviation of loge(FPG)
(σlog)

Subsequently, the still logarithmized lower limit of the reference range is calculated as:

and the upper limit of the reference range as:

Conversion back to non-logarithmized values are subsequently performed as:

Thus, the standard reference range for this example is estimated to be 4.4 to 6.4.

From arithmetic mean and variance

An alternative method of establishing a reference range with the assumption of log-normal distribution is to use the arithmetic mean and standard deviation. This is somewhat more tedious to perform, but may be useful in cases where a study presents only the arithmetic mean and standard deviation, while leaving out the source data. If the original assumption of normal distribution is less appropriate than the log-normal one, then, using the arithmetic mean and standard deviation may be the only available parameters to determine the reference range.

By assuming that the expected value can represent the arithmetic mean in this case, the parameters μlog and σlog can be estimated from the arithmetic mean (m) and standard deviation (s.d.) as:

Following the exampled reference group from the previous section:

Subsequently, the logarithmized, and later non-logarithmized, lower and upper limit are calculated just as by logarithmized sample values.

Directly from percentages of interest

Reference ranges can also be established directly from the 2.5th and 97.5th percentile of the measurements in the reference group. For example, if the reference group consists of 200 people, and counting from the measurement with lowest value to highest, the lower limit of the reference range would correspond to the 5th measurement and the upper limit would correspond to the 195th measurement.

This method can be used even when measurement values do not appear to conform conveniently to any form of normal distribution or other function.

However, the reference range limits as estimated in this way have higher variance, and therefore less reliability, than those estimated by an arithmetic or log-normal distribution (when such is applicable), because the latter ones acquire statistical power from the measurements of the whole reference group rather than just the measurements at the 2.5th and 97.5th percentiles. Still, this variance decreases with increasing size of the reference group, and therefore, this method may be optimal where a large reference group easily can be gathered, and the distribution mode of the measurements is uncertain.

Bimodal distribution

Bimodal distribution Bimodal.png
Bimodal distribution

In case of a bimodal distribution (seen at right), it is useful to find out why this is the case. Two reference ranges can be established for the two different groups of people, making it possible to assume a normal distribution for each group. This bimodal pattern is commonly seen in tests that differ between men and women, such as prostate specific antigen.

Interpretation of standard ranges in medical tests

In case of medical tests whose results are of continuous values, reference ranges can be used in the interpretation of an individual test result. This is primarily used for diagnostic tests and screening tests, while monitoring tests may optimally be interpreted from previous tests of the same individual instead.

Probability of random variability

Reference ranges aid in the evaluation of whether a test result's deviation from the mean is a result of random variability or a result of an underlying disease or condition. If the reference group used to establish the reference range can be assumed to be representative of the individual person in a healthy state, then a test result from that individual that turns out to be lower or higher than the reference range can be interpreted as that there is less than 2.5% probability that this would have occurred by random variability in the absence of disease or other condition, which, in turn, is strongly indicative for considering an underlying disease or condition as a cause.

Such further consideration can be performed, for example, by an epidemiology-based differential diagnostic procedure, where potential candidate conditions are listed that may explain the finding, followed by calculations of how probable they are to have occurred in the first place, in turn followed by a comparison with the probability that the result would have occurred by random variability.

If the establishment of the reference range could have been made assuming a normal distribution, then the probability that the result would be an effect of random variability can be further specified as follows:

The standard deviation, if not given already, can be inversely calculated by the fact that the absolute value of the difference between the mean and either the upper or lower limit of the reference range is approximately 2 standard deviations (more accurately 1.96), and thus:

Standard deviation (s.d.) ≈ | (Mean) - (Upper limit) |/2.

The standard score for the individual's test can subsequently be calculated as:

Standard score (z) = | (Mean) - (individual measurement) |/s.d..

The probability that a value is of a certain distance from the mean can subsequently be calculated from the relation between standard score and prediction intervals. For example, a standard score of 2.58 corresponds to a prediction interval of 99%, [12] corresponding to a probability of 0.5% that a result is at least such far from the mean in the absence of disease.

Example

Let's say, for example, that an individual takes a test that measures the ionized calcium in the blood, resulting in a value of 1.30 mmol/L, and a reference group that appropriately represents the individual has established a reference range of 1.05 to 1.25 mmol/L. The individual's value is higher than the upper limit of the reference range, and therefore has less than 2.5% probability of being a result of random variability, constituting a strong indication to make a differential diagnosis of possible causative conditions.

In this case, an epidemiology-based differential diagnostic procedure is used, and its first step is to find candidate conditions that can explain the finding.

Hypercalcemia (usually defined as a calcium level above the reference range) is mostly caused by either primary hyperparathyroidism or malignancy, [13] and therefore, it is reasonable to include these in the differential diagnosis.

Using for example epidemiology and the individual's risk factors, let's say that the probability that the hypercalcemia would have been caused by primary hyperparathyroidism in the first place is estimated to be 0.00125 (or 0.125%), the equivalent probability for cancer is 0.0002, and 0.0005 for other conditions. With a probability given as less than 0.025 of no disease, this corresponds to a probability that the hypercalcemia would have occurred in the first place of up to 0.02695. However, the hypercalcemia has occurred with a probability of 100%, resulting adjusted probabilities of at least 4.6% that primary hyperparathyroidism has caused the hypercalcemia, at least 0.7% for cancer, at least 1.9% for other conditions and up to 92.8% for that there is no disease and the hypercalcemia is caused by random variability.

In this case, further processing benefits from specification of the probability of random variability:

The value is assumed to conform acceptably to a normal distribution, so the mean can be assumed to be 1.15 in the reference group. The standard deviation, if not given already, can be inversely calculated by knowing that the absolute value of the difference between the mean and, for example, the upper limit of the reference range, is approximately 2 standard deviations (more accurately 1.96), and thus:

Standard deviation (s.d.) ≈ | (Mean) - (Upper limit) |/2 = | 1.15 - 1.25 |/2 = 0.1/2 = 0.05.

The standard score for the individual's test is subsequently calculated as:

Standard score (z) = | (Mean) - (individual measurement) |/s.d. = | 1.15 - 1.30 |/0.05 = 0.15/0.05 = 3.

The probability that a value is of so much larger value than the mean as having a standard score of 3 corresponds to a probability of approximately 0.14% (given by (100% 99.7%)/2, with 99.7% here being given from the 68–95–99.7 rule).

Using the same probabilities that the hypercalcemia would have occurred in the first place by the other candidate conditions, the probability that hypercalcemia would have occurred in the first place is 0.00335, and given the fact that hypercalcemia has occurred gives adjusted probabilities of 37.3%, 6.0%, 14.9% and 41.8%, respectively, for primary hyperparathyroidism, cancer, other conditions and no disease.

Optimal health range

Optimal (health) range or therapeutic target (not to be confused with biological target) is a reference range or limit that is based on concentrations or levels that are associated with optimal health or minimal risk of related complications and diseases, rather than the standard range based on normal distribution in the population.

It may be more appropriate to use for e.g. folate, since approximately 90 percent of North Americans may actually suffer more or less from folate deficiency, [14] but only the 2.5 percent that have the lowest levels will fall below the standard reference range. In this case, the actual folate ranges for optimal health are substantially higher than the standard reference ranges. Vitamin D has a similar tendency. In contrast, for e.g. uric acid, having a level not exceeding the standard reference range still does not exclude the risk of getting gout or kidney stones. Furthermore, for most toxins, the standard reference range is generally lower than the level of toxic effect.

A problem with optimal health range is a lack of a standard method of estimating the ranges. The limits may be defined as those where the health risks exceed a certain threshold, but with various risk profiles between different measurements (such as folate and vitamin D), and even different risk aspects for one and the same measurement (such as both deficiency and toxicity of vitamin A) it is difficult to standardize. Subsequently, optimal health ranges, when given by various sources, have an additional variability caused by various definitions of the parameter. Also, as with standard reference ranges, there should be specific ranges for different determinants that affects the values, such as sex, age etc. Ideally, there should rather be an estimation of what is the optimal value for every individual, when taking all significant factors of that individual into account - a task that may be hard to achieve by studies, but long clinical experience by a physician may make this method preferable to using reference ranges.

One-sided cut-off values

In many cases, only one side of the range is usually of interest, such as with markers of pathology including cancer antigen 19-9, where it is generally without any clinical significance to have a value below what is usual in the population. Therefore, such targets are often given with only one limit of the reference range given, and, strictly, such values are rather cut-off values or threshold values.

They may represent both standard ranges and optimal health ranges. Also, they may represent an appropriate value to distinguish healthy person from a specific disease, although this gives additional variability by different diseases being distinguished. For example, for NT-proBNP, a lower cut-off value is used in distinguishing healthy babies from those with acyanotic heart disease, compared to the cut-off value used in distinguishing healthy babies from those with congenital nonspherocytic anemia. [15]

General drawbacks

For standard as well as optimal health ranges, and cut-offs, sources of inaccuracy and imprecision include:

Also, reference ranges tend to give the impression of definite thresholds that clearly separate "good" or "bad" values, while in reality there are generally continuously increasing risks with increased distance from usual or optimal values.

With this and uncompensated factors in mind, the ideal interpretation method of a test result would rather consist of a comparison of what would be expected or optimal in the individual when taking all factors and conditions of that individual into account, rather than strictly classifying the values as "good" or "bad" by using reference ranges from other people.

In a recent paper, Rappoport et al. [16] described a novel way to redefine reference range from an electronic health record system. In such a system, a higher population resolution can be achieved (e.g., age, sex, race and ethnicity-specific).

Examples

See also

Related Research Articles

A histogram is a visual representation of the distribution of quantitative data. The term was first introduced by Karl Pearson. To construct a histogram, the first step is to "bin" the range of values— divide the entire range of values into a series of intervals—and then count how many values fall into each interval. The bins are usually specified as consecutive, non-overlapping intervals of a variable. The bins (intervals) are adjacent and are typically of equal size.

<span class="mw-page-title-main">Logarithm</span> Mathematical function, inverse of an exponential function

In mathematics, the logarithm is the inverse function to exponentiation. That means that the logarithm of a number x to the base b is the exponent to which b must be raised to produce x. For example, since 1000 = 103, the logarithm base  of 1000 is 3, or log10 (1000) = 3. The logarithm of x to base b is denoted as logb (x), or without parentheses, logbx. When the base is clear from the context or is irrelevant it is sometimes written log x.

<span class="mw-page-title-main">Normal distribution</span> Probability distribution

In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is The parameter is the mean or expectation of the distribution, while the parameter is the variance. The standard deviation of the distribution is . A random variable with a Gaussian distribution is said to be normally distributed, and is called a normal deviate.

A parameter, generally, is any characteristic that can help in defining or classifying a particular system. That is, a parameter is an element of a system that is useful, or critical, when identifying the system, or when evaluating its performance, status, condition, etc.

<span class="mw-page-title-main">Standard deviation</span> In statistics, a measure of variation

In statistics, the standard deviation is a measure of the amount of variation of a random variable expected about its mean. A low standard deviation indicates that the values tend to be close to the mean of the set, while a high standard deviation indicates that the values are spread out over a wider range. The standard deviation is commonly used in the determination of what constitutes an outlier and what does not.

<span class="mw-page-title-main">Box–Muller transform</span> Statistical transform

The Box–Muller transform, by George Edward Pelham Box and Mervin Edgar Muller, is a random number sampling method for generating pairs of independent, standard, normally distributed random numbers, given a source of uniformly distributed random numbers. The method was first mentioned explicitly by Raymond E. A. C. Paley and Norbert Wiener in their 1934 treatise on Fourier transforms in the complex domain. Given the status of these latter authors and the widespread availability and use of their treatise, it is almost certain that Box and Muller were well aware of its contents.

<span class="mw-page-title-main">Log-normal distribution</span> Probability distribution

In probability theory, a log-normal (or lognormal) distribution is a continuous probability distribution of a random variable whose logarithm is normally distributed. Thus, if the random variable X is log-normally distributed, then Y = ln(X) has a normal distribution. Equivalently, if Y has a normal distribution, then the exponential function of Y, X = exp(Y), has a log-normal distribution. A random variable that is log-normally distributed takes only positive real values. It is a convenient and useful model for measurements in the natural sciences, engineering, as well as medicine, economics and other fields. It can be applied to diverse quantities such as energies, concentrations, lengths, prices of financial instruments, and other metrics, while acknowledging the inherent uncertainty in all measurements.

In statistics, point estimation involves the use of sample data to calculate a single value which is to serve as a "best guess" or "best estimate" of an unknown population parameter. More formally, it is the application of a point estimator to the data to obtain a point estimate.

In statistics, interval estimation is the use of sample data to estimate an interval of possible values of a parameter of interest. This is in contrast to point estimation, which gives a single value.

<span class="mw-page-title-main">Beta distribution</span> Probability distribution

In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval [0, 1] or in terms of two positive parameters, denoted by alpha (α) and beta (β), that appear as exponents of the variable and its complement to 1, respectively, and control the shape of the distribution.

<i>Z</i>-test Statistical test

A Z-test is any statistical test for which the distribution of the test statistic under the null hypothesis can be approximated by a normal distribution. Z-test tests the mean of a distribution. For each significance level in the confidence interval, the Z-test has a single critical value which makes it more convenient than the Student's t-test whose critical values are defined by the sample size. Both the Z-test and Student's t-test have similarities in that they both help determine the significance of a set of data. However, the z-test is rarely used in practice because the population deviation is difficult to determine.

In probability theory and statistics, the geometric standard deviation (GSD) describes how spread out are a set of numbers whose preferred average is the geometric mean. For such data, it may be preferred to the more usual standard deviation. Note that unlike the usual arithmetic standard deviation, the geometric standard deviation is a multiplicative factor, and thus is dimensionless, rather than having the same dimension as the input values. Thus, the geometric standard deviation may be more appropriately called geometric SD factor. When using geometric SD factor in conjunction with geometric mean, it should be described as "the range from to, and one cannot add/subtract "geometric SD factor" to/from geometric mean.

In statistical inference, specifically predictive inference, a prediction interval is an estimate of an interval in which a future observation will fall, with a certain probability, given what has already been observed. Prediction intervals are often used in regression analysis.

<span class="mw-page-title-main">Standard error</span> Statistical property

The standard error (SE) of a statistic is the standard deviation of its sampling distribution or an estimate of that standard deviation. If the statistic is the sample mean, it is called the standard error of the mean (SEM). The standard error is a key ingredient in producing confidence intervals.

In probability theory and statistics, the coefficient of variation (CV), also known as normalized root-mean-square deviation (NRMSD), percent RMS, and relative standard deviation (RSD), is a standardized measure of dispersion of a probability distribution or frequency distribution. It is defined as the ratio of the standard deviation to the mean , and often expressed as a percentage ("%RSD"). The CV or RSD is widely used in analytical chemistry to express the precision and repeatability of an assay. It is also commonly used in fields such as engineering or physics when doing quality assurance studies and ANOVA gauge R&R, by economists and investors in economic models, and in psychology/neuroscience.

In statistics, the mode is the value that appears most often in a set of data values. If X is a discrete random variable, the mode is the value x at which the probability mass function takes its maximum value. In other words, it is the value that is most likely to be sampled.

In statistics and information theory, a maximum entropy probability distribution has entropy that is at least as great as that of all other members of a specified class of probability distributions. According to the principle of maximum entropy, if nothing is known about a distribution except that it belongs to a certain class, then the distribution with the largest entropy should be chosen as the least-informative default. The motivation is twofold: first, maximizing entropy minimizes the amount of prior information built into the distribution; second, many physical systems tend to move towards maximal entropy configurations over time.

This glossary of statistics and probability is a list of definitions of terms and concepts used in the mathematical sciences of statistics and probability, their sub-disciplines, and related fields. For additional related terms, see Glossary of mathematics and Glossary of experimental design.

In statistics, the frequency or absolute frequency of an event is the number of times the observation has occurred/recorded in an experiment or study. These frequencies are often depicted graphically or in tabular form.

<span class="mw-page-title-main">Cumulative frequency analysis</span>

Cumulative frequency analysis is the analysis of the frequency of occurrence of values of a phenomenon less than a reference value. The phenomenon may be time- or space-dependent. Cumulative frequency is also called frequency of non-exceedance.

References

Open Access logo PLoS transparent.svg This article was adapted from the following source under a CC0 license (2012) (reviewer reports): Mikael Häggström (2014). "Reference ranges for estradiol, progesterone, luteinizing hormone and follicle-stimulating hormone during the menstrual cycle" (PDF). WikiJournal of Medicine. 1 (1). doi: 10.15347/WJM/2014.001 . ISSN   2002-4436. Wikidata   Q44275619.

  1. Page 19 in: Stephen K. Bangert MA MB BChir MSc MBA FRCPath; William J. Marshall MA MSc MBBS FRCP FRCPath FRCPEdin FIBiol; Marshall, William Leonard (2008). Clinical biochemistry: metabolic and clinical aspects. Philadelphia: Churchill Livingstone/Elsevier. ISBN   978-0-443-10186-1.{{cite book}}: CS1 maint: multiple names: authors list (link)
  2. Dybkaer, R (November 1982). "International federation of clinical chemistry (IFCC)1),2) the theory of reference values. Part 6. Presentation of observed values related to reference values". Journal of Clinical Chemistry and Clinical Biochemistry. 20 (11): 841–5. PMID   7153721.
  3. Davis, C.Q.; Hamilton, R. (2021). "Reference ranges for clinical electrophysiology of vision". Doc Ophthalmol. 143 (2): 155–170. doi: 10.1007/s10633-021-09831-1 . PMC   8494724 . PMID   33880667.
  4. Page 48 in: Sterne, Jonathan; Kirkwood, Betty R. (2003). Essential medical statistics . Oxford: Blackwell Science. ISBN   978-0-86542-871-3.
  5. 1 2 Table 1. Subject characteristics in: Keevil, B. G.; Kilpatrick, E. S.; Nichols, S. P.; Maylor, P. W. (1998). "Biological variation of cystatin C: Implications for the assessment of glomerular filtration rate". Clinical Chemistry. 44 (7): 1535–1539. doi: 10.1093/clinchem/44.7.1535 . PMID   9665434.
  6. Page 65 in: Carl A. Burtis, David E. Bruns (2014). Tietz Fundamentals of Clinical Chemistry and Molecular Diagnostics (7 ed.). Elsevier Health Sciences. ISBN   9780323292061.
  7. Last page of Deepak A. Rao; Le, Tao; Bhushan, Vikas (2007). First Aid for the USMLE Step 1 2008 (First Aid for the Usmle Step 1) . McGraw-Hill Medical. ISBN   978-0-07-149868-5.
  8. 1 2 Reference range list from Uppsala University Hospital ("Laborationslista"). Artnr 40284 Sj74a. Issued on April 22, 2008
  9. MedlinePlus Encyclopedia : Glucose tolerance test
  10. Huxley, Julian S. (1932). Problems of relative growth. London. ISBN   978-0-486-61114-3. OCLC   476909537.
  11. Levitt H, Smith KG, Rosner MH (2009). "Variability in calcium, phosphorus, and parathyroid hormone in patients on hemodialysis". Hemodial Int. 13 (4): 518–25. doi:10.1111/j.1542-4758.2009.00393.x. PMID   19758299. S2CID   24963421.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  12. Page 111 in: Kirkup, Les (2002). Data analysis with Excel: an introduction for physical scientists. Cambridge, UK: Cambridge University Press. ISBN   978-0-521-79737-5.
  13. Table 20-4 in: Mitchell, Richard Sheppard; Kumar, Vinay; Abbas, Abul K.; Fausto, Nelson (2007). Robbins Basic Pathology. Philadelphia: Saunders. ISBN   978-1-4160-2973-1. 8th edition.
  14. Folic Acid: Don't Be Without It! by Hans R. Larsen, MSc ChE, retrieved on July 7, 2009. In turn citing:
    • Boushey Carol J.; et al. (1995). "A quantitative assessment of plasma homocysteine as a risk factor for vascular disease". Journal of the American Medical Association. 274 (13): 1049–57. doi:10.1001/jama.274.13.1049.
    • Morrison Howard I.; et al. (1996). "Serum folate and risk of fatal coronary heart disease". Journal of the American Medical Association. 275 (24): 1893–96. doi:10.1001/jama.1996.03530480035037. PMID   8648869.
  15. Screening for Congenital Heart Disease with NT-proBNP: Results By Emmanuel Jairaj Moses, Sharifah A.I. Mokhtar, Amir Hamzah, Basir Selvam Abdullah, and Narazah Mohd Yusoff. Laboratory Medicine. 2011;42(2):75–80. American Society for Clinical Pathology
  16. Rappoport, Nadav; Paik, Hyojung; Oskotsky, Boris; Tor, Ruth; Ziv, Elad; Zaitlen, Noah; Butte, Atul J. (2017-11-04). "Creating ethnicity-specific reference intervals for lab tests from EHR data". bioRxiv   10.1101/213892 .

Further reading