Confidence distribution

Last updated

In statistical inference, the concept of a confidence distribution (CD) has often been loosely referred to as a distribution function on the parameter space that can represent confidence intervals of all levels for a parameter of interest. Historically, it has typically been constructed by inverting the upper limits of lower sided confidence intervals of all levels, and it was also commonly associated with a fiducial [1] interpretation (fiducial distribution), although it is a purely frequentist concept. [2] A confidence distribution is NOT a probability distribution function of the parameter of interest, but may still be a function useful for making inferences. [3]

Contents

In recent years, there has been a surge of renewed interest in confidence distributions. [3] In the more recent developments, the concept of confidence distribution has emerged as a purely frequentist concept, without any fiducial interpretation or reasoning. Conceptually, a confidence distribution is no different from a point estimator or an interval estimator (confidence interval), but it uses a sample-dependent distribution function on the parameter space (instead of a point or an interval) to estimate the parameter of interest.

A simple example of a confidence distribution, that has been broadly used in statistical practice, is a bootstrap distribution. [4] The development and interpretation of a bootstrap distribution does not involve any fiducial reasoning; the same is true for the concept of a confidence distribution. But the notion of confidence distribution is much broader than that of a bootstrap distribution. In particular, recent research suggests that it encompasses and unifies a wide range of examples, from regular parametric cases (including most examples of the classical development of Fisher's fiducial distribution) to bootstrap distributions, p-value functions, [5] normalized likelihood functions and, in some cases, Bayesian priors and Bayesian posteriors. [6]

Just as a Bayesian posterior distribution contains a wealth of information for any type of Bayesian inference, a confidence distribution contains a wealth of information for constructing almost all types of frequentist inferences, including point estimates, confidence intervals, critical values, statistical power and p-values, [7] among others. Some recent developments have highlighted the promising potentials of the CD concept, as an effective inferential tool. [3]

History

Neyman (1937) [8] introduced the idea of "confidence" in his seminal paper on confidence intervals which clarified the frequentist repetition property. According to Fraser, [9] the seed (idea) of confidence distribution can even be traced back to Bayes (1763) [10] and Fisher (1930). [1] Although the phrase seems to first be used in Cox (1958). [11] Some researchers view the confidence distribution as "the Neymanian interpretation of Fisher's fiducial distributions", [12] which was "furiously disputed by Fisher". [13] It is also believed that these "unproductive disputes" and Fisher's "stubborn insistence" [13] might be the reason that the concept of confidence distribution has been long misconstrued as a fiducial concept and not been fully developed under the frequentist framework. [6] [14] Indeed, the confidence distribution is a purely frequentist concept with a purely frequentist interpretation, although it also has ties to Bayesian and fiducial inference concepts.

Definition

Classical definition

Classically, a confidence distribution is defined by inverting the upper limits of a series of lower-sided confidence intervals. [15] [16] [ page needed ] In particular,

For every α in (0, 1), let (−∞, ξn(α)] be a 100α% lower-side confidence interval for θ, where ξn(α) = ξn(Xn,α) is continuous and increasing in α for each sample Xn. Then, Hn(•) = ξn−1(•) is a confidence distribution for θ.

Efron stated that this distribution "assigns probability 0.05 to θ lying between the upper endpoints of the 0.90 and 0.95 confidence interval, etc." and "it has powerful intuitive appeal". [16] In the classical literature, [3] the confidence distribution function is interpreted as a distribution function of the parameter θ, which is impossible unless fiducial reasoning is involved since, in a frequentist setting, the parameters are fixed and nonrandom.

To interpret the CD function entirely from a frequentist viewpoint and not interpret it as a distribution function of a (fixed/nonrandom) parameter is one of the major departures of recent development relative to the classical approach. The nice thing about treating confidence distributions as a purely frequentist concept (similar to a point estimator) is that it is now free from those restrictive, if not controversial, constraints set forth by Fisher on fiducial distributions. [6] [14]

The modern definition

The following definition applies; [12] [17] [18] Θ is the parameter space of the unknown parameter of interest θ, and χ is the sample space corresponding to data Xn={X1, ..., Xn}:

A function Hn(•) = Hn(Xn, •) on χ × Θ  [0, 1] is called a confidence distribution (CD) for a parameter θ, if it follows two requirements:
  • (R1) For each given Xnχ, Hn(•) = Hn(Xn, •) is a continuous cumulative distribution function on Θ;
  • (R2) At the true parameter value θ = θ0, Hn(θ0)  Hn(Xn, θ0), as a function of the sample Xn, follows the uniform distribution U[0, 1].

Also, the function H is an asymptotic CD (aCD), if the U[0, 1] requirement is true only asymptotically and the continuity requirement on Hn(•) is dropped.

In nontechnical terms, a confidence distribution is a function of both the parameter and the random sample, with two requirements. The first requirement (R1) simply requires that a CD should be a distribution on the parameter space. The second requirement (R2) sets a restriction on the function so that inferences (point estimators, confidence intervals and hypothesis testing, etc.) based on the confidence distribution have desired frequentist properties. This is similar to the restrictions in point estimation to ensure certain desired properties, such as unbiasedness, consistency, efficiency, etc. [6] [19]

A confidence distribution derived by inverting the upper limits of confidence intervals (classical definition) also satisfies the requirements in the above definition and this version of the definition is consistent with the classical definition. [18]

Unlike the classical fiducial inference, more than one confidence distributions may be available to estimate a parameter under any specific setting. Also, unlike the classical fiducial inference, optimality is not a part of requirement. Depending on the setting and the criterion used, sometimes there is a unique "best" (in terms of optimality) confidence distribution. But sometimes there is no optimal confidence distribution available or, in some extreme cases, we may not even be able to find a meaningful confidence distribution. This is not different from the practice of point estimation.

A definition with measurable spaces

A confidence distribution [20] for a parameter in a measurable space is a distribution estimator with for a family of confidence regions for with level for all levels . The family of confidence regions is not unique. [21] If only exists for , then is a confidence distribution with level set . Both and all are measurable functions of the data. This implies that is a random measure and is a random set. If the defining requirement holds with equality, then the confidence distribution is by definition exact. If, additionally, is a real parameter, then the measure theoretic definition coincides with the above classical definition.

Examples

Example 1: Normal mean and variance

Suppose a normal sample Xi ~ N(μ, σ2), i = 1, 2, ..., n is given.

(1) Variance σ2 is known

Let, Φ be the cumulative distribution function of the standard normal distribution, and the cumulative distribution function of the Student distribution. Both the functions and given by

satisfy the two requirements in the CD definition, and they are confidence distribution functions for μ. [3] Furthermore,

satisfies the definition of an asymptotic confidence distribution when n→∞, and it is an asymptotic confidence distribution for μ. The uses of and are equivalent to state that we use and to estimate , respectively.

(2) Variance σ2 is unknown

For the parameter μ, since involves the unknown parameter σ and it violates the two requirements in the CD definition, it is no longer a "distribution estimator" or a confidence distribution for μ. [3] However, is still a CD for μ and is an aCD for μ.

For the parameter σ2, the sample-dependent cumulative distribution function

is a confidence distribution function for σ2. [6] Here, is the cumulative distribution function of the distribution .

In the case when the variance σ2 is known, is optimal in terms of producing the shortest confidence intervals at any given level. In the case when the variance σ2 is unknown, is an optimal confidence distribution for μ.

Example 2: Bivariate normal correlation

Let ρ denotes the correlation coefficient of a bivariate normal population. It is well known that Fisher's z defined by the Fisher transformation:

has the limiting distribution with a fast rate of convergence, where r is the sample correlation and n is the sample size.

The function

is an asymptotic confidence distribution for ρ. [22]

An exact confidence density for ρ is [23] [24]

where is the Gaussian hypergeometric function and . This is also the posterior density of a Bayes matching prior for the five parameters in the binormal distribution. [25]

The very last formula in the classical book by Fisher gives

where and . This formula was derived by C. R. Rao. [26]

Example 3: Binormal mean

Let data be generated by where is an unknown vector in the plane and has a binormal and known distribution in the plane. The distribution of defines a confidence distribution for . The confidence regions can be chosen as the interior of ellipses centered at and axes given by the eigenvectors of the covariance matrix of . The confidence distribution is in this case binormal with mean , and the confidence regions can be chosen in many other ways. [21] The confidence distribution coincides in this case with the Bayesian posterior using the right Haar prior. [27] The argument generalizes to the case of an unknown mean in an infinite-dimensional Hilbert space, but in this case the confidence distribution is not a Bayesian posterior. [28]

Using confidence distributions for inference

Confidence interval

CDinference1.png

From the CD definition, it is evident that the interval and provide 100(1  α)%-level confidence intervals of different kinds, for θ, for any α  (0, 1). Also is a level 100(1  α1  α2)% confidence interval for the parameter θ for any α1 > 0, α2 > 0 and α1 + α2 < 1. Here, is the 100β% quantile of or it solves for θ in equation . The same holds for a CD, where the confidence level is achieved in limit. Some authors have proposed using them for graphically viewing what parameter values are consistent with the data, instead of coverage or performance purposes. [29] [30]

Point estimation

Point estimators can also be constructed given a confidence distribution estimator for the parameter of interest. For example, given Hn(θ) the CD for a parameter θ, natural choices of point estimators include the median Mn = Hn1(1/2), the mean , and the maximum point of the CD density

Under some modest conditions, among other properties, one can prove that these point estimators are all consistent. [6] [22] Certain confidence distributions can give optimal frequentist estimators. [28]

Hypothesis testing

One can derive a p-value for a test, either one-sided or two-sided, concerning the parameter θ, from its confidence distribution Hn(θ). [6] [22] Denote by the probability mass of a set C under the confidence distribution function This ps(C) is called "support" in the CD inference and also known as "belief" in the fiducial literature. [31] We have

(1) For the one-sided test K0: θ  C vs. K1: θ  Cc, where C is of the type of (∞, b] or [b, ∞), one can show from the CD definition that supθ  CPθ(ps(C)  α) = α. Thus, ps(C) = Hn(C) is the corresponding p-value of the test.

(2) For the singleton test K0: θ = b vs. K1: θ  b, P{K0: θ = b}(2 min{ps(Clo), one can show from the CD definition that ps(Cup)}  α) = α. Thus, 2 min{ps(Clo), ps(Cup)} = 2 min{Hn(b), 1  Hn(b)} is the corresponding p-value of the test. Here, Clo = (∞, b] and Cup = [b, ∞).

See Figure 1 from Xie and Singh (2011) [6] for a graphical illustration of the CD inference.

Implementations

A few statistical programs have implemented the ability to construct and graph confidence distributions.

R, via the concurve, [32] [33] pvaluefunctions, [34] and episheet [35] packages

Excel, via episheet [36]

Stata, via concurve [32]

See also

Related Research Articles

In statistics, a location parameter of a probability distribution is a scalar- or vector-valued parameter , which determines the "location" or shift of the distribution. In the literature of location parameter estimation, the probability distributions with such parameter are found to be formally defined in one of the following equivalent ways:

The likelihood function is the joint probability of observed data viewed as a function of the parameters of a statistical model.

In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data. This is achieved by maximizing a likelihood function so that, under the assumed statistical model, the observed data is most probable. The point in the parameter space that maximizes the likelihood function is called the maximum likelihood estimate. The logic of maximum likelihood is both intuitive and flexible, and as such the method has become a dominant means of statistical inference.

In statistics, interval estimation is the use of sample data to estimate an interval of possible values of a parameter of interest. This is in contrast to point estimation, which gives a single value.

<span class="mw-page-title-main">Gamma distribution</span> Probability distribution

In probability theory and statistics, the gamma distribution is a two-parameter family of continuous probability distributions. The exponential distribution, Erlang distribution, and chi-squared distribution are special cases of the gamma distribution. There are two equivalent parameterizations in common use:

  1. With a shape parameter and a scale parameter .
  2. With a shape parameter and an inverse scale parameter , called a rate parameter.
<span class="mw-page-title-main">Confidence interval</span> Range to estimate an unknown parameter

Informally, in frequentist statistics, a confidence interval (CI) is an interval which is expected to typically contain the parameter being estimated. More specifically, given a confidence level , a CI is a random interval which contains the parameter being estimated % of the time. The confidence level, degree of confidence or confidence coefficient represents the long-run proportion of CIs that theoretically contain the true value of the parameter; this is tantamount to the nominal coverage probability. For example, out of all intervals computed at the 95% level, 95% of them should contain the parameter's true value.

In probability and statistics, an exponential family is a parametric set of probability distributions of a certain form, specified below. This special form is chosen for mathematical convenience, including the enabling of the user to calculate expectations, covariances using differentiation based on some useful algebraic properties, as well as for generality, as exponential families are in a sense very natural sets of distributions to consider. The term exponential class is sometimes used in place of "exponential family", or the older term Koopman–Darmois family. Sometimes loosely referred to as "the" exponential family, this class of distributions is distinct because they all possess a variety of desirable properties, most importantly the existence of a sufficient statistic.

In statistical inference, specifically predictive inference, a prediction interval is an estimate of an interval in which a future observation will fall, with a certain probability, given what has already been observed. Prediction intervals are often used in regression analysis.

In Bayesian statistics, a maximum a posteriori probability (MAP) estimate is an estimate of an unknown quantity, that equals the mode of the posterior distribution. The MAP can be used to obtain a point estimate of an unobserved quantity on the basis of empirical data. It is closely related to the method of maximum likelihood (ML) estimation, but employs an augmented optimization objective which incorporates a prior distribution over the quantity one wants to estimate. MAP estimation can therefore be seen as a regularization of maximum likelihood estimation.

In Bayesian probability, the Jeffreys prior, named after Sir Harold Jeffreys, is a non-informative prior distribution for a parameter space; its density function is proportional to the square root of the determinant of the Fisher information matrix:

Noncentral <i>t</i>-distribution Probability distribution

The noncentral t-distribution generalizes Student's t-distribution using a noncentrality parameter. Whereas the central probability distribution describes how a test statistic t is distributed when the difference tested is null, the noncentral distribution describes how t is distributed when the null is false. This leads to its use in statistics, especially calculating statistical power. The noncentral t-distribution is also known as the singly noncentral t-distribution, and in addition to its primary use in statistical inference, is also used in robust modeling for data.

In statistics, a pivotal quantity or pivot is a function of observations and unobservable parameters such that the function's probability distribution does not depend on the unknown parameters. A pivot quantity need not be a statistic—the function and its value can depend on the parameters of the model, but its distribution must not. If it is a statistic, then it is known as an ancillary statistic.

In estimation theory and decision theory, a Bayes estimator or a Bayes action is an estimator or decision rule that minimizes the posterior expected value of a loss function. Equivalently, it maximizes the posterior expectation of a utility function. An alternative way of formulating an estimator within Bayesian statistics is maximum a posteriori estimation.

Neyman construction, named after Jerzy Neyman, is a frequentist method to construct an interval at a confidence level such that if we repeat the experiment many times the interval will contain the true value of some parameter a fraction of the time.

In probability and statistics, a natural exponential family (NEF) is a class of probability distributions that is a special case of an exponential family (EF).

In statistics, additive smoothing, also called Laplace smoothing or Lidstone smoothing, is a technique used to smooth count data, eliminating issues caused by certain values having 0 occurrences. Given a set of observation counts from a -dimensional multinomial distribution with trials, a "smoothed" version of the counts gives the estimator:

Algorithmic inference gathers new developments in the statistical inference methods made feasible by the powerful computing devices widely available to any data analyst. Cornerstones in this field are computational learning theory, granular computing, bioinformatics, and, long ago, structural probability . The main focus is on the algorithms which compute statistics rooting the study of a random phenomenon, along with the amount of data they must feed on to produce reliable results. This shifts the interest of mathematicians from the study of the distribution laws to the functional properties of the statistics, and the interest of computer scientists from the algorithms for processing data to the information they process.

<span class="mw-page-title-main">Wrapped Cauchy distribution</span>

In probability theory and directional statistics, a wrapped Cauchy distribution is a wrapped probability distribution that results from the "wrapping" of the Cauchy distribution around the unit circle. The Cauchy distribution is sometimes known as a Lorentzian distribution, and the wrapped Cauchy distribution may sometimes be referred to as a wrapped Lorentzian distribution.

In particle physics, CLs represents a statistical method for setting upper limits on model parameters, a particular form of interval estimation used for parameters that can take only non-negative values. Although CLs are said to refer to Confidence Levels, "The method's name is ... misleading, as the CLs exclusion region is not a confidence interval." It was first introduced by physicists working at the LEP experiment at CERN and has since been used by many high energy physics experiments. It is a frequentist method in the sense that the properties of the limit are defined by means of error probabilities, however it differs from standard confidence intervals in that the stated confidence level of the interval is not equal to its coverage probability. The reason for this deviation is that standard upper limits based on a most powerful test necessarily produce empty intervals with some fixed probability when the parameter value is zero, and this property is considered undesirable by most physicists and statisticians.

In statistics, the variance function is a smooth function that depicts the variance of a random quantity as a function of its mean. The variance function is a measure of heteroscedasticity and plays a large role in many settings of statistical modelling. It is a main ingredient in the generalized linear model framework and a tool used in non-parametric regression, semiparametric regression and functional data analysis. In parametric modeling, variance functions take on a parametric form and explicitly describe the relationship between the variance and the mean of a random quantity. In a non-parametric setting, the variance function is assumed to be a smooth function.

References

  1. 1 2 Fisher, R.A. (1930). "Inverse probability." Proc. cambridge Pilos. Soc.26, 528–535.
  2. Cox, D.R. (1958). "Some Problems Connected with Statistical Inference", "The Annals of Mathematical Statistics", "29" 357-372 (Section 4, Page 363) doi:10.1214/aoms/1177706618
  3. 1 2 3 4 5 6 Xie, M. (2013). "Rejoinder of Confidence Distribution, the Frequentist Distribution Estimator of a Parameter – a Review". International Statistical Review, 81, 68-77. doi:10.1111/insr.12001
  4. Efron, B. (1998). "R.A.Fisher in the 21st Century" Statistical Science.13 95–122. JSTOR   2290557
  5. Fraser, D.A.S. (1991). "Statistical inference: Likelihood to significance." Journal of the American Statistical Association , 86, 258–265. JSTOR   2290557
  6. 1 2 3 4 5 6 7 8 Xie, M. and Singh, K. (2013). "Confidence Distribution, the Frequentist Distribution Estimator of a Parameter – a Review (with discussion)". International Statistical Review, 81, 3-39. doi:10.1111/insr.12000
  7. Fraser, D. A. S. (2019-03-29). "The p-value Function and Statistical Inference". The American Statistician. 73 (sup1): 135–147. doi: 10.1080/00031305.2018.1556735 . ISSN   0003-1305.
  8. Neyman, J. (1937). "Outline of a theory of statistical estimation based on the classical theory of probability." Phil. Trans. Roy. SocA237 333–380
  9. Fraser, D.A.S. (2011). "Is Bayes posterior just quick and dirty confidence?" Statistical Science26, 299-316. JSTOR   23059129
  10. Bayes, T. (1763). "An Essay towards solving a Problem in the Doctrine of Chances." Phil. Trans. Roy. Soc, London 53 370–418 54 296–325. Reprinted in Biometrika 45 (1958) 293–315.
  11. Cox, D. R. (June 1958). "Some Problems Connected with Statistical Inference". The Annals of Mathematical Statistics. 29 (2): 357–372. doi: 10.1214/aoms/1177706618 . ISSN   0003-4851.
  12. 1 2 Schweder, T. and Hjort, N.L. (2002). "Confidence and likelihood", Scandinavian Journal of Statistics.29 309–332. doi:10.1111/1467-9469.00285
  13. 1 2 Zabell, S.L. (1992). "R.A.Fisher and fiducial argument", Stat. Sci., 7, 369–387
  14. 1 2 Singh, K. and Xie, M. (2011). "Discussions of “Is Bayes posterior just quick and dirty confidence?” by D.A.S. Fraser." Statistical Science. Vol. 26, 319-321. JSTOR   23059131
  15. Cox, D. R. (2006). Principles of Statistical Inference, CUP. ISBN   0-521-68567-2. (page 66)
  16. 1 2 Efron, B. (1993). "Bayes and likelihood calculations from confidence intervals. Biometrika , 80 3–26.
  17. Singh, K. Xie, M. and Strawderman, W.E. (2001). "Confidence distributions—concept, theory and applications". Technical report, Dept. Statistics, Rutgers Univ. Revised 2004.
  18. 1 2 Singh, K. Xie, M. and Strawderman, W.E. (2005). "Combining Information from Independent Sources Through Confidence Distribution" Annals of Statistics , 33, 159–183. JSTOR   3448660
  19. Xie, M., Liu, R., Daramuju, C.V., Olsan, W. (2012). "Incorporating expert opinions with information from binomial clinical trials." Annals of Applied Statistics. In press.
  20. Taraldsen, Gunnar (2021). "Joint Confidence Distributions". doi:10.13140/RG.2.2.33079.85920.{{cite journal}}: Cite journal requires |journal= (help)
  21. 1 2 Liu, Dungang; Liu, Regina Y.; Xie, Min-ge (2021-04-30). "Nonparametric Fusion Learning for Multiparameters: Synthesize Inferences From Diverse Sources Using Data Depth and Confidence Distribution". Journal of the American Statistical Association. 117 (540): 2086–2104. doi:10.1080/01621459.2021.1902817. ISSN   0162-1459. S2CID   233657455.
  22. 1 2 3 Singh, K. Xie, M. and Strawderman, W.E. (2007). "Confidence Distribution (CD)-Distribution Estimator of a Parameter", in Complex Datasets and Inverse ProblemsIMS Lecture Notes—Monograph Series, 54, (R. Liu, et al. Eds) 132–150. JSTOR   20461464
  23. Taraldsen, Gunnar (2021). "The Confidence Density for Correlation". Sankhya A. 85: 600–616. doi: 10.1007/s13171-021-00267-y . ISSN   0976-8378. S2CID   244594067.
  24. Taraldsen, Gunnar (2020). "Confidence in Correlation". doi:10.13140/RG.2.2.23673.49769.{{cite journal}}: Cite journal requires |journal= (help)
  25. Berger, James O.; Sun, Dongchu (2008-04-01). "Objective priors for the bivariate normal model". The Annals of Statistics. 36 (2). arXiv: 0804.0987 . doi: 10.1214/07-AOS501 . ISSN   0090-5364. S2CID   14703802.
  26. Fisher, Ronald Aylmer, Sir (1973). Statistical methods and scientific inference ([3d ed., rev. and enl.] ed.). New York: Hafner Press. ISBN   0-02-844740-9. OCLC   785822.{{cite book}}: CS1 maint: multiple names: authors list (link)
  27. Eaton, Morris L.; Sudderth, William D. (2012). "Invariance, model matching and probability matching". Sankhyā: The Indian Journal of Statistics, Series A (2008-). 74 (2): 170–193. doi:10.1007/s13171-012-0018-4. ISSN   0976-836X. JSTOR   42003718. S2CID   120705955.
  28. 1 2 Taraldsen, Gunnar; Lindqvist, Bo Henry (2013-02-01). "Fiducial theory and optimal inference". The Annals of Statistics. 41 (1). arXiv: 1301.1717 . doi: 10.1214/13-AOS1083 . ISSN   0090-5364. S2CID   88520957.
  29. Cox, D. R.; Hinkley, D. V. (1979-09-06). Theoretical Statistics. Chapman and Hall/CRC. doi:10.1201/b14832. ISBN   978-0-429-17021-8.
  30. Rafi, Zad; Greenland, Sander (2020-09-30). "Semantic and cognitive tools to aid statistical science: replace confidence and significance by compatibility and surprise". BMC Medical Research Methodology. 20 (1): 244. arXiv: 1909.08579 . doi: 10.1186/s12874-020-01105-9 . ISSN   1471-2288. PMC   7528258 . PMID   32998683.
  31. Kendall, M., & Stuart, A. (1974). The Advanced Theory of Statistics, Volume ?. (Chapter 21). Wiley.
  32. 1 2 Rafi [aut, Zad; cre; Vigotsky, Andrew D. (2020-04-20), concurve: Computes and Plots Compatibility (Confidence) Intervals, P-Values, S-Values, & Likelihood Intervals to Form Consonance, Surprisal, & Likelihood Functions , retrieved 2020-05-05
  33. "Concurve plots consonance curves, p-value functions, and S-value functions « Statistical Modeling, Causal Inference, and Social Science". statmodeling.stat.columbia.edu. Retrieved 2020-04-15.
  34. Infanger, Denis (2019-11-29), pvaluefunctions: Creates and Plots P-Value Functions, S-Value Functions, Confidence Distributions and Confidence Densities , retrieved 2020-04-15
  35. Black, James; Rothman, Ken; Thelwall, Simon (2019-01-23), episheet: Rothman's Episheet , retrieved 2020-04-15
  36. "Modern Epidemiology, 2nd Edition". www.krothman.org. Archived from the original on 2020-01-29. Retrieved 2020-04-15.

Bibliography