Taylor diagram

Last updated

Taylor diagrams are mathematical diagrams designed to graphically indicate which of several approximate representations (or models) of a system, process, or phenomenon is most realistic. This diagram, invented by Karl E. Taylor in 1994 (published in 2001 [1] ) facilitates the comparative assessment of different models. It is used to quantify the degree of correspondence between the modeled and observed behavior in terms of three statistics: the Pearson correlation coefficient, the root-mean-square error (RMSE) error, and the standard deviation.

Contents

Although Taylor diagrams have primarily been used to evaluate models designed to study climate and other aspects of Earth's environment, [2] they can be used for purposes unrelated to environmental science (e.g., to quantify and visually display how well fusion energy models represent reality [3] ).

Taylor diagrams can be constructed with a number of different open source and commercial software packages, including: GrADS, [4] [5] IDL, [6] MATLAB, [7] [8] [9] NCL, [10] Python, [11] [12] R, [13] and CDAT. [14]

Sample diagram

The sample Taylor diagram shown in Figure 1 [15] provides a summary of the relative skill with which several global climate models simulate the spatial pattern of annual mean precipitation. Eight models, each represented by a different letter on the diagram, are compared, and the distance between each model and the point labeled “observed” is a measure of how realistically each model reproduces observations. For each model, three statistics are plotted: the Pearson correlation coefficient (gauging similarity in pattern between the simulated and observed fields) is related to the azimuthal angle (blue contours); the centered RMS error in the simulated field is proportional to the distance from the point on the x-axis identified as “observed” (green contours); and the standard deviation of the simulated pattern is proportional to the radial distance from the origin (black contours). It is evident from this diagram, for example, that for Model F the correlation coefficient is about 0.65, the RMS error is about 2.6 mm/day and the standard deviation is about 3.3 mm/day. Model F's standard deviation is clearly greater than the standard deviation of the observed field (indicated by the dashed contour at radial distance 2.9 mm/day).

Fig. 1: Sample Taylor diagram displaying a statistical comparison with observations of eight model estimates of the global pattern of annual mean precipitation. Primer fig1a.svg
Fig. 1: Sample Taylor diagram displaying a statistical comparison with observations of eight model estimates of the global pattern of annual mean precipitation.

The relative merits of various models can be inferred from Figure 1. Simulated patterns that agree well with observations will lie nearest the point marked "observed" on the x-axis. These models have relatively high correlation and low RMS errors. Models lying on the dashed arc have the correct standard deviation (which indicates that the pattern variations are of the right amplitude). In Figure 1 it can be seen that models A and C generally agree best with observations, each with about the same RMS error. Model A, however, has a slightly higher correlation with observations and has the same standard deviation as the observed, whereas model C has too little spatial variability (with a standard deviation of 2.3 mm/day compared to the observed value of 2.9 mm/day). Of the poorer performing models, model E has a low pattern correlation, while model D has variations that are much larger than observed, in both cases resulting in a relatively large (~3 mm/day) centered RMS error in the precipitation fields. Although models D and B have about the same correlation with observations, model B simulates the amplitude of the variations (i.e., the standard deviation) much better than model D, resulting in a smaller RMS error.

Theoretical basis

Taylor diagrams display statistics useful for assessing the similarity of a variable simulated by a model (more generally, the “test” field) to its observed counterpart (more generally, the “reference” field). Mathematically, the three statistics displayed on a Taylor diagram are related by the error propagation formula (which can be derived directly from the definition of the statistics appearing in it):

,

where ρ is the correlation coefficient between the test and reference fields, E′ is the centered RMS difference between the fields (with any difference in the means first removed), and and are the standard deviations of the reference and test fields, respectively. The law of cosines,

(where a, b, and c are the length of the sides of the triangle, and is the angle between sides a and b) provides the key to forming the geometrical relationship between the four quantities that underlie the Taylor diagram (shown in Figure 2).

Fig. 2: Geometrical relationship between statistics plotted on Taylor diagrams according to the law of cosines. Taylor diagram fig2.png
Fig. 2: Geometrical relationship between statistics plotted on Taylor diagrams according to the law of cosines.

The standard deviation of the observed field is side a, the standard deviation of the test field is side b, the centered RMS difference between the two fields (E′) is side c, and the cosine of the angle between sides a and b is the correlation coefficient (ρ).

The means of the fields are subtracted out before computing their second-order statistics, so the diagram does not provide information about overall biases (mean error), but solely characterizes the centered pattern error.

Taylor diagram variants

Among the several minor variations on the diagram that have been suggested are (see, Taylor, 2001 [1] ):

One of the main limitation of the Taylor diagram is the absence of explicit information about the prediction bias. A variant of the Taylor diagram that accounts for the prediction bias is provided by the solar diagram (see, Wadoux et al., 2022 [16] ).

Related Research Articles

<span class="mw-page-title-main">Normal distribution</span> Probability distribution

In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is

<span class="mw-page-title-main">Standard deviation</span> In statistics, a measure of variation

In statistics, the standard deviation is a measure of the amount of variation or dispersion of a set of values. A low standard deviation indicates that the values tend to be close to the mean of the set, while a high standard deviation indicates that the values are spread out over a wider range.

<span class="mw-page-title-main">Correlation</span> Statistical concept

In statistics, correlation or dependence is any statistical relationship, whether causal or not, between two random variables or bivariate data. Although in the broadest sense, "correlation" may indicate any type of association, in statistics it usually refers to the degree to which a pair of variables are linearly related. Familiar examples of dependent phenomena include the correlation between the height of parents and their offspring, and the correlation between the price of a good and the quantity the consumers are willing to purchase, as it is depicted in the so-called demand curve.

<span class="mw-page-title-main">Pearson correlation coefficient</span> Measure of linear correlation

In statistics, the Pearson correlation coefficient (PCC) is a correlation coefficient that measures linear correlation between two sets of data. It is the ratio between the covariance of two variables and the product of their standard deviations; thus, it is essentially a normalized measurement of the covariance, such that the result always has a value between −1 and 1. As with covariance itself, the measure can only reflect a linear correlation of variables, and ignores many other types of relationships or correlations. As a simple example, one would expect the age and height of a sample of teenagers from a high school to have a Pearson correlation coefficient significantly greater than 0, but less than 1.

<span class="mw-page-title-main">Standard score</span> How many standard deviations apart from the mean an observed datum is

In statistics, the standard score is the number of standard deviations by which the value of a raw score is above or below the mean value of what is being observed or measured. Raw scores above the mean have positive standard scores, while those below the mean have negative standard scores.

<span class="mw-page-title-main">Spearman's rank correlation coefficient</span> Nonparametric measure of rank correlation

In statistics, Spearman's rank correlation coefficient or Spearman's ρ, named after Charles Spearman and often denoted by the Greek letter (rho) or as , is a nonparametric measure of rank correlation. It assesses how well the relationship between two variables can be described using a monotonic function.

In statistics, an effect size is a value measuring the strength of the relationship between two variables in a population, or a sample-based estimate of that quantity. It can refer to the value of a statistic calculated from a sample of data, the value of a parameter for a hypothetical population, or to the equation that operationalizes how statistics or parameters lead to the effect size value. Examples of effect sizes include the correlation between two variables, the regression coefficient in a regression, the mean difference, or the risk of a particular event happening. Effect sizes complement statistical hypothesis testing, and play an important role in power analyses, sample size planning, and in meta-analyses. The cluster of data-analysis methods concerning effect sizes is referred to as estimation statistics.

In statistics and optimization, errors and residuals are two closely related and easily confused measures of the deviation of an observed value of an element of a statistical sample from its "true value". The error of an observation is the deviation of the observed value from the true value of a quantity of interest. The residual is the difference between the observed value and the estimated value of the quantity of interest. The distinction is most important in regression analysis, where the concepts are sometimes called the regression errors and regression residuals and where they lead to the concept of studentized residuals. In econometrics, "errors" are also called disturbances.

<span class="mw-page-title-main">Standard error</span> Statistical property

The standard error (SE) of a statistic is the standard deviation of its sampling distribution or an estimate of that standard deviation. If the statistic is the sample mean, it is called the standard error of the mean (SEM). The standard error is a key ingredient in producing confidence intervals.

In statistics, propagation of uncertainty is the effect of variables' uncertainties on the uncertainty of a function based on them. When the variables are the values of experimental measurements they have uncertainties due to measurement limitations which propagate due to the combination of variables in the function.

In statistics, a studentized residual is the dimensionless ratio resulting from the division of a residual by an estimate of its standard deviation, both expressed in the same units. It is a form of a Student's t-statistic, with the estimate of error varying between points.

Modern portfolio theory (MPT), or mean-variance analysis, is a mathematical framework for assembling a portfolio of assets such that the expected return is maximized for a given level of risk. It is a formalization and extension of diversification in investing, the idea that owning different kinds of financial assets is less risky than owning only one type. Its key insight is that an asset's risk and return should not be assessed by itself, but by how it contributes to a portfolio's overall risk and return. It uses the past variance of asset prices as a proxy for future risk.

<span class="mw-page-title-main">Fisher transformation</span> Statistical transformation

In statistics, the Fisher transformation of a Pearson correlation coefficient is its inverse hyperbolic tangent (artanh). When the sample correlation coefficient r is near 1 or -1, its distribution is highly skewed, which makes it difficult to estimate confidence intervals and apply tests of significance for the population correlation coefficient ρ. The Fisher transformation solves this problem by yielding a variable whose distribution is approximately normally distributed, with a variance that is stable over different values of r.

In probability theory and statistics, the coefficient of variation (CV), also known as Normalized Root-Mean-Square Deviation (NRMSD), Percent RMS, and relative standard deviation (RSD), is a standardized measure of dispersion of a probability distribution or frequency distribution. It is defined as the ratio of the standard deviation to the mean , and often expressed as a percentage ("%RSD"). The CV or RSD is widely used in analytical chemistry to express the precision and repeatability of an assay. It is also commonly used in fields such as engineering or physics when doing quality assurance studies and ANOVA gauge R&R, by economists and investors in economic models, and in psychology/neuroscience.

Linear discriminant analysis (LDA), normal discriminant analysis (NDA), or discriminant function analysis is a generalization of Fisher's linear discriminant, a method used in statistics and other fields, to find a linear combination of features that characterizes or separates two or more classes of objects or events. The resulting combination may be used as a linear classifier, or, more commonly, for dimensionality reduction before later classification.

This glossary of statistics and probability is a list of definitions of terms and concepts used in the mathematical sciences of statistics and probability, their sub-disciplines, and related fields. For additional related terms, see Glossary of mathematics and Glossary of experimental design.

In robust statistics, robust regression seeks to overcome some limitations of traditional regression analysis. A regression analysis models the relationship between one or more independent variables and a dependent variable. Standard types of regression, such as ordinary least squares, have favourable properties if their underlying assumptions are true, but can give misleading results otherwise. Robust regression methods are designed to limit the effect that violations of assumptions by the underlying data-generating process have on regression estimates.

<span class="mw-page-title-main">Intraclass correlation</span> Descriptive statistic

In statistics, the intraclass correlation, or the intraclass correlation coefficient (ICC), is a descriptive statistic that can be used when quantitative measurements are made on units that are organized into groups. It describes how strongly units in the same group resemble each other. While it is viewed as a type of correlation, unlike most other correlation measures, it operates on data structured as groups rather than data structured as paired observations.

<span class="mw-page-title-main">Standard deviation line</span>

In statistics, the standard deviation line marks points on a scatter plot that are an equal number of standard deviations away from the average in each dimension. For example, in a 2-dimensional scatter diagram with variables and , points that are 1 standard deviation away from the mean of and also 1 standard deviation away from the mean of are on the SD line. The SD line is a useful visual tool since points in a scatter diagram tend to cluster around it, more or less tightly depending on their correlation.

The Kling-Gupta efficiency (KGE) is a goodness-of-fit indicator widely used in the hydrologic sciences for comparing simulations to observations. It was created by hydrologic scientists Harald Kling and Hoshin Vijai Gupta. Its creators intended for it to improve upon widely used metrics such as the coefficient of determination and the Nash–Sutcliffe model efficiency coefficient.

References

  1. 1 2 Taylor, K.E. (2001). "Summarizing multiple aspects of model performance in a single diagram". J. Geophys. Res. 106: 7183–7192. Bibcode:2001JGR...106.7183T. doi:10.1029/2000JD900719.
  2. "Google Scholar". scholar.google.com.
  3. Terry, P.W.; et al. (2008). "Validation in fusion research: Towards guidelines and best practices". Phys. Plasmas . 15. arXiv: 0801.2787 . Bibcode:2008PhPl...15f2503T. doi:10.1063/1.2928909.
  4. "Calculate statistics used in Taylor diagram in GrADS".
  5. "Plot Taylor diagram in GrADS".
  6. "Creating a Taylor Diagram". www.idlcoyote.com.
  7. "Taylor Diagram". www.mathworks.com.
  8. "PeterRochford/SkillMetricsToolbox". www.mathworks.com.
  9. "SkillMetricsToolbox". October 27, 2021 via GitHub.
  10. "NCL Graphics: Taylor Diagrams". www.ncl.ucar.edu.
  11. "SkillMetrics Project". November 6, 2021 via GitHub.
  12. "geocat.viz.taylor.TaylorDiagram". geocat-viz.readthedocs.io. Retrieved 2023-11-16.
  13. Lemon, Jim; Bolker, Ben; Oom, Sander; Klein, Eduardo; Rowlingson, Barry; Wickham, Hadley; Tyagi, Anupam; Eterradossi, Olivier; Grothendieck, Gabor; Toews, Michael; Kane, John; Turner, Rolf; Witthoft, Carl; Stander, Julian; Petzoldt, Thomas; Duursma, Remko; Biancotto, Elisa; Levy, Ofir; Dutang, Christophe; Solymos, Peter; Engelmann, Robby; Hecker, Michael; Steinbeck, Felix; Borchers, Hans; Singmann, Henrik; Toal, Ted; Ogle, Derek; Baral, Darshan; Groemping, Ulrike; Venables, Bill (September 8, 2021). "plotrix: Various Plotting Functions" via R-Packages.
  14. "Taylor_Diagrams". cdat.llnl.gov.
  15. "Taylor diagram primer (2005), K.E. Taylor" (PDF).
  16. Wadoux, AMJ-C; Walvoort, DJJ; Brus, DJ (2022). "An integrated approach for the evaluation of quantitative soil maps through Taylor and solar diagrams". Geoderma. 405: 115332. doi:10.1016/j.geoderma.2021.115332.