Nicholas Fisher (statistician)

Last updated

Nicholas Irving Fisher (born on 28 May 1946[ citation needed ]) is an Australian statistician and entrepreneur. He was a statistical researcher at the CSIRO for over 30 years and has founded the analytics company ValueMetrics Australia. [1] He has contributed to the development and applications of directional statistics in geosciences, and statistical methods for quality improvement, specifically performance measurement for enterprises.

Contents

Education and career

Fisher was born in Melbourne, Australia. He studied at the University of Sydney, where he received his BSc in 1968 and MSc in 1970. He then went to University of North Carolina at Chapel Hill, where he studied under Wassily Hoeffding and obtained his PhD in statistics in 1975. Fisher joined CSIRO in 1969, before he started his PhD, and rose through the ranks over the years until 1997, when he became a chief research scientist. At CSIRO he led the development of the organisation's performance management system. He was awarded a DSc by the University of Sydney in 1994 for his work in directional statistics. He left CSIRO in 2001 to found the analytics research enterprise ValueMetrics Australia and also became a Visiting Professor of Statistics at the University of Sydney. [2] [3] He is professionally accredited by the Statistical Society of Australia and by the American Statistical Association.

Honors and awards

Fisher became a Fellow of the American Statistical Association in 1991 and an Honorary Life Member of the Statistical Society of Australia in 1997. He received the service award from the International Statistical Institute in 2015. He was named the W. Edwards Deming Lecturer by the American Statistical Association in 2019. [4]

Bibliography

Books

Selected papers

Related Research Articles

<span class="mw-page-title-main">Statistics</span> Study of the collection and analysis of data

Statistics is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a scientific, industrial, or social problem, it is conventional to begin with a statistical population or a statistical model to be studied. Populations can be diverse groups of people or objects such as "all people living in a country" or "every atom composing a crystal". Statistics deals with every aspect of data, including the planning of data collection in terms of the design of surveys and experiments.

<span class="mw-page-title-main">Statistical hypothesis test</span> Method of statistical inference

A statistical hypothesis test is a method of statistical inference used to decide whether the data sufficiently supports a particular hypothesis. A statistical hypothesis test typically involves a calculation of a test statistic. Then a decision is made, either by comparing the test statistic to a critical value or equivalently by evaluating a p-value computed from the test statistic. Roughly 100 specialized statistical tests have been defined.

<span class="mw-page-title-main">W. Edwards Deming</span> American engineer and statistician (1900–1993)

William Edwards Deming was an American business theorist, composer, economist, industrial engineer, management consultant, statistician, and writer. Educated initially as an electrical engineer and later specializing in mathematical physics, he helped develop the sampling techniques still used by the United States Census Bureau and the Bureau of Labor Statistics. He is also known as the father of the quality movement and was hugely influential in post-WWII Japan, credited with revolutionizing Japan's industry and making it one of the most dominant economies in the world. He is best known for his theories of management.

<span class="mw-page-title-main">Ronald Fisher</span> British polymath (1890–1962)

Sir Ronald Aylmer Fisher was a British polymath who was active as a mathematician, statistician, biologist, geneticist, and academic. For his work in statistics, he has been described as "a genius who almost single-handedly created the foundations for modern statistical science" and "the single most important figure in 20th century statistics". In genetics, Fisher was the one to most comprehensively combine the ideas of Gregor Mendel and Charles Darwin, as his work used mathematics to combine Mendelian genetics and natural selection; this contributed to the revival of Darwinism in the early 20th-century revision of the theory of evolution known as the modern synthesis. For his contributions to biology, Richard Dawkins declared Fisher to be the greatest of Darwin's successors. He is also considered one of the founding fathers of Neo-Darwinism. According to statistician Jeffrey T. Leek, Fisher is the most influential scientist of all time based off the number of citations of his contributions.

In statistical hypothesis testing, a result has statistical significance when a result at least as "extreme" would be very infrequent if the null hypothesis were true. More precisely, a study's defined significance level, denoted by , is the probability of the study rejecting the null hypothesis, given that the null hypothesis is true; and the p-value of a result, , is the probability of obtaining a result at least as extreme, given that the null hypothesis is true. The result is statistically significant, by the standards of the study, when . The significance level for a study is chosen before data collection, and is typically set to 5% or much lower—depending on the field of study.

<span class="mw-page-title-main">Confidence interval</span> Range to estimate an unknown parameter

Informally, in frequentist statistics, a confidence interval (CI) is an interval which is expected to typically contain the parameter being estimated. More specifically, given a confidence level , a CI is a random interval which contains the parameter being estimated % of the time. The confidence level, degree of confidence or confidence coefficient represents the long-run proportion of CIs that theoretically contain the true value of the parameter; this is tantamount to the nominal coverage probability. For example, out of all intervals computed at the 95% level, 95% of them should contain the parameter's true value.

<span class="mw-page-title-main">Prasanta Chandra Mahalanobis</span> Indian scientist and statistician (1893–1972)

Prasanta Chandra Mahalanobis OBE, FNA, FASc, FRS was an Indian scientist and statistician. He is best remembered for the Mahalanobis distance, a statistical measure, and for being one of the members of the first Planning Commission of free India. He made pioneering studies in anthropometry in India. He founded the Indian Statistical Institute, and contributed to the design of large-scale sample surveys. For his contributions, Mahalanobis has been considered the Father of statistics in India.

In information geometry, the Fisher information metric is a particular Riemannian metric which can be defined on a smooth statistical manifold, i.e., a smooth manifold whose points are probability measures defined on a common probability space. It can be used to calculate the informational difference between measurements.

<span class="mw-page-title-main">Statistical process control</span> Method of quality control

Statistical process control (SPC) or statistical quality control (SQC) is the application of statistical methods to monitor and control the quality of a production process. This helps to ensure that the process operates efficiently, producing more specification-conforming products with less waste scrap. SPC can be applied to any process where the "conforming product" output can be measured. Key tools used in SPC include run charts, control charts, a focus on continuous improvement, and the design of experiments. An example of a process where SPC is applied is manufacturing lines.

In mathematical statistics, the Fisher information is a way of measuring the amount of information that an observable random variable X carries about an unknown parameter θ of a distribution that models X. Formally, it is the variance of the score, or the expected value of the observed information.

In probability theory, Hoeffding's inequality provides an upper bound on the probability that the sum of bounded independent random variables deviates from its expected value by more than a certain amount. Hoeffding's inequality was proven by Wassily Hoeffding in 1963.

When classification is performed by a computer, statistical methods are normally used to develop the algorithm.

In statistics, the k-nearest neighbors algorithm (k-NN) is a non-parametric supervised learning method first developed by Evelyn Fix and Joseph Hodges in 1951, and later expanded by Thomas Cover. It is used for classification and regression. In both cases, the input consists of the k closest training examples in a data set. The output depends on whether k-NN is used for classification or regression:

Wassily Hoeffding was an American statistician and probabilist. Hoeffding was one of the founders of nonparametric statistics, in which Hoeffding contributed the idea and basic results on U-statistics.

<i>The Lady Tasting Tea</i> Book by David Salsburg about the history of modern statistics

The Lady Tasting Tea: How Statistics Revolutionized Science in the Twentieth Century (ISBN 0-8050-7134-2) is a book by David Salsburg about the history of modern statistics and the role it played in the development of science and industry.

<span class="mw-page-title-main">P. A. P. Moran</span> Australian statistician (1917–1988)

Patrick Alfred Pierce Moran FRS was an Australian statistician who made significant contributions to probability theory and its application to population and evolutionary genetics.

<span class="mw-page-title-main">A. David Buckingham</span> Australian chemist and cricketer (1930–2021)

Amyand David Buckingham born in Pymble, Sydney, New South Wales, Australia was a chemist, with primary expertise in chemical physics.

In statistics, mean absolute error (MAE) is a measure of errors between paired observations expressing the same phenomenon. Examples of Y versus X include comparisons of predicted versus observed, subsequent time versus initial time, and one technique of measurement versus an alternative technique of measurement. MAE is calculated as the sum of absolute errors divided by the sample size:It is thus an arithmetic average of the absolute errors , where is the prediction and the true value. Alternative formulations may include relative frequencies as weight factors. The mean absolute error uses the same scale as the data being measured. This is known as a scale-dependent accuracy measure and therefore cannot be used to make comparisons between predicted values that use different scales. The mean absolute error is a common measure of forecast error in time series analysis, sometimes used in confusion with the more standard definition of mean absolute deviation. The same confusion exists more generally.

In statistical theory, a U-statistic is a class of statistics defined as the average over the application of a given function applied to all tuples of a fixed size. The letter "U" stands for unbiased. In elementary statistics, U-statistics arise naturally in producing minimum-variance unbiased estimators.

V-statistics are a class of statistics named for Richard von Mises who developed their asymptotic distribution theory in a fundamental paper in 1947. V-statistics are closely related to U-statistics introduced by Wassily Hoeffding in 1948. A V-statistic is a statistical function defined by a particular statistical functional of a probability distribution.

References

  1. "About Us". ValueMetrics Australia. Retrieved 2022-12-31.
  2. "Nick Fisher". The Conversation . n.d. Retrieved 2023-01-23.
  3. "Nicholas Fisher, Sydney Mathematics & Statistics". www.maths.usyd.edu.au. Retrieved 2022-12-31.
  4. "2019 ASA Deming Lecture, Walking with Giants: A Research Odyssey - The W. Edwards Deming Institute" . Retrieved 2022-12-31.