Brian D. Ripley

Last updated

Brian David Ripley
Born (1952-04-29) 29 April 1952 (age 70)
CitizenshipBritish
Alma mater University of Cambridge (BA, MA, PhD)
Awards Smith's Prize (1975)
Davidson Prize (1976)
Adams Prize (1987)
Guy Medal (Silver, 2013)
Scientific career
Institutions Imperial College (197683)
University of Strathclyde (198390)
St Peter's College, Oxford (19902014)
Thesis Stochastic Geometry and the Analysis of Spatial Patterns (1976)
Doctoral advisor David George Kendall
Doctoral students Matthew Stephens
Jonathan Marchini

Brian David Ripley FRSE (born 29 April 1952) is a British statistician. From 1990, he was professor of applied statistics at the University of Oxford and is also a professorial fellow at St Peter's College. He retired August 2014 due to ill health. [1]

Contents

Biography

Ripley has made contributions to the fields of spatial statistics and pattern recognition. His work on artificial neural networks in the 1990s helped to bring aspects of machine learning and data mining to the attention of statistical audiences. [2] He emphasised the value of robust statistics in his books Modern Applied Statistics with S and Pattern Recognition and Neural Networks.

Ripley helped develop the S programming language and its implementations: S-PLUS [ citation needed ] and R. [3] He co-authored two books based on S, Modern Applied Statistics with S and S Programming. [4] [5] From 2000 to 2021 he was one of the most active committers to the R core. [6]

He was educated at the University of Cambridge, where he was awarded both the Smith's Prize (at the time awarded to the best graduate essay writer who had been undergraduate at Cambridge in that cohort) and the Rollo Davidson Prize. The university also awarded him the Adams Prize in 1987 for an essay entitled Statistical Inference for Spatial Processes, later published as a book. [7] He served on the faculty of Imperial College, London from 1976 until 1983, at which point he moved to the University of Strathclyde. [8]

Authored books

Related Research Articles

<span class="mw-page-title-main">Artificial neural network</span> Computational model used in machine learning, based on connected, hierarchical functions

Artificial neural networks (ANNs), usually simply called neural networks (NNs) or neural nets, are computing systems inspired by the biological neural networks that constitute animal brains.

<span class="mw-page-title-main">Statistical inference</span> Process of using data analysis

Statistical inference is the process of using data analysis to infer properties of an underlying distribution of probability. Inferential statistical analysis infers properties of a population, for example by testing hypotheses and deriving estimates. It is assumed that the observed data set is sampled from a larger population.

The theory of statistics provides a basis for the whole range of techniques, in both study design and data analysis, that are used within applications of statistics. The theory covers approaches to statistical-decision problems and to statistical inference, and the actions and deductions that satisfy the basic principles stated for these different approaches. Within a given approach, statistical theory gives ways of comparing statistical procedures; it can find a best possible procedure within a given context for given statistical problems, or can provide guidance on the choice between alternative procedures.

Geostatistics is a branch of statistics focusing on spatial or spatiotemporal datasets. Developed originally to predict probability distributions of ore grades for mining operations, it is currently applied in diverse disciplines including petroleum geology, hydrogeology, hydrology, meteorology, oceanography, geochemistry, geometallurgy, geography, forestry, environmental control, landscape ecology, soil science, and agriculture. Geostatistics is applied in varied branches of geography, particularly those involving the spread of diseases (epidemiology), the practice of commerce and military planning (logistics), and the development of efficient spatial networks. Geostatistical algorithms are incorporated in many places, including geographic information systems (GIS).

<span class="mw-page-title-main">Machine learning</span> Study of algorithms that improve automatically through experience

Machine learning (ML) is a field of inquiry devoted to understanding and building methods that 'learn', that is, methods that leverage data to improve performance on some set of tasks. It is seen as a part of artificial intelligence. Machine learning algorithms build a model based on sample data, known as training data, in order to make predictions or decisions without being explicitly programmed to do so. Machine learning algorithms are used in a wide variety of applications, such as in medicine, email filtering, speech recognition, and computer vision, where it is difficult or unfeasible to develop conventional algorithms to perform the needed tasks.

<span class="mw-page-title-main">Unsupervised learning</span> Machine learning task


Unsupervised learning is a machine learning paradigm for problems where the available data consists of unlabelled examples, meaning that each data point contains features (covariates) only, without an associated label. The goal of unsupervised learning algorithms is learning useful patterns or structural properties of the data. Examples of unsupervised learning tasks are clustering, dimension reduction, and density estimation.

<span class="mw-page-title-main">Information geometry</span> Technique in statistics

Information geometry is an interdisciplinary field that applies the techniques of differential geometry to study probability theory and statistics. It studies statistical manifolds, which are Riemannian manifolds whose points correspond to probability distributions.

<span class="mw-page-title-main">Density estimation</span> Estimate of an unobservable underlying probability density function

In probability and statistics, density estimation is the construction of an estimate, based on observed data, of an unobservable underlying probability density function. The unobservable density function is thought of as the density according to which a large population is distributed; the data are usually thought of as a random sample from that population.

<span class="mw-page-title-main">M. S. Bartlett</span> English statistician

Maurice Stevenson Bartlett FRS was an English statistician who made particular contributions to the analysis of data with spatial and temporal patterns. He is also known for his work in the theory of statistical inference and in multivariate analysis.

<span class="mw-page-title-main">Training, validation, and test data sets</span> Tasks in machine learning

In machine learning, a common task is the study and construction of algorithms that can learn from and make predictions on data. Such algorithms function by making data-driven predictions or decisions, through building a mathematical model from input data. These input data used to build the model are usually divided in multiple data sets. In particular, three data sets are commonly used in different stages of the creation of the model: training, validation and test sets.

Peter Whittle was a mathematician and statistician from New Zealand, working in the fields of stochastic nets, optimal control, time series analysis, stochastic optimisation and stochastic dynamics. From 1967 to 1994, he was the Churchill Professor of Mathematics for Operational Research at the University of Cambridge.

In robust statistics, robust regression seeks to overcome some limitations of traditional regression analysis. A regression analysis models the relationship between one or more independent variables and a dependent variable. Standard types of regression, such as ordinary least squares, have favourable properties if their underlying assumptions are true, but can give misleading results otherwise. Robust regression methods are designed to limit the effect that violations of assumptions by the underlying data-generating process have on regression estimates.

The median polish is a simple and robust exploratory data analysis procedure proposed by the statistician John Tukey. The purpose of median polish is to find an additively-fit model for data in a two-way layout table of the form row effect + column effect + overall median.

Algebraic statistics is the use of algebra to advance statistics. Algebra has been useful for experimental design, parameter estimation, and hypothesis testing.

<span class="mw-page-title-main">Stochastic geometry</span>

In mathematics, stochastic geometry is the study of random spatial patterns. At the heart of the subject lies the study of random point patterns. This leads to the theory of spatial point processes, hence notions of Palm conditioning, which extend to the more abstract setting of random measures.

Terrence L. Fine was an American scientist, engineer and philosopher. He is known especially for his contributions to the defense and development of alternatives to the classical calculus for probabilistic modeling and decision-making. Other contributions include Fine's theorem, the Fine numbers and the Fine–McMillan quantizer. He was the recipient of the first patent awarded in the area of statistical delta modulation.

<span class="mw-page-title-main">Stuart Geman</span> American mathematician

Stuart Alan Geman is an American mathematician, known for influential contributions to computer vision, statistics, probability theory, machine learning, and the neurosciences. He and his brother, Donald Geman, are well known for proposing the Gibbs sampler, and for the first proof of convergence of the simulated annealing algorithm.

<span class="mw-page-title-main">Convolutional neural network</span> Artificial neural network

In deep learning, a convolutional neural network is a class of artificial neural network (ANN), most commonly applied to analyze visual imagery. CNNs are also known as Shift Invariant or Space Invariant Artificial Neural Networks (SIANN), based on the shared-weight architecture of the convolution kernels or filters that slide along input features and provide translation-equivariant responses known as feature maps. Counter-intuitively, most convolutional neural networks are not invariant to translation, due to the downsampling operation they apply to the input. They have applications in image and video recognition, recommender systems, image classification, image segmentation, medical image analysis, natural language processing, brain–computer interfaces, and financial time series.

William Swain Cleveland II is an American computer scientist and Professor of Statistics and Professor of Computer Science at Purdue University, known for his work on data visualization, particularly on nonparametric regression and local regression.

<span class="mw-page-title-main">Outline of machine learning</span> Overview of and topical guide to machine learning

The following outline is provided as an overview of and topical guide to machine learning. Machine learning is a subfield of soft computing within computer science that evolved from the study of pattern recognition and computational learning theory in artificial intelligence. In 1959, Arthur Samuel defined machine learning as a "field of study that gives computers the ability to learn without being explicitly programmed". Machine learning explores the study and construction of algorithms that can learn from and make predictions on data. Such algorithms operate by building a model from an example training set of input observations in order to make data-driven predictions or decisions expressed as outputs, rather than following strictly static program instructions.

References

  1. Professor Ripley's Homepage at Oxford University. Accessed 2015-05-10.
  2. For instance, his book Pattern Recognition and Neural Networks was reviewed in the Journal of the American Statistical Association: Lange, Nicholas (December 1997), "Reviewed Works: Neural Networks for Pattern Recognition by C. M. Bishop; Pattern Recognition and Neural Networks by B. D. Ripley", Journal of the American Statistical Association , 92 (440): 1642–1645, doi:10.2307/2965437, JSTOR   2965437 .
  3. "R: Contributors" . Retrieved 17 April 2020.
  4. Venables, W.N.; Ripley, B.D. (2000). Modern applied statistics with S (4th ed.). Springer. ISBN   0-387-95457-0 . Retrieved 17 April 2020.
  5. Venables, W.N.; Ripley, B.D. (2000). S programming. Springer. ISBN   0-387-98966-8 . Retrieved 17 April 2020.
  6. Eddelbuettel, Dirk (20 March 2021). "An Ode to Stable Interfaces: R and R Core Deserve So Much Praise". Thinking Inside the Box. Archived from the original on 21 March 2021. Retrieved 21 March 2020.
  7. Ripley, B. D. (1988). Statistical Inference for Spatial Processes. Cambridge University Press. pp. iv, vii. ISBN   0-521-35234-7.
  8. Profile of Professor Brian D Ripley (Last edited 1 October 1999 by Brian Ripley). Accessed 2010-12-28.