Brian D. Ripley

Last updated

Brian David Ripley
Born (1952-04-29) 29 April 1952 (age 72)
CitizenshipBritish
Alma mater University of Cambridge (BA, MA, PhD)
Awards Smith's Prize (1975)
Davidson Prize (1976)
Adams Prize (1987)
Guy Medal (Silver, 2013)
Scientific career
Institutions Imperial College (197683)
University of Strathclyde (198390)
St Peter's College, Oxford (19902014)
Thesis Stochastic Geometry and the Analysis of Spatial Patterns  (1976)
Doctoral advisor David George Kendall
Doctoral students Matthew Stephens
Jonathan Marchini

Brian David Ripley FRSE (born 29 April 1952) is a British statistician. From 1990, he was professor of applied statistics at the University of Oxford and also a professorial fellow at St Peter's College. He retired August 2014 due to ill health. [1]

Contents

Biography

Ripley has made contributions to the fields of spatial statistics and pattern recognition. His work on artificial neural networks in the 1990s helped to bring aspects of machine learning and data mining to the attention of statistical audiences. [2] He emphasised the value of robust statistics in his books Pattern Recognition and Neural Networks and Modern Applied Statistics with S.

Ripley helped develop the S-PLUS programming language [3] [4] and its open source derivative R. [5] He co-authored two books based on S, S Programming and Modern Applied Statistics with S. [3] [4] Since mid-1997 he is a member of the "R Core Team" [6] and from 2000 to 2021 he was one of the most active committers to the R core. [7] The package MASS [8] is one of only fifteen "recommended packages" [9] for R (with June 2024 more than 20,900 [10] ).

He was educated at the University of Cambridge, where he was awarded both the Smith's Prize (at the time awarded to the best graduate essay writer who had been undergraduate at Cambridge in that cohort) and the Rollo Davidson Prize. The university also awarded him the Adams Prize in 1987 for an essay entitled Statistical Inference for Spatial Processes, later published as a book. [11] He served on the faculty of Imperial College, London from 1976 until 1983, at which point he moved to the University of Strathclyde. [12]

Authored books

Related Research Articles

Multivariate statistics is a subdivision of statistics encompassing the simultaneous observation and analysis of more than one outcome variable, i.e., multivariate random variables. Multivariate statistics concerns understanding the different aims and background of each of the different forms of multivariate analysis, and how they relate to each other. The practical application of multivariate statistics to a particular problem may involve several types of univariate and multivariate analyses in order to understand the relationships between variables and their relevance to the problem being studied.

<span class="mw-page-title-main">Neural network (machine learning)</span> Computational model used in machine learning, based on connected, hierarchical functions

In machine learning, a neural network is a model inspired by the structure and function of biological neural networks in animal brains.

Pattern recognition is the task of assigning a class to an observation based on patterns extracted from data. While similar, pattern recognition (PR) is not to be confused with pattern machines (PM) which may possess (PR) capabilities but their primary function is to distinguish and create emergent patterns. PR has applications in statistical data analysis, signal processing, image analysis, information retrieval, bioinformatics, data compression, computer graphics and machine learning. Pattern recognition has its origins in statistics and engineering; some modern approaches to pattern recognition include the use of machine learning, due to the increased availability of big data and a new abundance of processing power.

In machine learning, the perceptron is an algorithm for supervised learning of binary classifiers. A binary classifier is a function which can decide whether or not an input, represented by a vector of numbers, belongs to some specific class. It is a type of linear classifier, i.e. a classification algorithm that makes its predictions based on a linear predictor function combining a set of weights with the feature vector.

Machine learning (ML) is a field of study in artificial intelligence concerned with the development and study of statistical algorithms that can learn from data and generalize to unseen data, and thus perform tasks without explicit instructions. Recently, artificial neural networks have been able to surpass many previous approaches in performance.

Unsupervised learning is a method in machine learning where, in contrast to supervised learning, algorithms learn patterns exclusively from unlabeled data. Within such an approach, a machine learning model tries to find any similarities, differences, patterns, and structure in data by itself. No prior human intervention is needed.

<span class="mw-page-title-main">R (programming language)</span> Programming language for statistics

R is a programming language for statistical computing and data visualization. It has been adopted in the fields of data mining, bioinformatics, and data analysis.

<span class="mw-page-title-main">Information geometry</span> Technique in statistics

Information geometry is an interdisciplinary field that applies the techniques of differential geometry to study probability theory and statistics. It studies statistical manifolds, which are Riemannian manifolds whose points correspond to probability distributions.

<span class="mw-page-title-main">Density estimation</span> Estimate of an unobservable underlying probability density function

In statistics, probability density estimation or simply density estimation is the construction of an estimate, based on observed data, of an unobservable underlying probability density function. The unobservable density function is thought of as the density according to which a large population is distributed; the data are usually thought of as a random sample from that population.

<span class="mw-page-title-main">Linear discriminant analysis</span> Method used in statistics, pattern recognition, and other fields

Linear discriminant analysis (LDA), normal discriminant analysis (NDA), or discriminant function analysis is a generalization of Fisher's linear discriminant, a method used in statistics and other fields, to find a linear combination of features that characterizes or separates two or more classes of objects or events. The resulting combination may be used as a linear classifier, or, more commonly, for dimensionality reduction before later classification.

In machine learning, a common task is the study and construction of algorithms that can learn from and make predictions on data. Such algorithms function by making data-driven predictions or decisions, through building a mathematical model from input data. These input data used to build the model are usually divided into multiple data sets. In particular, three data sets are commonly used in different stages of the creation of the model: training, validation, and test sets.

In robust statistics, robust regression seeks to overcome some limitations of traditional regression analysis. A regression analysis models the relationship between one or more independent variables and a dependent variable. Standard types of regression, such as ordinary least squares, have favourable properties if their underlying assumptions are true, but can give misleading results otherwise. Robust regression methods are designed to limit the effect that violations of assumptions by the underlying data-generating process have on regression estimates.

<span class="mw-page-title-main">Spatial analysis</span> Formal techniques which study entities using their topological, geometric, or geographic properties

Spatial analysis is any of the formal techniques which studies entities using their topological, geometric, or geographic properties. Spatial analysis includes a variety of techniques using different analytic approaches, especially spatial statistics. It may be applied in fields as diverse as astronomy, with its studies of the placement of galaxies in the cosmos, or to chip fabrication engineering, with its use of "place and route" algorithms to build complex wiring structures. In a more restricted sense, spatial analysis is geospatial analysis, the technique applied to structures at the human scale, most notably in the analysis of geographic data. It may also be applied to genomics, as in transcriptomics data.

The median polish is a simple and robust exploratory data analysis procedure proposed by the statistician John Tukey. The purpose of median polish is to find an additively-fit model for data in a two-way layout table of the form row effect + column effect + overall median.

<span class="mw-page-title-main">Stochastic geometry</span>

In mathematics, stochastic geometry is the study of random spatial patterns. At the heart of the subject lies the study of random point patterns. This leads to the theory of spatial point processes, hence notions of Palm conditioning, which extend to the more abstract setting of random measures.

<span class="mw-page-title-main">Stuart Geman</span> American mathematician

Stuart Alan Geman is an American mathematician, known for influential contributions to computer vision, statistics, probability theory, machine learning, and the neurosciences. He and his brother, Donald Geman, are well known for proposing the Gibbs sampler, and for the first proof of convergence of the simulated annealing algorithm.

Convolutional neural network (CNN) is a regularized type of feed-forward neural network that learns feature engineering by itself via filters optimization. Vanishing gradients and exploding gradients, seen during backpropagation in earlier neural networks, are prevented by using regularized weights over fewer connections. For example, for each neuron in the fully-connected layer, 10,000 weights would be required for processing an image sized 100 × 100 pixels. However, applying cascaded convolution kernels, only 25 neurons are required to process 5x5-sized tiles. Higher-layer features are extracted from wider context windows, compared to lower-layer features.

This glossary of artificial intelligence is a list of definitions of terms and concepts relevant to the study of artificial intelligence, its sub-disciplines, and related fields. Related glossaries include Glossary of computer science, Glossary of robotics, and Glossary of machine vision.

RAMnets is one of the oldest practical neurally inspired classification algorithms. The RAMnets is also known as a type of "n-tuple recognition method" or "weightless neural network".

<span class="mw-page-title-main">R package</span> Extensions to the R statistical programming language

R packages are extensions to the R statistical programming language. R packages contain code, data, and documentation in a standardised collection format that can be installed by users of R, typically via a centralised software repository such as CRAN. The large number of packages available for R, and the ease of installing and using them, has been cited as a major factor driving the widespread adoption of the language in data science.

References

  1. Professor Ripley's Homepage at Oxford University. Accessed 8 June 2024.
  2. For instance, his book Pattern Recognition and Neural Networks was reviewed in the Journal of the American Statistical Association: Lange, Nicholas (December 1997), "Reviewed Works: Neural Networks for Pattern Recognition by C. M. Bishop; Pattern Recognition and Neural Networks by B. D. Ripley", Journal of the American Statistical Association , 92 (440): 1642–1645, doi:10.2307/2965437, JSTOR   2965437 .
  3. 1 2 Venables, W.N.; Ripley, B.D. (2000). S programming. Springer. ISBN   0-387-98966-8 . Retrieved 8 June 2024.
  4. 1 2 Venables, W.N.; Ripley, B.D. (2002). Modern applied statistics with S (4th ed.). Springer. ISBN   0-387-95457-0 . Retrieved 8 June 2024.
  5. "Contributors". The R Foundation. April 2024. Retrieved 9 June 2024.
  6. Kurt Hornik and the R Core Team (2023). "R FAQ: What is R?". CRAN . Retrieved 9 June 2024.
  7. Eddelbuettel, Dirk (20 March 2021). "An Ode to Stable Interfaces: R and R Core Deserve So Much Praise". Thinking Inside the Box. Archived from the original on 21 March 2021. Retrieved 8 June 2024.
  8. Ripley, Brian; Venables, Bill; Bates, Douglas M.; Hornik, Kurt; Gebhardt, Albrecht; Firth, David (26 April 2024). "MASS: Support Functions and Datasets for Venables and Ripley's MASS". CRAN . Retrieved 8 June 2024.
  9. Kurt Hornik and the R Core Team (2023). "R FAQ: Add-on packages from CRAN". CRAN . Retrieved 8 June 2024.
  10. "CRAN: Contributed Packages" . Retrieved 9 June 2024.
  11. Ripley, B. D. (1988). Statistical Inference for Spatial Processes. Cambridge University Press. pp. iv, vii. ISBN   0-521-35234-7.
  12. Profile of Professor Brian D Ripley (Last edited 1 October 1999 by Brian Ripley). Accessed 8 June 2024.