Adele Cutler

Last updated

Adele Cutler is a statistician known as one of the developers of archetypal analysis [1] and of the random forest technique for ensemble learning. [2] She is a professor of mathematics and statistics at Utah State University. [3]

Originally from England, Cutler moved to New Zealand as a child, [4] and studied mathematics at the University of Waikato and the University of Auckland. [3] She met her husband, statistician Richard Cutler, at the University of Auckland; the couple both went on to graduate study in statistics at the University of California, Berkeley, [4] where she earned a master's degree in 1984 and completed her doctorate in 1988. [3] Her dissertation, Optimization Methods in Statistics, was supervised by Leo Breiman. [5] Her doctoral work with Breiman concerned mathematical optimization techniques in statistics, and introduced archetypal analysis. [6]

After completing her doctorate she joined the faculty at Utah State University in 1988. [4] Her initial research there concerned mixture models, but shifted towards neural networks in the mid-1990s and from there to random decision trees, the basis of the random forest technique. [6]

Related Research Articles

Statistics Study of the collection, analysis, interpretation, and presentation of data

Statistics is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a scientific, industrial, or social problem, it is conventional to begin with a statistical population or a statistical model to be studied. Populations can be diverse groups of people or objects such as "all people living in a country" or "every atom composing a crystal". Statistics deals with every aspect of data, including the planning of data collection in terms of the design of surveys and experiments.

Boosting (machine learning) Method in machine learning

In machine learning, boosting is an ensemble meta-algorithm for primarily reducing bias, and also variance in supervised learning, and a family of machine learning algorithms that convert weak learners to strong ones. Boosting is based on the question posed by Kearns and Valiant : "Can a set of weak learners create a single strong learner?" A weak learner is defined to be a classifier that is only slightly correlated with the true classification. In contrast, a strong learner is a classifier that is arbitrarily well-correlated with the true classification.

Machine learning Study of algorithms that improve automatically through experience

Machine learning (ML) is the study of computer algorithms that improve automatically through experience and by the use of data. It is seen as a part of artificial intelligence. Machine learning algorithms build a model based on sample data, known as "training data", in order to make predictions or decisions without being explicitly programmed to do so. Machine learning algorithms are used in a wide variety of applications, such as in medicine, email filtering, speech recognition, and computer vision, where it is difficult or unfeasible to develop conventional algorithms to perform the needed tasks.

Mathematical statistics

Mathematical statistics is the application of probability theory, a branch of mathematics, to statistics, as opposed to techniques for collecting statistical data. Specific mathematical techniques which are used for this include mathematical analysis, linear algebra, stochastic analysis, differential equations, and measure theory.

Conference on Neural Information Processing Systems

The Conference and Workshop on Neural Information Processing Systems is a machine learning and computational neuroscience conference held every December. The conference is currently a double-track meeting that includes invited talks as well as oral and poster presentations of refereed papers, followed by parallel-track workshops that up to 2013 were held at ski resorts.

Bootstrap aggregating

Bootstrap aggregating, also called bagging, is a machine learning ensemble meta-algorithm designed to improve the stability and accuracy of machine learning algorithms used in statistical classification and regression. It also reduces variance and helps to avoid overfitting. Although it is usually applied to decision tree methods, it can be used with any type of method. Bagging is a special case of the model averaging approach.

Random forest Binary search tree based ensemble machine learning method

Random forests or random decision forests are an ensemble learning method for classification, regression and other tasks that operates by constructing a multitude of decision trees at training time. For classification tasks, the output of the random forest is the class selected by most trees. For regression tasks, the mean or average prediction of the individual trees is returned. Random decision forests correct for decision trees' habit of overfitting to their training set. Random forests generally outperform decision trees, but their accuracy is lower than gradient boosted trees. However, data characteristics can affect their performance.

Leo Breiman

Leo Breiman was a distinguished statistician at the University of California, Berkeley. He was the recipient of numerous honors and awards, and was a member of the United States National Academy of Science.

Donald Geman

Donald Jay Geman is an American applied mathematician and a leading researcher in the field of machine learning and pattern recognition. He and his brother, Stuart Geman, are very well known for proposing the Gibbs sampler and for the first proof of the convergence of the simulated annealing algorithm, in an article that became a highly cited reference in engineering. He is a Professor at the Johns Hopkins University and simultaneously a visiting professor at École Normale Supérieure de Cachan.

The philosophy of statistics involves the meaning, justification, utility, use and abuse of statistics and its methodology, and ethical and epistemological issues involved in the consideration of choice and interpretation of data and methods of statistics.

Emmanuel Candès

Emmanuel Jean Candès is a French statistician. He is a professor of statistics and electrical engineering at Stanford University, where he is also the Barnum-Simons Chair in Mathematics and Statistics. Candès is a 2017 MacArthur Fellow.

Svante Janson is a Swedish mathematician. A member of the Royal Swedish Academy of Sciences since 1994, Janson has been the chaired professor of mathematics at Uppsala University since 1987.

Bin Yu is a Chinese-American statistician. She is currently Chancellor's Professor in the Departments of Statistics and of Electrical Engineering & Computer Sciences at the University of California, Berkeley.

Archetypal analysis in statistics is an unsupervised learning method similar to cluster analysis and introduced by Adele Cutler and Leo Breiman in 1994. Rather than "typical" observations, it seeks extremal points in the multidimensional data, the "archetypes". The archetypes are convex combinations of observations chosen so that observations can be approximated by convex combinations of the archetypes.

Dorota Maria Dabrowska is a Polish statistician known for applying nonparametric statistics and semiparametric models to counting processes and survival analysis. Dabrowska's estimator, from her paper "Kaplan–Meier estimate on the plane" is a widely used tool for bivariate survival under random censoring.

Hélène Menexia Massam was a statistician known for her research on the Wishart distribution and on graphical models. She was a professor of mathematics and statistics at York University in Canada.

In mathematics and computer science, an algorithmic technique is a general approach for implementing a process or computation.

Kimberly Ann Flagg Sellers is an American statistician. She is an associate professor of statistics at Georgetown University, a principal researcher in the Center for Statistical Research and Methodology of the United States Census Bureau, the former chair of the Committee on Women in Statistics of the American Statistical Association, and an elected member of the International Statistical Institute. She specializes in count data and statistical dispersion, and is "the leading expert" on the Conway–Maxwell–Poisson distribution for count data. She has also worked in the medical applications of statistics, and in image analysis for proteomics. In 2018, she was recognized by Mathematically Gifted & Black as a Black History Month Honoree.

Cynthia Diane Rudin is an American computer scientist and statistician specializing in machine learning and known for her work in interpretable machine learning. She is the director of the Prediction Analysis Lab at Duke University, where she is a professor of computer science, electrical and computer engineering, and statistical science.

Sonja Petrović is a Serbian-American statistician and associate professor in the Department of Applied Mathematics, College of Computing, at Illinois Institute of Technology. Her research is focused on mathematical statistics and algebraic statistics, applied and computational algebraic geometry and random graph (network) models. She was elected to the International Statistics Institute in 2015.

References

  1. Morup, Morten; Hansen, Lars Kai (August 2010), "Archetypal analysis for machine learning", 2010 IEEE International Workshop on Machine Learning for Signal Processing (MLSP 2010), pp. 172–177, doi:10.1109/mlsp.2010.5589222
  2. Hastie, Trevor; Tibshirani, Robert; Friedman, Jerome (December 2008), "Random Forests", The Elements of Statistical Learning: Data Mining, Inference, and Prediction (2nd ed.), New York: Springer, pp. 587–604, doi:10.1007/978-0-387-84858-7_15
  3. 1 2 3 Adele Cutler Ph.D., Utah State University Department of Mathematics and Statistics, retrieved 2019-08-24
  4. 1 2 3 Cutler, Adele, Personal Information , retrieved 2019-08-24
  5. Adele Cutler at the Mathematics Genealogy Project
  6. 1 2 "Interview with Adele Cutler: Remembering Leo Breiman", Dan Steinberg's Blog, Salford Systems, retrieved 2019-08-24