Tim Hesterberg

Last updated

Tim Hesterberg is an American statistician. He is a Fellow of the American Statistical Association and currently works as a staff data scientist at Instacart.

Contents

Education and career

Tim Hesterberg graduated with a B.A in mathematics from St. Olaf College and received his Ph.D. in statistics from Stanford University. [1] He is a member of the National Institute of Statistical Sciences(NISS) and was previously on the NISS Board of Trustees. He is currently on the board of the Canadian Statistical Sciences Institute. [1] Hesterberg previously worked as a senior statistician at Google as well as at Franklin and Marshall College, Pacific Gas & Electric Co, and Insightful/MathSoft. [1] He is currently a senior statistician at Instacart.

Research

Hesterberg has contributed to statistical research by writing a textbook, articles, and a software package:

Related Research Articles

<span class="mw-page-title-main">Statistics</span> Study of the collection, analysis, interpretation, and presentation of data

Statistics is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a scientific, industrial, or social problem, it is conventional to begin with a statistical population or a statistical model to be studied. Populations can be diverse groups of people or objects such as "all people living in a country" or "every atom composing a crystal". Statistics deals with every aspect of data, including the planning of data collection in terms of the design of surveys and experiments.

The following outline is provided as an overview of and topical guide to statistics:

<span class="mw-page-title-main">Cross-validation (statistics)</span> Statistical model validation technique

Cross-validation, sometimes called rotation estimation or out-of-sample testing, is any of various similar model validation techniques for assessing how the results of a statistical analysis will generalize to an independent data set. Cross-validation includes resampling and sample splitting methods that use different portions of the data to test and train a model on different iterations. It is often used in settings where the goal is prediction, and one wants to estimate how accurately a predictive model will perform in practice. It can also be used to assess the quality of a fitted model and the stability of its parameters.

<span class="mw-page-title-main">Bootstrap aggregating</span> Ensemble method within machine learning

Bootstrap aggregating, also called bagging, is a machine learning ensemble meta-algorithm designed to improve the stability and accuracy of machine learning algorithms used in statistical classification and regression. It also reduces variance and helps to avoid overfitting. Although it is usually applied to decision tree methods, it can be used with any type of method. Bagging is a special case of the model averaging approach.

A permutation test is an exact statistical hypothesis test making use of the proof by contradiction. A permutation test involves two or more samples. The null hypothesis is that all samples come from the same distribution . Under the null hypothesis, the distribution of the test statistic is obtained by calculating all possible values of the test statistic under possible rearrangements of the observed data. Permutation tests are, therefore, a form of resampling.

In statistics, resampling is the creation of new samples based on one observed sample. Resampling methods are:

  1. Permutation tests
  2. Bootstrapping
  3. Cross validation
<span class="mw-page-title-main">Stepwise regression</span> Method of statistical factor analysis

In statistics, stepwise regression is a method of fitting regression models in which the choice of predictive variables is carried out by an automatic procedure. In each step, a variable is considered for addition to or subtraction from the set of explanatory variables based on some prespecified criterion. Usually, this takes the form of a forward, backward, or combined sequence of F-tests or t-tests.

Bootstrapping is any test or metric that uses random sampling with replacement, and falls under the broader class of resampling methods. Bootstrapping assigns measures of accuracy to sample estimates. This technique allows estimation of the sampling distribution of almost any statistic using random sampling methods.

<span class="mw-page-title-main">Bradley Efron</span> American statistician

Bradley Efron is an American statistician. Efron has been president of the American Statistical Association (2004) and of the Institute of Mathematical Statistics (1987–1988). He is a past editor of the Journal of the American Statistical Association, and he is the founding editor of the Annals of Applied Statistics. Efron is also the recipient of many awards.

<span class="mw-page-title-main">Anscombe's quartet</span> Four data sets with the same descriptive statistics, yet very different distributions

Anscombe's quartet comprises four data sets that have nearly identical simple descriptive statistics, yet have very different distributions and appear very different when graphed. Each dataset consists of eleven (xy) points. They were constructed in 1973 by the statistician Francis Anscombe to demonstrate both the importance of graphing data when analyzing it, and the effect of outliers and other influential observations on statistical properties. He described the article as being intended to counter the impression among statisticians that "numerical calculations are exact, but graphs are rough".

The topic of heteroskedasticity-consistent (HC) standard errors arises in statistics and econometrics in the context of linear regression and time series analysis. These are also known as heteroskedasticity-robust standard errors, Eicker–Huber–White standard errors, to recognize the contributions of Friedhelm Eicker, Peter J. Huber, and Halbert White.

Phillip I. Good is a Canadian-American mathematical statistician. He was educated at McGill University and the University of California at Berkeley.

<span class="mw-page-title-main">Computational statistics</span> Interface between statistics and computer science

Computational statistics, or statistical computing, is the bond between statistics and computer science, and refers to the statistical methods that are enabled by using computational methods. It is the area of computational science specific to the mathematical science of statistics. This area is also developing rapidly, leading to calls that a broader concept of computing should be taught as part of general statistical education.

<span class="mw-page-title-main">Least-angle regression</span>

In statistics, least-angle regression (LARS) is an algorithm for fitting linear regression models to high-dimensional data, developed by Bradley Efron, Trevor Hastie, Iain Johnstone and Robert Tibshirani.

In statistics and machine learning, lasso is a regression analysis method that performs both variable selection and regularization in order to enhance the prediction accuracy and interpretability of the resulting statistical model. The lasso method assumes that the coefficients of the linear model are sparse, meaning that few of them are non-zero. It was originally introduced in geophysics, and later by Robert Tibshirani, who coined the term.

<span class="mw-page-title-main">C. F. Jeff Wu</span> American statistician

Chien-Fu Jeff Wu is the Coca-Cola Chair in Engineering Statistics and Professor in the H. Milton Stewart School of Industrial and Systems Engineering at the Georgia Institute of Technology. He is known for his work on the convergence of the EM algorithm, resampling methods such as the bootstrap and jackknife, and industrial statistics, including design of experiments, and robust parameter design.

<span class="mw-page-title-main">Peter Rousseeuw</span> Belgian statistician (born 1956)

Peter J. Rousseeuw is a statistician known for his work on robust statistics and cluster analysis. He obtained his PhD in 1981 at the Vrije Universiteit Brussel, following research carried out at the ETH in Zurich, which led to a book on influence functions. Later he was professor at the Delft University of Technology, The Netherlands, at the University of Fribourg, Switzerland, and at the University of Antwerp, Belgium. Next he was a senior researcher at Renaissance Technologies. He then returned to Belgium as professor at KU Leuven, until becoming emeritus in 2022. His former PhD students include Annick Leroy, Hendrik Lopuhaä, Geert Molenberghs, Christophe Croux, Mia Hubert, Stefan Van Aelst, Tim Verdonck and Jakob Raymaekers.

<span class="mw-page-title-main">Robert Tibshirani</span> Canadian statistician

Robert Tibshirani is a professor in the Departments of Statistics and Biomedical Data Science at Stanford University. He was a professor at the University of Toronto from 1985 to 1998. In his work, he develops statistical tools for the analysis of complex datasets, most recently in genomics and proteomics.

In statistics and, in particular, in the fitting of linear or logistic regression models, the elastic net is a regularized regression method that linearly combines the L1 and L2 penalties of the lasso and ridge methods.

<span class="mw-page-title-main">Peter Bühlmann</span> Swiss mathematician

Peter Lukas Bühlmann is a Swiss mathematician and statistician.

References

  1. 1 2 3 "Tim Hesterberg". National Institute of Statistical Sciences. 2012-12-17. Retrieved 2021-11-04.
  2. "Mathematical Statistics with Resampling and R, 2nd Edition". Wiley. Retrieved 2021-11-04.
  3. Hesterberg, Tim C. (2015-10-02). "What Teachers Should Know About the Bootstrap: Resampling in the Undergraduate Statistics Curriculum". The American Statistician. 69 (4): 371–386. doi:10.1080/00031305.2015.1089789. ISSN   0003-1305. PMC   4784504 . PMID   27019512.
  4. 1 2 Hesterberg, Tim (2015-04-12), resample: Resampling Functions , retrieved 2021-11-04
  5. Hesterberg, Tim (2008-09-19), "Bootstrap", in D'Agostino, Ralph B.; Sullivan, Lisa; Massaro, Joseph (eds.), Wiley Encyclopedia of Clinical Trials, Hoboken, NJ: John Wiley & Sons, Inc., pp. eoct392, doi:10.1002/9780471462422.eoct392, ISBN   978-0-471-46242-2 , retrieved 2021-11-05
  6. Hesterberg, Tim; Choi, Nam Hee; Meier, Lukas; Fraley, Chris (2008-01-01). "Least angle and ℓ1 penalized regression: A review". Statistics Surveys. 2 (none). arXiv: 0802.0964 . doi: 10.1214/08-SS035 . ISSN   1935-7516. S2CID   13967045.
  7. Fraley, Chris; Hesterberg, Tim (March 2009). "Least Angle Regression and LASSO for Large Datasets". Statistical Analysis and Data Mining. 1 (4): 251–259. doi:10.1002/sam.10021. S2CID   14050406.