Experimentalist approach to econometrics

Last updated

The experimentalist approach to econometrics is a way of doing econometrics that, according to Angrist and Krueger (1999): … puts front and center the problem of identifying causal effects from specific events or situations. These events or situations are thought of as natural experiments that generate exogenous variations in variables that would otherwise be endogenous in the behavioral relationship of interest. An example from the economic study of education can be used to illustrate the approach. Here we might be interested in the effect of effect of an additional year of education (say X) on earnings (say Y). Those working with an experimentalist approach to econometrics would argue that such a question is problematic to answer because, and this is using their terminology, education is not randomly assigned. That is those with different education levels would tend to also have different levels of other variables. And these other variable, many of which would be unobserved (such as innate ability), also affect earnings. This renders the causal effect of extra years of schooling difficult to identify. The experimentalist approach looks for an instrumental variable that is correlated with X but uncorrelated with the unobservables. [1] [2] [3]

Joshua Angrist Israeli American economist

Joshua David Angrist is an Israeli American economist. He is Ford Professor of Economics at the Massachusetts Institute of Technology. Angrist has taught at the university level since 1989. He is known for his use of quasi-experimental research designs to study the effects of public policies and changes in economic or social circumstances. Angrist has also made contributions in theoretical econometrics. He co-founded and serves as a director of the School Effectiveness & Inequality Initiative based within the MIT Department of Economics. The center's work focuses on the economics of education, issues surrounding domestic poverty, and the connections between human capital and the American income distribution.

See also

Related Research Articles

Econometrics is the application of statistical methods to economic data in order to give empirical content to economic relationships. More precisely, it is "the quantitative analysis of actual economic phenomena based on the concurrent development of theory and observation, related by appropriate methods of inference". An introductory economics textbook describes econometrics as allowing economists "to sift through mountains of data to extract simple relationships". The first known use of the term "econometrics" was by Polish economist Paweł Ciompa in 1910. Jan Tinbergen is considered by many to be one of the founding fathers of econometrics. Ragnar Frisch is credited with coining the term in the sense in which it is used today.

A natural experiment is an empirical study in which individuals are exposed to the experimental and control conditions that are determined by nature or by other factors outside the control of the investigators. The process governing the exposures arguably resembles random assignment. Thus, natural experiments are observational studies and are not controlled in the traditional sense of a randomized experiment. Natural experiments are most useful when there has been a clearly defined exposure involving a well defined subpopulation such that changes in outcomes may be plausibly attributed to the exposure. In this sense, the difference between a natural experiment and a non-experimental observational study is that the former includes a comparison of conditions that pave the way for causal inference, but the latter does not.

Simultaneous equation models are a type of statistical model in the form of a set of linear simultaneous equations. They are often used in econometrics. One can estimate these models equation by equation; however, estimation methods that exploit the system of equations, such as generalized method of moments (GMM) and instrumental variables estimation (IV) tend to be more efficient.

In statistics, a spurious relationship or spurious correlation is a mathematical relationship in which two or more events or variables are associated but not causally related, due to either coincidence or the presence of a certain third, unseen factor.

Jerry Allen Hausman is the John and Jennie S. MacDonald Professor of Economics at the Massachusetts Institute of Technology and a notable econometrician. He has published numerous influential papers in microeconometrics. Hausman is the recipient of several prestigious awards including the John Bates Clark Medal in 1985 and the Frisch Medal in 1980.

Alberto Abadie is a Professor of Economics in the Department of Economics at MIT. He is also an Associate Director of the Institute for Data, Systems, and Society at MIT. He was born in the Basque Country, Spain. He received his PhD in Economics from M.I.T. in 1999.

In statistics, econometrics, epidemiology and related disciplines, the method of instrumental variables (IV) is used to estimate causal relationships when controlled experiments are not feasible or when a treatment is not successfully delivered to every unit in a randomized experiment. Intuitively, IVs are used when an explanatory variable of interest is correlated with the error term, in which case ordinary least squares and ANOVA give biased results. A valid instrument induces changes in the explanatory variable but has no independent effect on the dependent variable, allowing a researcher to uncover the causal effect of the explanatory variable on the dependent variable.

Granger causality

The Granger causality test is a statistical hypothesis test for determining whether one time series is useful in forecasting another, first proposed in 1969. Ordinarily, regressions reflect "mere" correlations, but Clive Granger argued that causality in economics could be tested for by measuring the ability to predict the future values of a time series using prior values of another time series. Since the question of "true causality" is deeply philosophical, and because of the post hoc ergo propter hoc fallacy of assuming that one thing preceding another can be used as a proof of causation, econometricians assert that the Granger test finds only "predictive causality".

External validity is the validity of applying the conclusions of a scientific study outside the context of that study. In other words, it is the extent to which the results of a study can be generalized to and across other situations, people, stimuli, and times. In contrast, internal validity is the validity of conclusions drawn within the context of a particular study. Because general conclusions are almost always a goal in research, external validity is an important property of any study. Mathematical analysis of external validity concerns a determination of whether generalization across heterogeneous populations is feasible, and devising statistical and computational methods that produce valid generalizations.

The Rubin causal model (RCM), also known as the Neyman–Rubin causal model, is an approach to the statistical analysis of cause and effect based on the framework of potential outcomes, named after Donald Rubin. The name "Rubin causal model" was first coined by Rubin's graduate school colleague, Paul W. Holland. The potential outcomes framework was first proposed by Jerzy Neyman in his 1923 Master's thesis, though he discussed it only in the context of completely randomized experiments. Rubin, together with other contemporary statisticians, extended it into a general framework for thinking about causation in both observational and experimental studies.

A causal model is a conceptual model that describes the causal mechanisms of a system. Causal models can improve study designs by providing clear rules for deciding which independent variables need to be included/controlled for.

In statistics, a mediation model is one that seeks to identify and explain the mechanism or process that underlies an observed relationship between an independent variable and a dependent variable via the inclusion of a third hypothetical variable, known as a mediator variable. Rather than a direct causal relationship between the independent variable and the dependent variable, a mediation model proposes that the independent variable influences the (non-observable) mediator variable, which in turn influences the dependent variable. Thus, the mediator variable serves to clarify the nature of the relationship between the independent and dependent variables.

In statistics, econometrics, political science, epidemiology, and related disciplines, a regression discontinuity design (RDD) is a quasi-experimental pretest-posttest design that elicits the causal effects of interventions by assigning a cutoff or threshold above or below which an intervention is assigned. By comparing observations lying closely on either side of the threshold, it is possible to estimate the average treatment effect in environments in which randomization is unfeasible. First applied by Donald Thistlethwaite and Donald Campbell to the evaluation of scholarship programs, the RDD has become increasingly popular in recent years.

The Blinder–Oaxaca decomposition is a statistical method that explains the difference in the means of a dependent variable between two groups by decomposing the gap into that part that is due to differences in the mean values of the independent variable within the groups, on the one hand, and group differences in the effects of the independent variable, on the other hand. The method was introduced by sociologist and demographer Evelyn M. Kitagawa in 1955. Ronald Oaxaca introduced this method in economics in his doctoral thesis at Princeton University and eventually published in 1973. The decomposition technique also carries the name of Alan Blinder who proposed a similar approach in the same year. Oaxaca's original research question was the wage differential between two different groups of workers, but his method has since been applied to numerous other topics.

The methodology of econometrics is the study of the range of differing approaches to undertaking econometric analysis.

Causal inference is the process of drawing a conclusion about a causal connection based on the conditions of the occurrence of an effect. The main difference between causal inference and inference of association is that the former analyzes the response of the effect variable when the cause is changed. The science of why things occur is called etiology. Causal inference is an example of causal reasoning.

Causation in economics has a long history with Adam Smith explicitly acknowledging its importance via his (1776) An Inquiry into the Nature and Causes of the Wealth of Nations and David Hume and John Stuart Mill (1848) both offering important contributions with more philosophical discussions. Hoover (2006) suggests that a useful way of classifying approaches to causation in economics might be to distinguish between approaches that emphasize structure and those that emphasise process and to add to this a distinction between approaches that adopt a priori reasoning and those that seek to infer causation from the evidence provided by data. He represented by this little table which useful identifies key works in each of the four categories.

Michael Keane (economist) American economist

Michael Patrick Keane is an American/Australian economist who is currently professor of economics and Australian Research Council laureate fellow at the University of New South Wales. From 2012 to 2017 he was the Nuffield Professor of Economics at the University of Oxford and a professorial fellow of Nuffield College. He is considered one of the world's leading experts in the fields of Choice Modelling, structural modelling, simulation estimation, and panel data econometrics.

Control functions are statistical methods to correct for endogeneity problems by modelling the endogeneity in the error term. The approach thereby differs in important ways from other models that try to account for the same econometric problem. Instrumental variables, for example, attempt to model the endogenous variable X as an often invertible model with respect to a relevant and exogenous instrument Z. Panel data use special data properties to difference out unobserved heterogeneity that is assumed to be fixed over time.

References

  1. Angrist, J. and A. Krueger, (1999) Empirical strategies in labor economics, in: O. Ashenfelter and D. Card, (Eds.), Handbook of labor economics, Vol. 3A. North-Holland, Amsterdam,
  2. Keane, M. P., & Structural Models of Optimization Behavior in Labor, Aging, and Health. (May 1, 2010). Structural vs. atheoretic approaches to econometrics. Journal of Econometrics, 156, 1, 3–May 01, 20.
  3. Angrist, Joshua D., and Jörn-Steffen Pischke. (2008) Mostly harmless econometrics: An empiricist's companion. Princeton university press