NLOGIT

Last updated
NLOGIT
NLOGIT logo.png
NLOGIT screenshot .png
NLOGIT 6 on Microsoft Windows
Original author(s) William H. Greene
Developer(s) Econometric Software, Inc.
Stable release
6 / September 7, 2016
Operating system Windows
Type statistical analysis econometric analysis
License proprietary software
Website nlogit.com

NLOGIT is an extension of the econometric and statistical software package LIMDEP. In addition to the estimation tools in LIMDEP, NLOGIT provides programs for estimation, model simulation and analysis of multinomial choice data, such as brand choice, [1] transportation mode and for survey and market data in which consumers choose among a set of competing alternatives. [2]

Contents

In addition to the economic sciences, NLOGIT has applications in biostatistics, noneconomic social sciences, physical sciences, and health outcomes research. [1]

History

Econometric Software, Inc. was founded in the early 1980s by William H. Greene. NLOGIT was released in 1996 with the development of the FIML nested logit estimator, originally an extension of the multinomial logit model in LIMDEP. The program derives its name from the Nested LOGIT model. With the additions of the multinomial probit model and the mixed logit model among several others, NLOGIT became a self standing superset of LIMDEP. [1]

Models

NLOGIT is a full information maximum likelihood estimator for a variety of multinomial choice models. NLOGIT includes the discrete estimators in LIMDEP plus model extensions for multinomial logit (many specifications), random parameters mixed logit, [3] random regret logit, WTP space specifications in mixed logit, scaled multinomial logit, nested logit, multinomial probit, heteroscedastic extreme value, error components, heteroscedastic logit and latent class models. [1] [4] [5]

Data Analysis

NLOGIT is typically used to analyze individual, cross section data on consumer choices and decisions from multiple alternatives. Analysis may also include market shares or frequency data, data on rankings of alternatives, and panel data from repeated observation of choice situations. [1]

The inference tools for hypothesis testing include the Wald, likelihood ratio and Lagrange multiplier tests and tools for discrete choice analysis, including built-in procedures for testing the IIA assumption of the multinomial logit model.

The models estimated by NLOGIT can be used in ‘what if’ analyses using the model simulation package. The base case model produces fitted probabilities data that aggregate to a prediction of the sample shares for the alternatives in the choice set. The simulator is then used, with the estimation data set or any other compatible data set, to recompute these shares under specified scenarios, such as a change in the price of a particular alternative or a change in household incomes. [5]

Notes

  1. 1 2 3 4 5 Hilbe, Joseph (2006). "A Review of LIMDEP 9.0 and NLOGIT 4.0". The American Statistician. 60 (2): 187–202. doi:10.1198/000313006x110492.
  2. Grebitus, Carola; Steiner, Bodo; Veeman, Michele (2013). "Personal Values and Decision Making: Evidence from Environmental Footprint Labeling in Canada". American Journal of Agricultural Economics. 95 (2): 397–403. doi:10.1093/ajae/aas109.
  3. Boomer, George. "Building a random parameters logit model using NLogit". StatWizards. Archived from the original on 2014-05-12.
  4. Boomer, George. "StatWizards Data Wizard Statistics Programs: LIMDEP's NLOGIT: Nested Logit". StatWizard.
  5. 1 2 McKenzie, Colin; Takaoka, Sumiko (2003). "2002: A LIMDEP Odyssey". Journal of Applied Econometrics. 18 (2): 241–247. doi:10.1002/jae.705.

See also

Related Research Articles

Logit Function in statistics

In statistics, the logit function is the quantile function associated with the standard logistic distribution. It has many uses in data analysis and machine learning, especially in data transformations.

In statistics, the logistic model is used to model the probability of a certain class or event existing such as pass/fail, win/lose, alive/dead or healthy/sick. This can be extended to model several classes of events such as determining whether an image contains a cat, dog, lion, etc. Each object being detected in the image would be assigned a probability between 0 and 1, with a sum of one.

Conjoint analysis

Conjoint analysis is a survey-based statistical technique used in market research that helps determine how people value different attributes that make up an individual product or service.

Heteroscedasticity Statistical property

In statistics, a vector of random variables is heteroscedastic if the variability of the random disturbance is different across elements of the vector. Here, variability could be quantified by the variance or any other measure of statistical dispersion. Thus heteroscedasticity is the absence of homoscedasticity. A typical example is the set of observations of income in different cities.

In statistics, a generalized linear model (GLM) is a flexible generalization of ordinary linear regression. The GLM generalizes linear regression by allowing the linear model to be related to the response variable via a link function and by allowing the magnitude of the variance of each measurement to be a function of its predicted value.

In statistics, a probit model is a type of regression where the dependent variable can take only two values, for example married or not married. The word is a portmanteau, coming from probability + unit. The purpose of the model is to estimate the probability that an observation with particular characteristics will fall into a specific one of the categories; moreover, classifying observations based on their predicted probabilities is a type of binary classification model.

In economics, discrete choice models, or qualitative choice models, describe, explain, and predict choices between two or more discrete alternatives, such as entering or not entering the labor market, or choosing between modes of transport. Such choices contrast with standard consumption models in which the quantity of each good consumed is assumed to be a continuous variable. In the continuous case, calculus methods can be used to determine the optimum amount chosen, and demand can be modeled empirically using regression analysis. On the other hand, discrete choice analysis examines situations in which the potential outcomes are discrete, such that the optimum is not characterized by standard first-order conditions. Thus, instead of examining "how much" as in problems with continuous choice variables, discrete choice analysis examines "which one". However, discrete choice analysis can also be used to examine the chosen quantity when only a few distinct quantities must be chosen from, such as the number of vehicles a household chooses to own and the number of minutes of telecommunications service a customer decides to purchase. Techniques such as logistic regression and probit regression can be used for empirical analysis of discrete choice.

RATS, an abbreviation of Regression Analysis of Time Series, is a statistical package for time series analysis and econometrics. RATS is developed and sold by Estima, Inc., located in Evanston, IL.

The topic of heteroskedasticity-consistent (HC) standard errors arises in statistics and econometrics in the context of linear regression and time series analysis. These are also known as heteroskedasticity-robust standard errors, Eicker–Huber–White standard errors, to recognize the contributions of Friedhelm Eicker, Peter J. Huber, and Halbert White.

Mixed logit is a fully general statistical model for examining discrete choices. It overcomes three important limitations of the standard logit model by allowing for random taste variation across choosers, unrestricted substitution patterns across choices, and correlation in unobserved factors over time. Mixed logit can choose any distribution for the random coefficients, unlike probit which is limited to the normal distribution. It is called "mixed logit" because the choice probability is a mixture of logits, with as the mixing distribution. It has been shown that a mixed logit model can approximate to any degree of accuracy any true random utility model of discrete choice, given appropriate specification of variables and the coefficient distribution.

In statistics and econometrics, the multinomial probit model is a generalization of the probit model used when there are several possible categories that the dependent variable can fall into. As such, it is an alternative to the multinomial logit model as one method of multiclass classification. It is not to be confused with the multivariate probit model, which is used to model correlated binary outcomes for more than one independent variable.

Choice modelling attempts to model the decision process of an individual or segment via revealed preferences or stated preferences made in a particular context or contexts. Typically, it attempts to use discrete choices in order to infer positions of the items on some relevant latent scale. Indeed many alternative models exist in econometrics, marketing, sociometrics and other fields, including utility maximization, optimization applied to consumer theory, and a plethora of other identification strategies which may be more or less accurate depending on the data, sample, hypothesis and the particular decision being modelled. In addition, choice modelling is regarded as the most suitable method for estimating consumers' willingness to pay for quality improvements in multiple dimensions.

The Heckman correction is a statistical technique to correct bias from non-randomly selected samples or otherwise incidentally truncated dependent variables, a pervasive issue in quantitative social sciences when using observational data. Conceptually, this is achieved by explicitly modelling the individual sampling probability of each observation together with the conditional expectation of the dependent variable. The resulting likelihood function is mathematically similar to the tobit model for censored dependent variables, a connection first drawn by James Heckman in 1974. Heckman also developed a two-step control function approach to estimate this model, which avoids the computational burden of having to estimate both equations jointly, albeit at the cost of inefficiency. Heckman received the Nobel Memorial Prize in Economic Sciences in 2000 for his work in this field.

Kenneth E. Train is an Adjunct Professor of Economics at the University of California, Berkeley, United States. He is also Vice President of NERA Economic Consulting, Inc. in San Francisco, California. He received a Bachelors in Economics at Harvard and PhD from UC Berkeley. He specializes in econometrics and regulation, with applications in energy, environmental studies, telecommunications and transportation.

The following outline is provided as an overview of and topical guide to regression analysis:

A limited dependent variable is a variable whose range of possible values is "restricted in some important way." In econometrics, the term is often used when estimation of the relationship between the limited dependent variable of interest and other variables requires methods that take this restriction into account. For example, this may arise when the variable of interest is constrained to lie between zero and one, as in the case of a probability, or is constrained to be positive, as in the case of wages or hours worked.

LIMDEP

LIMDEP is an econometric and statistical software package with a variety of estimation tools. In addition to the core econometric tools for analysis of cross sections and time series, LIMDEP supports methods for panel data analysis, frontier and efficiency estimation and discrete choice modeling. The package also provides a programming language to allow the user to specify, estimate and analyze models that are not contained in the built in menus of model forms.

In statistics and econometrics, the maximum score estimator is a nonparametric estimator for discrete choice models developed by Charles Manski in 1975. Unlike the multinomial probit and multinomial logit estimators, it makes no assumptions about the distribution of the unobservable part of utility. However, its statistical properties are more complicated than the multinomial probit and logit models, making statistical inference difficult. To address these issues, Joel Horowitz proposed a variant, called the smoothed maximum score estimator.

In statistics, linear regression is a linear approach for modelling the relationship between a scalar response and one or more explanatory variables. The case of one explanatory variable is called simple linear regression; for more than one, the process is called multiple linear regression. This term is distinct from multivariate linear regression, where multiple correlated dependent variables are predicted, rather than a single scalar variable.

References