Check yourself before you wreck yourself: Assessing discrete choice models through predictive simulations

06/06/2018
by   Timothy Brathwaite, et al.
0

Typically, discrete choice modelers develop ever-more advanced models and estimation methods. Compared to the impressive progress in model development and estimation, model-checking techniques have lagged behind. Often, choice modelers use only crude methods to assess how well an estimated model represents reality. Such methods usually stop at checking parameter signs, model elasticities, and ratios of model coefficients. In this paper, I greatly expand the discrete choice modelers' assessment toolkit by introducing model checking procedures based on graphical displays of predictive simulations. Overall, my contributions are as follows. Methodologically, I introduce a general and 'semi-automatic' algorithm for checking discrete choice models via predictive simulations. By combining new graphical displays with existing plots, I introduce methods for checking one's data against one's model in terms of the model's predicted distributions of P (Y), P (Y|X), and P (X | Y). Empirically, I demonstrate my proposed methods by checking the models from Brownstone and Train (1998). Through this case study, I show that my proposed methods can point out lack-of-model-fit in one's models and suggest concrete model improvements that substantively change the results of one's policy analysis. Moreover, the case study highlights a practical trade-off between precision and robustness in model checking.

READ FULL TEXT VIEW PDF
POST COMMENT

Comments

There are no comments yet.

Authors

page 19

11/24/2019

The Bouquet Algorithm for Model Checking Unbounded Until

The problem of verifying the "Unbounded Until" fragment in temporal logi...
05/12/2020

N-PAT: A Nested Model-Checker

N-PAT is a new model-checking tool that supports the verification of nes...
04/18/2018

Checking the Model and the Prior for the Constrained Multinomial

The multinomial model is one of the simplest statistical models. When co...
12/20/2017

Bayesian model checking: A comparison of tests

Two procedures for checking Bayesian models are compared using a simple ...
08/06/2019

Does Preliminary Model Checking Help With Subsequent Inference? A Review And A New Result

Statistical methods are based on model assumptions, and it is statistica...
10/03/2019

GRAVITAS: A Model Checking Based Planning and Goal Reasoning Framework for Autonomous Systems

While AI techniques have found many successful applications in autonomou...
01/05/2022

Systematic assessment of the quality of fit of the stochastic block model for empirical networks

We perform a systematic analysis of the quality of fit of the stochastic...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

“Check. Your. Work. It is okay to make mistakes, but it is not okay to turn in work that is riddled with mistakes that checking your work could have caught.” These sage words were spoken by my undergraduate mathematics professor, Dr. Kenneth Holladay. Unfortunately, somewhere between Dr. Holladay’s calculus classes and my graduate discrete choice studies, checking one’s work was demoted from a main actor to a supporting or missing role in the ever-unfolding data analysis drama.

The tragedy of this situation is that regardless of whether one is working on undergraduate calculus problems or building predictive models for governments and corporations, there is a need to check one’s work. From an econometric perspective, we do not typically have economic theory to guide all aspects of our model specification (Dagsvik, 2017, p.85). Accordingly, it would be wise to check the implications of our (likely) mis-specified models. And from a broader statistical view, it is often noted that, “all models [including discrete choice models] are false, but some are useful” (Box, 1976). Again, we find checking for correctness or truth of one’s model to be hopeless a priori, and the best we can do is check how our models misrepresent reality (Gelman et al., 1996, pp.734, 800).

For discrete choice models, many such checks can be broadly categorized as checking for “overfitting” or “underfitting.” overfitting is the degree to which one’s model performs better when evaluated on data used for estimation (i.e. training or in-sample data) as compared to data that was not used for estimation (i.e. testing or out-of-sample data) (Fariss and Jones, 2017, p.6). Colloquially, severe overfitting is characteristic of models that “learned” the patterns in the in-sample data at the expense of patterns that generalize to out-of-sample data. Conversely, underfitting is the degree to which one’s model has failed to learn patterns in one’s in-sample data (Fariss and Jones, 2017, p.6). Underfit models typically perform poorly on some problem-specific metric for both in-sample and out-of-sample data. See Figure 1 for an illustrative depiction of these two concepts in a binary discrete choice setting. Here, the data generating relationship is shown in dark blue. The underfit model, shown in dashed purple, is overly simple and has failed to capture observed patters in the data. Conversely, the overfit model is overly complex; in dashed and dotted light orange, the overfit model captures spurious patterns in the data that will not generalize to new samples from the data generating process.

Figure 1: Example of underfitting and overrfitting

Currently, numerous methods exist for diagnosing overfitting. Such techniques include cross-validation (Stone, 1974), resampling-methods such as the “0.632+” bootstrap (Efron and Tibshirani, 1997), and the use of held-out data (Steckel and Vanhonacker, 1993). Importantly, these techniques are highly general. Instead of testing for specific, hypothesized causes of overfitting, these methods check for overfitting’s presence and extent. This generality promotes use: researchers can use the methods regardless of the specific analyses111Of course, one’s analytic task can impact the way such checks are implemented. For instance, cross-validation is performed differently when dealing with cross-sectional panel data as compared to purely cross-sectional data. being performed. As a result, the overall quality of predictive analyses has increased. A mostly standardized tool-box and language for assessing overfitting has developed, enabling researchers to better evaluate predictive models.

Comparatively, checks for underfitting are far less common. For instance, within the discrete choice literature, few checks for the general presence and extent of underfitting have been proposed. As examples, consider two popular discrete choice textbooks: “Discrete Choice Methods with Simulation, 2nd Edition” by Kenneth Train (2009) and “Discrete Choice Analysis: Theory and Application to Travel Demand” by Ben-Akiva and Lerman (1985). In Train (2009)

, only 1 of the 90 sections in the book (Section 11.5) discusses model checking in a manner that can be described as a general check for underfitting. Even then, the checks discussed in that section are only applicable to a particular class of models—mixed logit models. Now consider

Ben-Akiva and Lerman (1985). In that text, only 1 of 78 sections (Section 7.6) describes procedures for directly checking one’s model against one’s data for lack-of-fit. Unfortunately, this paucity of methods for assessing underfitting goes beyond these two textbooks. It is characteristic of the field.

To improve the state of practice, I propose and demonstrate methods for using predictive simulations to perform general checks of underfitting in discrete choice models. In particular, I make the following contributions. Methodologically, I introduce a general and ‘semi-automatic’ algorithm for checking discrete choice models via predictive simulations. Moreover, by combining new graphical displays with existing techniques, I introduce methods for checking one’s data against one’s model in terms of the model’s predicted distributions of choices (), choices given explanatory variables (), and explanatory variables given choices (). Empirically, I demonstrate my proposed methods by checking the models from Brownstone and Train (1998). Through this case study, I show that my proposed methods can point out lack-of-model-fit in one’s models and suggest concrete model improvements that substantively change the results of one’s policy analysis.

Importantly, these model checking techniques complement rather than replace careful economic and behavioral scrutiny of one’s models. By ensuring adherence with sound economic principles while also diagnosing and reducing the amount of underfitting in our models, we increase our chances of creating maximally useful representations of reality.

The remainder of this paper is structured as follows. To ease discussion of this paper’s ideas, Section 2 immediately describes the proposed methodology and provides examples of the graphical displays introduced in this paper. Once readers are familiar with the proposed simulation-based checks for underfitting, Section 3 relates these methods to existing model tests and assessments in the discrete choice and statistics literature. Next, Section 4 will continue this paper’s case study, showing how the displays from Section 2 suggest avenues for model-improvement, not merely model criticism. Section 5 will then recap the paper’s contributions and conclude.

Finally, Appendix A contains answers to frequently asked questions about the interpretation and use of predictive simulations for model checking, and Appendix B contains model estimation results that are of secondary importance to the paper but perhaps of interest to some readers.

2 Methodology: Demonstration and Description

In this section, I demonstrate and describe this paper’s proposed methodology for assessing underfitting in discrete choice models. It cannot be reiterated strongly enough that the main focus of this paper’s techniques is to show and quantify the existence and extent of underfitting of one’s model. As a result, this section will not contain any information about how one might use the proposed techniques to help suggest model improvements. Such a discussion is of secondary importance to this paper and is deferred to Section 4. Moreover, not all techniques in this section will provide obvious clues to how one can improve one’s model. This is to be expected. As an analogy, consider thermometers. They are useful for detecting the presence and severity of a fever, even though they don’t suggest strategies for curing the underlying problem. Likewise, this paper’s methods are valuable for detecting and measuring underfitting, regardless of the degree to which they suggest ways to fix the underlying causes of the lack-of-fit.

Disclaimers aside, I use a case study of vehicle choice by Brownstone and Train (1998) to provide an initial illustration of the proposed techniques. Though published twenty years ago, I chose Brownstone and Train’s paper for three reasons: (1) because the data is freely available for download by the public, (2) because the article is well-known and well-cited within the discrete choice community, and (3) because the article represents standard best practices within the field. As such, by applying this paper’s proposed model checking techniques to the vehicle choice models of Brownstone and Train (1998), we gain insight into the types of underfitting that one might find, even in models built by the best of us.

Accordingly, Subsection 2.1.1 describes the vehicle choice data and model from Brownstone and Train (1998). Next, Subsections 2.1.2 - 2.1.8 show the seven types of plots that will be used in this paper to assess a model’s lack-of-fit. These plots are introduced in increasing order of granularity. For example, the first two plots assess whether one’s model fits one’s data in an aggregate sense; the second and third plots assess a model’s lack-of-fit for particular groups of observations for a given alternative; and the final three plots assess a model’s lack-of-fit with respect to a particular alternative and particular values of a chosen explanatory variable.

Finally, once specific examples of the various plots have been introduced, Subsection 2.1.9 will discuss the results and interpretations of these plots for diagnosing the character and extent of underfitting in Brownstone and Train (1998)’s multinomial logit (MNL) model. Subsection 2.2 will then abstract from the case study to present this paper’s general methodology for checking underfitting in discrete choice models. To make the methods as widely useful as possible, special attention will be paid to how the general method may be adapted for use with ‘unlabelled alternatives.’ Moreover, Subsection 2.2 will present a semi-automatic algorithm for applying this paper’s methodology, further decreasing the minimum efforts needed to apply this paper’s techniques.

2.1 A Case Study of Brownstone and Train (1998) (Part 1)

2.1.1 Data and modeling context

The data in this case study comes from a stated-preference study of household attitudes and preferences for alternative-fuel vehicles (Brownstone and Train, 1998; Brownstone et al., 1996). The sampling plan used random telephone-digit dialing followed by a postal survey to collect data from 4,654 households. The collected data included each household’s choice amongst six alternatives (three different fuel types and two body types per fuel type), the attributes of each alternative, and some limited characteristics of the household. For a detailed description of the collected attributes and characteristics, see Table 1. Together, Brownstone and Train used the collected222For further information about the data and collection protocol, see Brownstone and Train (1998); Brownstone et al. (1996). To access the raw data, visit the website of the Journal of Applied Econometrics and download the data associated with McFadden and Train (2000). data to estimate four vehicle choice models: a MNL model, a probit model, and two mixed logit models.

Variable Definition
Price/log(income) Purchase price (in thousands of dollars) divided by log(household income in thousands)
Range Hundreds of miles vehicle can travel between refuellings/rechargings
Acceleration Tens of seconds required to reach 30 mph from stop
Top speed Highest attainable speed in hundreds of MPH
Pollution Tailpipe emissions as fraction of those for new gas vehicle
Size 0=mini, 0.1 =subcompact, 0.2=compact, 0.3=mid-size or large
‘Big enough’ 1 if household size is over 2 and vehicle size is 3; 0 otherwise
Luggage space Luggage space as fraction of comparable new gas vehicle
Operating cost Cost per mile of travel (tens of cents): home recharging for electric vehicle; station refuelling otherwise
Station availability Fraction of stations that can refuel/recharge vehicle
Sports utility vehicle 1 if sports utility vehicle, 0 otherwise
Sports car 1 if sports car
Station wagon 1 if station wagon
Truck 1 if truck
Van 1 if van
EV 1 if electric vehicle (EV)
Commute & EV 1 if electric vehicle and commute miles/day
College & EV 1 if electric vehicle and some college education
CNG 1 if compressed natural gas (CNG) vehicle
Methanol 1 if methanol vehicle
College & Methanol 1 if methanol vehicle and some college education
Table 1: Variable Definitions

For part one of this case study, I demonstrate this paper’s methods with Brownstone and Train’s MNL model333However, I return to the assessment of Brownstone and Train’s chosen mixed logit model in Section 4.. I use this model because it is likely to be familiar to the largest spectrum of readers. Moreover, producing the predictive simulations for Brownstone and Train’s chosen mixed logit model seemed unnecessarily tedious. This would require writing custom code to estimate their model, yet the MNL can be used just as effectively to demonstrate this paper’s model checking techniques.

Justifications aside, Table 2, shows the results of reproducing Brownstone and Train’s MNL model. As expected, the results match those of the original study. Moreover, almost all variables in the model are statistically significant, and no variables have an obviously wrong sign. Brownstone and Train’s MNL model therefore seems to be a good ‘working model’ that one can begin refining by iteratively: (1) criticizing the working model, (2) positing and estimating a new model that addresses the earlier criticisms, and (3) repeating steps 1 and 2 (Box, 1976, p.795).

Variable Estimate Std. err
Price over log(income) -0.185** 0.027
Range (units: 100mi) 0.350** 0.027
Acceleration (units: 0.1sec) -0.716** 0.111
Top speed (units: 0.01mph) 0.261** 0.081
Pollution -0.444** 0.102
Size 0.934** 0.316
Big enough 0.143 0.077
Luggage space 0.501** 0.191
Operation cost -0.768** 0.076
Station availability 0.413** 0.096
Sports utility vehicle 0.820** 0.141
Sports car 0.637** 0.148
Station wagon -1.437** 0.062
Truck -1.017** 0.049
Van -0.799** 0.047
EV -0.179 0.172
Commute < 5 & EV 0.198* 0.084
College & EV 0.443** 0.109
CNG 0.345** 0.092
Methanol 0.313** 0.103
College & Methanol 0.228* 0.089
Log-likelihood -7,391.830
Note: * means and ** means .
Table 2: Brownstone and Train’s MNL model

Before seeing any data, Brownstone and Train were critical of the MNL model because its “independence from irrelevant alternatives” (IIA) property implies restrictive substitution patterns among the various alternatives being modeled (Brownstone and Train, 1998, pp.109-110). This a-priori criticism led Brownstone and Train to estimate the probit and mixed logit models mentioned above since those models do not have the IIA property. To complement such data-agnostic reasons for model refinement, this paper suggests a data-driven, graphical approach to model criticism by using predictive simulations to check for underfitting. In Subsections 2.1.2 - 2.1.8 that follow, I will use Brownstone and Train’s MNL model to showcase seven graphical model-checks that I believe can be widely useful for discrete choice models. Then, Subsection 2.2 will generalize these plots by detailing an overarching methodology for model checking via predictive simulations.

2.1.2 Log-predictive plots

This paper’s first type of graphical model check is the log-predictive plot. This plot compares a scalar summary of the predictive ability of a given model on the observed data against the distribution formed by evaluating the scalar measure on simulated datasets. Because this plot evaluates a single number, the observed log-predictive value, this plot is amongst the most aggregate checks of underfitting.

Figure 2: Log-predictive plot of Brownstone and Train’s MNL model

For a concrete example, consider Figure 2. Here, the scalar measure is the log-likelihood, . In this equation, is the design matrix for the observed data;

is a vector of outcomes with one element per observation (

) per available alternative (); denotes whether observation chose alternative ; contains the explanatory variables for observation and alternative ; is the choice set for individual ; is the maximum likelihood estimate (MLE) displayed in Table 2; and is calculated from Brownstone and Train’s MNL model.

The observed log-likelihood, , is compared against the distribution formed by computing for simulated outcomes , for draws . Here, is the total number of simulated datasets. In this case, we first simulate values of the choice model parameters () from the asymptotic sampling distribution of . In other words, if

is a multivariate normal distribution with mean

and covariance matrix , then I draw from where is the negative inverse of the Hessian matrix, evaluated at . Then for each draw of , I compute . Finally, for each observation , I simulate an outcome

from the categorical distribution with probabilities

. Once the simulated values of

have been calculated, their distribution can be displayed using histograms, kernel density estimates, and/or empirical cumulative density functions. In Figure

2, a kernel density estimate was used.

Mechanics aside, the rationale behind the log-predictive plots is as follows. If one’s model fits one’s data well, then the simulated outcomes from one’s model should be similar to one’s observed outcomes. The log-predictive plots measure similarity in terms of some log-predictive measure such as the log-likelihood evaluated at the MLE of the observed data. In a bayesian context, one might instead use the log pointwise-predictive density where is the total number of parameter samples from one’s posterior/prior distribution and is a single sample (Gelman et al., 2014, p.169). In both bayesian and frequentist analyses, one expects that, for a well-fitting model, the distribution of log-predictive values formed by will be symmetric and unimodal with its mode at the observed value . This would indicate that the log-predictive values of the observed outcomes are very likely, and that one’s model is not biased towards producing outcomes that have consistently lower or higher log-predictive values. Deviations from this ideal represent model misfit.

As exemplified by Figure 2, checking an estimated model with a log-predictive plot does not always reveal a lack of fit444On the other hand, log-predictive plots can reveal a lack of fit. The expectation of the distribution of simulated log-predictive values is not the observed log-predictive value. This is true despite the fact that the expectation of the simulated parameters is the point-estimate of the model parameters. As is well-known, . That is to say, the expectation of a function need not equal the function evaluated at an expectation. This is because frequentist analyses typically optimize for the log-likelihood, and even bayesian analyses revise their prior distributions to concentrate in areas of the parameter space with higher log-likelihoods (and hence high log pointwise-predictive values). However, if one’s log-predictive plot of an estimated model displays a marked lack of fit, this means that something is wrong. For example one’s posited model may be woefully inadequate, one’s estimation routine may not be converging, or one’s computations may be suffering from numeric errors. In this sense, log-predictive plots are basic, graphical sanity-checks.

Beyond checking one’s estimated models, log-predictive plots are most useful for checking one’s prior distributions in a bayesian analysis. In this setting, where analysts often choose their prior distributions for convenience rather than to reflect their true beliefs, the log-predictive plots can reveal large discrepancies between one’s implied prior beliefs about the outcomes for one’s dataset and the actual outcomes. If one notices such a conflict between one’s data and one’s prior after (1) observing a lack-of-fit between one’s posterior distribution and one’s data and (2) ruling out a lack-of-fit between one’s likelihood and one’s data, then one’s ‘convenience prior’ should be adjusted to reflect one’s true a-priori information and beliefs.

2.1.3 Market share plots

Market share plots are a more detailed check for lack-of-fit than log-predictive plots. Instead of using a scalar summary, market share plots judge one’s model using a vector of numbers, one for each alternative. In particular, market share plots compare the observed versus predicted numbers (or equivalently, percentages) of observations that choose each alternative. See Figure 3 for an example.

Figure 3: Market share plot of Brownstone and Train’s MNL model

Here, the measure of interest is where is the total number of alternatives in one’s dataset. As before, is compared against the distribution formed by computing for all simulated datasets, indexed by . For this plot, and all others that follow, the simulation procedure is the same as described in Subsection 2.1.2.

In general, one can use many different plots to visualize this model check. One can iterate through available alternatives, each time plotting the observed number of decision makers that choose the current alternative versus the predicted distribution of the number of decision makers choosing the current alternative. As with the log-predictive plots, the distribution can be visualized using histograms, kernel density estimates, and/or empirical cumulative distribution functions. The various comparisons may also be concisely displayed using boxplots, as in Figure

3. Here, the alternatives have been taken to be the various fuel types of each vehicle.

Overall, market share plots assess the degree of underfitting in the marginal distribution, , implied by one’s model. These plots check whether one’s model has learned the aggregate pattern of responses in one’s data. Note that such patterns may be of great importance for decision making. As an example, it would be counter-productive to use a vehicle choice model for inventory stocking decisions if the model consistently under-predicted the amount of electric vehicles by a large amount. Even beyond the decision-making use case, it is important to understand the ways that one’s model does not fit one’s data. For example, if one wants to understand a particular choice process, then market share plots can alert one to the need for caution when interpreting models that frequently under- or over-predict one’s alternatives of interest.

As a final caveat, note that a market share plot, as with all the other graphical checks in this paper, is most useful when it “measures some aspect of the data that might not be accurately fit by the model” (Gelman and Meng, 1996, p.191). Practically, this means that if the model being checked ensures that the predicted shares of each alternative match the observed shares, market share plots are not very useful. The most common of such models is an MNL model with alternative specific constants555This property is due to the first order conditions of the MNL model with alternative specific constants. See Train (2009, p.62) for more details.. Most other commonly used choice models (e.g. nested logit and mixed logit models) do not ensure that one’s predicted market shares match the observed shares (Donoso et al., 2011; Klaiber and von Haefen, 2018). As a result, market share plots remain a useful tool for checking generalized extreme value (GEV) models, mixed logit models, MNL models without alternative specific constants, and most other models.

2.1.4 Binned reliability plots

Conceptually, market share plots judge a model’s ability to predict for the various alternatives. In this subsection, binned reliability plots will instead be used to directly evaluate a model’s predictions of , thereby providing a more detailed assessment of a model’s predicted probabilities for each alternative. An example of a binned reliability plot for Brownstone and Train’s MNL model is shown in Figure 4 for methanol vehicles. At its core, binned reliability plots group (i.e. bin) observations according to their predicted probabilities of choosing a given alternative, and then for each group, the mean predicted probability is plotted on the x-axis while the proportion of observations that choose the alternative is plotted on the y-axis. Hopefully, one’s observed reliability curve would lie in the middle of the simulated reliability curves and exhibit the same visual trends as the simulated curves. See the following paragraphs for more information and for an explanation of the many principles, choices, and visual elements that comprise this diagram.

Figure 4: Binned reliability plot for methanol vehicles using Brownstone and Train’s MNL model

Starting with its name, a reliability plot tries to display the “reliability” of a set of probabilistic predictions, where “reliability refers to the degree of correspondence between forecast probabilities and observed relative frequencies” (Murphy and Winkler, 1977, p.41). If a model is perfectly reliable, then that model’s predicted probabilities should equal the “true probabilities”. As a result, a plot of true probabilities on the y-axis versus predicted probabilities on the x-axis should yield the straight line . Of course, one does not observe the true probabilities, one merely observes the discrete outcomes . Since discrete choice models generally yield distinct probability predictions for each observation, the sample average of the outcomes at each predicted probability will typically be a poor approximation of the true probabilities. Some smoothing of the observed responses (i.e. averaging outcomes across different predicted probabilities) will therefore be required to yield useful approximations of the desired reliability plot.

On one hand, continuous smooths such as locally weighted scatterplot smoothers (Cleveland, 1979), spline smoothers (Silverman, 1984), or kernel smoothers (Ghosh, 2018) might be preferred since the resulting smoothed curves usually avoid the sharp boundary effects that can occur with ad-hoc grouping. However, I have found such continuous smoothing procedures to be unhelpful for discrete outcome data, especially as the degree of class imbalance increases. As the proportion of observations choosing a given alternative moves away from 50%, the continuous smooths often led to reliability diagrams that were horizontal lines. This is due to the fact that if a given alternative is unlikely to be chosen, then unless one’s model has very high discriminatory power, observations with are likely surrounded by observations with . As a result, local averaging (as with continuous smooths) is likely to always yield values near zero, thus producing a horizontal line.

An alternative method of smoothing that avoids the problems just described is based on binning the predicted probabilities. This is the smoothing method used to create the reliability plots in this paper. Using this method, one decides on a number of partitions of the predicted probabilities. Then, the predicted probabilities are ordered from smallest to largest and divided into partitions of roughly equal size. The mean of each partition’s predicted probabilities is then plotted on the x-axis and the mean of the outcomes are plotted on the y-axis.

When determining the number of partitions, it should be noted that this choice is subject to a “precision/robustness” or “bias/variance” trade-off. As the number of partitions increases, one’s assessment of the quality of a particular probability prediction is less biased by predicted probabilities of different magnitudes. However, as the number of partitions increases, the variance of the mean outcome within each partition increases due to the reduction in the number of observations being averaged over. The negative effects of having too many partitions is especially acute in class-imbalanced scenarios. Here, each partition’s number of observations with

may rapidly decrease to zero as the number of partitions increase, thereby destroying one’s ability to accurately estimate the “true” probabilities. As a result, one can see that increased precision in one’s assessments of comes at the cost of decreased robustness of those assessments. As a general guide, I have empirically found that using ten partitions seems to be adequate for most visualizations666For example, Figure 4 uses ten partitions.. However, I suggest that one experiment with more and less partitions to see whether one’s conclusions about the fit of one’s model is sensitive to this choice.

Aside from the smoothing method and number of partitions for bin-based smooths, construction of a reliability plot also requires one to choose which predictions will be displayed on the x-axis. For ease of interpretation, I recommend choosing a single set of predicted probabilities for binning and plotting on the x-axis. Such a choice will ensure that each bin is associated with a unique and constant set of observations. With this constant association, one will be able to generate a reference distribution for where is the total number of partitions, is the set of observations in each partition, and is the number of observations in partition . If multiple sets of predicted probabilities were used (such as one set per simulated ), then the bin memberships of the observations might change with each draw, thus complicating the description of what we are simulating the reference distribution of.

Now, given that a single set of predicted probabilities is desired, is an obvious candidate in a frequentist setting. In a bayesian setting, the point estimate that seems most appropriate is the posterior mean where is a sample from one’s posterior distribution. Either way, once one has chosen a set of predicted probabilities, one can generate a distribution for using the simulated from one’s model. One will then be able to compare with this reference distribution using a plot such as Figure 4. The ‘observed’ curve, i.e. , will hopefully (1) lie near the line , indicating perfect reliability, (2) lie in the middle of the reference distribution, and (3) exhibit the same visual trends as the reference distribution. Deviations from any of these three ideals represent (1) ways that one’s model has failed to capture patterns in the data and (2) systematic over- or under-predictions of choice probabilities for the alternative of interest.

2.1.5 Binned marginal model plots

Like the binned reliability plots of the last subsection, binned marginal model plots provide a way to check one’s model of for underfitting. Binned marginal model plots display the binned value of a continuous variable on the x-axis and each bin’s average probabilities (both observed and simulated) for a given alternative on the y-axis. As a result, this plot checks the degree to which one’s model faithfully replicates the empirical relationships between the variable on the x-axis and the true (but unobserved) . For example, in Figure 5 the binned marginal model plot displays the predicted versus simulated relationship between and the relative cost of a sports utility vehicle (SUV).

Figure 5: Binned marginal model plot for sports utility vehicles using Brownstone and Train’s MNL model

As with reliability plots, the fact that the true probabilities are unobserved means that some smoothing is required. The original marginal model plots (Pardoe and Cook, 2002) used continuous smoothing techniques such as smoothing splines, but for the reasons discussed in Section 2.1.4, bin-based smoothing will be used in this paper. Accordingly, one must therefore determine the number of bins for the plot. To avoid repetition, see the previous subsection for a discussion of considerations for choosing the number of bins. In Figure 5, 10 partitions were used.

Beyond determining the number of partitions, one must make two additional choices. First, and most importantly, one must decide on the scalar variable to be plotted on the x-axis. Typical candidates to be plotted on the x-axis include the explanatory variables in one’s model and variables that are not currently included in one’s model. As noted in Pardoe and Cook (2002), marginal model plots can find variables that add useful information to one’s model, even when the p-values of a Wald-test might erroneously suggest such variables’ exclusion.

Lastly, one must decide whether to display the averages of the binned, predicted probabilities in addition to the simulated probabilities (i.e. the averages of the binned, simulated choices). For preciseness, note that for each sample (), one first predicts , and then one uses those probabilities to simulate a vector of choices . The averages of the binned, predicted probabilities are the averages of each bin’s , and the averages of the binned, simulated choices are the averages of each bin’s . Quantitatively, plotting the averages of the binned, predicted probabilities shows one the relationships that are to be asymptotically expected if one’s model is true. Plotting the simulated probabilities shows the sampling distribution (at one’s current sample size) of the empirical probabilities, assuming one’s model is true.

Personally, I have found that plotting the averages of the binned, predicted probabilities clarifies the relationship that one’s model expects asymptotically. For instance, examine Figure 5. Here, and in general, the averages of the binned, predicted probabilities (shown in light blue) are more tightly clustered than the simulated probabilities (shown in light purple). Quantitatively, this occurs because the averages of the binned, predicted probabilities do not have the inflated variance that the simulated probabilities incur due to the randomness that comes from simulation. Qualitatively, the more tightly clustered curves make it easier to see if one’s observed data differs from the (asymptotic) expectations of one’s model. These asymptotic judgements can then be qualitatively considered as one uses the binned, simulated probabilities to provide finite-sample assessments of the discrepancies between one’s observed data and data generated from one’s model. For instance, in Figure 5, the observed data (shown in dark blue) clearly differs from the asymptotic expectations of Brownstone and Train’s MNL model. This judgement reinforces the fact that the model’s simulated choices (of equal size as the observed data) show major differences from the observed data.

2.1.6 Simulated histograms

So far, the discussed model checking methods have all been aggregate checks of underfitting. To check a model’s implied distribution for , the log-predictive plots checked underfitting at the level of the entire dataset, and market share plots checked underfitting for each alternative. Becoming more disaggregate, binned, reliability plots and binned, marginal model plots checked for underfitting in a model’s predictions of for specific alternatives using observations that had been grouped according to values of a chosen univariate variable. In the reliability plots, observations were grouped according the their predicted values of or . For the binned, marginal model plots, observations were grouped according to whatever variable was chosen for the x-axis.

In this subsection (and the following two subsections), I introduce this paper’s most granular model checking methods. These techniques judge a model’s implied distribution of for underfitting of specific alternatives at each observed value of a chosen univariate variable. Here, I introduce simulated histograms to assess underfitting of a discrete explanatory variable. Figure 6 provides a concrete example. In this plot, I assess Brownstone and Train’s MNL model with respect to its ability to predict the number of choice observations that choose regular cars with an operating cost of 2 cents per mile. From the plot, we see that while 835 observations actually chose a regular car with an operating cost of 2 cents per mile, 96% of the simulated datasets had less than 835 observations choosing regular cars that operate at 2 cents per mile. The observed dataset is therefore unusual with respect to this feature, and the simulated histogram displays the model’s lack-of-fit at an operating cost of 2 cents per mile for regular cars.

Figure 6: Simulated histograms for regular cars using Brownstone and Train’s MNL model

Generalizing beyond this specific example, one constructs a simulated histogram by first choosing a given criteria. The number of observations meeting this criteria will be one’s ‘discrepancy’ measure, . Then, this discrepancy measure is tabulated in both the observed dataset and in each of one’s simulated datasets. Next, the number of simulated observations meeting the criteria is plotted in a histogram to show the distribution of the discrepancy measure, assuming that one’s model is correct. Finally, one draws a vertical line on the plot to indicate the value of the observed discrepancy . For the simulated histograms in this paper, the discrepancy measure is the number of observations that choose a given alternative when the discrete variable of interest takes on some specified value. As an example, in Figure 6, the alternative is regular cars; the variable of interest is the operating cost of the vehicle, and the specified value is 2 cents per mile.

2.1.7 Simulated kernel density estimates

In the last subsection, simulated histograms were used to judge underfitting for specific alternatives at each observed value of a discrete variable. This section introduces simulated kernel density estimates777

Note that kernel density estimates are non-parametric estimates of the probability density function of a random variable. For an econometrician’s and statistician’s introduction to KDEs, see

Racine (2008) and Wasserman (2006), respectively. (KDEs): an analogous model checking method that assesses underfitting for specific alternatives with respect to the individual values of a continuous variable. Figure 7 presents an example of simulated KDEs.

Figure 7: Simulated kernel density estimates for electric cars using Brownstone and Train’s MNL model

In this example, the alternative of interest is electric vehicles and the variable of interest is the relative price: the price of the electric vehicle, divided by the natural log of the purchaser’s household income. As a result, Figure 7 plots the kernel density estimates of the distribution of relative prices for decision makers that were observed or simulated to have chosen an electric vehicle. From the plot, we see that (1) around a relative price of 2, the observed KDE strays outside the band of simulated KDEs; (2) between a relative price of approximately 5 and 6.5, the observed KDE drifts to the bottom edge of the simulated KDEs; and (3) between a relative price of 6.5 and 7, the observed KDE is mostly flat whereas the simulated KDEs maintain a negative slope. As noted in Section 2.1.4, one wants the curves generated by the simulated data to envelop and straddle the curve generated by the observed data, and one wants the curves generated by the simulated data to display the same visual trends as the curve generated by the observed data. All three points mentioned above highlight values of the relative price where the simulated KDEs fail to envelop, straddle, or mimic the observed KDE. I.e., the simulated KDE plot reveals multiple ways in which Brownstone and Train’s MNL model fails to capture patterns in the observed relationship between relative price and the choice of electric vehicles.

Now, to construct a simulated KDE plot as used in this paper, the procedure is as follows. First, choose a continuous, univariate variable to be plotted on the x-axis. Next, choose the alternative that the KDEs will be plotted with respect to. Given the selected alternative and variable, construct the plot by iterating through each of the simulated vectors of choices () and the observed choices (). For each set of choices, select the observations that choose the alternative of interest, and from those selected observations, plot a KDE of the chosen variable. Note that in plotting the KDE of the chosen variable, one must choose a kernel and bandwidth for the KDE. In practice (and in Figure 7), I have found that the standard normal distribution kernel and Scott’s (1992) default bandwidth procedures are typically adequate for visualization purposes. Researchers are of course free, and encouraged, to experiment with alternative kernels and bandwidth selection procedures.

2.1.8 Simulated cumulative distribution functions

This paper’s final type of plot is a plot of simulated cumulative distribution functions (CDFs). Conceptually, simulated CDFs serve the same role as simulated KDEs: by visualizing , these plots are used to assess underfitting for specific alternatives with respect to the individual values of a continuous variable. Figure 8 gives an example of a simulated CDF plot.

Figure 8: Simulated CDFs for sport utility vehicles using Brownstone and Train’s MNL model

As in Section 2.1.5, the variable of interest is again the relative price, and the alternative of interest is again SUVs. This time, however, observations will not be binned according to their relative price values. Instead, for each vector of choices (both simulated and observed), the observations that choose SUVs will be sorted in ascending order of their relative price values. These relative prices will be plotted on the x-axis. Then, for each relative price (for the current vector of choices), the y-axis value will be the proportion of observations that chose SUVs and have relative prices less than or equal to the x-axis value. As always, we would like the CDF generated by the observed data to lie in the center of the CDFs generated by the simulated data, and we would like the simulated CDFs to display the same visual trends as the observed CDF. In the example given by Figure 8, the observed CDF is clearly at the bottom edge or outside the band of simulated CDFs for most of the observed relative price values. This indicates that Brownstone and Train’s MNL model is underfit with respect to the relative price variable and SUVs.

Now, given that simulated CDFs and simulated KDEs perform the same conceptual role in assessing underfitting, why should both types of plots be used? The answer is that the two types of plots are complementary. For instance, in Section 2.1.7, I noted that one has to choose a bandwidth and kernel in order to construct the simulated KDEs. Simulated CDF plots require no such ad-hoc choices. Moreover, while simulated KDEs highlight the exact values where one’s model expect more or less observations (i.e. where one’s model predicts a higher or lower density) than observed in one’s data, it can be difficult to visually identify situations when one’s model has consistently but only slightly over- or under-predicted the density for a sequence of values. Such situations are easily identified with a simulated CDF plot. Conversely, while simulated CDFs excel at exposing underfitting, simulated KDEs are potentially more useful for fixing it. Indeed, knowing a variable’s density with respect to a given alternative can suggest helpful transformations of that variable in one’s systematic utility functions (Kay and Little, 1987; Bergtold et al., 2009).

Lastly, the particular set of simulated CDFs in Figure 8 may seem redundant given that the relationship between the choice of SUVs and relative price was assessed in Figure 5. Again, these two plots are found to be complementary. The simulated CDF plot does not bin any observations, so we do not lose any information, and we are able to see that the relationship between relative price and SUVs is poorly captured for all relative prices below about 7.5. Because the simulated curves in Figure 5 do not exhibit the same visual trend as the observed curve, the marginal model plot also indicates the lack-of-fit in the MNL model’s estimated relationship between the relative price and the choice of SUV. However, Figure 5 also points out that the partitions with predicted probabilities between approximately 0.20 and 0.27 are actually predicted accurately. This highlights the fact that at least some of the predicted probabilities of choosing SUV are trustworthy, even if the overall estimated relationship is not. Depending on one’s use case, this may or may not be sufficient.

2.1.9 Discussion

In Subsections 2.1.2 - 2.1.8, I presented seven graphical checks, five of which (Figures 4 - 8) revealed patterns of underfitting in Brownstone and Train’s MNL model. In particular, the reliability plot in Figure 4 reveals that the MNL model’s probability predictions for methanol vehicles are generally unreliable. Analogous plots888These plots are not shown due to space considerations. However, they are available upon request and at https://github.com/timothyb0912/check-yourself/blob/master/notebooks/_06-tb-Make_introductory_plots.ipynb. for the other vehicle fuel and body types show that this pattern of inaccurate probability predictions is consistent across multiple dimensions. The conclusion from these plots is that overall, Brownstone and Train’s MNL model underfits numerous aspects of the observed data.

By itself, the fact that the MNL model does not adequately represent the data generating process is unsurprising. Indeed, Brownstone and Train (1998) began their paper with the presumption that the MNL model would be overly simple and would be better replaced with models such as the mixed logit and probit models. Instead, the checks in Figures 5 - 8 are most useful for the model-free way that they reveal lack-of-fit in the MNL model. Specifically, without relying on comparisons between two potentially flawed models, Figures 5 - 8 show that Brownstone and Train’s MNL model does not adequately relate individual choices to relative costs or operating costs for the various vehicle fuel and body types. All together, these checks allow one to directly assess whether key relationships have been captured in one’s model. As a result, model assessment is effectively separated from the (arguably less-important) task of model comparison: a task that often reduces to checking whether one potentially flawed model has a predictive score (e.g. log-likelihood) that is ‘sufficiently’ higher than another potentially flawed model.

2.2 General Methodology

Now that Section 2.1 has introduced specific examples of this paper’s model checking techniques, the general methodology can be introduced clearly.

In particular, the model checking methodology based on predictive simulations is as follows. First, let the observed data be represented by . Note that not all of the variables in need be part of one’s choice model specification. Next, specify one or more statistics of the data. For each statistic, :

  1. Calculate for the observed data, yielding a value . Note that may be a vector, e.g. an empirical cumulative density function.

  2. From some distribution, draw samples of one’s model parameters999Note, we sample from a distribution of parameters because we are interested in assessing our model, not in assessing a particular point estimate for our model. We draw parameter samples to account for our uncertainty about the true parameter value if one is a bayesian or about the possible datasets if one is a frequentist.. For bayesian analyses, the distribution could be one’s prior or posterior distribution (depending on whether one wants to check the prior or posterior). For frequentist analyses, the distribution may be the sampling distribution of one’s model. The sampling distribution can be approximated using the asymptotic sampling distribution or by using the paired bootstrap (Brownstone and Valletta, 2001).

    1. For each sampled vector of parameters , calculate the predicted probabilities of each available alternative being chosen by the decision makers in one’s dataset.

    2. For each set of predicted probabilities, simulate a choice for each decision maker.

    3. For each set of simulated choices , calculate .

  3. Compare with the distribution formed by the samples .

  4. The more ‘surprising’ is given one’s simulated values of , the more one’s model underfits one’s data (in terms of the statistic ).

As examples, I specified as a function with vector outputs for the simulated KDEs and CDFs; for the binned, marginal model plots and binned, reliability plots; and for the market share plots. Conversely, I specified as having scalar outputs for the simulated histograms and log-predictive plots.

Regardless of how one specifies , the model-checking procedure described above requires an analyst to choose one or more statistics to be checked. In general, researchers should always specify statistics that are of substantive importance in one’s study. For example, if building a bicycle mode choice model using revealed preference data, one might check the observed number of cyclists in different geographic regions of one’s study, even if one had not explicitly built a ‘spatial choice model.’

However, despite the admonition to thoughtfully define the statistics that are being checked, I readily acknowledge that analysts may not always have a set of statistics in mind. Moreover, it is useful to know, in many specific but common ways, whether or not one’s model fits one data. To this end, I propose the following new method for “automatically” checking the fit of one’s model to one’s data.

  1. In the case of “labeled” alternatives (e.g. shopping brand or travel mode choice) choose one or more alternatives of interest.

    1. For each alternative of interest, construct market share plots and binned reliability plots to perform a basic assessment one’s probability predictions.

    2. For each alternative of interest, cycle through the related, available, explanatory variables. For instance, continuing with bicycle mode choice, if one is looking at cyclists, then examine bicycle travel time instead of automobile travel time.

    3. If the explanatory variable being examined is discrete, iterate through the distinct values of the variable.

    4. For each value of the discrete explanatory variable being examined, define to be the number of individuals with the current value of the current explanatory variable that choose the current alternative.

    5. If the explanatory variable being examined is continuous, then define to be the empirical cumulative density function or a kernel density estimate of the explanatory variable, conditional on choosing the current alternative.

  2. In the case of “unlabeled” alternatives,(e.g. travel route choice or vehicle choice), if one has a discrete explanatory variable of great importance (e.g. vehicle type in a vehicle choice model), then treat the values of that discrete explanatory variable as one’s “alternative labels” and proceed as in Step 1.

  3. In the case of unlabeled alternatives without a discrete explanatory variable of great importance, one can treat the “labels” as “chosen or not”, and define the alternatives of interest to be the “chosen” alternatives. Then proceed as in Step 1.

This proposed method is nearly automatic because it merely requires one to define “proxy labels” for unlabelled alternatives and to define the variables of interest that are related to each distinct alternative. Note that since random utility models are the most common type of discrete choice model, each alternative’s systematic utility function is likely to specify which variables relate to which alternatives.

3 Relations to Existing Literature

Section 2 introduced seven specific model checks and a general method for using predictive simulations to assess underfitting in one’s discrete choice model. In this section, I clarify how my proposed methods deliver increased benefits compared to pre-existing model checking and model comparison techniques. First, Section 3.1 shows how my proposed methods continue the progression of graphical model checks in statistics from residual plots to posterior predictive checks. Here, I highlight how my proposed methods facilitate the routine use of checks for underfitting and how they offer new ways to check one’s model specifications. Next, Section 3.2 contrasts my proposed methods with model comparison techniques—the more common class of methods in discrete choice. In this subsection, I explain how my proposed model checking techniques avoid the limitations imposed by model comparison methods when checking for underfitting.

3.1 Graphical Model Checks

Historically, much research has focused on the development of graphical methods for checking predictive models. Dating back to at least the 1960s (Anscombe and Tukey, 1963; Cox and Snell, 1968)

, the most widely known graphical model checks are residual plots for linear regression. Conceptually, residual plots are similar to the plots described in Section

2: they implicitly compare an observed vector of residuals against the vectors that would be expected if one’s model was true. Specifically, if one’s data are truly generated by one’s model, then under common linear regression assumptions101010For e.g., the assumptions of the Gauss-Markov theorem., one’s residuals (i.e. the observed data minus one’s predictions) are independently and identically distributed according to a normal distribution with mean zero and constant variance. Accordingly, one’s residual plots would be expected to look like a series of random draws from a (scaled) standard normal distribution. This simplicity of the residual plot’s expected patterns enabled one, without simulation, to visually determine when the observed residuals were markedly different from residuals that could be generated by one’s model. Depending on the type of residual plots being examined, marked differences between expected and observed patterns of residuals could then be used to identify when and how one’s linear function was misspecified, to identify when transformations of the outcome variable would be useful, to identify when major assumptions such as independence across observations were violated, etc. (Anscombe and Tukey, 1963; Cox and Snell, 1968).

Following the tremendous success of graphical model checks for linear regression, and encouraged by the unifying framework of generalized linear models, researchers in the 1980’s sought to extend graphical methods like residual plots to discrete choice models (e.g. binary and multinomial logistic regression). Unfortunately, such efforts were frustrated by the discrete dependent variables inherent to discrete choice models

(Albert and Siddhartha, 1995, pp. 747-748). In particular, discrete dependent variables complicate the expected patterns in one’s residual plots, thereby precluding easy, implicit assessment of one’s model. To cope with these challenges, even the earliest efforts leveraged many modern features of graphical model checks. For example, Landwehr et al. (1984) is one of the first papers to introduce graphical checks for underfitting in (binary) discrete choice models. As with my proposed methods in Section 2, Landwehr et al. (1984, p. 64) use predictive simulations to create reference distributions that are explicitly displayed on one’s plots. Moreover, Landwehr et al. (p. 66) also use smoothing to deal with the discrete nature of the quantities being plotted. Even the idea of using the observed distributions of to check one’s model was already noted in Donald B. Rubin’s discussion (pp. 79 - 80) of Landwehr et al. (1984). Though differing from Rubin in implementation, his idea directly inspired my creation of the simulated histograms, KDEs, and CDFs.

In summary, the “classical” diagnostic plots for discrete choice models compare and contrast with this paper’s proposed techniques as follows. As a point of commonality, the classical plots pioneered many of the techniques upon which my proposed methods rest: using predictive simulations to construct reference distributions, using smoothing to plot discrete quantities, and checking as well as . However, research on classical diagnostic plots was more limited than this paper’s proposed techniques. In particular, classical diagnostic plots were produced in a piecemeal fashion: one stroke of scholarly genius at a time. No unifying connections were made between these few distinct sets of graphical checks, and no unifying methodology was proposed for creating tailored graphical checks for one’s unique modeling problems.

For example, Murphy and Winkler (1977, 1987) introduced reliability plots to judge one’s probability predictions, and a reference distribution based on predictive simulations was developed for these plots in Bröcker and Smith (2007). In the intervening years between these efforts, Landwehr et al. (1984) introduced empirical probability plots, local mean deviance plots (a graphical form of subgroup analyses), and partial residual plots. Here, empirical probability plots (indirectly) assess the same relationship as reliability plots, but no connection between the two techniques was made111111For completeness, note that reliability plots can be considered graphical representations of the numeric Hosmer-Lemeshow goodness-of-fit statistic (Lemeshow and Hosmer Jr, 1982). Reliability plots are also similar in spirit to the Horowitz-Louviere test (Horowitz, 1985; Horowitz and Louviere, 1993) in the sense that they directly compare observed choices versus predicted probabilities.. Finally, model checking graphs such as the “Separation Plot” (Greenhill et al., 2011) and the “Heat Map Plot” (Esarey and Pierce, 2012) have been recently created by researchers in quantitative political science. Of particular interest is the heat map plot which is based on familiar techniques such as smoothing and creating reference distributions based on predictive simulations. Unfortunately, although this plot is the exact analogue of the reliability plot (simply using continuous smoothing instead of binning), the connection between the two types of plots has been ignored.

In contrast to these classical diagnostic plots, a second generation of graphical model checks has emerged since the mid-1990s. Beginning in earnest with the posterior predictive checks of Gelman et al. (1996), this framework unifies the various model checking plots. For example, building on the approaches described in Gelman (2004) and Buja et al. (2009), all the graphs in Section 2.1

correspond to different choices of a test statistic

and different ways of plotting the reference distribution that is created by the predictive simulations. Aside from these two choices, graphical model checks basically follow the seven-step procedure detailed in Section 2.2. This clarity and abstractness of the procedure for creating graphical model checks has numerous benefits. For instance, the abstractness of the procedure has allowed bayesians and frequentists to make use of a single methodology, merely substituting their own preferred methods for sampling from their estimated models. Moreover, by appropriately defining , the common methodology makes it easy to define custom checks that are tailored to one’s modeling context.

Clearly, the model checking techniques proposed in this paper draw heavily upon the posterior predictive checking framework. For instance, Evans and Moshonov (2006, p. 896) note that

“[…] the prior induces a probability distribution on the set of possible likelihood functions via the prior predictive probability measure

. If the observed likelihood […] is a surprising value from this distribution, then this would seem to indicate that a prior-data conflict exists.”

The log-predictive plots merely re-purpose this observation by substituting the posterior distribution for the prior distribution. Similarly, the market share plots are simply graphical depictions of the market segment prediction tests of Ben-Akiva and Lerman (1985, Section 7.6), alongside a reference distribution created by predictive simulations. Continuing through the remaining plots, the binned reliability plots are the same as their classical counterparts, but enhanced with a graphical depiction of the reference distribution created by the predictive simulations from one’s model. Marginal model plots are explicit descendants of the posterior predictive checking framework (Pardoe and Cook, 2002, pp.266-267), and this paper has only modified these plots to use bin-based smoothing instead of continuous smooths. Lastly, I noted above that as far back as 1984, Donald Rubin [in Landwehr et al. (1984, pp. 79-80)] suggested that

“[…] for diagnostic checking of estimates of , my preference is reversed. I believe that we should examine the consequences of a proposed estimate of using the discriminant analysis formulation [i.e. ].”

My proposed plots of simulated histograms, KDEs, and CDFs are merely new applications of Rubin’s insight to examine the implied class-conditional distributions , complemented by reference distributions that are simulated from one’s estimated model.

Given that the plots of Section 2.1 are firmly grounded in the posterior predictive framework, some readers may wonder if there is anything truly new in this paper’s proposed methodology. There is. Thus far, practitioners who use posterior predictive checks have restricted themselves to developing custom graphical checks for each new choice modeling endeavor. See Gelman et al. (2000) and Carlin et al. (2001) for examples from statisticians, and see Jedidi et al. (2003), Musalem et al. (2008), and Gilbride and Lenk (2010) for examples from the discrete choice literature. Except for Jedidi et al. (2003), the checks in these papers are mostly specific to their particular studies/models and would not be immediately applicable to analysts in general.

In Section 2.2, however, I proposed a new approach for automatic model checking of discrete choice models. Compared to the existing literature, my model checking recommendations are more broadly applicable and more comprehensive. First, my proposed methods are applicable whenever one has disaggregate choice data available. The applicability of my methods does not depend on the specific choice model being used. Secondly, my proposed checks are more comprehensive than checks provided in most other papers because they assess model fit on multiple dimensions. They allow one to assess the fit of one’s model overall (using log-predictive plots), on particular alternatives (using market share plots and binned reliability plots), and on particular variables (using binned marginal model plots, simulated histograms, simulated KDEs, and simulated CDFs).

Of course, I am not claiming that my proposed model checks account for all types of checks that one may wish to do. For references to other model checks that may be of use, see the last paragraph of Appendix A. Moreover, I do not make the claim that my proposed methods will always uncover an underfit model, nor that my proposed methods are the best automatic model checking procedures that one can follow. Finally, I do not think there is any problem with using model checking techniques that are tailored to one’s problem and chosen modeling technique. On the contrary, I believe that such targeted model checks are complementary to the checks described in this paper. I simply note the lack of commonly-used or agreed-upon checks for discrete choice models, and I propose my methodology as one useful approach to facilitate the routine use of graphical model checks for underfitting.

3.2 Model Comparison versus Model Checking

Thus far, most of the articles cited in this literature review have been examples of graphical model checking from the statistics literature. This is by necessity as opposed to choice. Overall, there have been few examples of graphical model checking within the discrete choice literature. The papers of Dunn et al. (1987) and Nagel and Hatzinger (1992) are notable exceptions. Instead, most model diagnostics within the discrete choice literature are instances of model comparison as opposed to model checking.

For concreteness, take “A Diagnostic Test for the Multinomial Logit Model” by Tse (1987). This diagnostic test does not directly compare the observed data with the estimated model to check for lack-of-fit. Instead, Tse proposes “a Lagrange multiplier test for the multinomial logit model against the dogit model as the alternative hypothesis” (1987). In general, discrete choice model checks for the “independence from irrelevant alternatives” (IIA) property are all model comparisons. Some, as in Tse (1987), test the MNL model against discrete choice models that do not have the IIA property. Other tests such as the Hausman-McFadden test (Hausman and McFadden, 1984), compare a given MNL model against the same MNL model estimated on a subset of alternatives. Likewise, model comparisons are implicitly or explicitly behind Wald, Lagrange Multiplier, and Likelihood Ratio tests for omitted variables or parameter heterogeneity (both systematic and unobserved heterogeneity as embodied in mixed logit models). All of these tests can be viewed as tests of a restricted model against an unrestricted model.

The major problem with such comparisons is that both models being compared may be misspecified. Accordingly, the fact that one’s model passes a model comparison test may not mean anything more than the fact that one’s entertained model was better than an even worse model. As noted by Zheng (2008), “[…] the performance of these parametric tests rely on correct specification of the alternative model. If the alternative is misspecified, those tests may have low or no power.” Model checks as described in this paper avoid the issue of mis-specifying one’s alternative model by directly testing one’s entertained model against the observed data.

4 Case Study (Part 2)

With a greater understanding of how this paper’s proposed methodology is positioned with respect to the relevant literature, this section returns to the case study began earlier. In Section 2.1, I used this paper’s proposed model checks to reveal serious underfitting in Brownstone and Train’s MNL model. In this section, I will show how the proposed checks can be used to reduce the observed lack-of-fit.

Overall, the model checking procedure of Section 2.1 provided two main insights. First, underfitting was revealed by all of the plots involving the relative price variable (i.e. price over log income). This can be seen, for instance, in Figures 5, 7, and 8. To partially address this issue, the relative price variable was re-specified as piecewise linear: one term accounted for relative prices less than 3, and the other term accounted for relative prices greater than 3. Three was chosen as the knot location for the piecewise linear specification based on plots such as Figures 5 and 8. These graphs showed marked underfitting at the relative price value of 3. For a more formal method of model improvement, one could have instead re-specified the relative price variable using techniques such as generalized additive models (Hastie and Tibshirani, 1986; Abe, 1999).

Secondly, the “automatic” model checking procedure of Section 2.2 revealed that for many variables, the pattern of underfitting often differed by both vehicle fuel type and body type. Such differing patterns indicate that there is heterogeneity in how the original MNL model differs from the true data generating process, and as result, there is likely heterogeneity in the true relationship between the various explanatory variables and the probability of a given alternative being chosen. To accommodate such heterogeneity, I re-specified the MNL by including an interaction between the body type variable and the piecewise linear price variable, as well as an interaction between the body type and the range, acceleration, top speed, pollution, vehicle size, and operating cost variables. Similarly, I also added an interaction term between the fuel type variable and the piecewise linear price variables, range, top speed, pollution, vehicle size, and operating cost variables. Acceleration was not interacted with fuel type since the simulated histograms of acceleration by vehicle fuel type did not suggest severe underfitting.

In summary, I changed the specification of Brownstone and Train

’s MNL model to include a piecewise linear specification of the relative price variable and to include interactions between many of the explanatory variables and the vehicle body type and fuel type. These changes were not exhaustive of all improvements suggested by my proposed model checking procedure. For instance, simulated histograms of the range, top speed, pollution, vehicle size and operating cost variables all suggest that these variables should not be entered using a single linear term in the model. These variables have few unique values, and given that the patterns of underfitting vary across these values, the variables in question should be represented as categorical (e.g. using dummy-variables) rather than continuous. Despite recognizing these issues, I restricted myself to the changes described in this paragraph’s opening sentence for simplicity

121212Using dummy variables for the vehicle attributes mentioned above would have necessitated the use of regularization to ensure monotonicity in the (ordered) attribute values.. In particular, my aim in this section is to show that the proposed model checking procedures can suggest concrete model improvements, not to demonstrate how one can make all possible improvements and not to demonstrate the best possible way to make model improvements.

As a result, the parameter estimates are not of major importance in this study. For space reasons, interested readers will therefore have to find the typical table of parameter estimates in Appendix B.

Summary Measure Original Expanded
In-sample Log-Likelihood -7,391.830 -7,311.634
McFadden’s 0.111 0.113
AIC 14,825 14,787
Out-of-sample Log-Likelihood -741.183 -739.723
Table 3: Model Summary Statistics

Instead, Table 3 provides general summary statistics about the fits of the original and expanded MNL models. This table’s main message is that there are multiple points of corroborating, classical evidence that the additional parameters in the expanded MNL are useful and not simply overfitting. With all the changes described two paragraphs ago, the number of estimated parameters increased from 21 coefficients in Brownstone and Train’s MNL model to 89 in the expanded model. Accordingly, the in-sample log-likelihood of the expanded model increased to -7,311.634—higher than that of the MNL, probit, or mixed logit models estimated in Brownstone and Train (1998). Even after penalizing the expanded model for its extra parameters, McFadden’s was higher in the expanded MNL (0.113) versus the original MNL model (0.111). Likewise, the Akaike Information Criterion (AIC) was lower in the expanded MNL (14,787) versus the original MNL (14,825). Finally, the out-of-sample log-likelihood131313These values were computed using the average out-of-sample log-likelihood from ten-fold cross-validation. was greater for the expanded MNL model (-739.7) as compared to the original model (-741.2).

Beyond classical measures of fit, the expanded MNL model was subjected to the same graphical model checking procedures as the original model. As expected, and as shown in Figures 9 - 13, many of the observed instances of underfitting from the original MNL model have been resolved. In particular, Figures 10, 12, and 13 show that the underfitting of the relative price variable has been greatly reduced. For instance, the observed KDE no longer lies outside the simulated KDEs at any point, and the observed CDF now lies in the middle of the simulated CDFs. Likewise, the marginal model plots have improved: in the expanded model, only two points lie outside the blue band of asymptotically expected probabilities, whereas 7 points lied outside this band in the Brownstone and Train’s MNL model. In terms of the operating cost, Figure 11 shows that the observed data is less extreme in the expanded MNL model as compared to the original: 86% of predicted samples in the expanded model are less than the observed value as compared to 96% with the original MNL model. Finally, by comparing Figure 9 to Figure 4, one can see that the expanded MNL model has better calibrated (but still unreliable) probability predictions. For methanol vehicles, the area between the dark blue curve and the dashed reference line is lower in Figure 9 with the expanded MNL model as compared to Figure 4 with Brownstone and Train’s MNL model. This finding is also confirmed by the reliability plots (not shown due to space considerations) for the other fuel types.

Figure 9: Binned reliability plot for methanol vehicles using the ‘expanded’ MNL model
Figure 10: Binned marginal model plot for sports utility vehicles using the ‘expanded’ MNL model
Figure 11: Simulated histograms for regular cars using the ‘expanded’ MNL model
Figure 12: Simulated kernel density estimates for electric cars using the ‘expanded’ MNL model
Figure 13: Simulated CDFs for sport utility vehicles using the ‘expanded’ MNL model

Finally, the changes inspired by this paper’s techniques do more than merely improve the fit of the original MNL model. The changes also lead to substantively different policy implications when compared to Brownstone and Train’s chosen mixed logit model (“Mixed logit B”). In particular, one main policy that was analyzed in Brownstone and Train (1998) was a 20% increase in the price of large gasoline cars. Brownstone and Train’s chosen model predicted that the market share of large gas cars would have a relative decrease of approximately 12% percent141414Forecast results for Brownstone and Train’s Mixed logit B were not numerically reported in their paper. The forecast results in this text were manually produced using custom written python code and Brownstone and Train’s estimated parameters.. Similarly, the expanded MNL model predicted an 8% relative decrease in the market share of large gas cars.

Mixed logit B (%) Expanded MNL (%)
Large gasoline truck
(8.16)
Sub-compact electric van
(3.03)
Large gasoline station wagon
(6.15)
Compact CNG station wagon
(3.02)
Large gasoline van
(5.72)
Large gasoline truck
(2.88)
Compact CNG station wagon
(4.14)
Large gasoline station wagon
(2.42)
Table 4: Top-Four Vehicle Categories Forecasted to Increase in Market Share

Despite similar predictions for the market share of large gas cars, the two model’s forecasts differ most regarding the vehicle classes that are predicted to grow appreciably in market share. These differences are shown in Table 4. On one hand, Brownstone and Train’s chosen model predicts an environmentally “hellish” scenario where people buy other, even larger gasoline powered vehicles. On the other hand, my expanded MNL model predicts an environmentally “heavenly” scenario where individuals buy smaller, electric and compressed natural gas vehicles.

Specifically, Brownstone and Train’s chosen model predicts that large gasoline station wagons, trucks, and vans would see relative increases in their market share of approximately 6%, 8%, and 6% respectively. Moreover, Brownstone and Train’s chosen model does not predict that any alternative fuel vehicle will have a relative market share increase greater than approximately 4%. In contrast, my expanded MNL model predicts that subcompact electric vans and compact, compressed natural gas station wagons will have the largest relative market share increases, at 3% each. After this, the expanded MNL model predicts relative market share increases of 2% for both large gasoline vehicles as well as many alternative fuel vehicles. These differing forecasts between Mixed logit B and the expanded MNL model imply conflicting qualitative lessons for policy makers that may have been considering policies such as environmentally motivated taxes on large gasoline cars.

5 Conclusion

In this paper, I introduced and demonstrated a procedure for checking one’s discrete choice models using predictive simulations. Conceptually, this method compares features of one’s observed data to the distribution of those features in the datasets formed using simulated outcomes. To operationalize these ideas, I introduced seven graphical model checks for underfitting. These seven plots are designed to be of use in routine model checking situations, ideally throughout and at the conclusion of any discrete choice analysis. In addition to these specific plots, I also propose a general and nearly-automatic procedure for designing custom checks for underfitting. To demonstrate the proposed procedures, I used a vehicle choice case study from Brownstone and Train (1998). In this case study, the seven plots and the proposed model checking procedure uncovered serious underfitting in Brownstone and Train’s MNL model—underfitting that was not addressed by their final mixed logit model. I then demonstrated how the insights derived from my proposed plots lead to an expanded MNL model with greater levels of fit and substantively different forecasts from Brownstone and Train’s final mixed logit model. In particular, this paper’s proposed model checks inspired a more detailed representation of the systematic heterogeneity in individual sensitivities to variables such as price, operating cost, etc. Once this systematic heterogeneity was better captured, the resulting model fit better than either Brownstone and Train’s MNL or mixed logit models. Moreover, the new model had aggregate forecasts that were much more hopeful about the ability of pricing policies to push individuals away from large gas cars to smaller, alternative fuel vehicles.

Beyond this specific example, the goal of this paper is to introduce a new set of model checking tools to the discrete choice community. The case study of Brownstone and Train (1998)

was used to show how models estimated by the best of us may still be underfitting our data. In my own work, I have used these techniques to spot errors even more egregious than those detailed in this paper. These errors have been found everywhere: from simple binary logit models to complex machine learning techniques. No model is safe. However, with the model checking methods introduced in this paper, we can better see how our models are misrepresenting reality, thereby allowing us to raise the quality of models used within our papers, our companies, and our governments.

Acknowledgements

Thanks to Albert Yuen, Joan Walker, Andrew Campbell, and two reviewers for providing helpful comments on an earlier draft of this manuscript. Any remaining errors are my own. Thanks also to Andrew Campbell for inspiring this research by asking the fundamental question: “so, how do you check your models after you’ve estimated them?”

Replication Materials

All replication data and code for this paper can be found online at https://github.com/timothyb0912/check-yourself.

Additionally, stand-alone software for producing the posterior predictive simulations and plots mentioned in this paper can be found in the “pylogit.viz” module of the PyLogit open-source package for the python programming language. This software will work regardless of the software that one initially used for model estimation. One simply needs to be able to store one’s data and simulated probability predictions in a software-agnostic format such as a csv file. The package is available through the Python Packaging Index (PyPI) as well as through the Anaconda distribution. See https://github.com/timothyb0912/pylogit for detailed installation instructions.

Appendix A Frequently Asked Questions

This paper introduced the concept of assessing underfitting in discrete choice models by visually comparing predictive simulations from one’s model to one’s observed data. This introduction has included concrete examples of such graphical model checks in Section 2.1, a general algorithm for use with labeled and unlabeled alternatives in Section 2.2, relations to the broader literature on model checking in statistics and discrete choice in Section 3, and examples of how these model checks could be used to guide one’s model improvement efforts in Section 4.

However, despite all the topics discussed in this introduction, some related issues have still been excluded in order to minimize the article’s length. This appendix on Frequently Asked Questions attempts to briefly address the most common questions that reviewers, friends, and colleagues have had after reading earlier drafts of this paper.

  1. How do predictive simulations help us make decisions about how to improve our models?

    Essentially, predictive simulations are like thermometers. Thermometers are useful for assessing a patient’s body temperature, and through repeated use, one can assess how a patient’s temperature is changing alongside a caretaker’s actions. Likewise, predictive simulations are most useful for assessing the discord between one’s model and reality, and through repeated use, one can assess how this discord has been altered by changes to one’s model. Taking the analogy further, note that thermometers do not tell doctors the cause of a fever or how a fever can be reduced. Similarly, predictive simulations do not tell researchers the cause of the problems they uncover, nor do they suggest particular remedies to those problems. Once aware of the extent to which one’s current model misrepresents reality, it is the analyst’s responsibility to diagnose their model and brainstorm ways to increase its realism. Afterwards, predictive simulations will be there to show the analyst how well their changes worked.

    Figure 14: Example histogram where the predictive p-value is misleading
  2. Why should we use graphical model checks instead of simply looking at point measures or purely numeric summaries?

    In general, solely looking at scalar values (such as p-values) leads to a loss of information. Far from being harmless, this information loss can potentially lead to incorrect assessments of model fit. Imagine one is looking at a simulated histogram such as the one shown in Figure 14. Here, one’s predictive distribution is bimodal and is roughly symmetric about the observed value. As a result, one’s p-value may be close to 0.5, even though few simulated datasets in this example have test statistics that are similar to the observed test statistic (the dashed vertical line in the plot). To facilitate correct assessments of model fit and avoid being unnecessarily misled, I recommend using graphical model checks in addition to purely numeric summaries.

    Beyond the insufficiency of predictive p-values for model checking, creating other scalar surrogates for the graphical check is non-trivial. For instance, even if one’s predictive distribution is unimodal, it is not valid to think that the value of a test statistic, evaluated using the probabilities generated at a point estimate of one’s parameters, will serve as a useful surrogate for the mean of one’s predictive distribution of that test statistic. This is because with respect to a random variable and arbitrary non-linear function . In this paper’s model checking context, this fact implies that even though is the mean of the asymptotic sampling distribution, and therefore the mean of the simulated parameters , . The probabilities at our point estimate of the parameters are not the mean of the simulated probabilities. Given this inequality, even simple graphical checks such as simulated histograms cannot be replaced by scalar comparisons based on point estimates. More complicated graphical checks such as simulated KDEs use an additional layer of non-linear functions in their calculation of the test statistic. These additional non-linearities rule out the use of a point estimate of the model’s parameters for purposes of trivially computing the mean of the predictive distributions of one’s test statistics.

  3. What are some ways that predictive simulations should NOT be used?

    Predictive simulations should not be used as hypothesis tests that are used to accept or reject models. A-priori, we know all models are wrong. Predictive simulations should therefore be used to understand the ways and extent to which one’s model differs from reality. If desired, this understanding can then be used increase the realism of one’s model or to inform the way that the model is used. In some cases, this informing may mean completely rejecting the use of a particular model. However, in other cases, one may decide to only make inferences from the model when the model performs well on checks related to the inference in question. The decision is up to the “stakeholders” of the model (e.g. peer reviewers, company or organization managers, etc.). For example, if a travel mode choice model performs well on underfitting checks related to transit travel cost but not travel time, then perhaps one may use the model to forecast the impact of fare changes but not for forecasting the impact of adding an express service that would reduce individual travel times.

  4. How do we interpret the proposed graphical model checks?

    In general, the proposed graphical model checks are ways to visually assess the similarity between data generated by one’s model and one’s observed data. As mentioned in the question “How do predictive simulations help us make decisions about how to improve our models?,” the proposed checks are to be used like a thermometer. One checks one’s temperature, and any suspiciously high temperature measurements must then be investigated for their root cause and effective remedies. Similarly, any discrepancies uncovered by the proposed model checks have to be investigated for their root cause and for appropriate model improvements. Such investigations may not be easy or straightforward. The visually diagnosed problems may be related to one’s overall model specification (e.g. using an MNL model instead of a nested logit model), to omitted variables, or even to misspecification of the relationship between one or more correlated variables and the outcome variable. As usual, an analyst must still make use of their domain knowledge, intuition, and thoroughness to resolve any discovered problems. Similar to the question of interpreting graphical model checks is the question of how one interprets their associated predictive p-values. Mathematically, the predictive p-value is the proportion of simulated datasets for which a given scalar statistic was lower than the observed value of that scalar statistic. These predictive p-values can be helpful (though insufficient, in my opinion) for evaluating the extent to which one’s observed data “looks like” one’s simulated data. In particular, they may be especially helpful for assessing small model changes that are hard to distinguish visually in a graphical model check.

  5. When should one be concerned by the results of a particular graphical model check?

    I do not think that there is a universal answer to this question. It is equivalent to asking “how good is good enough?” One’s answer to these questions will almost assuredly be driven by problem specific considerations such as the purpose for which the model is being built and the specific statistic being checked. However, there are some general guidelines that one can use. Overall, if one’s model is close to the data generating process, then one’s observed data should look like a random draw of the data simulated from one’s model. Accordingly, one should expect the observed statistics to lie near the center of the distribution of simulated statistics. For scalar statistics, I would be concerned if the distribution of simulated statistics was not clustered around a line for the observed statistic. For vector statistics, I would be concerned if the observed vector, at all points, was not near the center of the simulated vectors. Additionally I would be concerned if the observed vector did not follow the same visual trends as the simulated vectors. (After all, the observed vector should “look like” a random draw of the simulated vectors). How far away from this ideal a model is allowed to exist is up to the “stakeholders” of the model to decide and to defend.

  6. Besides the examples shown in this paper, what are other specific ways that one can use predictive simulations to check our models?

    What follows is by no means an exhaustive list, but predictive simulations can also be used in the following four ways to check one’s models. First, as demonstrated in Pardoe and Cook (2002), one can use predictive simulations to determine whether variables should be added or removed from one’s model. Secondly, one can use the proposed graphical checks to visualize and understand differences between competing model extensions. For instance, one can check the data generated from using a common specification of one’s systematic utilities but different distributions of the random utility (i.e. different link functions), different mixing distributions in a mixed logit model, different class membership models in a latent class model, etc. Third, a closely related use of predictive simulations is to check one’s mixed logit models by comparing one’s inferences for the distribution individual tastes with one’s assumed mixing distribution of tastes (see Allenby and Rossi (1998) and Gilbride and Lenk (2010) for examples). Finally, as demonstrated and explained in Kucukelbir et al. (2017), one could also use predictive simulations to discover groups of observations that one’s model fits particularly poorly.

Appendix B Expanded MNL Estimation Results

Variable Estimate Std. err z P > |z|
Price over log(income) <= 3 (sports_utility_vehicle) 0.2328 0.257 0.905 0.366
Price over log(income) <= 3 (sports_car) -0.1069 0.253 -0.423 0.673
Price over log(income) <= 3 (station_wagon) -0.3201 0.109 -2.937 0.003
Price over log(income) <= 3 (truck) -0.2541 0.089 -2.863 0.004
Price over log(income) <= 3 (van) -0.2418 0.088 -2.741 0.006
Price over log(income) <= 3 (electric) -0.1319 0.100 -1.324 0.186
Price over log(income) <= 3 (compressed_natural_gas) -0.0938 0.092 -1.024 0.306
Price over log(income) <= 3 (methanol) 0.0629 0.089 0.707 0.480
Price over log(income) <= 3 -0.2519 0.113 -2.225 0.026
Price over log(income) 3 (sports_utility_vehicle) 0.2714 0.077 3.508 0.000
Price over log(income) 3 (sports_car) 0.2039 0.080 2.534 0.011
Price over log(income) 3 (station_wagon) -0.0341 0.047 -0.722 0.470
Price over log(income) 3 (truck) -0.0087 0.035 -0.249 0.804
Price over log(income) 3 (van) -0.0400 0.036 -1.110 0.267
Price over log(income) 3 (electric) -0.0941 0.035 -2.655 0.008
Price over log(income) 3 (compressed_natural_gas) -0.0534 0.033 -1.616 0.106
Price over log(income) 3 (methanol) -0.0699 0.031 -2.253 0.024
Price over log(income) 3 -0.1326 0.037 -3.553 0.000
Range (units: 100mi) (sports_utility_vehicle) -0.0964 0.100 -0.965 0.334
Range (units: 100mi) (sports_car) -0.0721 0.117 -0.616 0.538
Range (units: 100mi) (station_wagon) -0.0725 0.086 -0.848 0.397
Range (units: 100mi) (truck) 0.0125 0.070 0.178 0.859
Range (units: 100mi) (van) 0.0628 0.067 0.939 0.348
Range (units: 100mi) (electric) 0.1430 0.130 1.098 0.272
Range (units: 100mi) (compressed_natural_gas) 0.2916 0.132 2.203 0.028
Range (units: 100mi) (methanol) 0.3219 0.135 2.393 0.017
Range (units: 100mi) 0.1028 0.115 0.893 0.372
Acceleration (units: 0.1sec) (sports_utility_vehicle) -0.1927 0.497 -0.388 0.698
Acceleration (units: 0.1sec) (sports_car) -1.0425 0.602 -1.732 0.083
Acceleration (units: 0.1sec) (station_wagon) -0.3109 0.446 -0.698 0.485
Acceleration (units: 0.1sec) (truck) -0.1371 0.360 -0.381 0.704
Acceleration (units: 0.1sec) (van) -0.6375 0.323 -1.977 0.048
Acceleration (units: 0.1sec) -0.5157 0.146 -3.538 0.000
Top speed (units: 0.01mph) (sports_utility_vehicle) -0.0701 0.331 -0.212 0.832
Top speed (units: 0.01mph) (sports_car) 0.2942 0.399 0.737 0.461
Top speed (units: 0.01mph) (station_wagon) 0.1550 0.282 0.550 0.582
Top speed (units: 0.01mph) (truck) 0.1298 0.224 0.579 0.562
Top speed (units: 0.01mph) (van) -0.0136 0.218 -0.063 0.950
Top speed (units: 0.01mph) (electric) 0.5526 0.325 1.698 0.089
Table 5: Expanded MNL model
Variable Estimate Std. err z P > |z|
Top speed (units: 0.01mph) (compressed_natural_gas) 0.0994 0.286 0.348 0.728
Top speed (units: 0.01mph) (methanol) -0.1352 0.269 -0.503 0.615
Top speed (units: 0.01mph) 0.1290 0.194 0.665 0.506
Pollution (sports_utility_vehicle) 0.0771 0.332 0.232 0.816
Pollution (sports_car) 0.2570 0.376 0.683 0.495
Pollution (station_wagon) -0.5810 0.280 -2.077 0.038
Pollution (truck) 0.0841 0.214 0.394 0.694
Pollution (van) -0.1365 0.216 -0.633 0.527
Pollution (compressed_natural_gas) 0.0576 0.314 0.184 0.854
Pollution (methanol) 0.0379 0.266 0.143 0.887
Pollution -0.4618 0.164 -2.813 0.005
Size (sports_utility_vehicle) 2.4675 0.943 2.617 0.009
Size (sports_car) -2.5383 0.988 -2.569 0.010
Size (station_wagon) 0.9575 0.851 1.126 0.260
Size (truck) -0.2520 0.719 -0.351 0.726
Size (van) 2.7231 0.825 3.299 0.001
Size (electric) -0.2671 0.769 -0.347 0.728
Size (compressed_natural_gas) -0.4512 0.759 -0.595 0.552
Size (methanol) 0.4433 0.721 0.615 0.539
Size 0.7798 0.616 1.266 0.205
Big enough 0.1208 0.078 1.543 0.123
Luggage space 0.4516 0.194 2.324 0.020
Operation cost (sports_utility_vehicle) 0.5759 0.313 1.842 0.065
Operation cost (sports_car) 0.5041 0.377 1.338 0.181
Operation cost (station_wagon) -0.5383 0.280 -1.926 0.054
Operation cost (truck) 0.3537 0.212 1.672 0.095
Operation cost (van) 0.0033 0.205 0.016 0.987
Operation cost (electric) 0.2615 0.231 1.133 0.257
Operation cost (compressed_natural_gas) -0.1772 0.273 -0.648 0.517
Operation cost (methanol) -0.3987 0.242 -1.650 0.099
Operation cost -0.7469 0.177 -4.217 0.000
Station availability 0.3862 0.099 3.920 0.000
Sports utility vehicle -0.7066 0.838 -0.843 0.399
Sports car 0.8799 0.852 1.032 0.302
Station wagon -0.1506 0.477 -0.316 0.752
Truck -0.5589 0.408 -1.370 0.171
Van -0.6099 0.426 -1.433 0.152
EV -0.8867 0.596 -1.489 0.137
Commute < 5 & EV 0.2088 0.084 2.490 0.013
College & EV 0.4586 0.110 4.175 0.000
CNG -0.1985 0.605 -0.328 0.743
Methanol -0.5804 0.599 -0.969 0.332
College & Methanol 0.2075 0.090 2.304 0.021
Table 6: Expanded MNL model (cont’d)

References

References

  • Abe (1999) Abe, M., 1999. A Generalized Additive Model for Discrete-Choice Data. Journal of Business & Economic Statistics 17, 271–284.
  • Albert and Siddhartha (1995) Albert, J., Siddhartha, C., 1995. Bayesian residual analysis for binary response regression models. Biometrika 82, 747–759.
  • Allenby and Rossi (1998) Allenby, G.M., Rossi, P.E., 1998. Marketing models of consumer heterogeneity. Journal of econometrics 89, 57–78.
  • Anscombe and Tukey (1963) Anscombe, F.J., Tukey, J.W., 1963. The examination and analysis of residuals. Technometrics 5, 141–160.
  • Ben-Akiva and Lerman (1985) Ben-Akiva, M.E., Lerman, S.R., 1985. Discrete Choice Analysis: Theory and Application to Travel Demand. MIT Press.
  • Bergtold et al. (2009) Bergtold, J.S., Spanos, A., Onukwugha, E., 2009. Bernoulli Regression Models: Revisiting the Specification of Statistical Models with Binary Dependent Variables. Journal of Choice Modelling 3, 1–28.
  • Box (1976) Box, G.E., 1976. Science and Statistics. Journal of the American Statistical Association 71, 791–799.
  • Bröcker and Smith (2007) Bröcker, J., Smith, L., 2007. Increasing the Reliability of Reliability Diagrams. Weather and Forecasting 22, 651–661.
  • Brownstone et al. (1996) Brownstone, D., Bunch, D.S., Golob, T.F., Weiping, R., 1996. Transactions choice model for forecasting demand for alternative-fuel vehicles. Research in Transportation Economics 4, 87–129.
  • Brownstone and Train (1998) Brownstone, D., Train, K., 1998. Forecasting new product penetration with flexible substitution patterns. Journal of Econometrics 89, 109–129.
  • Brownstone and Valletta (2001) Brownstone, D., Valletta, R., 2001.

    The bootstrap and multiple imputations: harnessing increased computing power for improved statistical tests.

    The Journal of Economic Perspectives 15, 129–141.
  • Buja et al. (2009) Buja, A., Cook, D., Hofmann, H., Lawrence, M., Lee, E.K., Swayne, D.F., Wickham, H., 2009. Statistical Inference for Exploratory Data Analysis and Model Diagnostics. Philosophical Transactions: Mathematical, Physical and Engineering Sciences 367, 4361–4383.
  • Carlin et al. (2001) Carlin, J.B., Wolfe, R., Brown, C.H., Gelman, A., 2001. A case study on the choice, interpretation and checking of multilevel models for longitudinal binary outcomes. Biostatistics 2, 397–416.
  • Cleveland (1979) Cleveland, W.S., 1979. Robust Locally Weighted Regression and Smoothing Scatterplots. Journal of the American Statistical Association 74, 829–836.
  • Cox and Snell (1968) Cox, D.R., Snell, E.J., 1968. A general definition of residuals. Journal of the Royal Statistical Society. Series B (Methodological) , 248–275.
  • Dagsvik (2017) Dagsvik, J.K., 2017. Invariance axioms and functional form restrictions in structural models. Mathematical Social Sciences .
  • Donoso et al. (2011) Donoso, P., De Grange, L., González, F., 2011. A maximum entropy estimator for the aggregate hierarchical logit model. Entropy 13, 1425–1445.
  • Dunn et al. (1987) Dunn, R., Longley, P., Wrigley, N., 1987. Graphical Procedures For Identifying Functional Form In Binary Discrete Choice Models. Regional Science and Urban Economics 17, 151–171.
  • Efron and Tibshirani (1997) Efron, B., Tibshirani, R., 1997. Improvements on Cross-Validation: The .632+ Bootstrap Method. Journal of the American Statistical Association 92, 548–560. URL: http://www.jstor.org/stable/2965703, doi:10.2307/2965703.
  • Esarey and Pierce (2012) Esarey, J., Pierce, A., 2012. Assessing Fit Quality and Testing for Misspecification in Binary-Dependent Variable Models. Political Analysis 20, 480–500.
  • Evans and Moshonov (2006) Evans, M., Moshonov, H., 2006. Checking for Prior-Data Conflict. Bayesian Analysis 1, 893–914.
  • Fariss and Jones (2017) Fariss, C.J., Jones, Z.M., 2017. Enhancing Validity in Observational Settings when Replication is Not Possible. Political Science Research and Methods , 1–16.
  • Gelman (2004) Gelman, A., 2004. Exploratory Data Analysis for Complex Models (with discussion). Journal of Computational and Graphical Statistics 13, 755–787.
  • Gelman et al. (2014) Gelman, A., Carlin, J.B., Stern, H.S., Dunson, D.B., Vehtari, A., Rubin, D.B., 2014. Bayesian Data Analysis. 3 ed., Taylor & Francis Group, Boca Raton, FL, USA.
  • Gelman et al. (2000) Gelman, A., Goegebeur, Y., Tuerlinckx, F., Van Mechelen, I., 2000. Diagnostic checks for discrete data regression models using posterior predictive simulations. Applied Statistics 49, 247–268.
  • Gelman and Meng (1996) Gelman, A., Meng, X.L., 1996.

    Model checking and model improvement, in: Markov chain Monte Carlo in practice. Springer, pp. 189–201.

  • Gelman et al. (1996) Gelman, A., Meng, X.L., Stern, H., 1996. Posterior Predictive Assessment of Model Fitness Via Realized Discrepancies (with discussion). Statistica Sinica 6, 733–807.
  • Ghosh (2018) Ghosh, S., 2018. Kernel Smoothing: Principles, Methods and Applications. John Wiley & Sons.
  • Gilbride and Lenk (2010) Gilbride, T.J., Lenk, P.J., 2010. Posterior predictive model checking: An application to multivariate normal heterogeneity. Journal of Marketing Research 47, 896–909.
  • Greenhill et al. (2011) Greenhill, B., Ward, M.D., Sacks, A., 2011. The Separation Plot: A New Visual Method for Evaluating the Fit of Binary Models. American Journal of Political Science 55, 990–1002.
  • Hastie and Tibshirani (1986) Hastie, T., Tibshirani, R., 1986. Generalized Additive Models (with discussion). Statistical Science 1, 297–318.
  • Hausman and McFadden (1984) Hausman, J., McFadden, D., 1984. Specification Tests for the Multinomial Logit Model. Econometrica 52, 1219–1240.
  • Horowitz (1985) Horowitz, J.L., 1985. Testing probabilistic discrete choice models of travel demand by comparing predicted and observed aggregate choice shares. Transportation Research Part B 19, 17–38.
  • Horowitz and Louviere (1993) Horowitz, J.L., Louviere, J.J., 1993. Testing Predicted Choices Against Observations in Probabilistic Discrete-Choice Models. Marketing Science 12, 270–279.
  • Jedidi et al. (2003) Jedidi, K., Jagpal, S., Manchanda, P., 2003. Measuring heterogeneous reservation prices for product bundles. Marketing Science 22, 107–130.
  • Kay and Little (1987) Kay, R., Little, S., 1987. Transformations of the Explanatory Variables in the Logistic Regression Model for Binary Data. Biometrika 74, 495–501.
  • Klaiber and von Haefen (2018) Klaiber, H.A., von Haefen, R.H., 2018. Do Random Coefficients and Alternative Specific Constants Improve Policy Analysis? An Empirical Investigation of Model Fit and Prediction. Environmental and Resource Economics , 1–17.
  • Kucukelbir et al. (2017) Kucukelbir, A., Wang, Y., Blei, D.M., 2017. Evaluating Bayesian models with posterior dispersion indices, in: Precup, D., Teh, Y.W. (Eds.), Proceedings of the 34th International Conference on Machine Learning, PMLR, International Convention Centre, Sydney, Australia. pp. 1925–1934. URL: http://proceedings.mlr.press/v70/kucukelbir17a.html.
  • Landwehr et al. (1984) Landwehr, J.M., Pregibon, D., Shoemaker, A.C., 1984. Graphical methods for assessing logistic regression models (with discussion). Journal of the American Statistical Association 79, 61–80.
  • Lemeshow and Hosmer Jr (1982) Lemeshow, S., Hosmer Jr, D.W., 1982. A review of goodness of fit statistics for use in the development of logistic regression models. American journal of epidemiology 115, 92–106.
  • McFadden and Train (2000) McFadden, D., Train, K., 2000. Mixed MNL Models for Discrete Response. Journal of Applied Econometrics 15, 447–470.
  • Murphy and Winkler (1977) Murphy, A.H., Winkler, R.L., 1977. Reliability of Subjective Probability Forecasts of Precipitation and Temperature. Journal of the Royal Statistical Society: Series C (Applied Statistics) 26, 41–47.
  • Murphy and Winkler (1987) Murphy, A.H., Winkler, R.L., 1987. A General Framework for Forecast Verification. Monthly Weather Review 115, 1330–1338.
  • Musalem et al. (2008) Musalem, A., Bradlow, E.T., Raju, J.S., 2008. Who’s got the coupon? estimating consumer preferences and coupon usage from aggregate information. Journal of Marketing Research 45, 715–730.
  • Nagel and Hatzinger (1992) Nagel, H., Hatzinger, R., 1992. Diagnostics in some discrete choice models. Computational Statistics & Data Analysis 13, 407–414.
  • Pardoe and Cook (2002) Pardoe, I., Cook, R.D., 2002. A graphical method for assessing the fit of a logistic regression model. The American Statistician 56, 263–272.
  • Racine (2008) Racine, J.S., 2008. Nonparametric Econometrics: A Primer. now Publishers.
  • Scott (1992) Scott, D., 1992. Multivariate Density Estimation: Theory, Practice, and Visualization. John Wiley & Sons.
  • Silverman (1984) Silverman, B.W., 1984. Spline Smoothing: The Equivalent Variable Kernel Method. The Annals of Statistics 12, 898–916.
  • Steckel and Vanhonacker (1993) Steckel, J.H., Vanhonacker, W.R., 1993. Cross-Validating Regression Models in Marketing Research. Marketing Science 12, 415–427.
  • Stone (1974) Stone, M., 1974. Cross-Validatory Choice and Assessment of Statistical Predictions (with discussion). Journal of the Royal Statistical Society. Series B (Methodological) 36, 111–147.
  • Train (2009) Train, K., 2009. Discrete Choice Methods With Simulation. 2 ed., Cambridge University Press, New York, NY, USA.
  • Tse (1987) Tse, Y.K., 1987. A Diagnostic Test for the Multinomial Logit Model. Journal of Business & Economic Statistics 5, 283–286.
  • Wasserman (2006) Wasserman, L., 2006. All of Nonparametric Statistics. Springer.
  • Zheng (2008) Zheng, X., 2008. Testing for discrete choice models. Economics Letters 98, 176–184.