Multinomial Logit is a form of regression analysis that models a discrete and nominal dependent variable with more than two outcomes (Yes/No/Maybe, Red/Green/Blue, Brand A/Brand B/Brand C, etc.). It is also known as a multinomial logistic regression and multinomial logistic discriminant analysis.
Note: this model is not the same as the conditional logit model commonly used in choice modeling; see Choice Modeling  Hierarchical Bayes and Choice Modeling  Latent Class Analysis.
Data format
The key requirement for a multinomial logit regression is that the dependent variable is discrete with more than two outcomes. In Displayr, the best data format for this type is “Nominal: Mutually exclusive categories”.
The independent variables can be continuous, categorical, or binary — just as with any regression model. However, it is recommended that the predictor variables be scaled to lie between [0,1] to improve the speed and convergence properties of the algorithm.
Interpretation
Variable statistics measure the impact and significance of individual variables within a model, while overall statistics apply to the model as a whole. Both are shown in the multinomial logit output.
Variable statistics
The format of the default summary output only shows the coefficient values. To see a more detailed output with standard errors, tstatistics, and pvalues, select Multiple imputation in Inputs > Missing data.
Estimate the magnitude of the coefficient indicates the size of the change in the independent variable as the value of the dependent variable changes. A positive number indicates a direct relationship (y increases as x increases), and a negative number indicates an inverse relationship (y decreases as x increases).
The coefficient is colored and bolded if the variable is statistically significant at the 5% level.
Standard Error measures the accuracy of an estimate. The smaller the standard error, the more accurate the predictions.
tstatistic the estimate divided by the standard error. The magnitude (either positive or negative) indicates the significance of the variable. The values are highlighted based on their magnitude.
pvalue expresses the tstatistic as a probability. A pvalue under 0.05 means that the variable is statistically significant at the 5% level; a pvalue under 0.01 means that the variable is statistically significant at the 1% level. Pvalues under 0.05 are shown in bold.
Overall statistics
n the sample size of the model
Rsquared & McFadden’s rhosquared assess the goodness of fit of the model. A larger number indicates that the model captures more of the variation in the dependent variable.
AIC Akaike information criterion is a measure of the quality of the model. When comparing similar models, the AIC can be used to identify the superior model.
See also Regression Diagnostics.
Example
The example below is a model that predicts a survey respondent’s preferred fastfood chain based on characteristics like age, gender, and work status.
Object Inspector Options
Outcome The variable to be predicted by the predictor variables.
Predictors The variable(s) to predict the outcome.
Algorithm The fitting algorithm. Defaults to Regression but may be changed to other machine learning methods.
Type: You can use this option to toggle between different types of regression models, but note that certain types are not appropriate for certain types of outcome variables.
 Linear Appropriate for a continuous outcome variable. See Linear Regression.
 Binary Logit Appropriate if the outcome is binary (i.e. falls in one of two categories). See Binary Logit.
 Ordered Logit Appropriate for a discrete outcome where the categories have a natural order (e.g. Low, Medium, High). See Ordered Logit Regression.
 Multinomial Logit Appropriate for a discrete outcome with unordered categories. See Multinomial Logit Regression.
 Poisson Appropriate for count outcomes (i.e. outcomes that take only positive integer values). See Poisson Regression.
 QuasiPoisson Appropriate for count outcomes. See Quasi Poisson Regression.
 NBD Appropriate for count outcomes. See NBD Regression.
Robust standard errors Computes standard errors that are robust to violations of the assumption of constant variance (i.e., heteroscedasticity). See Robust Standard Errors. This is only available when Type is Linear.
Missing data See Missing Data Options.
Output
 Summary The default; as shown in the example above.
 Detail Typical R output, some additional information compared to Summary, but without the pretty formatting.
 ANOVA Analysis of variance table containing the results of Chisquared likelihood ratio tests for each predictor.
 Relative Importance Analysis The results of a relative importance analysis (also known as Johnson's relative weights). See here and the references for more information. This option is not available for Multinomial Logit. Note that categorical predictors are not converted to be numeric, unlike in Driver (Importance) Analysis  Relative Importance Analysis.
 Shapley Regression See here and the references for more information. This option is only available for Linear Regression. Note that categorical predictors are not converted to be numeric, unlike in Driver (Importance) Analysis  Shapley.
 Jaccard Coefficient Computes the relative importance of the predictor variables against the outcome variable with the Jaccard Coefficients. See Driver (Importance_ Analysis  Jaccard Coefficient. This option requires both binary variables for the outcome variable and the predictor variables.
 Correlation Computes the relative importance of the predictor variables against the outcome variable via the bivariate Pearson product moment correlations. See Driver (Importance) Analysis  Correlation and references therein for more information.
 Effects Plot Plots the relationship between each of the Predictors and the Outcome. Not available for Multinomial Logit.
Correction The multiple comparisons correction applied when computing the pvalues of the posthoc comparisons.
Variable names Displays Variable Names in the output instead of labels.
Absolute importance scores Whether the absolute value of Relative Importance Analysis scores should be displayed.
Auxiliary variables Variables to be used when imputing missing values (in addition to all the other variables in the model).
Weight. Where a weight has been set for the R Output, it will automatically applied when the model is estimated. By default, the weight is assumed to be a sampling weight, and the standard errors are estimated using Taylor series linearization (by contrast, in the Legacy Regression, weight calibration is used). See Weights, Effective Sample Size and Design Effects.
Filter The data is automatically filtered using any filters prior to estimating the model.
Crosstab Interaction Optional variable to test for interaction with other variables in the model. The interaction variable is treated as a categorical variable. Coefficients in the table are computed by creating separate regressions for each level of the interaction variable. To evaluate whether a coefficient is significantly higher (blue) or lower (red), we perform a ttest of the coefficient compared to the coefficient using the remaining data as described in Driver Analysis. Pvalues are corrected for multiple comparisons across the whole table (excluding the NET column). The Pvalue in the subtitle is calculated using a likelihood ratio test between the pooled model with no interaction variable, and a model where all predictors interact with the interaction variable.
Automated outlier removal percentage A numeric value between 0 and 50 (including 0 but not 50) is used to specify the percentage of the data that is removed from analysis due to outliers. All regression types except for the case of Multinomial Logit support this feature. If a zero value is selected for this input control then no outlier removal is performed and a standard regression output for the entire (possibly filtered) dataset is applied. If a nonzero value is selected for this option then the regression model is fitted twice. The first regression model uses the entire dataset (after filters have been applied) and identifies the observations that generate the largest residuals. The userspecified percent of cases in the data that have the largest residuals are then removed. The regression model is refitted on this reduced dataset and output is returned. The specific residual used varies depending on the regression Type.
 Linear: The studentized residual in an unweighted regression and the Pearson residual in a weighted regression. The Pearson residual in the weighted case adjusts appropriately for the provided survey weights.
 Binary Logit and Ordered Logit: A type of surrogate residual from the sure R package (see Greenwell, McCarthy, Boehmke and Liu (2018)^{[1]} for more details). In Binary Logit it uses the resids function with the jitter parametrization. In Ordered Logit it uses the resids function with the latent parametrization to exploit the ordered logit structure.
 NBD Regression, Poisson Regression: A studentized deviance residual in an unweighted regression and the Pearson residual in a weighted regression.
 QuasiPoisson Regression: A type of quasideviance residual via the rstudent function in an unweighted regression and the Pearson residual in a weighted regression.
The studentized residual computes the distance between the observed and fitted value for each point and standardizes (adjusts) based on the influence and an externally adjusted variance calculation. The studentized deviance residual computes the contribution the fitted point has to the likelihood and standardizes (adjusts) based on the influence of the point and an externally adjusted variance calculation (see rstudent function in R and Davison and Snell (1991)^{[2]} for more details). The Pearson residual in the weighted case computes the distance between the observed and fitted value and adjusts appropriately for the provided survey weights. See rstudent function in R and Davison and Snell (1991) for more details of the specifics of the calculations.
Note that this feature is not supported when using the Multiple imputation option for handling Missing data.
Stack data Whether the input data should be stacked before analysis. Stacking can be desirable when each individual in the data set has multiple cases and an aggregate model is desired. More information is available at Stacking Data Files. If this option is chosen then the Outcome needs to be a single Question that has a Multi type structure suitable for regression such as a Pick One  Multi, Pick Any or Number  Multi. Similarly, the Predictor(s) need to be a single Question that has a Grid type structure such as a Pick Any  Grid or a Number  Grid. In the process of stacking, the data reduction is inspected. Any constructed NETs are removed unless comprised of source values that are mutually exclusive to other codes, such as the result of merging two categories.
Random seed Seed used to initialize the (pseudo)random number generator for the model fitting algorithm. Different seeds may lead to slightly different answers, but should normally not make a large difference.
Increase allowed output size Check this box if you encounter a warning message "The R output had size XXX MB, exceeding the 128 MB limit..." and you need to reference the output elsewhere in your document; e.g., to save predicted values to a Data Set or examine diagnostics.
Maximum allowed size for output (MB). This control only appears if Increase allowed output size is checked. Use it to set the maximum allowed size for the regression output in Megabytes. The warning referred to above about the R output size will state the minimum size you need to increase to return the full output. Note that having very many large outputs in one document or page may slow down the performance of your document and increase load times.
Additional options are available by editing the code.
DIAGNOSTICS
Plot  Cook's Distance Creates a line/rug plot showing Cook's Distance for each observation.
Plot  Cook's Distance vs Leverage Creates a scatterplot showing Cook's distance vs leverage for each observation.
Plot  Influence Index Creates index plots of studentized residuals, hat values, and Cook's distance.
Multicollinearity Table (VIF) Creates a table containing variance inflation factors (VIF) to diagnose multicollinearity.
Plot  Normal QQ Creates a normal QuantileQuantile (QQ) plot to reveal departures of the residuals from normality.
PredictionAccuracy Table Creates a table showing the observed and predicted values, as a heatmap.
Test Residual Heteroscedasticity Conducts a heteroscedasticity test on the residuals.
Test Residual Normality (ShapiroWilk) Conducts a ShapiroWilk test of normality on the (deviance) residuals.
Plot  Residuals vs Fitted Creates a scatterplot of residuals versus fitted values.
Plot  Residuals vs Leverage Creates a plot of residuals versus leverage values.
Plot  ScaleLocation Creates a plot of the square root of the absolute standardized residuals by fitted values.
Test Residual Serial Correlation (DurbinWatson) Conducts a DurbinWatson test of serial correlation (autocorrelation) on the residuals.
SAVE VARIABLE(S)
Fitted Values Creates a new variable containing fitted values for each case in the data.
Predicted Values Creates a new variable containing predicted values for each case in the data.
Residuals Creates a new variable containing residual values for each case in the data.
Additional Properties
When using this feature you can obtain additional information that is stored by the R code which produces the output.
 To do so, select Create > R Output.
 In the R CODE, paste: item = YourReferenceName
 Replace YourReferenceName with the reference name of your item. Find this in the Report tree or by selecting the item and then going to Properties > General > Name from the object inspector on the right.
 Below the first line of code, you can paste in snippets from below or type in str(item) to see a list of available information.
For a more indepth discussion on extracting information from objects in R, check out our blog post here.
Properties which may be of interest are:
 Summary outputs from the regression model:

 item$summary$coefficients # summary regression outputs
Acknowledgments
Uses the multinom function from the nnet^{[3]} R package. See also Regression  Generalized Linear Model.
References
 Greenwell, B., M., McCarthy, A., J., Boehmke, B., C., & Liu, D. (2018). Residuals and Diagnostics for Binary and Ordinal Regression Models: An Introduction to the sure Package. The R Journal, 10(1), 381. https://doi.org/10.32614/rj2018004
 Davison, A. C. and Snell, E. J. (1991) Residuals and diagnostics. In: Statistical Theory and Modelling. In Honour of Sir David Cox, FRS, eds. Hinkley, D. V., Reid, N. and Snell, E. J., Chapman & Hall.
 Venables WN, Ripley BD (2002). Modern Applied Statistics with S https://cran.rproject.org/web/packages/nnet/index.html