How to Develop a MaxDiff Typing Tool to Assign New Cases into Meaningful Segments

Similar documents
Statistical Innovations Online course: Latent Class Discrete Choice Modeling with Scale Factors

Using Mixture Latent Markov Models for Analyzing Change with Longitudinal Data

A Procedure for Classifying New Respondents into Existing Segments Using Maximum Difference Scaling

Tutorial #6: Analyzing Partial Rank Data

Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus

Weight of Evidence Module

LATENT GOLD CHOICE 4.0

Binary Logistic Regression

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.

Facebook Friend Suggestion Eytan Daniyalzade and Tim Lipus

ASSIGNMENT 4 PREDICTIVE MODELING AND GAINS CHARTS

This chapter will demonstrate how to perform multiple linear regression with IBM SPSS

Data Mining Part 5. Prediction

STATISTICA Formula Guide: Logistic Regression. Table of Contents

Stepwise Regression. Chapter 311. Introduction. Variable Selection Procedures. Forward (Step-Up) Selection

New Work Item for ISO Predictive Analytics (Initial Notes and Thoughts) Introduction

CALCULATIONS & STATISTICS

Tutorial #6A: Comparing Segments obtained from LC Cluster and DFactor Models in a Consumer Preference Study DemoData = crackers0.

Session 7 Bivariate Data and Analysis

A Study to Predict No Show Probability for a Scheduled Appointment at Free Health Clinic

Chapter Seven. Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS

Model selection in R featuring the lasso. Chris Franck LISA Short Course March 26, 2013

UNDERSTANDING THE TWO-WAY ANOVA

Predictive Dynamix Inc

Data Mining - Evaluation of Classifiers

Sawtooth Software. Which Conjoint Method Should I Use? RESEARCH PAPER SERIES. Bryan K. Orme Sawtooth Software, Inc.

Statistics Graduate Courses

Center for Advanced Studies in Measurement and Assessment. CASMA Research Report

SPSS Explore procedure

Market Simulators for Conjoint Analysis

Lasso on Categorical Data

The primary goal of this thesis was to understand how the spatial dependence of

Image MD. A superior way to measure Brand Imagery. Paul Richard McCullough. A MACRO White Paper, 2013

Review of Fundamental Mathematics

not possible or was possible at a high cost for collecting the data.

Dimensionality Reduction: Principal Components Analysis

Risk pricing for Australian Motor Insurance

LOGISTIC REGRESSION ANALYSIS

Advanced analytics at your hands

Regression with a Binary Dependent Variable

Predict the Popularity of YouTube Videos Using Early View Data

Mode and Patient-mix Adjustment of the CAHPS Hospital Survey (HCAHPS)

Tutorial Customer Lifetime Value

CS 2750 Machine Learning. Lecture 1. Machine Learning. CS 2750 Machine Learning.

The Probit Link Function in Generalized Linear Models for Data Mining Applications

Covariance and Correlation

A Comparison of Decision Tree and Logistic Regression Model Xianzhe Chen, North Dakota State University, Fargo, ND

Multiple Imputation for Missing Data: A Cautionary Tale

Unit 9 Describing Relationships in Scatter Plots and Line Graphs

CHAPTER 3 EXAMPLES: REGRESSION AND PATH ANALYSIS

4. Multiple Regression in Practice

Applied Data Mining Analysis: A Step-by-Step Introduction Using Real-World Data Sets

Connecting English Language Learning and Academic Performance: A Prediction Study

Generalizing Random Forests Principles to other Methods: Random MultiNomial Logit, Random Naive Bayes, Anita Prinzie & Dirk Van den Poel

Using Data Science & Predictive Models to Produce Foresight: The case of the presumptuous assumptions

Building risk prediction models - with a focus on Genome-Wide Association Studies. Charles Kooperberg

Formal Languages and Automata Theory - Regular Expressions and Finite Automata -

Silvermine House Steenberg Office Park, Tokai 7945 Cape Town, South Africa Telephone:

Keep It Simple: Easy Ways To Estimate Choice Models For Single Consumers

LOGIT AND PROBIT ANALYSIS

Analyzing Structural Equation Models With Missing Data

Two Correlated Proportions (McNemar Test)

Sample Size Issues for Conjoint Analysis

How to Get More Value from Your Survey Data

A Property and Casualty Insurance Predictive Modeling Process in SAS

Variable Selection in the Credit Card Industry Moez Hababou, Alec Y. Cheng, and Ray Falk, Royal Bank of Scotland, Bridgeport, CT

hp calculators HP 50g Trend Lines The STAT menu Trend Lines Practice predicting the future using trend lines

Association Between Variables

Analysis of Bayesian Dynamic Linear Models

CHAPTER 14 ORDINAL MEASURES OF CORRELATION: SPEARMAN'S RHO AND GAMMA

We discuss 2 resampling methods in this chapter - cross-validation - the bootstrap

Missing Data: Part 1 What to Do? Carol B. Thompson Johns Hopkins Biostatistics Center SON Brown Bag 3/20/13

OUTLIER ANALYSIS. Data Mining 1

Comparison of machine learning methods for intelligent tutoring systems

Covered California CAHPS Ratings Fall 2014 Scoring Health Plan s Historical CAHPS Results

Regularized Logistic Regression for Mind Reading with Parallel Validation

Additional sources Compilation of sources:

Beating the MLB Moneyline

Regression III: Advanced Methods

Introduction to nonparametric regression: Least squares vs. Nearest neighbors

Canonical Correlation Analysis

PS 271B: Quantitative Methods II. Lecture Notes

(AA12) QUANTITATIVE METHODS FOR BUSINESS

There are six different windows that can be opened when using SPSS. The following will give a description of each of them.

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12

REPORT DOCUMENTATION PAGE

Speaker First Plenary Session THE USE OF "BIG DATA" - WHERE ARE WE AND WHAT DOES THE FUTURE HOLD? William H. Crown, PhD

Multiple Choice Models II

Missing Data. A Typology Of Missing Data. Missing At Random Or Not Missing At Random

Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression

THE HYBRID CART-LOGIT MODEL IN CLASSIFICATION AND DATA MINING. Dan Steinberg and N. Scott Cardell

It is important to bear in mind that one of the first three subscripts is redundant since k = i -j +3.

Constructing a TpB Questionnaire: Conceptual and Methodological Considerations

A Hybrid Modeling Platform to meet Basel II Requirements in Banking Jeffery Morrision, SunTrust Bank, Inc.

CONTENTS PREFACE 1 INTRODUCTION 1 2 DATA VISUALIZATION 19

Application of discriminant analysis to predict the class of degree for graduating students in a university system

Student evaluation of degree programmes: the use of Best-Worst Scaling

The Artificial Prediction Market

REFLECTIONS ON THE USE OF BIG DATA FOR STATISTICAL PRODUCTION

How to Design and Interpret a Multiple-Choice-Question Test: A Probabilistic Approach*

Transcription:

How to Develop a MaxDiff Typing Tool to Assign New Cases into Meaningful Segments A white paper by: Jay Magidson, Ph.D., Statistical Innovations Inc. Gary Bennett, The Stats People Ltd jay@statisticalinnovations.com garyb@statspoeple.com ABSTRACT Use of latent class modeling to analyze MaxDiff response data has been found to be a good way to obtain meaningful needs-based segments. In order to classify new cases into the most appropriate segment, we describe and illustrate an innovative approach involving simulation to develop a typing tool where a relatively small number of binary items is included in future surveys, each item requesting a simple choice between a specified attribute pair. Both the latent class analysis and simulation are performed using the Latent GOLD program (Vermunt and Magidson, 2014). BACKGROUND Best-worst scaling (Louviere et. al, 2015), referred to as MaxDiff by marketing researchers, is an accurate and efficient way to measure needs-based preferences. Similarly, latent class analysis (LC) is generally recognized as the best approach for obtaining meaningful segments from different kinds of response data (Magidson, et. al, 2003, 2009). Together, LC applied to MaxDiff response data has been found to be a powerful way to identify segments with different needs (Cohen, 2003; Magidson, 2003; Louviere, et. al, 2015). When managers incorporate a segmentation solution into their strategic thinking, they naturally want to profile their customers as well as other respondents to future surveys into these same segments. However, it is not feasible to repeat the entire MaxDiff experiment in future surveys to obtain the necessary response data. In fact, inclusion of even a reduced number of MaxDiff sets may also be too difficult to implement without substantial respondent fatigue. Therefore, marketers often find it valuable to include a relatively small number of Golden Questions in future surveys to classify respondents into the most likely of these needs-based segments. To date, the challenge of identifying such questions for a MaxDiff segmentation has not been accomplished in a way that consistently yields high classification accuracy.

NEW APPROACH FOR CREATING A MAXDIFF TYPING TOOL In this paper we propose a new approach to develop a relatively small number of golden questions to incorporate in future surveys, along with a typing tool algorithm that is likely to yield good classification accuracy. Specifically, our approach selects a relatively small number of binary items for inclusion in future surveys, each item eliciting a simple choice between a specified attribute pair. We propose a specific method for selecting such items and for developing the associated classification equations. Since the new questions are to be included in future surveys, the fewer and simpler the questions the better, subject only to an acceptable correct classification rate. A key feature of our approach involves the use of LC class-specific utility estimates, to simulate all pairwise responses separately for each segment. These parameters characterize the distinct preferences of each LC segment. The simulated binary responses are used as candidate predictors from which a small but predictive subset of binary items can be selected. We illustrate our approach using real data, based on a design that accommodates measurement of the incremental benefit of using the simulated response data. Results suggest that, based on 14 selected binary items, a relatively high percentage of new respondents (85%) could be expected to be assigned to the correct segments. Without use of the simulated data, the expected accuracy from the resulting model would only be expected to be about 70%. Our proposed typing tool is not only easy to implement, but it also improves over earlier approaches, by allowing a customization which can provide higher accuracy for those segments judged most important to target. The approach is applicable to both simple, as well as complex MaxDiff designs, where options in each choice set are displayed 3-5 (or more) at a time and the total options can be large (e.g. 30+ options), with several versions. DATA The approach is illustrated using response data obtained from the Sydney Transport study (see Louviere, Flynn, & Marley, 2015), which consists of N=200 respondents. 2

The design is based on 9 funding options (Table 1), which are presented in 12 sets of triples. TABLE 1: THE 9 IMPROVEMENT OPTIONS INVESTIGATED IN THE SYDNEY TRANSPORT STUDY For example, choice set 1 (see Table 2) consists of options 2, 4 and 8, where respondents were asked: Which improvement should receive a) the highest priority to fund? b) the lowest priority to fund? TABLE 2: CHOICE SET #1 All 12 choice sets are displayed in Figure 1. Note that each of the ( 9 ) = 36 pairwise comparisons 2 of the 9 options occurs exactly once in this design. FIGURE 1: THE 12 CHOICE SETS USED IN THIS MAXDIFF DESIGN 3

METHODOLOGY The proposed method consists of 4 steps: STEP 1: SEGMENTATION. Use LC to Segment MaxDiff response data to obtain segments. We use the sequential best worst model to analyze the data. One of the advantages of this model is that the Bayesian Information Criteria (BIC) can be used to determine the number of segments (see Technical Appendix). The BIC identified 8 LC segments. For the class-specific utility parameter estimates for these 8 segments, see Table 3. Class1 Class2 Class3 Class4 Class5 Class6 Class7 Class8 class size 0.2083 0.177 0.136 0.123 0.1183 0.1123 0.0849 0.0401 Class-Specific Utility Parameters Alternatives Class1 Class2 Class3 Class4 Class5 Class6 Class7 Class8 1 More frequent off peak trains bet 1.2 1.0 0.4-0.1-0.3-0.7-0.1-0.4 2 Improved peak rail capacity 1.3 3.1 0.8 3.6 0.7 2.2 1.1 1.3 3 More frequent bus services on ma 1.4 1.8 0.2 2.6 0.6 2.2 0.8 0.6 4 Extensions of light rail services 0.5-1.9-1.2-1.8-0.3-1.4-2.4 1.3 5 Integrated fares -0.5 0.4 0.3-1.0-0.3-0.3 2.0 2.0 6 Integrated ticketing -0.7 0.4 0.1-1.2-0.4-0.4 2.7 2.6 7 Real-time arrival information -0.8-0.1 0.4-1.9-2.0-1.1-1.3 0.4 8 New cycleways; more bike and sco -1.3-2.6-1.6-2.0 0.4 0.9-1.9-3.6 9 Trains use green power -1.0-2.2 0.6 1.7 1.6-1.4-0.8-4.1 TABLE 3: CLASS SIZES AND UTILITY ESTIMATES FOR EACH LC SEGMENT Note, for example, that segment 5 (class size about 12%) is somewhat distinct from the other segments in that it has as its highest utility option 9 Trains use green power. In contrast, this option is one of the lowest priorities (negative utility estimate) for most of the other segments. For more details on these data see Magidson and Vermunt (2014). STEP 2: SIMULATE DATA. Use these LC model parameter estimates to define the segments (populations) and simulate a sample from each of these segment populations, where the simulated data consists of responses to all possible pairs (to be used as predictors). Here, we simulated N = 1,000 cases per segment 1. Since each segment is simulated separately, the dependent variable (true segment membership) is known, so the resulting data set contains both the dependent variable as well as the 36 candidate predictors as shown in Figure 2. The simulation approach allows us to work directly with true segment membership as the 1 Since the size of each segment obtained from the new cases will not necessarily be the same as that obtained from the original 200 respondents, equal numbers of respondents were generated for each LC segment. 4

dependent variable, rather than the classifications of the original 200 respondents, which contains about 8% misclassification error as estimated by the 8-class model. As mentioned above, each segment (population k) is defined by its associated MaxDiff model parameters. With these parameters, N[k]=1,000 cases were simulated using the MNL formula to generate responses to all 36 paired comparisons. The resulting simulated data (N=8,000 for the total 8 segments) are then analyzed (Step 3) to determine the subset of pairs that best predicts segment membership, and parameter estimates for these predictors to be used in a typing tool. FIGURE 2: SIMULATED DATA FOR ALL POSSIBLE (36) PAIRED PREFERENCES (PREDICTORS= CHOICE.12, CHOICE.13,, CHOICE.89) AS WELL AS THE TRUE SEGMENT (DEPENDENT VARIABLE = CLASS#) STEP 3: VARIABLE REDUCTION. Use a multinomial logit regression (MNL) modeling approach (such as stepwise MNL) that allows for variable reduction, to obtain the reduced number of pairs for inclusion in future surveys. Ideally, a regularized MNL modeling approach such as lasso (Friedman et. al, 2010) or Correlated Component Regression (Magidson, 2013) would be used, but since the sample size is relatively large and the number of predictors here is relatively small, the stepwise MNL may be sufficient. A large sample size is needed to prevent the analysis from being overly affected by sampling fluctuation. That is the reason that we use simulation rather than analyzing the relatively small sample of the MaxDiff respondents (N = 200). 5

Alternatively, a tree approach such as CHAID could be used to obtain an adaptive typing tool. Our expectation is that approach might not work as well because of reduced power, but since there is no limit on the size of the sample obtained from the simulation, this alternative is worth investigating. We plan to compare the use of such an adaptive approach in a follow-up study 2. An additional 9,000 respondents were simulated and held out from the MNL estimation for use as validation data, to validate the classification accuracy of the MNL equations. STEP 4: OBTAIN SCORING EQUATIONS. Use the resulting predictors and MNL parameter estimates to obtain the scoring equations. RESULTS The 14 pairs selected are provided in Table 4 below in the order in which they entered into the model, along with their p-value. Pair p-value (6,9) 1.9E-198 (7,8) 4.4E-217 (4,9) 2.0E-280 (3,6) 2.3E-167 (7,9) 9.2E-204 (1,8) 6.9E-222 (2,6) 1.4E-183 (5,9) 4.0E-170 (4,7) 3.0E-158 (3,9) 7.5E-171 (1,6) 2.4E-156 (6,8) 2.7E-151 (4,5) 2.4E-142 (3,7) 9.3E-134 TABLE 4: BINARY PREDICTORS SELECTED BY THE STEPWISE MNL MODEL. 2 We also plan to compare the current simulation approach with the use of a D-optimal design to select the binary items. 6

Tables 5A and 5B show that overall the 14-predictor MNL model correctly classifies 86.5% of the N= 8000 simulated respondents into the appropriate segment. Table 5B shows that the accuracy ranges from 82% - 94% correct classification rates across the 8 segments. TABLES 5A AND 5B: ACCURACY OF THE 14-PREDICTOR MNL MODEL (RESULTS FROM SIMULATED DATA) Tables 6A and 6B show that there is very little falloff in the correct classification rates when applied to the simulated validation data. TABLES 6A AND 6B: ACCURACY OF THE 14-PREDICTOR MNL MODEL (RESULTS FROM VALIDATION DATA) 7

The results are summarized in Columns 2 and 3 of Table 7 and in Figure 4 below for MNL results with 1 through 14 binary responses included as model predictors. It is clear that accuracy increases as the number of binary predictors increases up through 14 predictors at which point the accuracy achieves an overall rate of 85.2% based on the simulated validation sample. Number of Pairs Accuracy- In Sample Accuracy- Validation Sample Accuracy- Original Sample 1 24.2% 24.2% 30.5% 2 37.6% 37.6% 39.5% 3 46.3% 45.8% 47.0% 4 58.5% 58.3% 52.5% 5 64.1% 63.1% 56.0% 6 68.3% 68.0% 64.0% 7 71.5% 71.1% 68.5% 8 74.2% 73.9% 70.0% 9 76.9% 76.5% 71.5% 10 79.4% 79.3% 73.5% 11 81.6% 80.9% 74.5% 12 82.9% 82.5% 78.0% 13 84.8% 83.9% 80.5% 14 86.5% 85.2% 84.0% TABLE 7: SUMMARY OF CORRECT CLASSIFICATION RATES RESULTS FOR MODELS WITH 1-14 BINARY PREDICTORS. Results of Analysis with Simulation Data 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 Accuracy-In Sample Accuracy-Validation Sample FIGURE 4: ACCURACY IN ORIGINAL SAMPLE AND VALIDATION SAMPLE FOR MODELS WITH 1-20 BINARY PREDICTORS Table 7 also contains an additional column which summarizes how well each model recovers the LC class assignments of the original N= 200 sample. We would expect that these rates would be similar to those obtained in the validation sample. As can be seen by comparing the 8

two rightmost columns in Table 7, this expectation is confirmed, further supporting the validity of the proposed approach. ASSESSING THE VALUE OF THE SIMULATION A standard methodology is to append data to an observed sample in an attempt to improve prediction of some dependent variable on that observed data. A natural question is whether this approach would perform equally well in the current application. As mentioned earlier, the 12 sets utilized in this MaxDiff design allows comparison of each of the 36 option pairs exactly once. For example, the pairwise comparison of options 2 and 4 occurs in choice set 1 (recall Figure 1). Thus, it is possible to utilize the inferred responses from these 36 pairwise comparisons as predictors, and repeat the MNL modeling using the original N=200 respondents, as an alternative to using the N= 8,000 simulated respondents. This comparison was done, and results are plotted in Figures 5A and 5B below. While a MNL model based on the original N= 200 respondents containing 12 binary predictors predicts correctly the LC assignments for 91.5% of the respondents, use of the validation data shows that a substantial falloff would be expected when applying this model to new cases, resulting in only 69% of the cases being correctly classified. Comparing Figures 5A and 5B we see that the simulation approach would be expected to work much better than modeling based on the observed data. In particular, with 14 binary predictors we would expect an 85% correct classification rate with the simulation approach compared to only about 70% 3 if the MNL modeling were based on the original sample. Results of Analysis on Simulated Data Results of Analysis on Original Data (N=200) 100% 80% 60% 40% 20% 0% 1 3 5 7 9 11 13 15 17 19 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% 1 2 3 4 5 6 7 8 9 10 11 12 Accuracy-In Sample Accuracy-In Sample Accuracy-Validation Sample Accuracy-Validation Sample FIGURE: 5A OBSERVED VS. EXPECTED ACCURACY FOR MODELS BASED ON SIMULATED DATA FIGURE: 5B OBSERVED VS. EXPECTED ACCURACY FOR MODELS BASED ON ORIGINAL DATA 3 On the original sample stepwise MNL resulted in 12 predictors. Adding 2 additional predictors yields 70% accuracy. 9

DISCUSSION In this presentation we proposed a methodological approach to develop a typing tool based on MaxDiff response data consisting of only a small number of dichotomous items included in future surveys without much respondent fatigue. We proposed a 4-step approach for accomplishing this: 1. Segment the MaxDiff response data using the latent class (LC) Sequential Best-Worst Model. 2. For each resulting LC segment use the model parameter estimates from Step 1 (the class-specific utilities) to simulate N=1,000 best choices from all possible pairwise comparisons. 3. Use a variable reduction method to analyze the data obtained in Step 2. The dependent variable is the class label (LC segment) and the 36 binary responses are candidate predictors. 4. Use the selected pairs in future surveys and apply the prediction scoring equations obtained in Step 3 to get class predictions for the new cases. The results obtained from a real sample MaxDiff experiment suggested that a typing tool based on 14 binary response items would be expected to yield overall correct classification rates of over 85%. Repeating the approach without the use of the simulated data resulted in only 70% correct classification, supporting the value of the simulation. While the example used here contained only 9 options displayed in 12 sets of triples and consisted of only a single version, the proposed approach can be applied with much larger MaxDiff designs that include multiple versions. The reason for this is that the simulation requires only the class-specific parameters obtained from the LC segmentation analysis of the original data. Note that for such larger designs, it is much more difficult to perform the variable reduction task (step 3) without simulated data. The approach requires the ability to create simulated pairwise comparisons for each segment. This is straightforward with LC models 4 but simulation would be very difficult or impossible if HB were used to produce individual utilities. Moreover, we generally advise against using 2-step 4 Here we use the plain-vanilla version of latent class MaxDiff, but the approach can be generalized easily to LC models extended to include different scale factors for best and worst responses and/or latent scale classes (Magidson and Vermunt, 2014), inclusion of covariates, and fused models where the segmentation is based on combination of MaxDiff responses and ratings. 10

(tandem) approaches such as Hierarchical Bayes (HB) to get individual utilities followed by clustering, since the normal prior assumed by HB is tantamount to the assumption that segments do not exist (Magidson, Dumont and Vermunt (2015). The particular LC MaxDiff model used here was the sequential logit model as implemented in Latent GOLD, for which the BIC statistic can be used to determine the number of segments. With the standard LC MaxDiff implementation (Louviere, 1993), BIC and related fit statistics cannot be used to determine the number of segments since they always suggest too many segments. See Technical Appendix for further details. Simulated data were generated using Latent GOLD. As mentioned above (recall footnote 3), the approach may also be generalized to extended LC models estimated using Latent GOLD, which can account for within-class heterogeneity, and simulated data can also be generated for these models using Latent GOLD. In summary, we were able to develop a typing tool with only 14 binary response pairs that would be expected to correctly classify 85% of new cases. By increasing the number of binary items to 20 we could increase the overall classification rate to 92% at the cost of some additional respondent fatigue. Use of all 36 binary items in this example would be expected to produce 96% correct classification, but at the expense of additional fatigue that might be expected to reduce this rate somewhat. Options Correct Classification Fatigue All 36 binary pairs <96% High 20 selected pairs 92% Medium Only 14 selected pairs 85% Low For comparison MaxDiff with 12 sets of triples <92% Very High Best response only w original 12 sets 83% TABLE 8: COMPARISON OF APPROACHES 11

FUTURE STUDIES We plan to conduct future studies to investigate the following: 1. Whether a regularized regression approach such as Correlated Component Regression (Magidson, 2013) outperform stepwise MNL 2. Whether an adaptive typing tool obtained using CHAID in place of stepwise MNL can outperform the current approach 3. Whether the proposed simulation strategy can also be used to reduce the number of variables in current typing tools, by applying the known class option in Latent GOLD to an existing segmentation to obtain the parameters needed for the simulation 4. How the use of D-optimal design to select the subset of predictors compares with the current approach 12

References Cohen, Steve (2003), Maximum Difference Scaling: Improved Measures of Importance and Preference for Segmentation, 2003 Sawtooth Software Conference Proceedings, Sequim, WA., pages 61-74. Friedman, J., Hastie, T. and Tibshirani, R. (2010) Regularization Paths for Generalized Linear Models via Coordinate Descent J Stat Softw. 2010; 33(1): 1 22. Louviere (1993) The Best-Worst or Maximum Difference Measurement Model: Applications to Behavioral Research in Marketing, The American Marketing Association's 1993 Behavioral Research Conference, Phoenix. Louviere, J.J., Marley, A.A.J., & Flynn, T. (2015) Best-Worst Scaling: Theory, Methods and Applications, forthcoming, Cambridge University Press. Magidson, J. (2013). Correlated component regression: Re-thinking regression in the presence of near collinearity. In: New perspectives in partial least squares and related methods. Springer Verlag. Magidson, Jay (2003), Discussant Comments on Steve Cohen s Maximum Difference Scaling: Improved Measures of Importance and Preference for Segmentation, 2003 Sawtooth Software Conference Proceedings, Sequim, WA., pages 75-76. Magidson, Dumont, and Vermunt (2015) A New Modeling Tool for Identifying Meaningful Segments and their Willingness to Pay: Improving Validity by Reducing the Confound between Scale and Preference Heterogeneity, 2015 ART Forum. Magidson, J. Thomas, D., Vermunt, J.K. (2009). A new model for the fusion of MaxDiff scaling and ratings data. 2009 Sawtooth Software Proceedings, 83-103. Magidson, J. and Vermunt J.K. (2014) Statistical Innovations MaxDiff Tutorial 8A: Analyzing MaxDiff Data with Scale Factors: Applications of Latent GOLD Basic + Choice, http://www.statisticalinnovations.com/wp-content/uploads/lgchoice_tutorial_8a.pdf Magidson, J., Eagle, T., and Vermunt, J.K. (2003). New developments in latent class choice models. April 2003 Sawtooth Software Conference Proceedings, 89-112. (pdf) Vermunt, J.K. and J. Magidson (2014). Upgrade Manual for Latent GOLD Choice 5.0: Basic, Advanced, and Syntax. Belmont Massachusetts: Statistical Innovations Inc. http://www.statisticalinnovations.com/wp-content/uploads/lgchoice5.0_upgrademanual.pdf 13

Technical Appendix Model Consistent Margins BIC suggest # classes Parameters Unbiased MaxDiff NO YES YES MaxDiff-Ind YES NO -- too many NO Best-Worst YES YES YES There are 3 primary MaxDiff (Best-Worst) modeling approaches, all implemented in Latent GOLD: We used the Sequential logit method ( Best-Worst above) as implemented in the Latent GOLD program (recommended). The MaxDiff model is described by Marley and Louviere (2005), and The MaxDiff-Ind model was proposed earlier by Louviere (1993) and implemented by Sawtooth Software. Our Best-Worst model with 8 classes contains 71 parameters. Additional parameters can be included in the models: Scale parameters account for additional within-class variability Note: Different scale factors for Best and Worst choices (generalizes the +1, -1 coding) appears to almost always be supported with real data.