Sawtooth Software. Sample Size Issues for Conjoint Analysis Studies RESEARCH PAPER SERIES. Bryan Orme, Sawtooth Software, Inc.

Similar documents
Sawtooth Software. Which Conjoint Method Should I Use? RESEARCH PAPER SERIES. Bryan K. Orme Sawtooth Software, Inc.

Sawtooth Software. Brief Managerial Overview of Conjoint (Trade-Off) Analysis RESEARCH PAPER SERIES. Bryan Orme, Sawtooth Software, Inc.

Sawtooth Software. Using Utility Constraints to Improve the Predictability of Conjoint Analysis RESEARCH PAPER SERIES

Sawtooth Software. ACA, CBC or Both? Effective Strategies for Conjoint Research RESEARCH PAPER SERIES. Bryan K. Orme, Sawtooth Software 1997

Sawtooth Software. Learning Effects in Preference Tasks: Choice-Based Versus Standard Conjoint RESEARCH PAPER SERIES

Sawtooth Software. A Perspective on Adaptive CBC (What Can We Expect from Respondents?) RESEARCH PAPER SERIES. Rich Johnson, Sawtooth Software, Inc.

Gush vs. Bore: A Look at the Statistics of Sampling

Chapter 12. Sample Surveys. Copyright 2010 Pearson Education, Inc.

Sawtooth Software. The Benefits of Accounting for Respondent Heterogeneity in Choice Modeling RESEARCH PAPER SERIES

Sawtooth Software. The Importance Question in ACA: Can It Be Omitted? RESEARCH PAPER SERIES

Chapter 4: Foundations for inference. OpenIntro Statistics, 2nd Edition

International Program for Development Evaluation Training (IPDET)

Secondary Math Margin of Error

Sawtooth Software Conference

Equipment and preparation required for one group (2-4 students) to complete the workshop

Choice Modeling Analytics Benefits Of New Methods

If you are using a survey: who will participate in your survey? Why did you decide on that? Explain

MATHEMATICAL MODELLING

AP Stats ~ Lesson 8A: Confidence Intervals OBJECTIVES:

A Method for Handling a Large Number of Attributes in Full Profile Trade-Off Studies

Chapter 19. Confidence Intervals for Proportions. Copyright 2012, 2008, 2005 Pearson Education, Inc.

BOOTSTRAPPING AND CONFIDENCE INTERVALS

Sawtooth Software. The CBC System for Choice-Based Conjoint Analysis. Version 9 TECHNICAL PAPER SERIES. Sawtooth Software, Inc.

Marketing Research to Support the Stage Gate New Product Development Process

Confidence Intervals

What Is Conjoint Analysis? DSC 410/510 Multivariate Statistical Methods. How Is Conjoint Analysis Done? Empirical Example

Sawtooth Software. Comparing Hierarchical Bayes Draws and Randomized First Choice for Conjoint Simulations RESEARCH PAPER SERIES

Confidence Intervals for Large Sample Means

Contact: Version: 2.0 Date: March 2018

An Introduction to Choice-Based Conjoint

Managerial Overview of Conjoint Analysis

An Introduction to Choice-Based Conjoint

Guest Concepts, Inc. (702)

Chapter 12 Module 3. AMIS 310 Foundations of Accounting

Managing a Successful CBC Project

Lecture (chapter 7): Estimation procedures

Make sure to listen to this audio: as you go through this handout, to get maximum value.

Monica H. Swahn, Ph.D., MPH Associate Vice President for Research Professor, Public Health Partnership for Urban Health Research Georgia State

Sawtooth Software. The ACA/Web v6.0 Technical Paper TECHNICAL PAPER SERIES

Suppose we wanted to use a sample of 25 students to estimate the average GPA of all students. Now suppose we choose our sample by random sampling and

Click to return to In this Lesson

Statistics 201 Spring 2018 Exam 2 Practice Exam (from Fall 2016)

Understanding and Managing Uncertainty in Schedules

Thus, there are two points to keep in mind when analyzing risk:

Displaying Bivariate Numerical Data

Operations and Supply Chain Management Prof. G. Srinivisan Department of Management Studies Indian Institute of Technology, Madras

Gasoline Consumption Analysis

Forecasting Introduction Version 1.7

DRAFT NON-BINDING BEST PRACTICES EXAMPLES TO ILLUSTRATE THE APPLICATION OF SAMPLING GUIDELINES. A. Purpose of the document

Data Collection Instrument. By Temtim Assefa

Watson-Glaser III Critical Thinking Appraisal (US)

Sampling and Investigating Hard Data

FAQ: Collecting and Analyzing Data

Capability on Aggregate Processes

The Meaningful Hospitality Smart Hiring Guide

Daily Optimization Project

The Envision process Defining tomorrow, today

Discriminant Analysis Applications and Software Support

THE MUTUAL GAINS APPROACH TO NEGOTIATION: A FOUR-STEP PROCESS

Watson Glaser III (US)

// How Traditional Risk Reporting Has Let Us Down

Why Do So Many Online Businesses Fail?

Survey Question Analysis (Draft )

Case 2.1. American Idol: A Big Hit for Marketing Research?

Practical Sampling for Impact Evaluations

BUSINESS Operating in a local business environment (Component 1)

Webinar Advanced Topics in CBC Part 1

Sawtooth Software. Political Landscape 2008: Segmentation Using MaxDiff and Cluster Ensemble Analysis RESEARCH PAPER SERIES

provided that the population is at least 10 times as large as the sample (10% condition).

Cut Your PPC Campaigns Down To Size

Chapter 18: Sampling Distribution Models

Utilizing Optimization Techniques to Enhance Cost and Schedule Risk Analysis

Day 1: Confidence Intervals, Center and Spread (CLT, Variability of Sample Mean) Day 2: Regression, Regression Inference, Classification

Scope Creep: need strong change control to prevent or mitigate. Must establish the change control system early in the project.

VIDEO 1: WHY ARE FORMS IMPORTANT?

Software for Typing MaxDiff Respondents Copyright Sawtooth Software, 2009 (3/16/09)

Three steps to joining and participating in unions

8 ways to make your survey useless if you are making any of these mistakes with your surveys, you might as well throw away the data and start all over

Multiple Regression. Dr. Tom Pierce Department of Psychology Radford University

A GUIDE TO GETTING SURVEY RESPONSES

The 10 Big Mistakes People Make When Running Customer Surveys

Fifteen Undeniable Truths About Project Cost Estimates, or Why You Need an Independent Cost Estimate

Coaching Underperforming Employees

Getting Started with OptQuest

Chapter 2 Data Collection

Managing Your Backlog

Build Trust in Survey Responses Sample Selection & Sample Size

Using Natural Language Processing and Machine Learning to Forecast Legislative Outcomes

How digital is transforming retail: The view from ebay

Five Interview Phases To Identify The Best Software Developers

Making a Bid For Profit. Profit-Driven Bidding Tactics in AdWords

32 BETTER SOFTWARE JULY/AUGUST 2009

Definitive Guide for Better Pricing. Build a solid pricing foundation that will help you create consistent sales and profit growth.

Which Method Should I Use?

TickITplus Implementation Note

GUIDANCE ON CHOOSING INDICATORS OF OUTCOMES

Bitcoin can be used to buy things electronically. In that sense, it s like conventional dollars, Euros, or yen, which are also traded digitally.

SCHEDULING EFFICIENCY: FINDING THE RIGHT MEASURE

STOP. COLLABORATE & LISTEN. EIGHT BEST PRACTICES FOR IMPROVING COLLABORATION IN THE PROPOSAL PROCESS

Lecture 9. Income disparity among countries Endogenous growth: a model of human capital accumulation

Transcription:

Sawtooth Software RESEARCH PAPER SERIES Sample Size Issues for Conjoint Analysis Studies Bryan Orme, Sawtooth Software, Inc. 1998 Copyright 1998-2001, Sawtooth Software, Inc. 530 W. Fir St. Sequim, WA 98382 (360) 681-2300 www.sawtoothsoftware.com

Sample Size Issues for Conjoint Analysis Studies Bryan Orme, Sawtooth Software, Inc. Copyright Sawtooth Software, 1998 Introduction I m about to conduct a conjoint analysis study. How large a sample size do I need to get statistically significant results? This is a common question. Unfortunately, it is one of the most difficult questions to answer, as so many issues come to play: What is it exactly that you are trying to measure to get a statistically significant result : a specific part-worth, preference for a product, or differences in preference between groups of people? Do you expect that the differences between features/products/groups you are trying to detect is a subtle or a strong one? What level of certainty do you need to be able to place upon your conclusions: 99% confidence, 90% confidence, or what? How large is the total population in the market for your product? What conjoint methodology do you plan to use? How many conjoint questions will each respondent answer? Do you need to compare subsets of respondents, or are you going to be looking at results only as a whole? How homogenous is your market? Do people tend to think alike, or are there at lot of differences in preferences? How do you plan to select your sample? Random? Convenience Sample? How much budget do you have to work with? The answers to each of these questions play a role in determining the appropriate sample size for a conjoint study. This paper will provide both tools and theory to help conjoint researchers make sample size decisions. Though most of the principles that influence the sample size determination are based on statistics, successful researchers develop heuristics for quickly determining sample sizes based on experience, rules-of-thumb and budget constraints.

Sampling Error vs. Measurement Error Errors are deviations from truth. In marketing research, we are always concerned with reducing error in cost-efficient ways. Assuming that you have selected the appropriate modeling method, there are two sources of error that cause preference data to deviate from truth. The first is Sampling Error. Sampling error occurs when samples of respondents deviate from the underlying population. If we have drawn a random sample (each population element has an equal probability of being selected), any sampling error is due to chance. If, on the other hand, our sample is not random (for example, a convenience sample), the sampling errors may be systematic. With random samples, we reduce sampling error by simply increasing the sample size. If a sample is not random, we may be able to reduce sampling error by increasing sample size only to a point, after which we can never get closer to truth due to our sample reflecting systematic differences from the general population. To illustrate sampling error, assume we wanted to figure out how far the average adult can throw a baseball. If we drew a random sample of 30 people, and by chance happened to include Ken Griffey, Jr. (center fielder, Seattle Mariners), our estimate would likely be farther than the true distance for the average adult. It is important to note that the samples we use in marketing research are rarely truly random. Some respondents resist being interviewed, and by selecting themselves out of our study are a source of Non-Response Bias. A second source of error in conjoint data is Measurement Error. Measurement error is reduced by having more (and/or better) data from each respondent. Consider again the example of the baseball toss. Suppose you are one of the study participants. You throw the ball, but you accidentally step into an uneven spot on the ground, and the ball doesn t go as far as you typically could throw it. If we asked you to take another few tosses, and averaged the result, we would reduce the measurement error and get a better idea of how far you truly could throw a baseball. In conjoint analysis, we reduce measurement error by including more conjoint questions. We recognize, however, that respondents get tired, and there is a limit beyond which we can no longer get reliable responses, and therefore a limit to the amount we can reduce measurement error. Confidence Intervals for Proportions We re all familiar with the practice of reporting the results of opinion polls. Typically, a report may say something like: If the election were held today, Mike Jackson is projected to capture 50% of the vote. The survey was conducted by the XYZ company and has a Margin of Error of +/- 3%. A margin of error of +/- 3% is standard for opinion research, and is usually based on a 95% confidence interval. One can achieve a margin of error for proportions based on a 95% confidence interval with a sample size of just over 1000 respondents. In more technical terms, the report meant that if the XYZ company were to repeat the poll a large number of times (with a different random sample each time), the confidence interval associated with each estimate (in

this case, 47% to 53%) would contain the true proportion that would vote for Jackson 95% of the time. It also means that the poll results would be wrong 5% of the time: the true vote for Jackson wouldn t fall within the reported confidence interval. The Standard Error of a Proportion is equal to: PQ (n-1) where: P = expected proportion Q = (1 - P) n = sample size And the margin of error for a proportion is equal to +/- 1.96 times the standard error, where 1.96 is the Z-score associated with the 95% confidence interval (two-tail test). Suppose we interviewed 500 respondents (randomly selected) and asked whether they approved of the president s job performance, and 65% said yes. What is the margin of error at the 95% confidence level? 196. (. 65 )(. 35 ) ( 500-1) = 0.042 The margin of error is +/- 4.2%, and we are 95% confident that the true proportion that approve of the president s job performance is somewhere between 60.8% and 69.2%. The reason we have spent time here talking about confidence intervals for proportions is that these same principles can be applied to choice-based conjoint results and Shares of Choice for First Choice market simulations. Confidence Intervals for Continuous Variables Given the assumption of a random sample and a particular confidence level, it is easy for a marketing research analyst to calculate the sample size needed to achieve a certain margin of error for a proportion. It is simple, because the formula for calculating the Standard Error for proportions is always the same. With continuous variables (ratings-based responses to conjoint profiles), one cannot know the standard error before fielding a study. The Standard Error of the Mean is directly related to how much respondents agree in their responses, which differs from study to study.

Let s say, however, that we were conducting an ACA study and had already collected 40 interviews. Furthermore, assume we were interested in being able to simulate the purchase likelihood (on a 100-pt scale) of our client s planned product introduction with a margin of error of +/- 3 at the 95% confidence level. The ACA market simulator reports the Margin of Error next to the purchase likelihood estimate. Suppose the report was as follows: Total Respondents = 40 Purchase Likelihood Std Error Product A 78.34 3.06 The margin of error is equal to +/- 1.96 times the standard error of the mean, or +/- 6.00. We need to cut this margin of error in half to achieve our target level of precision. Sampling theory tells us that the standard error of either a proportion or a mean is always figured by dividing by the square root of the sample size (minus one). To decrease the error by a factor of two, one must increase sample size by a factor of four. Therefore, we need to interview about 120 more respondents to reduce the margin of error to +/- 3. Sampling for Small Populations The examples we ve presented thus far have assumed infinite (very large) population sizes. However, let s suppose that instead of estimating the job performance rating of the president by the US population at large, we instead wanted to estimate (with a margin of error of +/- 3%) the job performance rating of a school principal by members of the PTA. Suppose there are only 100 members of the PTA. We previously stated that one needs to sample roughly 1000 respondents to achieve a +/- 3% margin of error. That assumed an infinite sample. But in this case there are only 100 potential respondents in the entire population. How many PTA members do we need to interview to achieve a margin of error of +/- 3% for our estimate? First, we should introduce a new term called a Finite Population Correction (FPC), equal to (N - n)/(n - 1), where n = sample size and N = population size. The FPC is often simplified to (1 - f), where f = (n/n), which is approximately equivalent to (N - n)/(n - 1) for all except the smallest of populations. Once a population reaches about 5000 individuals or more, one can generally ignore the finite population correction factor, as it has a very small impact on sample size decisions for typical market research studies. The margin of error (95% confidence interval) for a finite sample (using the simplified FPC) is equal to: + / - 196. ( 1-f ) PQ (n -1 )

With a population of 100, one can solve for n (equation not shown here), assuming an expected proportion. The worst case scenario (largest standard error) is for a 50% proportion, so it is standard to set P = 0.50. After solving for n, we discover that we would need to interview 92 PTA members, or 92% of the population to achieve a margin of error of +/- 3%. The important point to be made is that with small populations, you may have to interview a significant proportion of the population to achieve stable estimates. Suppose your client produces a very expensive, highly specialized piece of machinery, for which there were only 100 total potential customers in the world. Given many peoples unwillingness to complete surveys, it is likely much more difficult to complete surveys with 92 out of 100 potential buyers of this product than to interview roughly 1000 potential buyers of something like office chairs, for which there are so many buyers as to approximate an infinite population. Even so, in terms of estimating a proportion, both scenarios lead to the same margin of error. Many researchers and dozens of data sets have demonstrated that conjoint utilities do a good job in predicting individual respondents preferences for products. Holdout choice sets (choice tasks not used to estimate utilities) are often included in conjoint questionnaires. Using the conjoint data, a respondent s holdout choices usually can be predicted with a hit rate of roughly 75 to 85%. These choice tasks typically include between three to five different product concepts, so by chance we expect a success rate of between 20% to 33%. The hit rates with conjoint are significantly greater than chance; and significantly better than the marketer s best guess even if the marketer knows each customer very well. In fact, conjoint predictions at the individual level frequently match or even exceed test-retest reliability, suggesting that a good set of conjoint utilities are just as reliable at predicting choices to repeated holdout tasks as the respondents themselves. This suggests that if there was even only one buyer of your product in the world, you could learn a great deal about that individual s preferences from a conjoint interview. The utility data would be reasonably accurate for predicting his preferences and the weights he placed upon different attributes. It seems reasonable to use conjoint analysis among even the smallest of populations, provided you interview enough respondents to represent the population adequately. Sample Size Issues for ACA An ACA interview results in a set of utilities for each individual. Therefore, in the limit, the minimum sample size is just one person. Of the three conjoint methods discussed in this paper, ACA is generally the best at reducing measurement error. ACA s interviews adapt to the respondent, asking questions designed to be maximally relevant and efficient for refining utility estimates. The priors section helps significantly in stabilizing the utility estimates at the individual level. One generally sees fewer reversals in part worths (out-of-order utilities) for ordered attributes like price in ACA than for CVA (traditional conjoint) and CBC (with ICE estimation).

In ACA, one needs to decide how many pairs questions to ask. The number of pairs each respondent completes has a significant role in reducing measurement error. The suggested number of pairs is: where, 3 (N - n - 1) - N N = total number of levels n = total number of attributes If respondents answer as many pairs as suggested, a total of three times the number of observations as parameters are available at the individual level for computing utilities (this includes information from the priors matrix). Sometimes the suggested number of pairs is greater than respondents can reasonably do. You should make sure not to overburden respondents, as this can lead to poor results. Please see a paper entitled Accuracy of Utility Estimation in ACA available for downloading from our home page at http://www.sawtoothsoftware.com in our technical papers library for examples of how the number of pairs affects the precision of estimates. If your sample size is particularly small and the number of attributes to measure is relatively large (greater than about 6), ACA is likely the best tool to use. In fact, I have heard of an entire research study designed to learn about the preferences of just one respondent (an important buyer of an expensive industrial product). There are many other considerations for determining whether ACA is appropriate for your study. I d recommend an article entitled Which Conjoint Method Should I Use? also available for downloading from our home page. Sample Size Issues for Traditional Conjoint Like ACA, traditional conjoint (CVA) usually leads to individual-level utilities. Again, the minimum sample size is one. However, because the CVA methodology does not include a priors section, its utilities tend to have greater variability (larger standard errors) at the individual level relative to ACA. One should include enough conjoint questions (cards) to help control measurement error. The CVA manual suggests asking enough questions to obtain three times the number of observations as parameters to be estimated, or a number equal to: 3 (N - n + 1) where, N = total number of levels n = total number of attributes

Again, sometimes respondents will not have the energy or patience to answer so many questions. You must strike a good balance between overworking the respondent (and getting noisy data) and not asking enough questions to reasonably stabilize the estimates. Sample Size Issues for CBC Though generally considered more realistic than traditional conjoint, choice-based questions are a relatively inefficient way to learn about preferences. As a result, sample sizes are typically larger than with ACA or CVA, and CBC results have traditionally been analyzed by aggregating respondents. Lately, Hierarchical Bayes and ICE (Individual Choice Estimation) analysis have permitted individual-level estimation of utilities from CBC data. However, to compute individual-level models, both of these techniques use information from many respondents to refine the utilities of each individual. Therefore, one usually doesn t calculate utilities using a sample size of one. (It should be noted, however, that logit analysis can be run at the individuallevel, if the number of parameters to be estimated is few, the design is highly efficient, and the number of observations is many.) If you plan to estimate individual utilities using ICE, you should probably include at least 20 tasks to reduce measurement error to a reasonable level. While the lower limits on sample size are not known, you begin to tread on shaky ground when computing ICE utilities using small sample sizes. There are rules-of-thumb for determining sample sizes for CBC if we are willing to think about aggregate estimation of effects. Like proportions, choices reflect 0/1 data, and the rules for computing confidence intervals for proportions are well defined and known prior to collecting data. Consider a design with three brands and three prices. Assume each person completes 10 tasks, and each task displays three products (each brand and price occurs once per task). If we interview 100 respondents, each brand will have been available for choice (100 respondents) x (10 tasks) x (3 concepts) / (3 brands) = 1000 times. It is our experience that having each respondent complete 10 tasks is about as good at reducing error as having 10 times as many respondents complete one task. Of course, in the limit this logic falls apart: it doesn t make sense to say that having one respondent complete 1000 tasks is as good as having 1000 respondents complete one task. But, if you include enough respondents to represent the population adequately, doubling the number of tasks they complete is about as good as doubling the sample size in terms of reducing error. (Johnson and Orme demonstrate this is their paper entitled How Many Questions Should You Ask in Choice Based Conjoint Studies?, also available for downloading from our home page.) Therefore, it makes sense from a cost-benefit standpoint to have respondents complete many choice tasks. Johnson, author of the CBC System, has recommend the following rule-of-thumb when deciding sample size for aggregate-level CBC modeling: nta / c >= 500

where: n = number of respondents t = number of tasks a = number of alternatives per task (not including the None ) c = number of analysis cells When considering main-effects, c is equal to the largest number of levels for any one attribute. If you are also considering all two-way interactions, c is equal to the largest product of levels of any two attributes. Simulation Methods and Sample Sizes Conjoint utilities can be used within a choice simulator to predict preference for different product concepts in competitive scenarios. There are various simulation methods, including the simple First Choice (maximum utility rule) and the logit/btl models. First Choice simulations assume that each respondent can vote for only one product, and that one alternative captures 100% of the share for each respondent. Shares of preference under the First Choice rule are really just the familiar proportions. In contrast, logit/btl models let respondents choose products in a probabilistic manner. Assume three products in a market scenario (A, B, C). A respondent might reflect the following choice probabilities (0.6, 0.3, 0.1), whereas the First Choice rule for that same respondent would reflect (1, 0, 0). The probabilistic model captures more information from each respondent to stabilize share estimates. The standard errors for share predictions from logit/btl simulations are always smaller than under the First Choice rule. Therefore, if you plan to use the First Choice model, you will need larger sample sizes to stabilize share of choice estimates relative to probabilistic simulation models. Typical Sample Sizes and Practical Guidelines All of the recommendations below assume infinite (very large) populations. They are based on the theories above and this author s observations of common practices in the market research community. Sample sizes for conjoint studies generally range from about 150 to 1200 respondents. If the purpose of your research is to compare groups of respondents and detect significant differences, you should include enough sample size to accommodate a minimum of about 200 per group. Therefore, if you are conducting a segmentation study and plan to divide respondents into as many as 4 groups (i.e. through cluster analysis) it would be wise to include at a minimum (4)(200) = 800 respondents. This of course assumes your final group sizes will be about equal, so one would usually want more data. The stronger segmentation studies I m aware of include about 800 or more respondents.

For robust quantitative research where one does not intend to compare subgroups, I d generally recommend at least 300 respondents. For investigational work and developing hypotheses about a market, between 30 to 60 respondents may do. These suggestions always have to be weighed against research costs. These are difficult decisions to be made based on experience, the application of statistical principles, and sound judgment. If after the fact you find yourself questioning whether you really needed to have collected such a large sample size for a particular project, it is an interesting exercise to delete (randomly) some of the data and see how having fewer respondents would have affected your findings.