Mixed-Mode Methods for Conducting Survey Research

Similar documents
Probability of Selecting Response Mode Between Paper- and Web-Based Options: Logistic Model Based on Institutional Demographics

Response Rate and Measurement Differences in Mixed Mode Surveys. Using Mail, Telephone, Interactive Voice Response and the Internet 1

Introduction to Survey Methodology. Professor Ron Fricker Naval Postgraduate School Monterey, California

IVR: Interactive Voice Response

THE FUTURE OF INTERNET-BASED SURVEY METHODS

Questionnaire Design in Telephone Surveys: Interviewers and Research Call Center Managers Experience

Jon A. Krosnick and LinChiat Chang, Ohio State University. April, Introduction

KNOWLEDGEPANEL DESIGN SUMMARY

Lessons Learned From an NHIS Web- Followback Feasibility Study

5. Survey Samples, Sample Populations and Response Rates

Americans and text messaging

17% of cell phone owners do most of their online browsing on their phone, rather than a computer or other device

Get it together: six ways to effectively integrate phone, web, and surveys

The application of alternative modes of data collection in UK Government social surveys. A report for the Government Statistical Service

Optimising survey costs in a mixed mode environment

Income and Health Insurance Coverage: Comparing the American Community Survey (ACS) and the Current Population Survey (CPS)

BACKGROUND ON THE SURVEY PROCESS

Americans and their cell phones

Designing & Conducting Survey Research

Surveys: Why, When and How. Gavin Henning Pamela Bagley

Evaluating Mode Effects in the Medicare CAHPS Fee-For-Service Survey

Keywords personal interview; interviewer bias; interviewer effects; data quality; response bias

Use of Text Messaging to Increase Response Rates

Good Practices in Survey Design Step-by-Step

LITERATURE REVIEW OF WEB AND SURVEYS

Multimode Strategies for Designing Establishment Surveys

Internet Surveys. Examples

MBA 6001, Organizational Research and Theory Course Syllabus. Course Description. Prerequisites. Course Textbook. Course Learning Objectives.

Amy F Subar, PhD, MPH, RD Applied Research Program Risk Factor Monitoring and Methods Branch

Consumer s Guide to Research on STEM Education. March Iris R. Weiss

MERIDITH PEASE SELDEN, PH.D. ASSISTANT PROFESSOR OF PSYCHOLOGY

Using Online Surveys to Evaluate Distance Education Programs

Eye Tracking on a Paper Survey: Implications for Design

Five steps to improving the customer service experience

Online versus traditional teaching evaluation: mode can matter

WEB-BASED ORIGIN-DESTINATION SURVEYS: AN ANALYSIS OF RESPONDENT BEHAVIOUR

Self-administered Questionnaire Survey

A Comparison of Level of Effort and Benchmarking Approaches for Nonresponse Bias Analysis of an RDD Survey

71% of online adults now use video sharing sites

YUNTAO DONG. Assistant Professor, School of Business, University of Connecticut

Using Surveys for Data Collection in Continuous Improvement

Response Rates and Mode Preferences in Web-Mail Mixed-Mode Surveys: A Meta-Analysis

Chat Enhancements Optimize Customers Web Experience

The Own Your Future Long-Term Care Awareness Campaign: Implications for CLASS

LAWRENCE MORLEY HANSER

Interactive Intelligence

CardSwap. Daniel Watrous. Marketing. Northwest Nazarene University

Evava S. Pietri Curriculum Vitae

James E. Bartlett, II is Assistant Professor, Department of Business Education and Office Administration, Ball State University, Muncie, Indiana.

Leadership Studies, Ph.D. University of San Diego Exit Survey Results

Objectives 3/27/12. Process Improvement: The Customer is Not the Problem. Rosa West PhD, MBA, LMHC, LMFT. Westcare Wellness & ConsulHng, LLC

Consumer research into use of fixed and mobile internet

EcoPinion Consumer Survey No. 19 The Conflicted Consumer Landscape in the Utility Sector. May 2014

The Role of Feedback Management in Becoming Customer Centric

Department of Marketing

Transcription:

Mixed-Mode Methods for Conducting Survey Research Herbert M. Baum, Ph.D.; Anna Chandonnet M.A.; Jack Fentress M.S., M.B.A.; and Colleen Rasinowich, B.A. www.datarecognitioncorp.com

In November 2010 the Federal government held The Workshop on the Future of Federal Household Surveys where Dr. Don Dillman presented a paper entitled Making New Data Collection Modes Effective in Household Surveys and in 2011 Public Opinion Quarterly published an article The Future Modes of Data Collection (Couper). Both these distinguished authors note that there is growing interest in mixed-mode surveys, because of the limitations posed by each of the major existing modes (paper/pencil, personal, telephone, and web). The typical limitations cited by mode are: Cost paper/pencil and personal Time paper/pencil and personal Coverage telephone and web Decreasing response rates paper/pencil and telephone Different preferences of respondents all modes To overcome these limitations purchasers of surveys as well as the survey research community have been seeking innovative solutions. In addition to exploring mixedmode methods from the perspective of its impact on the quality of data being gathered, this topic is important because with declining response rates there is greater interest in offering respondents options of how they want to respond to a survey. To increase response rates, survey research needs to move from our needs to those of the respondent. For example, people are accessing information via the Internet, but increasingly doing so via smartphones. Can we use that to invite people to participate in a survey? What is a mixed-mode survey? Couper (2011) presents a lengthy discourse on what is meant by mode and notes that there can be as much variation within mode as between modes. For example, when conducting telephone interviewing is the survey being conducted using computer-assisted telephone interviewing (CATI) or by interactive voice response (IVR)? The former has a live person conducting the interview while the latter does not. Between CATI surveys there may be differences in how the sample was selected random digit dialing (RDD) versus dual-frame sampling; thus we need to acknowledge that intra-mode variations do need to be considered as well. Intra-mode variation is not covered in this paper. For this paper, mode refers to the four general survey methods: paper/pencil, telephone (CATI or IVR), personal (including Computer Assisted Personal Interviewing CAPI) and web. We define mixed-mode as using one survey instrument with two or more data collection modes (e.g., gathering data via paper/pencil and the web). By this definition, conducting multiple surveys of a population by different modes is not mixed-mode. That definition leads to the questions: When is it best to use a mixed-mode survey? What challenges are posed when fielding a survey using mixed-modes? The answer to the former relates to the survey instrument, coverage, access, and respondent preferences. The latter to formatting the instrument and the introduction of bias in one mode that is not present in the other mode (e.g., satisficing or social desirability). In this white paper we review what a mixed-mode method survey means, when it is best to use that approach, the challenges posed in conducting a mixed-mode survey, the potential benefits of a mixed-mode survey, and DRC s experience in conducting mixed-mode surveys. 1

When should you use a mixed-mode approach? There is no general answer other than letting the potential survey instrument dictate how the survey is best conducted. As a result, there are instances when the length and complexity of the survey dictate only using one mode. Specifically, it would not be advisable to conduct adaptive conjoint or complex interviews on paper. The complicated skip patterns are best suited for an interviewer conducting either CATI or CAPI or possibly a web mode that leverages available analytic software. Cost and time would dictate which one of these methods you would select. Similarly, there are instances when the mode selected requires a mixed mode approach. For a general population web survey, there is potential bias related to coverage issues. As of 2010 Internet access at home was limited to 71% of the population and this percentage varies by both age and race (Smith, 2010). To correct for this differential in access, potential respondents without Internet access must be offered an alternative means of responding (e.g., paper/pencil); making the survey dual mode by design. If however you had a closed population, one where all the potential respondents are identified including email addresses (e.g., college students or employees of a company), a single mode web-based survey might be the most appropriate choice. In addition to theoretic reasons for using either a single or mixed-mode approach, there are the pragmatic reasons of cost and time. To boost response rates and keep the cost of survey reasonable, it might be decided to follow up with respondents of a paper/pencil survey with CATI. By conducting the mail response first, you limit the number of calls that need to be made. What are the challenges? The challenges relate to not introducing bias in the results by using different modes for gathering the data; thus mode effects become an important design consideration with the goal of minimizing their impact. The literature on mixed-modes primarily focuses on differences in paper/pencil versus the web; probably because these two modes deliver the survey instrument via visual stimuli. In 2000, Young et al. found virtually no differences between paper/pencil and web modes of a survey. Church (2001) went beyond the paper/web comparision and also considered another respondent-administered mode (IVR). He found that two dual-mode surveys, one utilizing paper-and-pencil and web and the other using paper-and-pencil and IVR, showed a small percentage of unique variance in the data when comparing modes. Church recommends that with equivalent data quality results across mode, one should choose the methodology based on the project parameters (e.g., culture fit or ease of implementation). A mixed mode study that used both self-administered and interviewer administered questions would be prone to bias being introduced from the use of an interviewer. Dillman, et al. (2009) reported that for the satisfactiondissatisfaction questions respondents to the aural modes are significantly more likely than are respondents to the visual modes to give extreme positive responses, a difference that cannot be accounted for by a tendency toward recency effects with the telephone. This type of effect is known as social desirability bias (i.e., providing a positive response when there is somebody asking the question). Chang and Krosnick (2009) compared an internet survey version with a telephone interview and found that the telephone data manifested more random measurement error, more survey satisficing, and more social desirability response bias than did the Internet data. Kraut, Oltrogge, and Block (1998) observed social 2

desirability bias when they found the telephone mode of an Employee Opinion survey to have significantly more favorable results than the paper/pencil mode of the survey. These studies indicate a measurement effect that is present when questions are asked by an interviewer as opposed to self-administered. Social desirability bias, can be minimized to the extent that questions are factual or by using the phone as a mode without an interviewer (IVR). The latter is substantiated by Knapp and Kirk (2003) who found no significant differences in responses to sensitive questions between paper/pencil, web, and automated touchtone phone response modes. As noted, mixed methods might be used to increase response rate. Dillman, et al. (2009) utilized different combinations of modes for the initial phase compared to the follow up to non respondents (second phase). Phase 1 data collection was conducted by telephone interview, mail, interactive voice response, and the Internet, while Phase 2 consisted of nonrespondents to Phase 1, and was conducted by telephone or mail. In general, results from [the] study suggest that switching to a second mode is an effective means of improving response. If conducting a mixed-mode survey, should there be different surveys for each? Doing so could introduce response bias associated with the survey instrument. Dillman and Christian (2003) recommend using unimode construction where the instrument is written for the more restrictive environment. This is the practice employed by DRC. If conducting a survey via paper/pencil and using CATI to increase the response rate, the latter would be considered the more restrictive environment. CATI being administered aurally requires that questions be fairly short and answer options brief so that respondents can remember both the question and answers as they respond. A paper/pencil survey can have longer questions and answer options, because the respondent can go back and review them. Though not a challenge directly related to mixed-mode, many organizations are considering changing the mode they use to conduct a survey. If longitudinal trends are important, this can introduce bias. If a test of both modes is not made, then at least all reports should indicate when the change occurred. What are the benefits of conducting a mixed-mode survey? As noted earlier there are pragmatic benefits, usually cost and time. A mail survey with CATI follow up may increase response rate and it will be much less expensive than conducting the entire survey via CATI. De Leeuw (2005) notes Survey designers choose a mixed-mode approach because mixing modes gives an opportunity to compensate for the weaknesses of each individual mode at affordable cost. Offering a paper survey option with a web survey may help to alleviate some concerns about web surveys, such as implementation issues (bandwidth), confidentiality, perceived versus actual accessibility, respondent burden, tolerance level with technological glitches, perceived versus actual cost savings, data representativeness and accuracy, and lower response rate (Church, 2001). This is not meant to imply that there are not benefits in conducting a web survey. The benefits of include: a quicker turnaround time; lower item nonresponse; use of longer open-ended questions (Kwak & Radler, 2002); the time and cost savings associated with not printing/mailing; and less transcription or scanning errors because the data are returned in an electronic format (Kaplowitz, Hadlock, & Levine, 2004). As noted above, there is some evidence that a mixedmode approach does increase response rates. This has been demonstrated in general population surveys, special 3

populations, among professionals, and in establishment surveys (de Leeuw, 2005). A mixed mode approach also demonstrates to respondents that you respect their time and are offering them options that may allow them to respond when it is most convenient for them. This is particularly relevant for branded surveys (customer satisfaction or loyalty). If you are using technology in innovative ways (e.g., IVR or web surveys), clients will view your approach as leading edge but by also offering a traditional option they see that you are attempting to appeal to a spectrum of respondents. Increasing the response rate has two benefits. It reduces the likelihood of nonresponse bias and it decreases the standard errors, if you are weighting the data. The latter occurs because when making population estimates there has to be a weight accounting for nonresponse, and a larger weight usually means larger standard errors. Opinion Research (AAPOR) Annual Conference (Randolf, 2001). And as a supplement to this white paper, an additional DRC white paper is available with the specific focus on comparing paper-and-pencil and web survey modes (Fenlason & Christianson DeMay, 2002). For additional information on this topic or working with DRC s professional staff on your mixed-mode survey please contact Jon Leinen, Vice President, Business Development, Survey Services group by email (JLeinen@DataRecognitionCorp.com) or by telephone (763.268.2542) Overall, the benefits of using a mixed-mode approach, when feasible, are many. The challenges, with proper attention and expertise can be overcome or minimized. DRC s experience in conducting mixed-mode surveys DRC has a long history of providing mixed-mode expertise on a variety of survey types (e.g., 360 feedback, employee opinion, customer satisfaction, parent satisfaction with school services) for both commercial and federal clients. As thought leaders on this topic, DRC has presented or supported several multi-mode research papers at the following conferences: The Society for Industrial and Organizational Psychology (SIOP) Annual Conference (Christianson DeMay & Armstrong, 2000; Christianson DeMay & Toquam-Hatten, 2001), The International Military Testing Association Annual Conference (Elig & Waller, 2001; Quigley, Riemer, Cruzen, & Rosen, 2000), and the American Association for Public 4

References (Those in bold are written or supported by DRC staff) Chang, L., & Krosnick, J. A. (2009). National surveys via RDD telephone interviewing versus the internet: Comparing sample representativeness and response quality. Public Opinion Quarterly, 73(4), 641-678. Christianson DeMay, C., & Armstrong, S. (2000, April). Dual process surveying: A technical perspective. In W. H. Macey (chair), Making the move to internet-based surveys: Lessons from experience. Practitioner Forum presented at the annual conference of the Society for Industrial and Organizational Psychology, New Orleans, LA. Christianson DeMay, C., & Toquam-Hatten, J. (2001). 360-degree assessment at The Boeing Company: A dual-process (web/paper) approach. In M. M. Harris (chair), The internet and I-O Psychology: Applications and issues. Practitioner Forum presentation at the 16th annual meeting of the Industrial and Organizational Psychology, San Diego, CA, April 2001. Church, A. H. (2001). Is there a method to our madness? The impact of data collection methodology on organizational survey results. Personnel Psychology, 54, 937-969. Couper, M.P. (2011). The future modes of data collection. Public Opinion Quarterly, 75(5), 889-908. de Leeuw, E. D. (2005). To mix or not to mix data collection modes in surveys. Journal of Official Statistics, 21(2), 233-255. Dickinson, J. R., Faria, A. J., & Friesen, D. (1994). Live vs. automated telephone interviewing: Study finds automated interviews yield comparable results for less money. Marketing Research, 6(1), 28-34. Dillman, D. A., & Christian, L. M. (2003). Survey mode as a source of instability in responses across surveys. Retrieved 01/13/05 from http://survey.sesrc.wsu. edu/dillman/papers.htm Dillman, D. A., Phelps, G., Tortora, R., Swift, K., Kohrell, J., Berck, J., & Messer, B. L. (2009). Response rate and measurement differences in mixed-mode surveys using mail, telephone, interactive voice response, and the internet. Social Science Research, 38(1), 1-18. Dillman D.A. (2010) Making New Data Collection Modes Effective in Household Surveys. Presented at the Workshop on the future of Federal Household Surveys, November 4-5. Washington, DC. Elig, T., & Waller, V. (2001). Internet versus paper survey administration: Impact on qualitative responses. Paper presented at the 43rd Annual Conference of The International Military Testing Association, Canberra, Australia. Fenlason, K. J. & Christianson DeMay, C. (2002). Paper-and-pencil and web surveys: Different methods, same results? Published electronically on http://www. datarecognitioncorp.com/survey/websur.htmlhttp:// www.datarecognitioncorp.com/survey/websur.html. Kaplowitz, M. D., Hadlock, T. D., & Levine, R. (2004) A comparison of web and mail survey response rates. Public Opinion Quarterly, 68, 94-101. Knapp, H., & Kirk, S. A. (2003). Using pencil and paper, Internet and touch-tone phones for self-administered surveys: Does methodology matter? Computers in Human Behavior, 19, 117-134. 5

Kraut, A. I., Oltrogge, C. G., & Block, C. J. (1998, April). Written vs. telephone surveys of employees: Are the data really comparable? Paper presented at the 13th annual conference of the Society for Industrial and Organizational Psychology, Dallas, TX. Kwak, N., & Radler, B. (2002). A comparison between mail and web surveys: Response pattern, respondent profile, and data quality. Journal of Official Statistics, 18(2), 257-273. Quigley, B., Riemer, R. A., Cruzen, D. E., & Rosen, S. (2000). Internet versus paper survey administration: Preliminary findings on response rates. Paper presented at the 42nd Annual Conference of the International Military Testing Association, Edinburgh, U.K. Randolph, J. (2001). Internet versus paper survey administration: Impact on qualitative responses. Presented at the American Association for Public Opinion Research Annual Conference, Montreal, Quebec. Smith, A. (2010). Home Broadband 2010. Pew Research Center s Internet & American Life Project. August 11, 2010. Washington DC. Young, S. A., Daum, D. L., Robie, C., & Macey, W. H. (2000, April). Paper versus Web survey administration: Do different methods yield different results? Paper presented at the 15th annual conference of the Society for Industrial and Organizational Psychology, New Orleans, LA. 6

www.datarecognitioncorp.com