Eye Tracking on a Paper Survey: Implications for Design

Similar documents
Improving Government Websites and Surveys With Usability Testing and User Experience Research

Does In-Store Marketing Work? Effects of the Number and Position of Shelf Facings on Brand Attention and Evaluation at the Point of Purchase

Guidelines for Using the Retrospective Think Aloud Protocol with Eye Tracking

Keeping an eye on recruiter behavior

Usability Testing Jeliot 3- Program Visualization Tool: Evaluation Using Eye-Movement Tracking

white paper EYETRACKING STUDY REPORT: Clamshells vs Paperboard Boxes CUshop Research, Clemson University

New Findings on Questionnaire Design from 2012 AAPOR Conference Papers AAPOR conference recap mini-conference 6/22/2012 San Francisco, CA Su Li

MEASURING THE TRUE VALUE OF ADVERTISING IN PRINT AGAINST ONLINE - AN EYE TRACKING EXPERIMENT

Eye-Tracking Methodology and Applications in Consumer Research 1

Google Search Mapped Results: Eye Tracking Insights

Data Collection and Sampling OPRE 6301

Online Media Research. Peter Diem/Vienna

USING EYE TRACKING TECHNOLOGY FOR WEB SITE USABILITY ANALYSIS: THE APPLICATION OF ERICA TO GEFANUC.COM

Understanding Consumers Quality Evaluation of Online Health Information: Using a Mixed-Method Approach

Tracking translation process: The impact of experience and training

USABILITY TESTING OF DEPARTMENTAL WEB SITES: A CASE STUDY WITH AUTHENTIC USERS AND AUTHENTIC TASKS

Differences in Length of Survey Administration between Spanish-Language and English-Language Survey Respondents

Eye tracking in usability research: What users really see

Keywords Banner Ad Position; Congruence; Advertising Objective; Banner Ad Fixation; Brand Awareness; Product Knowledge

SUPERIOR EYE TRACKING TECHNOLOGY. Totally Free Head Motion Unmatched Accuracy State-Of-The-Art Analysis Software.

Measuring success on Facebook

Usability Testing. Credit: Slides adapted from H. Kiewe

What TV Ratings Really Mean

Tue 5.1 Shoppers Attention to Packaging and In-Store Media

GIM Capital Goods / B2B. Heidelberg, April 2015

Using Software to Collect Data Electronically for the Economic Census

Keywords: time use, nonresponse

Best Practices in Data Visualizations. Vihao Pham January 29, 2014

Best Practices in Data Visualizations. Vihao Pham 2014

GAZETRACKERrM: SOFTWARE DESIGNED TO FACILITATE EYE MOVEMENT ANALYSIS

Web Site Design Preferences of Middle School Youth Sarita Nair, Project Director Jennifer Peace, Research Associate. Introduction

Using Surveys for Data Collection in Continuous Improvement

Research Overview: Telephone versus Online Research Advantages and Pitfalls

STUDENTS PERCEPTIONS OF ONLINE LEARNING AND INSTRUCTIONAL TOOLS: A QUALITATIVE STUDY OF UNDERGRADUATE STUDENTS USE OF ONLINE TOOLS

Metrics by design A practical approach to measuring internal audit performance

Eye-tracking. Benjamin Noël

Evaluating Survey Questions

Eye Tracking in User Experience Design

Billing Code: P DEPARTMENT OF HEALTH AND HUMAN SERVICES. Agency for Toxic Substances and Disease Registry. [30Day ]

Effect of Using Human Images in Product Presentation of E-Commerce Website on Trust, Fixation and Purchase Intention: A Design of Experiment

Comparing Recommendations Made by Online Systems and Friends

How to Write a Great Marketing Offer Test

Visual Design, Order Effects, and Respondent Characteristics in a Self-Administered Survey

The Effect of Interactive Program Loyalty Banners on Television Advertising Avoidance

Towards Inferring Web Page Relevance An Eye-Tracking Study

Jon A. Krosnick and LinChiat Chang, Ohio State University. April, Introduction

National Survey of Franchisees 2015

Marketing Dept. to guide a study towards its objectives) marketing problems / research objectives? expected, but is not sure about the reasons.

One Size Doesn t Fit All: How a Segmentation Approach Can Help Guide CE Product Strategy


MOBILE EYETRACKING IN MARKETING

Winning Marketing Claims

MCC Branding Overview and Survey Results. Meeting of the Board of Trustees Thursday, July 23, 2015

Principles of Measuring Advertising Effectiveness

05/10/2015. Chapter 3 - Marketing Research. Marketing Project Plan

Choosing Colors for Data Visualization Maureen Stone January 17, 2006

GO SOCIAL: SEEING SEARCH. An Eye Tracking Analysis on Social Networking Sites

Chapter 2: Research Methodology

The Student Experience of Student Loan Counseling

Questionnaire Testing for Business Surveys at Statistics Canada: Practices and Considerations

How do I: Conduct Market Research?

Best practices for effective marketing. Inboxes are crowded places. Here are some proven tips to help your message rise to the top.

Applying Knowledge of Business Survey Response Processes to the Design of Data Collection Software at the U.S. Census Bureau

HMRC Tax Credits Error and Fraud Additional Capacity Trial. Customer Experience Survey Report on Findings. HM Revenue and Customs Research Report 306

Interpretive Report of WMS IV Testing

Simple Predictive Analytics Curtis Seare

Using Eye Tracking to Compare Web Page Designs: A Case Study

KNOWLEDGEPANEL DESIGN SUMMARY

Evaluation of Sensors as Input Devices for Computer Music Interfaces

STUDY SERIES (Survey Methodology # )

A Picture Is Worth a Thousand Words: Results from Cognitive Testing of Graphic Show Cards

Software Customer Satisfaction

The American Time Use Survey: Test of Coding Procedures and Lexicon

White Paper Combining Attitudinal Data and Behavioral Data for Meaningful Analysis

Content Marketing in 2014:

Question Bank FACTSHEET 2. Methods of Data Collection in Social Surveys

Survey Research: Designing an Instrument. Ann Skinner, MSW Johns Hopkins University

Usability Test Plan Docutek ERes v University of Texas Libraries

METHODOLOGIES FOR STUDIES OF PROGRAM VISUALIZATION

Visualizing Data from Government Census and Surveys: Plans for the Future

Chapter 14: Usability testing and field studies. The aims: Usability testing

Usability Testing for Beginners on a Tight Budget

Citizen participation has become an increasingly

How To Find Out How Fast A Car Is Going

Unresolved issues with the course, grades, or instructor, should be taken to the point of contact.

CSU, Fresno - Institutional Research, Assessment and Planning - Dmitri Rogulkin

Marketing Research Core Body Knowledge (MRCBOK ) Learning Objectives

71% of online adults now use video sharing sites

Numbers as pictures: Examples of data visualization from the Business Employment Dynamics program. October 2009

Outside the Answer Boxes: Messages from Respondents

Knowledge Visualization: A Comparative Study between Project Tube Maps and Gantt Charts

Usability Evaluation with Users CMPT 281

Function: Admissions Marketing. Function: Public Relations. ensuring the MVC brand is recognizable on all materials. Using the same logo, etc.

Revisiting the Survey Form: The Effects of Redesigning the Current Employment Statistics Survey s Iconic 1-Page Form with a Booklet Style Form

RESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 29 (2008) Indiana University

APPENDIX V METHODOLOGY OF 2011 MONTANA HEALTH INSURANCE SURVEYS

EcoPinion Consumer Survey No. 19 The Conflicted Consumer Landscape in the Utility Sector. May 2014

Direct Integration: Training Software Developers and Designers to Conduct Usability Evaluations

Online Course Delivery - A Santa Barbara City College Perspective

Market research. Chapter 4 Market research. The objectives of market research

Transcription:

Eye Tracking on a Paper Survey: Implications for Design Lauren Walton 1, Jennifer Romano Bergstrom 2, David Charles Hawkins 2, Christine Pierce 1 1 The Nielsen Company, Tampa, Florida {lauren.walton, christine.pierce}@nielsen.com 2 Fors Marsh Group, Arlington, Virginia {jbergstrom, dhawkins}@forsmarshgroup.com Abstract. Asking respondents to record their activity in a diary can be a difficult task due to retrospective reporting and cognitive burden as well as the complexity of the data collection tool. Diary questionnaires typically require multiple pieces of information including demographics, activities, and duration over a data collection period. Like other questionnaire types, visual design principles can be used to help people perceive and understand what is being asked of them during diary measurement. Eye tracking, a technology that allows us to passively study people s eye movements, has been used mostly for questionnaire testing within the survey research field. This study focuses on using eye tracking and other user experience measures to analyze how respondents perceive, understand and experience different designs of the paper Nielsen TV Diary. We used eye tracking to gain insights into visual elements that draw attention, the amount of text that respondents read (e.g., terms/instructions), and how respondents complete the survey. This paper centers on the collecting and analyzing of qualitative and quantitative measures of the user experience, including eye-tracking data (e.g., fixation count, time to fixate), participants verbalizations, self-reported satisfaction, and performance data (e.g., accuracy, steps to complete). We also provide recommendations about the design of the paper diary based on the user experience and eye-tracking results. Keywords: Eye tracking, survey, diary, visual design, usability 1 Introduction 1.1 Diary Research Survey researchers are continuously trying to understand how design decisions impact data quality and with it, total survey error [1]. Changes to a survey s design may affect the representativeness of the data collected because design may encourage or discourage people from responding. Each survey is unique, as are the visual stimuli given to respondents. The survey response process for self-administered surveys begins with the respondent s perception of the survey materials. Traditional pretesting methods like cognitive interviews, think-aloud interviews, and focus groups are de- adfa, p. 1, 2011. Springer-Verlag Berlin Heidelberg 2011

pendent on the participant being able to tell the researcher what they are thinking or feeling. Passive methods present data and allow researchers to make inferences without the data suffering from social desirability or recall bias. Eye tracking is a passive technology used to study people s eye movements. Recording eye movements allows researchers to infer where and for how long a participant s attention is directed. In market research, eye tracking has been used to study products that participants are drawn to while looking at grocery store shelves as well as what parts of websites people examine the most. Because people have difficulties recalling everything they look at or the order in which they examine objects, collecting eyetracking data removes the measurement error associated with asking respondents directly. Instead, eye-tracking data can show objectively what was looked at first and how long the eyes fixated a given element. Within the survey research literature, the library of eye-tracking research is relatively small. Eye tracking has been used to study how respondents perceive question stems, response options, and images within self-administered surveys. Current eye-tracking research mostly focuses on internet surveys and is limited to a few key papers (see [2], [3], and [4]). Redline and Lankford [2] found that participants did not read linearly, and they skipped around looking at the survey. Galesic, Tourangeau, Couper, and Conrad [4] used eye tracking as a method to understand how respondents perceive survey questions and found that participants spend more time looking at options on the top of a list of responses, and participants who took the time to read the entire response list also took the time to read the instructions and other parts of the questionnaire. Lenzner, Kaczmirek, and Galesic [5] examined if respondent comprehension is impacted by text features going beyond response time as an indicator and found that longer fixation time on question stems equate to comprehension issues. Libman and Smyth [6] looked at smiley faces as symbolic language and found that participants with low literacy levels moved faster to look at the smiley faces in the web survey. No research to date, to our knowledge, has used eye tracking to understand how respondents process and work through a paper diary survey and how much attention is given to various visual elements. This study aims to fill this gap. We used eye tracking to understand how participants perceived, processed and used various versions of a paper diary. 1.2 The Nielsen Company The Nielsen Company is the world s leading provider of global marketing information, consumer insights, business media products, and services. The mission of Nielsen is to provide clients with the most complete understanding of consumers and markets worldwide. Nielsen has measured the television audience of the United States for over 60 years producing the Nielsen TV Ratings.

Nielsen uses both electronic and paper measurement techniques to collect viewing data. The focus of this paper is the Nielsen TV Diary (shown in Figure 1), which is a week-long survey sent to randomly selected households across the United States, using an address based sampling approach, during data collection windows known as TV sweeps. Fig. 1. Nielsen TV Diary Cover Nielsen strives to improve the diary product through, including but not limited to, testing the optimization of household incentives, advanced sampling techniques, and improvements to survey visual design. In February 2012, Nielsen tested newly designed diary materials (Figure 2) that did not perform as well as anticipated. The new household communications utilized motivational messaging for hard to reach groups following best practices of organizations including the Hispanic / Latino Advisory Council, African American Advisory Council, U.S. Census Bureau, Scarborough Research, internal diary data analysis, and Nielsen internal best practices on recruitment and compliance for long-term panels. The communications were vetted via focus groups for confirmation with these demographic groups that the approach/messaging was appropriate, and for red flags with the non-targeted demographics. There was a high level of due diligence in the development of the new communications, but a significant decrease in cooperation was seen for these New household communications materials (16.7% vs. 18.4%, p.001). Fig. 2. Example of Old (left), Bilingual New (middle), and English New (right) household communication diary.

These new materials were used by Nielsen during February, May, and July, and were retested in November 2012. Figure 3 presents Nielsen Diary response rates over time showing a decline in response when these New materials were used. In order to understand these Diary materials further, Nielsen partnered with Fors Marsh Group to design an innovative study to examine these paper diaries using eye tracking. Results of this usability study are presented in this paper. Fig. 3. Nielsen TV Diary Response Rates. *Diary measurements where New household communications materials were used. 2 Eye-Tracking Usability Study Researchers used a combination of metrics to understand users experiences while working on the diary. To assess the behavioral impact of additional changes to the diary, a fourth Prototype diary was also tested. The Prototype had the same cover as the Old diary (Fig. 2), did not have any motivational language, and contained some alterations to the TV recording grid. 2.1 Participants and Procedure Seventy-four participants (29 male, 45 female) with an average age of 37 were recruited from the greater Washington DC area to take part in the usability study. Sessions took place in the Fors Marsh Group User Experience Laboratory, and each session lasted approximately one hour, during which time participants were randomly assigned to work with the Old diary being mailed to households, the New February 2012 household communications test paper diary, or the redesigned Prototype diary. Participants were instructed to work on the diary as if they were at home and to think aloud as they worked. The diary was mounted on a stand above a Tobii X2-60 eye tracker. Upon completion of the sessions, researchers identified Areas of Interest (AOIs) and examined eye-tracking patterns for these AOIs. AOIs included various visual compo-

nents of the diary that were hypothesized to increase user engagement (e.g., logo on cover, motivational language). In addition, we examined the order in which participants interacted with various components, such as reading the instructions and entering TV viewing. Attention to the Cover Images. We used Tobii Studio to create mean fixation count heat maps to examine attention to the covers of the different diaries (shown in Figure 4). We found that the Nielsen logo was fixated less often on the cover of the Bilingual New and English New diaries than on the Old diary. The logo on the Old diary is located in the middle of the page, while on the New diaries it is on the bottom left. Additionally, participants statements during debriefing interviews indicated that the television set on the New diaries was hard to identify specifically, many participants said they thought the modern flat screen TV on the cover was a picture frame. When shown the Old diary cover, participants said it was easier to see that the Old cover had a TV. Thus, possible explanations for the drop in response rates for the new diary are a loss of brand recognition and lack of recognition of the TV image. Both of these issues may have led to people being less inclined to respond. Fig. 4. Left to right: Old, Bilingual New, English New diaries. Mean fixation count heat maps show that fixations to the logo on the New covers were less frequent than on the Old cover. Motivational Language. Much time and effort was spent making sure that motivational language was relevant for targeted demographics. However, eye-tracking data of the New diaries showed that the motivational language was not regularly attended to. Figure 5 shows that the motivation language (in blue circles) was not fixated as participants completed various sections. These results suggest that the messages may have been overlooked in the fielded survey, and thus, the motivational language may not have had its intended effect because it was not seen or read.

Fig. 5. Mean fixation count heat maps show that the motivational language in the blue circles was not fixated. Processing the Diary Grid. It is important to test proposed design changes to evaluate whether or not the alterations will have a positive impact on performance. In this study, a redesigned Prototype diary (in addition to the Old and New ) was tested to examine a new layout of the viewing grid pages. We found that participants viewed and used the Prototype grid in a left-to-right progression, while attention to the Old Diary and New diary grids was less uniform and more scattered, as shown in the gaze plots in Figure 6. Eye-tracking data provided insight into the user experience of the column ordering. The first column in the Prototype grid asks for the name of the program; the first column in the New and Old diary grids asks for the channel number. Participants mentioned in the debriefing interviews that when recalling previous TV viewing, the name of the program was one the first things that came to mind. Results suggest that the Prototype grid was easier to interpret than the Old and New diary grids. Based on these findings, we concluded that the Prototype grid may be more successful in the field and may lead to increased accuracy and response rate.

Fig. 6. Participant gaze plots for the Old diary grid (left) and New diary grid (middle) are more scattered and do not follow a left-to-right pattern like in the Prototype diary grid (right). Examples are Overlooked In Step 3 of the diary, the directions instruct respondents to report the people living in their household. An example is provided to aid in the proper and thorough reporting of all household members and any potential guests. Participants reported that the example was helpful; however, because the example is located below the form, 45% of participants did not notice the example until after they had recorded (often incorrectly) all household members (see example gaze plots in Figure 7). Results from eye tracking informed researchers that on average, 55 fixations were made on the Step 3 pages before participants noticed the example. Part of the reason why it was so common for participants to miss the Step 3 example was because Steps 1 and 2 followed a linear progression. For example, in Steps 1 and 2, survey items were located on top of each other. Participants worked their way down the page as they responded to various demographic questions. The layout of instructions in Step 3 did not meet their expectations. Participants became used to the line-by-line progression and had no reason to suspect that for Step 3 they needed to examine the entire page before getting started.

Fig. 7. Example gaze plots demonstrating the linear progression in Step 1 (left image) and the lack of linear progression in Step 3 (right image) in which the participant did not see the example at the bottom of the page until after recording housemates at the top of the page. Satisfaction. Each participant only worked with one version of the diary. After working with the diary and recording up to three previously viewed TV shows, participants completed a seven-item satisfaction questionnaire. We conducted paired t-tests comparing satisfaction among the diaries and found no difference in satisfaction between the Old and New diaries and between the Old and Prototype diaries. However, two differences in satisfaction were discovered when comparing the New diary to the Prototype diary. Users of the Prototype diary (N = 24) rated the instructions as significantly more clear than users of the New diary (N = 28) (µ = 3.45, µ = 2.96, respectively on a scale of 1 5; p =.007). Additionally, the organization and location of instructions in the Prototype diary (N = 24) were rated as significantly more clear than in the New diary (N = 28) (µ = 3.3, µ = 3.03, respectively; p =.025). The participants perceptions of instruction clarity and organization may be related to the organization of the grid as mentioned, the Prototype diary grid promoted a greater degree of left-to-right processing and visual search than the other grids. 3 Conclusion Visual design in self-administered surveys is important because there is no interviewer present to clarify confusing information or to encourage response. Visual design has the potential to increase engagement; however it also has the capability of distracting and interfering with the survey and ultimately may affect data quality and response rates. In this study, we used eye tracking to understand the user experience of several versions of a self-administered diary, and specifically, we aimed to understand how participants perceived, processed and used the various versions of a paper diary. We aimed to understand differences in the user experience of the different ver-

sions to understand more about the decline in response rate that previously occurred in the field. Eye tracking and user feedback enabled us to learn that images (e.g., logo, TV) and motivational language did not lead to increased engagement, as had been intended in the New diary. Rather, participants did not pay attention to these elements and/or did not recognize them (e.g., the TV). These findings suggest that graphics and images should only be used when they supply meaning, and that including them to encourage motivation and engagement [7] is not always successful. We learned that participants process TV viewing information in a very specific way, and the Prototype diary TV viewing grid that was organized in a way that matched users mental models was processed more efficiently. Based on these findings, Nielsen is testing the Prototype diary grid design, mailing it to thousands of homes during the February 2014 diary sweep in the hopes of minimizing respondent burden while optimizing respondent reporting. We are looking forward to results in Spring 2014. References 1. Groves, R. M. & Lyberg, L.: Total survey error: Past, present, and future. Public Opinion Quarterly, 74(5), 849-879 (2010 2. Redline, C. D., & Lankford, C. P.: Eye-movement analysis: A new tool for evaluating the design of visually administered instruments (paper and web). In: Proceedings of the Section on Survey Research Methods, American Statistical Association. Paper presented at 2001 AAPOR Annual Conference, Montreal, Quebec, Canada (2001) 3. Graesser, A. C., Cai, Z., Louwerse, M.M., & Daniel, F.: Question understanding aid (quaid) a web facility that tests question comprehensibility. Public Opinion Quarterly, 70(1), 3-22 (2006) 4. Galesic, M., Tourangeau, R., Couper, M. P., & Conrad, F.G.: Eye-tracking data new insights on response order effects and other cognitive shortcuts in survey responding. Public Opinion Quarterly, 72(5), 892-913 (2008) 5. Lenzner, T., Kaczmirek, L., & Galesic, M.: Seeing through the eyes of the respondent: An eye-tracking study on survey question comprehension. International Journal of Public Opinion Research. 23(3), 361-373 (2011) 6. Libman, A. & Smyth, J.: Turn that frown up-side down: The use of smiley faces as symbolic language in self-administered surveys. Paper presented at 2012 AAPOR Annual Conference, Orlando, Florida, May (2012) 7. Manfreda, K. L., Batagelj, Z., & Vehovar, V. Design of web survey questionnaires: Three basic experiments. Journal of Computer-Mediated Communications. 7(3) (2002)