Top 10 Customer Satisfaction Best Practices



Similar documents
WHITE PAPER. The Five Fundamentals of a Successful FCR Program

Satisfaction with Contact Centers Drives Customer Loyalty

Ask the Customer Experience Experts

How Omnichannel is Delivering on the Promise of Multichannel

MEASURING EMPLOYEE EXPERIENCE TO DRIVE POSITIVE EMPLOYEE ENGAGEMENT A FORESEE WHITE PAPER

The Evolving Role of Process Automation and the Customer Service Experience

Reducing Customer Churn

Using a Multichannel Strategy to Deliver an Exceptional Customer Experience

Uniphore Software Systems Contact: Website: 1

Schedule Adherence: WHITE PAPER. 7 Not- So- Best Practices and How to Make Them Better. Authored by: Rebecca Wise Girson Wise Workforce Strategies

7 Best Practices for Speech Analytics. Autonomy White Paper

How to Design an Employee Engagement Survey

NICE MULTI-CHANNEL INTERACTION ANALYTICS

Small Business Checkup

NINE WAYS TO GET YOUR SALESPEOPLE TO FOLLOW UP ON MORE LEADS by Mari Anne Vanella The Vanella Group, Inc

SUSTAINING COMPETITIVE DIFFERENTIATION

CONTACT CENTER SCHEDULE ADHERENCE

The Customer Experience:

10 Steps to a Multichannel Strategy and an Exceptional Customer Experience

A New Foundation For Customer Management

business brief How Mobility Transforms the Art of Customer Interaction jabra.com

How to choose the best CRM implementation partner for your call center

CONTACT CENTER REPORTING Start with the basics and build success.

GUIDE TO PURCHASING A PHONE SYSTEM FOR YOUR CALL CENTER

CONTACT CENTER 09: Five Steps to a Lean, Customer-Centric Service Organization

Speech Analytics: Best Practices for Analytics- Enabled Quality Assurance. Author: DMG Consulting LLC

Benchmarking in the contact center: Tips for managing what you measure

Unleashing your power through effective 360 feedback 1

How to improve service quality through enterprise feedback management?

Achieving High Performance: The Value of Benchmarking

UX analytics to make a good app awesome

Customer Experience Management

WHITE PAPER Speech Analytics for Identifying Agent Skill Gaps and Trainings

CX Trends Forecast: A Summary of the Top 10 Disrupters

Accenture Business Intelligence for Fashion and Luxury. Creating a Differentiated Customer Experience for Long-term Brand Loyalty

Designing and Implementing Your Communication s Dashboard: Lessons Learned

Seven ways to boost customer loyalty and profitability through an empowered contact center

Hotel Marketing Automation

Workforce Insights Employee Satisfaction Surveying

20 Customer Service Best Practices SELL. SERVICE. MARKET. SUCCEED.

Concur Customer Experience 2015 REPORT. Concur // Customer Experience 2015 Report

THE 8 STEP ROADMAP TO LEAD GENERATION

Turning an enriched customer experience into a valuable asset

Employee Surveys: Four Do s and Don ts. Alec Levenson

Target and Acquire the Multichannel Insurance Consumer

What s Trending in Analytics for the Consumer Packaged Goods Industry?

Setting smar ter sales per formance management goals

Location Analytics for Financial Services. An Esri White Paper October 2013

The Performance Marketer s Guide to Marketing: Engaging Your Subscribers

Call Quality Practices 2012

Introduction to Integrated Marketing: Lead Scoring

Leverage Customer Data 5 to Improve Customer Experience About Us 7. Whitepaper

Table of Contents. Introduction... 3 Post-Call Analytics vs. Real-Time Monitoring How Real-Time Monitoring Works... 4

20 CUSTOMER SERVICE BEST PRACTICES HOW TO DELIVER EXCEPTIONAL SERVICE

Creating Real Value for Your Customers and Your Organization in the Call Center

Customer Relationship Management - a strategic approach

Effectively Using Your Data for Customer Retention

TURN YOUR COMPANY S GOALS INTO AN ACTIONABLE PLAN

Creating and Monitoring Customer Satisfaction

Root Causes of Why Customer Experience Programs Fail. and what you can do to ensure your program is successful

RFM Analysis: The Key to Understanding Customer Buying Behavior

Its all about the Customer Experience CX

Best Practice in Customer Experience Management

9 Reasons Your Product Needs. Better Analytics. A Visual Guide

Measuring the Effectiveness of Your Content Marketing

Best Practices for Planning and Budgeting. A white paper prepared by PROPHIX Software October 2006

AnswerNow Guides Why Live Service Matters!

Best Practices for Monitoring: Reduce Outages and Downtime. Develop an effective monitoring strategy with the right metrics, processes and alerts.

Workforce Management: Controlling Costs, Delivering Results

Best practices for planning and budgeting. A white paper prepared by Prophix

Adobe Campaign Touchpoint Marketing Guide

Multi-channel Marketing

Improving Customer Contact Quality

37 Marketing Automation Best Practices David M. Raab Raab Associates Inc.

Measure the Value of Customer Experience Improvements

the big key criteria for a great SaaS company

Keeping up with the KPIs 10 steps to help identify and monitor key performance indicators for your business

TAKING CONTROL OF YOUR SHORT CODE. twilio.com

4.5% 2014 Digital Marketing Optimization Survey results > 4.5% Top lessons learned from the leaders

Which Online 360? A 10-step checklist

5 KEYS [ ] to Successfully Tracking Customer Experience. Are you delivering a top-notch customer experience that <keeps them coming back?

RTM Consulting. Practical Knowledge Management. The Keys to Customer Satisfaction. Randy Mysliviec CEO

HOW A CRM SOLUTION CAN HELP YOUR BUSINESS Zyprr E-Book Series. 1

Customer Success Platform Buyer s Guide

Transcription:

Interactive Intelligence, Inc. Top 10 Customer Satisfaction Best Practices For Contact Center IVR Surveys

Table of Contents Introduction... 3 1. Use the scientific questionnaire design... 4 2. Define the goal... 4 3. Keep IVR surveys short and sweet... 5 4. Measure what matters... 6 5. Use the right scale... 6 6. Don t strive for a benchmark find the rightmark... 7 7. Coordinate with IT... 8 8. Beware of using survey results to evaluate individuals... 8 9. Report often and make results accessible... 9 10. If you ve invested, make it work... 9 Take Action... 10 Copyright 2008 2012 CFI Group and Interactive Intelligence Group, Inc. All rights reserved. Brand and product names referred to in this document are the trademarks or registered trademarks of their respective companies. Interactive Intelligence Group, Inc. 7601 Interactive Way Indianapolis, Indiana 46278 Telephone/Fax (317) 872-3000 www.inin.com Update 8/12, version 2 Interactive Intelligence Group, Inc. 2 Best Practices for Contact Center IVR Surveys

Introduction Now more than ever, what customers think matters. Whether in response to economic uncertainty, competitive pressures, or board-level initiatives around customer experience management, companies are looking for ways to gauge customer perceptions to better retain valued customers. CFI Group s Top 10 Customer Satisfaction Best Practices for Contact Center IVR Surveys 1. Use a scientific questionnaire design 2. Define the goal 3. Keep surveys short and sweet 4. Measure what matters 5. Use the right scale 6. Don t strive for a benchmark find the rightmark 7. Coordinate with IT 8. Don t use survey results to evaluate individuals 9. Report often and make results accessible 10. If you ve invested, make it work The contact center, as the central hub for customer interactions, is optimally positioned to collect customer feedback at the point of opportunity. The best way to understand a customer s relationship with a company is to ask the customer directly in a survey. Moreover, it is important to ask these questions sooner rather than later with respect to the timing of the interaction a post-call survey. And, to make the survey even more powerful, survey data should be incorporated with operational data from the contact center easily achievable with an all-in-one communications platform. Advances in contact center technology have increased available post-call survey tools to capture customer feedback and gain insight into customer perceptions. However, technology alone does not assure effective post-call surveys. The makeup and delivery of post-call satisfaction surveys themselves can significantly influence the reliability (and usability) of survey data. Only by using the right technology in combination with survey best practices will contact centers be able to gain actionable information from the postcall customer satisfaction surveys they conduct. To that end, Interactive Intelligence is proud to sponsor this CFI Group whitepaper in conjunction with our Interaction Feedback module, an integrated application for the Interactive Intelligence Customer Interaction Center (CIC) IP application suite that allows contact centers to seamlessly create and manage automated post-call satisfaction surveys using the all-in-one CIC platform. International consulting firm CFI Group is a recognized thought leader in customer satisfaction analysis and founder of the American Customer Satisfaction Index (ACSI) published annually by the University of Michigan. CFI Group has worked with Interactive Intelligence to design built-in ACSI question templates for the Interaction Feedback solution that enable contact centers to author more effective surveys by applying industry-proven survey methodology. (Read more about the ACSI at www.theacsi.org.) Interactive Intelligence Group, Inc. 3 Best Practices for Contact Center IVR Surveys

1. Use the scientific questionnaire design You probably already have a very good idea of what drives customer satisfaction. Customer service professionals can generally create surveys that, on the surface, might closely resemble those created by survey methodologists. However, the words generally, might, and closely are crucial. Familiarity with the day-to-day needs of customers can well inform the content of a survey, but even the best intentions can have unintended consequences on question design. The nation s only uniform cross-industry measure of customer satisfaction is the American Customer Satisfaction Index (ACSI). CFI Group, the developers of the ACSI, has determined the precise questions, phrasing, and scales to accurately measure customer satisfaction. Rigorous scientific testing and decades of empirical evidence enable this methodology to link satisfaction to financial performance 1 you won t need to reinvent the wheel. Without training in the science (and art!) of survey design and writing, survey data will be unreliable and possibly invalid. For example, a common survey mistake is to ask: Did you find the agent knowledgeable and experienced? While knowledge and experience are important, by attempting to address both issues at once, this question fails to capture meaningful information for either. This error is called a double-barreled question; it should be two questions. Another complicated area of questionnaire design is response categories. If a survey asks how often you shop, what s wrong with these response options: Never, Occasionally, Sometimes, Often, and Regularly? The problem is that although these are natural answers, they have different meanings to different people. In the aggregate, sometimes is devoid of meaning. These response categories are too vague to be useful for analysis. A better (but perhaps counterintuitive) way to phrase response categories is on a mutually-exclusive, numerical scale, with the endpoints of a continuum anchored with text. 2. Define the goal Understanding how you will use survey results is critical to good survey design. Unfortunately, most surveys are fielded without clear research objectives. Setting objectives up front means thinking about how you want to analytically extract useful information from the data. It s very easy to ask for one type of information when you actually want another. Not thinking about this in the design phase may mean not being able to use the data once it has been collected to address the questions one wants to answer. For example, when word gets out that the contact center is conducting a survey, various departments often request input. Marketing would like to add a question about promotions. Product Development wants to ask about packaging. And the Sales department has an idea for up-selling at the end of the survey. Everyone has just one little question to add. Too often, the result is survey data that none of the stakeholders find useful. 1 Beating the market with Customer Satisfaction, Christopher W. Hart, Harvard Business Review, March 2007. Interactive Intelligence Group, Inc. 4 Best Practices for Contact Center IVR Surveys

Keep the survey focused on the goal. A clearly focused survey makes success more likely for several reasons. A focused survey supports the desired analysis, allowing for deeper coverage of critical topics rather than shallow coverage of many. But most importantly, it respects the customer s time and helps avoid respondent fatigue. If the objective is to improve the contact center experience, survey questions should be limited to those interactions. Each question will serve a specific purpose and capture information that can be acted on and analyzed. Keep surveys focused on the goal Support the desired analysis allow for deeper coverage of critical topics Respect the customer s time avoid respondent fatigue 3. Keep IVR surveys short and sweet The more questions there are, the lower the completion rate. This is particularly true in an IVR context, where there is limited commitment to participate. Interacting with a machine quickly becomes tiresome, and it s much easier to hang up on a machine than on a live interviewer. It s more useful to measure survey length in terms of time, rather than the number of questions. How questions are phrased has a huge impact on the time it takes to administer a survey. The goal is not completion rate per se, but to collect actionable information. While there are no hard and fast rules, post-call surveys of 2-3 minutes collect enough information to be useful, but not so much that respondents drop out. A common mistake is to ask customers to rate the importance of each item on the survey. The customer is first asked to rate the agent s courtesy, and then again to rate the importance of courtesy. Aside from doubling the length of the survey, it is a flawed measure of what drives customer behavior and does not help prioritize improvements. Interactive Intelligence Group, Inc. 5 Best Practices for Contact Center IVR Surveys

4. Measure what matters What customers say is important is often different from what actually drives their business. CFI Group reveals an extreme based on airline safety. If asked to rank characteristics of an airline by importance, every customer will agree that safety is the single most important function of an airline. However, safety is not the driving force behind airline choice. Using cause-and-effect analytics, CFI Group identifies what actually drives customers to choose a specific airline. Price, schedule, and loyalty programs have the most effect on ticket purchase though none of these are more important than safety. Safety is a requirement, the minimum requirement, a price of entry into the market. Safety is not the driving force behind ticket purchase, and knowing its importance doesn t help influence customer behavior. A far better way to understand the drivers of customer satisfaction is to empirically pinpoint the elements of the customer experience that drive behavior and predict the change in customer behavior that would result if these elements were improved. Cause-and-effect analytics will bring out the more salient implications of price, schedule, and loyalty programs and their impact on the airline customers choose. 5. Use the right scale In order for managers to execute on survey results, measurement must be precise. Many firms want to report customer satisfaction in percentage terms (85% of our customers are satisfied or our customer satisfaction is 85%). This is a costly mistake. In statistical terms, the problem with top-box, or percentages is that the resulting number has a large margin of error. Satisfaction (like intelligence) is not as simple as a binary satisfied or dissatisfied. Without a much greater investment in substantially larger sample sizes to reach the same level of precision as a mean score, the imprecision of percentages makes it impossible to compare across products, regions, outlets, etc. With an approach like this, valuable information is lost in favor of simplicity. How many scale points should there be in a customer satisfaction questionnaire? In order to use customer satisfaction data to make better business decisions, scales need granularity. A five-point scale isn t enough. Even a seven-point scale is questionable. This is where customer satisfaction becomes counterintuitive especially for those with statistical knowledge. Most things are normally distributed (and it makes statistical analysis a lot easier when they are), but customer satisfaction is not. It s only when dissatisfied customers have nowhere else to go that we find a normal distribution. 2 2 Claes Fornell, The Satisfied Customer: Winners and Losers in the Battle for Buyer Preference, New York, Palgrave Macmillan, 2007 Interactive Intelligence Group, Inc. 6 Best Practices for Contact Center IVR Surveys

Expected Change in Satisfaction The ACSI finds that a ten-point scale provides the precision and resolution necessary to link customer satisfaction to operations and financial performance. However, a tenpoint scale is not practical in the IVR context due to the potential for respondents to commit slow-finger coding errors (where respondents have difficulty typing 1-0 on the phone s keypad fast enough to be captured by the IVR system). A nine-point scale using the numbers 1 through 9 resolves this issue while still preserving much of the needed precision and resolution of the ACSI standard ten-point scale 3. Scores are then converted to a 100-point scale for reporting and benchmarking purposes. 6. Don t strive for a benchmark find the rightmark One of CFI Group s clients planned to build a new call center in order to meet some very aggressive targets for abandonment rates targets set by the industry s best-in-class. However, by using this client s customer satisfaction metrics to calculate optimal service levels, it soon became clear that for this individual company, the best-in-class target service level was overkill. Striving for the best-in-class benchmark would take valuable time, money, and effort, and this company s customers were not going to reward that level of service. An effective customer satisfaction methodology discerns profitable improvement initiatives from those that will divert resources from more important pursuits. Rightmarking is the integration of operational targets and customer satisfaction data to find the sweet spot where service and delivery converge. Knowing the exact point where service levels will most satisfy customers can prevent you from over- or underinvesting. The best-in-class goal didn t make sense for this client s customer base; it simply wasn t their rightmark. Knowing this, the client was able to reallocate these resources to service improvements that really mattered to customers and profitability. The rightmark is not the industry average, and not the average of the current best performers, but the target that makes your unique customer base happy. <4.0% Original Target Revised Target 6.6% 7.1% Initial Measure Setting a target too aggressively results in overspending. To reach original target, client would have had to build an additional contact center. Client opted not to make that investment. 9.77% 3 4 5 6 7 8 9 10 11 12 13 14 % Calls Abandoned while in Queue at Contact Center 3 While the use of a 0 to 9 scale would preserve the ten-point scaling of the ACSI, the placement of the 0 key on the telephone touchpad is awkward and leads to slow-finger coding errors as well. Interactive Intelligence Group, Inc. 7 Best Practices for Contact Center IVR Surveys

7. Coordinate with IT Check with your IT department before you take this whitepaper and start to plan for your glorious future. While the operational data, customer usage, and demographics information exists, oftentimes getting at it falls in the lap of IT, and they are going to need advanced warning and you will need their buy-in. Many company systems have been built up over time. As such, integration can be a challenge. This is the single biggest challenge to linkage modeling in our experience. Simple customer and representative IDs that allow one kind of data to be linked to another are quite often missing and take considerable IT investment to rectify. It s much better to involve IT up front in this process since success will be in teamwork with them. If you talk to IT before you talk to the world, you ll get a better sense of what is possible in your environment, and how much effort will be involved. 8. Beware of using survey results to evaluate individuals Organizations that use survey results to punish performance are setting themselves up for failure. When people are punished or rewarded based on customer survey responses, they will find a way to thwart the process. There is also a less obvious link back to scaling and precision. When top box percentages are used, the margin of error is often so large that it is not possible to discern good performance from bad performance. This opens the company up to legal action if compensation has been based on imprecise measurement. A colleague of mine called the help desk of a large cell phone provider (who shall remain nameless). At the end of the call, the agent told my colleague, You ll probably get a survey about this You ll probably get a survey about this please give me a 5, or I could get fired. please give me a 5, or I could get fired and remember, the survey is not about this company, it s about me. Needless to say, my colleague did not bother to participate. In this same vein, agents must not have the responsibility of transferring a customer to the survey. This must be done automatically by the IVR system for immediate postcall surveys. Feedback is most valuable when used in a positive way. Integrated tools like supervisory monitoring allow users to request real-time help, and supervisors can listen in and coach during an interaction. If the technology platform records surveyed calls, high-scoring agents can reveal best practices. A low-scoring agent can be monitored more closely and given additional training. In the future, coaching can be tailored to leverage survey findings and address the issues most important to providing a good customer experience. Interactive Intelligence Group, Inc. 8 Best Practices for Contact Center IVR Surveys

9. Report often and make results accessible Satisfaction data often lags the collection window by a month or more. But in the highvolume, fast paced environment of the call center, timing becomes even more important, and valuable information and customers can be lost in a short amount of time. The more quickly survey results are available in the call center, the greater the opportunity for action. Ideally, survey results are available immediately after the survey is completed. Additionally, the greater the option for dashboards, heads-up displays, and other types of real-time (or near-time) alerts, the better. Too often, satisfaction results are reported at too high of a level to allow any meaningful action by the people looking to use the information to make better decisions. It is useful for managers to have access to continuous reporting, and to be able to look at key segments of interest, such as reason for calling, issue resolution, time/date range, product supported, and customer type. Advanced reporting can provide insight into potential gaps between internal perceptions (how do we think we performed?) and external perceptions (how does the customer think we performed?) across key segments. This, coupled with satisfaction by agent and trend data, enables managers to improve contact center performance in a much more meaningful way. 10. If you ve invested, make it work Don t lose sight of why you want to understand customer satisfaction: to improve business. Satisfaction data is the roadmap to success. Left as a number in some annual report, it is an opportunity lost. You ve spent the time and money to collect the data, now put it to work for you! Interactive Intelligence Group, Inc. 9 Best Practices for Contact Center IVR Surveys

Take Action Since 1994 Interactive Intelligence has offered a fully integrated, standards-based all-inone software platform for multichannel business communications. Today, more than 4,500 organizations around the world utilize our platform and preintegrated application suites to manage interactions and customer information in the contact center, in the enterprise, and for the mobile workforce. Visit us at www.inin.com to learn more about our solutions including Interaction Feedback for complete customer feedback management in one intelligent application. CFI Group offers your contact center 20 years of experience using the world s most precise, reliable, and actionable science to turn voice of the customer" data into strategies for measurable results. No other survey and analysis methodology can capture the feedback you need and project the financial impacts of improvements and change. With one system, CFI Group delivers key performance metrics and finds the best opportunities for your business to increase customer satisfaction for the greatest return on investment. Use CFI Group to quantify the value of your initiatives and make sound, informed, strategic business decisions that protect your greatest asset: your customer. www.cfigroup.com Interactive Intelligence Group, Inc. 10 Best Practices for Contact Center IVR Surveys