Errors in Operational Spreadsheets: A Review of the State of the Art



Similar documents
Decision Support Systems


White Paper May Spreadsheet-based planning in today s economy: Why you re up against the wall

Quality Control in Spreadsheets: A Software Engineering-Based Approach to Spreadsheet Development

Module 3. Ways of Finding Answers to Research Questions

WVU STUDENT EVALUATION OF INSTRUCTION REPORT OF RESULTS INTERPRETIVE GUIDE

CFSD 21 ST CENTURY SKILL RUBRIC CRITICAL & CREATIVE THINKING

Identifying and Managing Project Risk, Second Edition 2008 Tom Kendrick. The PERIL Database

Data Quality Assurance

Preventing Duplicate Payments for Lawson Customers

Measurement and Metrics Fundamentals. SE 350 Software Process & Product Quality

This Module goes from 9/30 (10/2) 10/13 (10/15) MODULE 3 QUALITATIVE RESEARCH TOPIC 1. **Essential Questions**

Tableau for Robotics: Storing Data

ANALYSIS OF HR METRICS for the Northern Non Profit Service Providers

The Lookup Technique to Replace Nested-IF Formulas in Spreadsheet Programming

A Guide To Writing Measurable Goals and Objectives For

This report provides the project s findings, including its three practical, actionable products:

Automating Spreadsheet Discovery & Risk Assessment

EQR PROTOCOL 4 VALIDATION OF ENCOUNTER DATA REPORTED BY THE MCO

Accurately and Efficiently Measuring Individual Account Credit Risk On Existing Portfolios

STOP THAT SUBVERSIVE SPREADSHEET!

Chapter 1: Data and Statistics GBS221, Class January 28, 2013 Notes Compiled by Nicolas C. Rouse, Instructor, Phoenix College

Using Excel for Statistics Tips and Warnings

The Use of Spreadsheets: Considerations for Section 404 of the Sarbanes-Oxley Act*

DATA AUDIT: Scope and Content

Risk Assessment For Spreadsheet Developments: Choosing Which Models to Audit

Exhibit memory of previously-learned materials by recalling facts, terms, basic concepts, and answers. Key Words

Chapter 17 Qualitative Data Analysis

Drawing a histogram using Excel

(Refer Slide Time: 01:52)

Department: Speech Pathology and Audiology. Number of students enrolled in the program in Fall, 2011: 75

ANOMALIES IN FORM 5500 FILINGS: LESSONS FROM SUPPLEMENTAL DATA FOR GROUP HEALTH PLAN FUNDING

Webb s Depth of Knowledge Guide

The importance of using the right tools

Quantrix & Excel: 3 Key Differences A QUANTRIX WHITE PAPER

Risk Management Studio:

If your course is part of a similar assessment effort in your program, you probably need to assess the following outcomes in your course:

Computer Science Teachers Association Analysis of High School Survey Data (Final Draft)

Prepared Remarks by Richard Cordray Director of the Consumer Financial Protection Bureau. Credit Reporting Field Hearing

White Paper from Global Process Innovation. Fourteen Metrics for a BPM Program

Pantone 376C. Risk and Reward: Are you still using Spreadsheets for Budgeting and Forecasting? Whitepaper.

Solvency II Data audit report guidance. March 2012

Basic Concepts in Research and Data Analysis

The extent to which the use of Accounting Software and Excel is incorporated in HE Accountancy Courses

Haulsey Engineering, Inc. Quality Management System (QMS) Table of Contents

Practical Implementation of Data Analysis. IIA OC Chapter Meeting February 18, 2014 Edwards Lifesciences Global Internal Audit

An Activity-Based Costing Assessment Task: Using an Excel Spreadsheet

Internal Audit. Audit of HRIS: A Human Resources Management Enabler

Measurement in ediscovery

Alignment of Taxonomies

Data Analysis, Statistics, and Probability

IBM SPSS Direct Marketing 23

Analyzing and interpreting data Evaluation resources from Wilder Research

EFFECTIVE ROOT CAUSE ANALYSIS AND CORRECTIVE ACTION PROCESS

IBM SPSS Direct Marketing 22

To create a histogram, you must organize the data in two columns on the worksheet. These columns must contain the following data:

Spreadsheet Auditing Software

Writing the Empirical Social Science Research Paper: A Guide for the Perplexed. Josh Pasek. University of Michigan.

Soft Skills Requirements in Software Architecture s Job: An Exploratory Study

Data entry and analysis Evaluation resources from Wilder Research

Lab 11: Budgeting with Excel

Non-Emergency Medical Transportation Reviews

APPLYING BENFORD'S LAW This PDF contains step-by-step instructions on how to apply Benford's law using Microsoft Excel, which is commonly used by

THE ANALYTIC HIERARCHY PROCESS (AHP)

Creating A Grade Sheet With Microsoft Excel

Spreadsheets have become the principal software application for teaching decision models in most business

Chapter 1 Introduction to the Study

Software Quality Assurance Plan. Introduction

Advanced Excel 10/20/2011 1

MS Health Care Administration ( )

In-Depth Guide Advanced Spreadsheet Techniques

CREATING LEARNING OUTCOMES

The purpose of Capacity and Availability Management (CAM) is to plan and monitor the effective provision of resources to support service requirements.

Software Solutions Appendix B. B.1 The R Software

Junk Research Pandemic in B2B Marketing: Skepticism Warranted When Evaluating Market Research Methods and Partners By Bret Starr

Bloom s Taxonomy. List the main characteristics of one of the main characters in a WANTED poster.

TRAINING NEEDS ANALYSIS

Software Engineering for Spreadsheets

This chapter provides information on the research methods of this thesis. The

Methods in Case Study Analysis

IMPROVING THE DESIGN-CONSTRUCTION INTERFACE

WHITE PAPER. The Five Fundamentals of a Successful FCR Program

White Paper Resource Management Across IT

Basic Data Analysis. Stephen Turnbull Business Administration and Public Policy Lecture 12: June 22, Abstract. Review session.

The Risks of Using Spreadsheets for Statistical Analysis

Finding and Evaluating Evidence: Systematic Reviews and Evidence-Based Practice

Predictor Coef StDev T P Constant X S = R-Sq = 0.0% R-Sq(adj) = 0.

ACCESS Importing and Exporting Data Files. Information Technology. MS Access 2007 Users Guide. IT Training & Development (818)

Steps to Data Analysis

Microsoft Office 2010: Access 2010, Excel 2010, Lync 2010 learning assets

Risk and Reward. Are You Still Using Spreadsheets for Budgeting and Forecasting? WHITEPAPER

Intended Use of the document: Teachers who are using standards based reporting in their classrooms.

Center for Effective Organizations

From Our Classroom Strategy Library During Reading

Self-checks and Controls in spreadsheets

ASSESSMENT CENTER FOR IDENTIFYING POTENTIAL PROJECT MANAGERS: A CHANCE FOR SYSTEMATIC HUMAN RESOURCE DEVELOPMENT

CHAPTER 1 INTRODUCTION

Almost all spreadsheet programs are based on a simple concept: the malleable matrix.

Transcription:

Errors in Operational Spreadsheets: A Review of the State of the Art Stephen G. Powell Tuck School of Business Dartmouth College sgp@dartmouth.edu Kenneth R. Baker Tuck School of Business Dartmouth College Kenneth.R.Baker@tuck.dartmouth.edu Barry Lawson Tuck School of Business Dartmouth College Lawson384@charter.net Abstract Spreadsheets are thought to be highly prone to errors and misuse. In some documented instances, spreadsheet errors have cost organizations millions of dollars. Given the importance of spreadsheets, little research has been done on how they are used in organizations. We review the existing state of understanding of spreadsheet errors, concentrating on two recent studies. One analyzes errors in 50 operational spreadsheets; the other studies the quantitative impacts of errors in 25 spreadsheets from five organizations. These studies suggest that counts of error cells are not sufficient to understand the problem of errors. Average error cell counts reported in the literature range from 1 percent to 5 percent depending on definitions and methods used. However, some errors are benign while others are fatal. Furthermore, spreadsheets in some organizations appear to be error-free. Several types of new research are needed to understand the spreadsheet error problem more fully. 1. Introduction Spreadsheets are the dominant desktop application in contemporary organizations. Millions of ordinary workers acting as end-user programmers design, build, and use spreadsheets every day. Whether these computer systems serve the organizations that use them in a positive way is therefore a question of enormous practical significance. Many authors and spreadsheet practitioners have warned about the ubiquity and dire consequences of errors in spreadsheets. A widely cited estimate [3] is that 5% of all formulas contain errors. But the most relevant question is not how common errors are but how significant they are to their users. After all, there is strong evidence that humans make frequent errors in many areas but few of these errors are consequential [9]. This paper describes the current state of the art in spreadsheet error research. After briefly summarizing the existing literature, we concentrate on two recent large-scale studies of errors and their impacts. While these studies alter our understanding of spreadsheet errors in some ways, they also highlight many areas in which our knowledge is lacking. The paper closes with a discussion of some promising lines of research. 2. Existing Literature We recently summarized the research literature on spreadsheet errors [5]. That paper organized the literature into five areas: categories of errors impact of errors frequency of errors creation and prevention of errors detection of errors We summarized the existing literature on categories of errors as follows: Classifications are generally offered without specifying the context or purpose for which the classification is intended. The existing classifications do not include sufficient examples of specific errors that satisfy each category. Classifications are generally not rigorously tested to demonstrate that multiple users can consistently classify actual errors into the proper categories. The boundary between quantitative errors and qualitative errors remains vague. A number of studies have attempted to measure the frequency of errors in spreadsheets. Panko [3] summarized the results of seven field audits in which operational spreadsheets were examined. His results show that 94% of spreadsheets have errors and that the average cell error rate (the ratio of cells with errors to all cells with formulas) is 5.2%. However, in our analysis of these studies we concluded that these estimates are unreliable given that the studies are not sufficiently documented or comparable. A better estimate of the cell error rate based on these studies would be 1.3% of cells, but only 13 spreadsheets can be included in this average (Note that corrected cell and spreadsheet error rates now appear on Pankos website [4]). The least studied aspect of spreadsheet usage is the quantitative impact that errors have on the results. What little evidence we have on the impact of errors 978-0-7695-3450-3/09 $25.00 2009 IEEE 1

in audited spreadsheets is largely anecdotal. For example, Hicks (cited in [3]) reported that the errors found in the audit of a single large spreadsheet caused the results to be off by 1.2%. Clermont [1] found errors in an average of 3.03% of cells in three large spreadsheets but reported that we did not find any tremendous erroneous result values that might have had severe negative effects on the company. Lukasic (personal communication cited in [3]) found a 16% error in the results of one of two spreadsheets audited. Panko s two interviewees [3] suggested that 5% of spreadsheets had serious errors. Another source of information on the impact of errors is the stories that have appeared in the press documenting losses due to mistakes involving spreadsheets. The European Spreadsheet Risks Interest Group (EUSPRIG) maintains a web page [2] that documents dozens of these cases. Many of these incidents involve errors in formulas, but others involve errors in the use of spreadsheets, such as in sorting, or even errors in interpreting results. These types of errors, while potentially serious, are beyond the scope of the current research. We can summarize the literature on the impacts of errors in spreadsheets as follows: No systematic studies have been reported that measure the quantitative impacts of errors in spreadsheets. Studies of the frequency of errors suggests around 1% of formulas are erroneous. Errors can be classified in many ways and no one classification has become accepted. 3. Errors in 50 Operational Spreadsheets As we have seen, the previous studies of spreadsheet errors have been small in scale and in many cases are not well documented. Several years ago we set out to study a large sample of operational spreadsheets using a well-tested and fullydocumented method [6, 7]. We stress that our subject was errors in completed, operational spreadsheets that had been in use for some time, not errors discovered in laboratory experiments. We used competent Excel users, not professional programmers, to conduct our study. And we did not have access to the spreadsheet developers. This study allowed us to identify a wide range of errors, but it had its limitations as well. One limitation was that we could only detect errors in the spreadsheets themselves, not errors in formulating the underlying problem or in the actual use of the spreadsheet. And, of course, we cannot guarantee that we found all the errors actually in these spreadsheets. Our sample of 50 spreadsheets came from a variety of organizations and from several websites. While not a strict random sample, ours is typical of the general population of spreadsheets. No one organization or application area contributed more than five spreadsheets to the sample. To assist us in analyzing these spreadsheets we used two software tools: XLAnalyst [11] and Spreadsheet Professional [10]. These software applications are add-ins to Excel that assist in auditing and building spreadsheets. XL Analyst evaluates 28 aspects of a spreadsheet, for example, formulas that evaluate to an error. Spreadsheet Professional includes tools for building, testing, analyzing, and using spreadsheets. One of its most useful features for our purposes is the mapping tool, which translates each worksheet in a workbook into a map showing which formulas have been copied. Our analysts were undergraduates, graduates, and recent alumni from our business or engineering schools. All of them had several years of experience with Excel, but none were professional programmers, even in Excel. Our training program took about 10 hours, which includes time to read the study procedure, analyze several sample workbooks, and receive feedback. The protocol we used involves eleven steps: 1. Run the two software tools XLAnalyst and Spreadsheet Professional. 2. Transfer selected results from the software tools to a data record sheet. 3. Record the purpose of the workbook and each worksheet. 4. Examine the workbook for use of Excel functions. 5. Review the results of XL Analyst and use them to locate errors. 6. Review the Spreadsheet Professional maps and use them to locate errors. 7. Review the Spreadsheet Professional calculation tests and use them to locate errors. 8. Review all formulas not already reviewed for errors. 9. Conduct sensitivity analyses to uncover errors. 10. Rate the workbook on various aspects of spreadsheet design (e.g., use of modules). 11. Record the total time taken by the audit and record comments on any special situations encountered. Since no generally-accepted taxonomy of spreadsheet errors exists we experimented with several before settling on the following six types of errors: Logic error - a formula is used incorrectly, leading to an incorrect result Reference error - a formula contains one or more incorrect references to other cells Hard-coding error - one or more numbers appear in formulas 2

Copy/Paste error - a formula is wrong due to inaccurate use of copy/paste Data input error - an incorrect data input is used Omission error - a formula is wrong because one or more of its input cells is blank. The senior researchers reviewed each workbook after the analysis was completed and made final judgments as to which errors would be accepted. together accounted for less than 5% of instances and cells. Not all of the 483 error instances we identified involved wrong results. In fact, roughly 40 percent of these were Hard-coding errors, in which a number was included in a formula rather than a cell reference; most of these did not produce wrong results. In order to determine how many errors actually gave rise to wrong results we categorized each instance as a Wrong Result or a Poor Practice. With this more restrictive definition, we found that 7 of the 50 spreadsheet workbooks we audited were error-free, for a spreadsheet error rate of 86 percent. Table 1 Errors categorized by type Error Type Instances Cells Hard-coding 182 (37.7% of total) 2,111 (43.5% of total) Reference 159 (32.9%) 1,074 (22.1%) Logic 106 (21.9%) 1,389 (28.6%) Copy/Paste 17 (3.4%) 206 (4.2% Omission 13 (2.7%) 65 (1.3%) Data Input 6 (1.2%) 10 (0.2%) Total 483 4,855 Although our inter-rater reliability was 92 percent, we found that errors come in hundreds of subtle variations that require thought and debate to categorize correctly. This experience deepened our skepticism of previous studies of errors in which the taxonomy of errors and the procedures used to find them were not spelled out. Of the 50 spreadsheets we studied, three were found to have no errors of the types included in our protocol. This gives a spreadsheet error rate (percent of spreadsheets with at least one error) of 94 percent, which is the same as Panko s estimate [3]. In the 47 spreadsheets with at least one error we found a total of 483 error instances involving 4,855 cells. Thus each error instance involves on average 10 cells. The average cell error rate implied by these results is 1.79 percent of all formula cells, considerably lower than the 5.2 percent error rate given by Panko [3]. Table 1 shows how the error instances and error cells were distributed by error type. Hard-coding errors (placing numbers in cells) were the most common (37.7% of instances and 43.5% of cells), followed by Reference errors (32.9% of instances and 22.1% of cells) and Logic errors (21.9% of instances and 28.6% of cells). The remaining three categories, Copy/Paste, Omission, and Data Input errors, distribution of cell error rates using the more restrictive definition. In this case 68 percent had cell In the remaining 43 spreadsheets we found a total of 281 instances of Wrong Results involving a total of 2,353 cells. Using this definition, the average cell error rate over all 270,722 formulas audited was 0.87 percent. This is, of course, substantially lower than the 1.79 percent cell error rate we estimated using our more inclusive definition of error, and much less than Panko s estimate of 5.2 percent. Our results suggest that the average cell error rate across all spreadsheets is in the range of 1 percent when we use a restrictive definition and 2 percent when we use an inclusive definition. But the average error rate does not tell us how the error rate varies across spreadsheets. The received wisdom is that errors are unavoidable in spreadsheets, as they are in most facets of life. This suggests that we should expect most spreadsheets to have a cell error rate close to the average. Figure 1 shows the distribution of cell error rates using our original, inclusive definition. Using that definition, 46 percent of our sample spreadsheets had error rates below 2 percent; 70 percent had error rates below 5 percent. However, several spreadsheets had error rates above 10 percent, in fact one had a cell error rate of 28 percent. Figure 2 shows the error rates below 2 percent and 90 percent were below 5 percent. The highest single error rate was 17. 3

Figure 1 Distribution of cell error rates across spreadsheets Number of Spreadsheets 16 14 12 10 8 6 4 2 0 0.00% 5.00% 10.00% 15.00% 20.00% 25.00% 30.00% Cell Error Rate Figure 2 Distribution of cell error rates across spreadsheets (wrong results) Number of Spreadsheets 20 18 16 14 12 10 8 6 4 2 0 0.00% 5.00% 10.00% 15.00% 20.00% Cell Error Rate - Wrong Results This study leads to the following tentative conclusions: Moderately-skilled Excel users can successfully analyze operational spreadsheets for errors with minimal training when provided with an explicit protocol to follow. Cell error rates average 1 or 2 percent, depending on the definition of error. 4

The distribution of error rates across spreadsheets is highly skewed, with many falling below the mean and a few far above the mean. 4. Impacts of Errors in 25 Spreadsheets While the previous study revealed new and interesting patterns in the prevalence of errors in spreadsheets, it did not address one critical question: do errors have a substantial quantitative impact on the results? As our review of previous studies showed [5], this is a question that has received almost no attention in the past. Accordingly, we undertook a new study to estimate the quantitative impacts of errors [8]. Our sample in this study consisted of 25 spreadsheets, five each from organizations that allowed us access to their spreadsheet developers. This included two consulting companies, a large financial services firm, a manufacturing company, and an educational institution. These organizations identified five volunteers, each of whom provided one spreadsheet. We provided the following specifications to the volunteers for help in choosing a spreadsheet: -contains 3-10 worksheets -contains 250-10,000 formulas -occupies 200-1,000 kb of memory -has been in use for no more than a year -contains no complex Visual Basic code -is well understood by the developer -has no broken external links Not all the spreadsheets we audited conform to these specifications. In fact, the average number of sheets in our sample was 15.2 (the range was from 2 to 135 sheets) and the average size in kilobytes was 1,463 (the range was 45 to 7,450kb). Many of the spreadsheets in our sample were larger on one or more of the dimensions than specified above. We used the same error taxonomy as in the previous study and the same general auditing protocol, although we eliminated many of the data gathering steps from the previous study as they were not necessary for our purposes. Determining the quantitative impact of an error in a spreadsheet is not as straightforward as it might appear. First, some errors occupy a single cell while others occupy many cells. Do we consider each cell as a separate error and measure the impact of correcting it alone, or do we correct all the cells with a similar error and measure the overall impact? Second, some error cells influence many other cells while others impact no other cells. When a cell impacts many other cells it is not always obvious which of the impacted cells to use to measure the effect. Third, it is not always clear how to correct an error. For example, if erroneous inputs were used do we replace them with average inputs or extreme inputs? Finally, it is necessary to decide whether to measure errors in absolute or relative terms, and whether to combine all the errors in a given workbook into one overall error or to treat them separately. We chose to measure the effect of each error separately and then to report the maximum error in a workbook. In most cases we corrected all the cells with a given type of error, treating this as one error with a single (overall) impact. When such an error impacted only the erroneous cells themselves, we computed the maximum change from the base case and took that as our error estimate. When such an error impacted a single dependent cell, we measured the impact of correcting all the error cells on that one cell. When an error cell had many dependent cells, we identified the most important dependent cell and measured the impact of correcting the error on that cell. Obviously, the magnitude of the error impacts we identified are dependent to some extent on the measurement system used, although we believe our choices are justifiable and our results fairly reflect the quantitative accuracy of the workbooks tested. Table 2 summarizes our results. In column 1 we have used a two-digit code to label each spreadsheet. For example, spreadsheet 3.4 is the fourth spreadsheet from organization 3. The table gives the following information for each spreadsheet: number of issues raised in our audits number of errors confirmed in interviews number of errors with non-zero quantitative impact maximum percentage impact maximum absolute impact Many different kinds of issues turned out not to be errors after we discussed them with the spreadsheet developers. For example, if we found one odd formula in a column of identical formulas we flagged it as an issue. Typically the developers would then tell us that this case was calculated differently from all the rest and that it was correct, although there was no such indication in the spreadsheet itself. Most of these cases, in fact, involved inconsistent formulas in a column or row. Errors with zero quantitative impact also arose in a variety of ways. Sometimes a formula referred to the wrong cell for an input, but the numerical value in that cell was the same as in the correct input cell. We categorized this case as an error, but one with zero impact. We recognize that this is a dangerous practice, which could lead to a nonzero error during 5

the use of the spreadsheet. However, for the purposes of this study we categorized it as an error with zero impact. Within this sample of 25 spreadsheets we identified a total of 381 issues. After we discussed these issues with the developers we found that nine spreadsheets had no errors; among the remaining 16 spreadsheets we found a total of 117 errors. Of these 177 errors, 47 had zero quantitative impact, leaving 70 errors with non-zero impact. Table 2 Errors and impacts in 25 spreadsheets Organization- Workbook Size (KB) # Issues # Errors Errors with No Impact Maximum Percentage Impact Maximum Absolute Impact 1.1 842 7 3 3 0.0% $0 1.2 3,291 50 6 1 28.8% $32,105,400 1.3 3,556 18 7 4 137.5% $110,543,305 1.4 1,376 4 1 1 0.0% $0 1.5 1,678 0 0 0 NA NA 2.1 3,180 19 6 1 3.6% $13,909,000 2.2 520 27 11 4 16.0% $74,000,000 2.3 394 6 0 0 NA NA 2.4 156 30 4 1 416.5% $10,650,000 2.5 1,734 40 2 0 NA $0 3.1 389 19 2 0 5.3% $238,720 3.2 177 1 1 1 0.0% $0 3.3 47 11 2 0 15.6% $4,930,000 3.4 990 6 1 1 0.0% $0 3.5 286 23 1 1 0.0% $0 4.1 606 27 22 10 116.7% $13,355,445 4.2 576 8 4 2 141.8% $272,000 4.3 1,571 0 0 0 NA NA 4.4 2,116 1 0 0 NA NA 4.5 1,932 79 44 17 39.1% $216,806 5.1 7,450 2 0 0 NA NA 5.2 415 2 0 0 NA NA 5.3 1,900 0 0 0 NA NA 5.4 309 0 0 0 NA NA 5.5 427 1 0 0 NA NA Totals 381 117 47 As we pointed out above, there are two ways to measure the impact of errors: absolute and relative. Absolute impacts are important because they tell us how large errors are in the units of the spreadsheet. However, they cannot be compared easily across workbooks, since a million dollar error may be trivial in one spreadsheet and catastrophic in another. Relative (or percentage) errors more accurately reflect the significance of an error, but they have their shortcomings as well. One problem with relative errors is that percentage changes cannot be determined when the initial value is zero; another is that percentage changes in percentages are not generally as meaningful as percentage changes in 6

dollar amounts. Accordingly, we present our results here in both absolute and relative terms. Recall that 47 of the errors we found had zero impact on the spreadsheet. This often came about Figure 3 Issues, errors and errors with impact, by organization 0 10 20 30 40 50 60 70 80 90 100 110 120 130 I II III IV V Issues Errors Errors with Non-Zero Impact when a formula had an erroneous reference, but both the erroneous and the correct input cells had the same value. Thus when the error was fixed the results did not change. Twelve of the errors involved percentages; among these the average absolute change was 22 percent. Twenty-four of the remaining 58 errors involved absolute errors less than $10,000. However, some errors were huge: the largest single absolute error we found was over $100 million! Our evidence suggests that spreadsheet practice is very different among the five organizations we studies. Figure 3 shows the distribution of issues, confirmed errors, and errors with non-zero impact in the five organizations we studied. In the five spreadsheets from Organization 5 we could identify only five issues to discuss with the developers and no errors were identified among those five issues. Organization 5 is a small consulting company with highly educated employees and a culture that demands excellence. The spreadsheets we analyzed from this firm were works of art: thoughtfully designed, well documented, easy to understand, and error free. Organization 4 had two spreadsheets with no errors and two with 22 and 44 errors, respectively. The quality of the spreadsheet practice in this organization clearly depends on just where one looks. In the case we found both the best of practice and the worse of practice in offices just a few miles apart. In Organization 3, which is another consulting company, all the spreadsheets we analyzed had errors but in three cases no error had a measurable impact on the results. Even in the remaining two spreadsheets the errors were few in number and fairly small in terms of impact. Organizations 1 and 2 are both very large. One is a financial firm and the other is a manufacturing firm. Some of the spreadsheets we analyzed from these companies were astonishingly large and complex. Perhaps for this reason, only two of the ten we analyzed were error-free (although four had no errors with impact). The quality of spreadsheet practice in both of these companies was inconsistent, with inadequate attention paid to spreadsheet design, simplicity, ease of use, documentation, and consistency. We can summarize our findings as follows: 7

Some organizations have spreadsheets that are essentially error-free. Within a single organization, spreadsheet practice can range from excellent to poor. Some organizations use spreadsheets that are rife with errors and some of these errors are of substantial magnitude. Many errors have zero impact, or impact unimportant calculations. There is little correlation between the importance of the application or the risk involved and the quality of the spreadsheet. Few spreadsheets contain errors that, in the eyes of their developers, would destroy their usefulness. 5. Directions for Research Of all desktop applications for business, the spreadsheet must rank second behind word processing as the most commonly used. While errors made in word processing may be embarrassing or annoying, they probably rarely lead to financial losses. By contrast, errors made in spreadsheets quite obviously can and do lead to such losses. It is therefore surprising how little research has been conducted on the subject of spreadsheet errors. While our research has introduced some new data and hypotheses into the literature, we believe its value lies in raising better questions rather than in settling them. Our studies suggest that more groundwork is needed to define useful categories of errors and to develop more effective methods for identifying errors in operational spreadsheets. We believe that current estimates of the average cell error rate are sufficient, since it is evident that error rates vary widely among spreadsheets and organizations. Much more work is needed to refine our understanding of how error rates vary across the universe of spreadsheets and on the impacts of errors. We suggest that four types of studies are needed at this point: Retrospective studies: study the impact that spreadsheet use had on a single decision in an organization Causal studies: analyze the determinants of spreadsheet quality at the individual and organizational levels. For example, how do training, support, and culture influence quality? Best practices studies: how did exemplary individuals and organizations develop their spreadsheet skills? Action studies: study the impact of a spreadsheet quality control program in an organization. 6. References [1] M. Clermont. A Spreadsheet Auditing Tool Evaluated in an Industrial Context, Proceedings of the European Spreadsheet Risks Interest Group Conference Cardiff Wales (2002) 35-46. [2] EUSPRIG: http://www.eusprig.org/stories.htm [3] Panko, R. (2005). What We Know About Spreadsheet Errors. http://panko.cba.hawaii.edu/ssr/mypapers/whatknow.htm, accessed September 2, 2006. [4] Panko, R. (2008). Errors in Field Audits of Operational and Preoperational Spreadsheets. http://panko.shidler.hawaii.edu/ssr/, accessed August 22, 2008. [5] S. Powell, K. Baker, and B. Lawson. A Critical Review of the Literature on Spreadsheet Errors, forthcoming Decision Support Systems, 2008a. [6] S. Powell, K. Baker, and B. Lawson. An Auditing Protocol for Spreadsheet Models, forthcoming Information & Management, 2008b. [7] S. Powell, K. Baker, and B. Lawson. Errors in Operational Spreadsheets, forthcoming Journal of End User and Organizational Computing, 2008c. [8] S. Powell, K. Baker, and B. Lawson. Impacts of Errors in Operational Spreadsheets, submitted to Decision Support Systems, 2008d. [9] J. Reason. Human Error, Cambridge University Press: Cambridge, UK (1990). [10] Spreadsheet Professional: http://www.spreadsheetinnovations.com/. [11] XLAnalyst: http://www.codematic.net/default.htm 8