Decision Support Systems
|
|
|
- Philip Davis
- 10 years ago
- Views:
Transcription
1 Decision Support Systems 46 (2008) Contents lists available at ScienceDirect Decision Support Systems journal homepage: A critical review of the literature on spreadsheet errors Stephen G. Powell, Kenneth R. Baker, Barry Lawson Tuck School of Business, Dartmouth College, Hanover, NH 03755, USA article info abstract Article history: Received 8 December 2006 Received in revised form 20 May 2008 Accepted 2 June 2008 Available online 7 June 2008 Keywords: Spreadsheets Error classification Decision support End-user computing Among those who study spreadsheet use, it is widely accepted that errors are prevalent in operational spreadsheets and that errors can lead to poor decisions and cost millions of dollars. However, relatively little is known about what types of errors actually occur, how they were created, how they can be detected, and how they can be avoided or minimized. This paper summarizes and critiques the research literature on spreadsheet errors from the viewpoint of a manager who wishes to improve operational spreadsheet quality. We also offer suggestions for future research directions that can improve the state of knowledge about spreadsheet errors and mitigate spreadsheet risks Elsevier B.V. All rights reserved. 1. Introduction The problem of eliminating errors from software has been around since the beginning of the computer era. The discipline of software engineering [38] arose out of a need for error-free software code. With the advent of the personal computer in the 1980s and the rapid rise of end-user computing, control of software development passed out of the hands of professionals and into the hands of millions of spreadsheet users, few of whom had any formal training for the task. As spreadsheets have diffused throughout business, evidence has accumulated that many spreadsheets contain errors [16,24,25,29] and that errors can be costly to the organizations that use them (European Spreadsheet Risks Interest Group: Nevertheless, end users and organizations that rely on spreadsheets have generally not recognized the risks of spreadsheet errors [21]. In fact, spreadsheets are somewhat ignored, both as corporate assets and as sources of risk. Although research has suggested that errors are prevalent in spreadsheets, there is much we don't know about the types of This work was performed under the sponsorship of the U.S. Department of Commerce, National Institute of Standards and Technology. Reproduction of this article, with the customary credit to the source, is permitted. Corresponding author. address: [email protected] (S.G. Powell). errors that occur, why they occur, and how to avoid them. We believe that a critical review of the relevant literature can inform future research on this topic. This paper provides such a review. Rather than give a chronological account of the literature on spreadsheet errors, we organize the discussion around the following topics: Classification: what types of errors occur? Impact: what are the consequences of errors? Frequency: how common are errors? Creation and prevention: how can we build trustworthy spreadsheets? Detection: how can we audit spreadsheets to correct errors when they occur? We devote a section of the paper to each of these topics in turn and conclude by suggesting guidelines for future research. 2. Types of errors Any classification system allows us to understand the commonalities among individual instances. The Linnaean system, for example, classifies living things into species, genera, families, and so on. This hierarchy allows us to infer that individuals in the same species are more alike than those in different species, and species in the same genera are more alike than those in different genera. An effective classification of spreadsheet errors would allow us to compare errors across /$ see front matter 2008 Elsevier B.V. All rights reserved. doi: /j.dss
2 S.G. Powell et al. / Decision Support Systems 46 (2008) studies and application areas. It would also allow us to diagnose the causes and prescribe the cures for errors. Spreadsheet errors can be placed into categories in a variety of ways. For example, errors could be distinguished by cause, effect, form, stage, or risk. Some examples of causes of errors are typing mistakes, copy/paste errors, and lack of knowledge. A second type of classification would be based on the effects of errors. Some errors lead to erroneous numerical results, while others may simply make the spreadsheet hard to use correctly. Errors could also be distinguished by their form. For example, one form of error is a mistaken formula; another is an input parameter that appears in multiple cells; a third (perhaps) is poor documentation. A fourth classification scheme would be based on the stage at which the error was created. Some errors first appear during the conceptual design of a spreadsheet, while others appear during the building process. Some do not appear until a spreadsheet is actually used. Finally, errors could be categorized by the risks they pose. In a multi-user spreadsheet, for example, failing to use cell protection to ensure that formulas are not overwritten would be a significant error. There is no single accepted classification scheme for spreadsheet errors. Perhaps that is unavoidable, because errors can be classified in different ways for different purposes. Any well-developed classification should have three characteristics. First, it should specify the purpose for which it was created and/or the context in which it is meant to be used. Second, each category should be clearly defined and examples provided that is, some errors that fall into the given category and some that do not. Third, the classification should be tested in the relevant context and evidence provided that different people classify errors consistently. Classifications are created for a purpose, and a suitable classificationforonepurposemaynotbesuitableforanother.in the spreadsheet realm, we can imagine that the classification a lab experimenter might find helpful would not be suitable for field audits. For example, in a lab environment, it might be possible to observe an incorrect formula being entered with a typing error. However, when discovered during a field audit, that same error might be indistinguishable from an error in logic. Second, categories should be clearly defined and illustrated with examples. Broad categories such as qualitative errors are difficult to define precisely, and subcategories such as constants in formulas are not well defined without specificexamples.for example, the following formula includes the constant 1, but few spreadsheet users would consider it erroneous: Sales 06 ¼ Sales 05T 1 þ growth rate Third, the classification should be tested and proven in its area of application. Thus, if a classification is intended for use in auditing operational spreadsheets, it should be tested on spreadsheets from the field. To pass this test, researchers must show that individuals can categorize actual errors successfully and that multiple researchers place the same error in the same category a high percentage of the time Early classifications The earliest studies concerned with errors either gave examples or listed overlapping types of errors. Brown and Gould [2], for example, observed 17 errors, of which 15 led to wrong numbers. Of these, 11 involved errors in formulas, three involved mistyping, and one involved a rounding error. The remaining two errors involved misuse of cell protection and an error in logic. Ronen et al. [36] cited eight error types mentioned in the practitioner literature: mistakes in logic incorrect ranges in formulas incorrect cell references confused range names incorrectly copied formulas incorrect use of formats and column widths accidentally overwritten formulae misuse of built-in functions Cragg and King [7] offered a different list of error types: erroneous formulae incorrect ranges omitted factors data input errors incorrect use of functions duplication of effort Note that most of these sources focus on errors in formulas, although other types of errors (such as data input errors and duplication of effort) are mentioned. Galletta et al. [11] were the first to offer conceptual distinctions between classes of errors. They distinguished domain errors and device errors. Domain refers to the spreadsheet's application area (e.g., accounting), while device refers to the spreadsheet technology itself. For example, a mistake in logic due to a misunderstanding of depreciation is a domain error, but entering the wrong reference in the depreciation function SLN is a device error. Saarilouma and Sajaniemi [37] distinguished between location errors and formula errors. Location errors occur in formulas that are conceptually correct but in which one or more of the cell references are incorrect. Formula errors include misuse of operators or the wrong number of operators. Panko and Halverson [26] offered two more distinctions. One distinction is between eureka and Cassandra errors. Eureka errors are logic errors that are easily proven, while Cassandra errors are difficult to prove (even if detected). The other distinction is between pure and domain errors. Pure errors arise from a lapse in general logic while domain errors arise from a lack of specific domain knowledge Development of taxonomies The first serious attempt to offer a complete classification of errors came from Panko and Halverson [26]. They distinguished between quantitative and qualitative errors and further categorized quantitative errors. Quantitative errors lead to wrong numbers in the current version of the spreadsheet. An example is the wrong range in a SUM calculation, resulting in an incorrect total. Qualitative errors refer to practices that don't lead to wrong numbers in the current version of the spreadsheet but are risky practices that could lead to errors in subsequent uses. An example is hardcoding input parameters into a formula.
3 130 S.G. Powell et al. / Decision Support Systems 46 (2008) Quantitative errors are thus well defined and unambiguous, although not necessarily easy to find. Qualitative errors are not always so clear. For example, hard-coding numbers into formulas is not a quantitative error if the current version of the spreadsheet is correct, but it could lead to quantitative errors in subsequent uses if the user does not change inputs appropriately. Qualitative errors can also include complex formulas, confusing layout, or insufficient documentation. Panko and Halverson further divided quantitative errors into three subcategories: mechanical errors, due to mistakes in typing or pointing logic errors, due to choosing the wrong function or creating the wrong formula omission errors, due to misinterpreting the situation to be modeled This classification, which is based on the causes of errors, raises a number of questions. The first is: what is the context in which errors are being studied? If the context is laboratory experiments, we may be able to observe the spreadsheet developer in action and detect how each error was committed. On the other hand, if the context is evaluating operational spreadsheets in the field, we can only observe cells in the spreadsheet, not the process that created them. An appropriate classification of errors for one context may not be appropriate for the other. A second question is: what is the purpose of the classification? If the purpose is theoretical understanding, then we can usefully list all possible causes of errors. But if the purpose is a practical one, then an appropriate classification should include only errors we can observe and distinguish from one another. In field audits, we must concentrate not on causes but on outcomes. For example, the ultimate cause of an erroneous financial formula could be that the inputs were not typed correctly, the developer did not understand the necessary financial theory, the developer misunderstood the function, or the developer omitted an important factor from the model itself. In theory, these causes can be distinguished, but in practice we can observe only the error itself, so the finer categories may not be relevant. One of the shortcomings of existing classifications is the lack of examples. However, Panko [28] provided a detailed listing of the errors made by subjects in several spreadsheetdesign experiments. A total of 130 errors were briefly described and categorized into logic errors, omission errors, or mechanical errors. Teo and Tan [40] built on the classification scheme of Panko and Halverson, adding two additional types of qualitative errors: jamming errors and duplication errors. A jamming error occurs when more than one parameter is placed in a single cell, as when a volume is calculated by multiplying parameters for length, height and width, but the individual parameters are not isolated elsewhere. A duplication error occurs when the same parameter appears in two or more cells. Rajalingham et al. [33] and Purser and Chadwick [32] developed the most elaborate taxonomy of errors available to date. The first distinction in this hierarchy is between application-identified errors and developer/user-identified errors. Excel has eight categories of errors that it displays in a cell when it cannot resolve a formula. Some examples are #DIV/0! for dividing by zero and #VALUE! for the wrong type of argument in a cell reference. These are application-identified errors. Most such errors are likely to be caught by spreadsheet developers as they build a spreadsheet. However, in our own error research [30,31] we have occasionally observed these errors even in operational spreadsheets. The hierarchy for developer/user-identified errors can be presented in an outline form, as in Table 1. At the highest level, quantitative errors are distinguished from qualitative errors. Qualitative errors are divided into structural errors, which arise from flaws in the design, and temporal errors, which arise from the use of non-current data. Structural errors are further divided into visible and hidden errors. Presumably, a visible structural error is one that can be observed in the current spreadsheet while a hidden one arises only during use. Note that these authors defined qualitative errors somewhat differently from Panko and Halverson. As best we can tell, structural errors in Rajalingham et al. correspond to Panko and Halverson's category of qualitative errors. But the former also classify as a qualitative error the use of noncurrent data, which presumably leads to quantitative errors in the current spreadsheet. Quantitative errors are divided first into reasoning and accidental errors. Reasoning errors arise from a lack of knowledge, whereas accidental errors are created while entering a formula. Reasoning errors arise from lack of domain knowledge or from implementation errors. Lack of domain knowledge itself can involve lack of real world knowledge or lack of mathematical knowledge. Implementation errors can involve problems with syntax (the rules for creating proper formulas in Excel) or with logic. Accidental errors arise either during insertion or updating. Updating itself can involve modification or deletion. The taxonomy developed by Rajalingam et al. [33] uses at least three different criteria to distinguish categories of errors. The first distinction, between application-identified errors and developer/user-identified errors, is based on the detection method. The distinction between errors due to lack of real world knowledge and lack of mathematical knowledge is based on the cause of the error. Finally, the distinction between an error occurring during insertion or updating is based on the stage at which it was caused. Although we acknowledge that a theoretical taxonomy may be based on Table 1 Developer/user errors (after Rajalingham [33] and Purser and Chadwick [32]) A. Qualitative errors 1. Structural a. Visible b. Hidden 2. Temporal B. Quantitative errors 1. Reasoning a. Domain knowledge 1. Real world knowledge 2. Mathematical representation b. Implementation 1. Syntax 2. Logic 2. Accidental a. Insertion b. Update 1. Modification 2. Deletion
4 S.G. Powell et al. / Decision Support Systems 46 (2008) such disparate criteria, we question the usefulness of such a taxonomy in practice. Although this hierarchy is the most detailed categorization of errors in the literature, as far as we know, it has not been tested on spreadsheets from the field. Thus, we have no empirical evidence on how well the classification categorizes errors found in practice. Common sense suggests that because this hierarchy is focused in large part on the process by which an error occurred, it may not be possible for an auditor (or even the original developer) to reconstruct the cause and thereby categorize it when only the completed artifact is available. For example, if a complex formula involving nested IF, VLOOKUP, and NPV functions gives a wrong result, we cannot determine whether the cause was incorrect specification of the problem, lack of domain knowledge, or an accident in typing. A more serious concern with this taxonomy is that its categories overlap. For example, an error caused by lack of real world knowledge (category B1a1) could occur during modification (category B2b1). Likewise, an error due to lack of mathematical knowledge (B1a2) could surface as an application-identified error (e.g., #VALUE!). In the end, this taxonomy may therefore be of more theoretical than practical value Other approaches Most of the work cited to this point concerns itself with errors in completed spreadsheets. Panko [24,25] correctly pointed out, however, that different kinds of errors can occur at different stages in the life cycle of a spreadsheet, and the incidence of errors also may vary by stage. Errors can occur in the formulation of a model, before an actual spreadsheet is constructed. They can also occur during cell entry, during testing, and during use. Studies of errors committed during cell entry [20,22] found, not surprisingly, that developers commit a large number of errors while actually building a spreadsheet, but most are corrected immediately. Ayalew et al. [1] took a somewhat different approach to categorizing errors. They avoided categorizing based on the cause of the error but based their three-level schema on the spreadsheet modeling concept involved. Their three categories are physical-area related errors, logical-area related errors, and general errors. Physical-area related errors involve formulas that refer to blank cells, cells with values of the wrong type, or cells in the wrong range. Logical-area related errors involve overwriting a formula with a constant or copying a formula with incorrect references. Finally, the category of general errors includes all other possible errors in formulas, including typographical errors and domain or mathematical errors. Finally, we mention a study by Caulkins et al. [5] based on field interviews of 45 managers whose organizations rely on spreadsheets. These subjects mentioned observing the following errors, in descending order of frequency: inaccurate data errors inherited from reusing spreadsheets model error error in use of functions misinterpretation of output/report link broken/failed to update copy/paste lost file/saved over file Note that some of these categories, such as link broken/ failed to update are not explicitly represented in any of the existing classifications Summary Spreadsheet errors can be categorized in many different ways, for different contexts and purposes. No generallyaccepted taxonomy of spreadsheet errors has yet been established. In our survey we have identified several shortcomings of the existing classifications: Classifications are offered without specifying the context or purpose for which the classification is intended. The existing classifications do not include sufficient examples of specific errors that satisfy each category. Classifications are not rigorously tested to demonstrate that multiple users can consistently classify actual errors into the proper categories. The boundary between quantitative errors and qualitative errors remains vague. No studies are available that compare the types of errors by development stage. Our own research into errors in operational spreadsheets [30,31] suggests that a wide variety of types of errors occur. Furthermore, there are numerous borderline cases, where careful researchers can disagree as to whether a particular cell is an error. If classification of actual errors is as difficult as our experience suggests, any useful classification must be based on extensive analysis of actual spreadsheets, and must provide many examples to help researchers use the classification effectively. Future research on error classifications should clearly state the purpose for which the classification is designed. Then a set of examples should be provided, both of errors that fall in a given category and those that do not. Finally, any proposed classification should be tested to determine if it fits its purpose. Such tests should at a minimum establish three things: Errors can be identified in the subject population that fall into each category. No errors are encountered that cannot be classified. Independent examiners agree on the proper classification of the majority of errors. 3. Impact of errors Ironically, the impact of errors on spreadsheet results is the least studied of all the topics we address. Impact can be measured in several ways. An obvious measure is the percentage error in the outputs of the spreadsheet. But a 1% error in one spreadsheet could be devastating, while a 10% error in another could be inconsequential. A more telling measure of impact would be the actual dollar losses from erroneous or poor decisions resulting from spreadsheet errors. To estimate this impact would require tracking a problem within an organization from recognition through to implementation of a solution, and determining to what extent a suboptimal outcome was due to spreadsheet errors. This would be an ambitious undertaking, to say the least.
5 132 S.G. Powell et al. / Decision Support Systems 46 (2008) The evidence we do have on the impact of errors in audited spreadsheets is largely anecdotal. In a detailed audit of a single operational spreadsheet, Hicks (cited in Panko [25]) reported that the errors found in the audit caused the results to be off by 1.2%. Clermont et al. [6] found errors in an average of 3.03% of cells in three large spreadsheets but reported that we did not find any tremendous erroneous result values that might have had severe negative effects on the company. Lukasic (1998, personal communication cited in Panko [25]) found a 16% error in the results of one of two spreadsheets audited. Panko's interviewees (Panko [25]) suggested that 5% of spreadsheets had serious errors. Although we have little data on the impact of errors in audited spreadsheets, many stories have been published documenting losses due to mistakes involving spreadsheets. The European Spreadsheet Risks Interest Group (EUSPRIG) maintains a web page ( that documents dozens of these cases. Here is a small selection. Some candidates for police officer jobs are told they passed the test when in fact they had failed. Reason: the spreadsheet was sorted improperly. A school loses 30,000 because its budget is underestimated. Reason: numbers entered as text in a spreadsheet. Benefits of unbundling telecommunication services are understated by $50 million. Reason: incorrect references in a spreadsheet formula. A wide variety of errors has been documented by EUSPRIG, including formula errors, data entry errors, sorting errors, copy and paste errors, errors due to inadequate internal controls, formatting errors, and so on. The great majority of these reports, perhaps 80%, suggest that the error was actually in a formula. But there are also cases where the error was due to poor naming conventions, interaction with other software, and even fraud. Such examples remind us that the existing literature on errors focuses on what must be only a fraction of the actual errors surrounding the use of spreadsheets. Caulkins et al. [5] also asked their survey subjects to evaluate the risk of spreadsheets in decision making. Of the 44 who responded, 25 (57%) agreed with the statement Spreadsheet errors are a significant threat to decisions. Ten respondents (23%) said they used spreadsheets to inform decisions but were not particularly concerned about errors. The remainder did not use spreadsheets directly to inform decisions. Essentially no research has been conducted on the impact of errors in spreadsheets. Research in this area should be based on an explicit taxonomy of errors and a well-specified auditing procedure. In addition, it will be necessary to define carefully what is meant by impact. An error in a single formula can affect thousands of other cells in a spreadsheet. The impact can be small on some cells and large on others. To understand the overall impact will require a deep understanding of the purpose and uses of the spreadsheet, and probably the close collaboration of the developer. 4. Frequency of errors How common are errors in spreadsheets? Not surprisingly, the answer to this question depends on the definition of errors, the lifecycle stage, and the setting (operational or laboratory). About the only general conclusion we can draw from the literature is that no studies have suggested that errors are not a problem in spreadsheets, with the exception of Nardi and Miller [22] who concluded that users devote considerable effort to debugging their spreadsheet models they are very self-conscious about the probability of error and routinely track down errors before they can do any real harm. That conclusion, however, was based on self-reports by spreadsheet developers, not on an actual audit of the spreadsheets in question. Panko and Halverson [26] and Panko [24,25] summarized several dozen studies on the frequency of errors. We focus here on the latest of these studies, Panko [25]. Panko's summary is divided into four categories that correspond roughly to the spreadsheet development life cycle: cell entry experiments development experiments code inspection experiments field audits The first three of these categories involve experiments in which subjects develop or debug spreadsheets. Only the last category directly addresses the frequency of errors in operational spreadsheets Cell entry experiments Panko reported three cell entry experiments, in which subjects were observed as they created formulas. The percentage of cells with errors ranged from 11.3% to 21%. These errors were counted as they occurred, whether or not they were subsequently corrected. Thus, some of these errors were essentially typing errors, not errors in finished spreadsheets. This research may be of interest to those who wish to improve the efficiency with which spreadsheets are created, but it should not be taken as indicative of the error rate in finished spreadsheets. (After all, the error rate for keystrokes in word processing might be 5 10%, depending on the typist, but the error rate in finished documents would typically be much less than 1% of characters entered.) 4.2. Development experiments In development experiments, subjects are typically given a word problem and asked to create an appropriate spreadsheet model. Panko cited eleven of these studies. Cell error rates ranged from a low of 2% to a high of 17%. (In one study, individuals had a cell error rate of 4.6% while groups of three had a cell error rate of 1.0% on the same task.) The percentage with at least one error ranged from 24% to 86%. In a later study, Reinhardt and Pillay [35] reported on the errors made by computer literacy students. In their experiment, 82% made errors when using absolute addressing, 87% made errors using an IF function, and 93% made errors when using a financial function. Laboratory experiments on spreadsheet errors may be useful for certain purposes, but we should be extremely careful in making inferences about the error rate in operational spreadsheets from errors rates in laboratory experiments. The two contexts differ in a host of ways. Operational spreadsheets are typically large, developed by a team over a significant period of time, and subjected to repeated use.
6 S.G. Powell et al. / Decision Support Systems 46 (2008) Spreadsheets in laboratory experiments, on the other hand, are typically small, built over a short period of time by one person, and rarely used. Reason [34] provides two warnings that apply here about investigating errors in the laboratory: First, the need to establish precise control over the possible determinants of the error often forces investigators to focus upon rather trivial phenomena. Second, it is usually the case that the greater the measure of control achieved by the experimenter, the more artificial and unnatural are the conditions under which the error is elicited. Certainly laboratory experiments on spreadsheets are to some extent artificial and unnatural, when compared to spreadsheet use in organizations. In a later section, we provide some perspective on the appropriate use of laboratory experiments in spreadsheet error research Code inspection experiments Code inspection experiments measure the ability of subjects to find errors in spreadsheets. We discuss these results in a later section on detecting errors Field audits Panko reported on thirteen field audits in which operational spreadsheets were examined (typically by an outsider to the organization). Since this class of studies is most directly relevant to the subject of this paper we review these studies in more detail. Panko summarized this literature by reporting that 94% of spreadsheets have errors, with an average cell error rate of 5.2%. We will revisit these estimates after reviewing the studies individually. Davies and Ikin [8] examined 19 Lotus spreadsheets from a variety of firms. Five of the 19 had no quantitative or qualitative errors, while four had major errors. Presumably the remaining 10 had minor errors. Ditlea [9] reported that one Price-Waterhouse consultant found 128 errors in four large spreadsheets. Butler (personal communication 1992, cited in Panko [25]) reported errors requiring additional tax payments in 10.7% of 273 small-business tax returns audited by HM Customs and Excise (the tax authority in the UK). Cragg and King [7] reviewed 20 spreadsheets from 10 companies. A single auditor spent an average of 2 h examining each spreadsheet. Five of the 20 contained errors. Dent (1995, personal communication with Panko [25]), described an audit in a mining company that found errors in about 30% of spreadsheets. Hicks (1995, cited in Panko [25]) reported on an extensive code inspection process for one large spreadsheet at a major firm. Errors were found in five of 19 modules. In all, 45 errors were identified, with a cell error rate of 1.2%. Freeman [10] reported on the experience of auditors at Coopers and Lybrand, who investigated 23 spreadsheets of which 91% had errors. KPMG (1997, cited in Panko [25]) investigated 22 spreadsheets and again, 91% had major errors. Butler [3,4] reported on audits of 7 tax spreadsheets using the same procedure as in Butler (1992, cited in Panko [25]) and 86% had errors. Lukasik (1998, personal communication with Panko [25]) audited two spreadsheets and found cell errors rates of 2.2 and 2.5%. Clermont et al. [6] audited three large workbooks used by a single industrial firm. These workbooks involved 78 worksheets and 60,446 cells. The average cell error rate was 3.03% (0.08% in one spreadsheet, 1.3% in a second, and 6.7% in the third). Of the total of 1,832 error cells identified, 241 (15%) were quantitative errors. In two sets of interviews in 2003 with experienced consultants who perform spreadsheet audits professionally, Panko was told that approximately 5% of spreadsheets had serious errors [25]. Lawrence and Lee [19] reported on 30 audits conducted by Mercer Finance and Risk Consulting on project financing models. These models involved an average of 2182 unique formulas. On average, the auditor raised issues with the developer concerning 6.9% of all unique formulas in the spreadsheet; however, the range was from 3.1% to 22.5% of unique formulas. These sources range from personal communications to well-documented research studies. Perhaps the most important conclusion to draw from them is that no source has maintained that operational spreadsheets are error-free. But what do they suggest is the average error rate? Panko attempted to answer this question by averaging over the seven most recent studies, on the grounds that auditing procedures improved after These studies are summarized in Table 2. In total, 88 spreadsheets are represented in the table. For all 88, the weighted average percentage of spreadsheets with errors is 94%. Data on cell error rates were available on 43 of these spreadsheets, and the weighted average for this sample is 5.2%. There are several reasons to question the reliability of these estimates. First, three of the seven sources are unpublished. Second, the majority of the sources gave little or no information on their definition of errors or on the methods used to find errors. Third, Lawrence and Lee [19], whose observations account for 70% of the sample used to estimate the cell error rate, did not actually report a cell error Table 2 Spreadsheet error rates (after Panko [25]) Source Number Percent audited with errors Cell error rate (% of cells) 1. Coopers and N/A Lybrand (1997) 2. KPMG (1998) N/A 3. Hicks (1995) Lukasic (1998) Butler (2000) Clermont (2002) Lawrence and Lee (2004) Average 94 a 5.2 b a Weighted average of 88 spreadsheets in sources 1 7. b Weighted average of 43 spreadsheets in sources 3 7.
7 134 S.G. Powell et al. / Decision Support Systems 46 (2008) rate of 6.9% on completed spreadsheets. Rather, they reported that auditors had issues on the initial review of a model that they subsequently discussed with the developers concerning 6.9% of the cells. The authors provided no details on the definition of an issue or its relation to an actual error. Moreover, the initial model version that was reviewed was revised an average of six times before the model was complete. Thus the estimate of a 6.9% error rate applies only to issues raised by auditors on initial model versions, not to errors in completed spreadsheets. To be clear, we do not mean to suggest that the true error rates in operational spreadsheets are negligible. Our own studies [30,31] suggest the opposite. However, the evidence behind the estimates in the literature does not provide sufficient grounds for a reliable, quantitative estimate. Managers who understand the importance of spreadsheet analysis to their business want to know how prevalent errors are in spreadsheets. Unfortunately, this is simply not a question we can answer accurately at this time, despite a number of studies in the area. The problem is a lack of standardization regarding: the definition of errors the methods used to detect errors the sample of spreadsheets studied Previous studies have used different definitions of errors, when they have provided a definition at all. Many studies have not revealed details about the detection methods used; those that provided details used different methods. Finally, the characteristics of the spreadsheets used in the experimental sample have rarely been reported, so we cannot know whether the results apply to any particular area of interest, such as stage of development, application domain, or size of spreadsheet. What we do know from this literature can be summarized as follows: Laboratory experiments have shown high error rates, whether measured by cell error rates or percent of spreadsheets with errors. Field audits have generally shown high cell and spreadsheet error rates, but methods and results vary widely. No studies of spreadsheets themselves (rather than interviews of developers) have shown errors to be rare or inconsequential. To be effective, future research on the prevalence of errors must satisfy three criteria: The procedures used to select spreadsheets and the characteristics of those spreadsheets (such as the number of formulas) must be described. A well-tested taxonomy of errors must be used. The procedure used to audit the spreadsheets for errors must be clearly described. 5. Creation and prevention of errors We know remarkably little about how errors are created by end users. Not surprisingly, perhaps, no one has attempted to study spreadsheet development in the field at a level of detail that would permit observation of developers making errors. What little we do know comes from laboratory experiments, yet as previously stated, the relevance of laboratory results to the field is questionable. In a study mentioned earlier, Brown and Gould [2] performed experiments in which their subjects created three spreadsheets from written descriptions. Subjects were videotaped during this process and their keystrokes recorded, thus allowing the causes of errors to be determined. The researchers observed typing mistakes, misuse of cell protection, errors in logic, erroneous cell references, copying mistakes, and misplacement of data. Olson and Nilson [23] and Lerch [20] tracked actual keystrokes while subjects built spreadsheets. In 56 formulas written by these subjects, 12 had at least one extra keystroke compared to the optimal sequence. (Extra keystrokes are those beyond the minimum required for the purpose.) Although these studies reported on errors made during formula creation, not errors in finished spreadsheets, they suggest that the methods used by spreadsheet developers may have a significant impact on the efficiency with which formulas are created. Lerch also observed that mechanical errors increased when the cells being referenced were in different columns and rows, which suggests that a logical layout may reduce the incidence of errors. Many papers have offered advice on how to build errorfree spreadsheets, but very few studies provide concrete evidence. Brown and Gould [2] observed that participants in their study did not spend a lot of time planning before launching into creating a spreadsheet. That observation raises the question whether planning might reduce the incidence of errors. Nardi and Miller [22] noted that in all the cases they studied, more than one person was involved in the creation or debugging of a spreadsheet. That observation raises the question of whether groups can develop models more successfully than individuals and what group mechanisms might contribute to success. Janvrin and Morrison [15] carried out two experiments in which subjects were taught to use a structured design method (Data Flow Diagrams) when developing a spreadsheet. Subjects were given 1 h of instruction on spreadsheet development and then asked to develop a workbook with ten worksheets and 51 formulas requiring links among the worksheets. In the first study, subjects who used the structured design approach had a 7% error rate in linking formulas to inputs, whereas subjects who used an ad hoc approach had a 10% error rate. That work suggests that structured methods reduce errors, at least modestly. In a second study, conceptual and oversight errors were also observed, but although the structured approach still reduced the incidence of linking errors, it had no impact on the other types. Kruck and Sheetz [18] and Kruck [17] first developed and then tested a theory of spreadsheet accuracy. In the first paper, they identified three beliefs or hypotheses about practices that reduce errors: reduce problem complexity, test for errors, and use an explicit design process. In the second paper, they tested whether providing information on these practices to student subjects reduces the number of errors committed. The information that was provided to subjects is summarized in Table 3 below. The experimental results show that all three aids improve performance: a control group committed 6.4
8 S.G. Powell et al. / Decision Support Systems 46 (2008) Table 3 Aids for spreadsheet accuracy (after Kruck [17]) Planning and design aids 1. Use a planned layout so that movement of the data after entering will be minimized. 2. Use an organized layout to isolate data and computation areas. 3. Use descriptive labels. 4. Repeat the input data near the output. Formula complexity aids 1. Formulas should contain only cell references. 2. Formulas can be used to repeat data in other locations. 3. Split complex formulas. 4. Relative and absolute cell addressing should be used where appropriate. Test and debugging aids 1. Use Excel's auditing tools to verify formulas. 2. Review worksheet for error messages. 3. Determine if the numbers look reasonable. 4. Verify all calculations with a calculator or test with known models. errors on average, but those given one of the aids committed only 4.2 to 4.6 errors on average. Finally, an expert was asked to evaluate each of these spreadsheets for ease of use and proper segmentation of inputs and outputs. Spreadsheets that rated best on these criteria had an average of 4 errors, while those rated worst had an average of 24 errors. The literature does not tell us a great deal about how errors occur. A few laboratory experiments have observed developers making errors, but these experiments may not provide much insight into the causes of errors in operational spreadsheets. Most spreadsheet users create large numbers of errors as they enter data and formulas, but many of these are corrected immediately, and some are eventually corrected during testing or use. We still know very little about why the remaining errors escape detection. Preventing errors in spreadsheets is a complex undertaking. Designing a serious research program to determine ways to improve the process would be equally complex. Research in this area can focus either on individual or organizational practices. At the individual level, laboratory research could establish whether practices such as use of explicit design processes or use of auditing software reduces errors. The same question could be studied in the field by surveying or observing developers, but the difficulty in deriving clean results would be substantial. At the organizational level, it would be helpful to know whether background education, in-house training, or spreadsheet standards have a measurable effect on spreadsheet errors. 6. Detection of errors Two types of studies have examined detection of errors in completed spreadsheets. In laboratory experiments, subjects are asked to find errors placed in spreadsheets by the researcher; in field audits, experts try to find errors in operational spreadsheets. In the great majority of laboratory studies, the subjects have been given no specific training or instruction on how to identify errors. Thus, these studies may have limited implications for detecting errors. One exception is Teo and Tan [40], who performed a two-part experiment in which subjects first built a spreadsheet from a written problem description and then performed what-if analysis by changing some parameters. This experiment provided insight into the possibility that some errors can be detected during the use of a spreadsheet. Their results showed that only 13 of 70 subjects actually corrected errors in their spreadsheets during the second exercise. Of all mechanical errors committed in the first exercise, 39% were corrected in the second exercise. A somewhat smaller percentage, 30%, of logic and omission errors was corrected at the second stage. However, offsetting this improvement was an increase in the overall number of errors during the second exercise: 49% new mechanical errors, 30% new logic errors and 65% new omission errors. Therefore, it appears that during use, errors made during development can be corrected, but new errors can also be introduced in the process of modifying the spreadsheet. Teo and Lee-Partridge [39] conducted a related set of experiments to test the dependence of error detection on the nature of the error, individual factors (such as expertise), and prior practice. Their subjects were not taught how to look for errors in spreadsheets. The results suggest that success in locating errors increases with practice, although more for logic and omission errors than for mechanical, jamming or duplication errors. Panko and Halverson [27] conducted an experiment in which both individuals and groups examined a spreadsheet for errors. The groups found about two-thirds of all errors, while individuals found only one-third, suggesting that groups may be more effective than individuals in certain tasks related to spreadsheets. Galletta et al. [11,12] performed two studies in which subjects searched for errors that had previously been seeded in spreadsheets. In the first study, about 55% of the errors were identified by subjects, and expertise in both accounting and spreadsheets contributed to success. In the second study, the subjects were shown either a printed copy or an electronic copy of the spreadsheet, with or without formulas. Again, about 50% of the errors were identified and subjects who could refer to the formulas did not find more errors than those who could only check numbers. Howe and Simkin [14] similarly found that experimental subjects found 67% of errors seeded in a spreadsheet. Their regression analysis failed to identify any significant determinants of error detection rates by error type. Purser and Chadwick [32] administered web-based surveys to professionals and students asking them to identify errors in spreadsheets. Both groups were given a first survey in which no information was provided on the types of errors to look for and a second survey in which the error classification scheme discussed above was provided. The results showed experienced users detecting a much higher percentage of errors than novices, but knowledge of error types did not always improve performance. In fact, the number of qualitative errors identified went up as expected, but the number of quantitative errors identified went down. In addition to laboratory experiments, a few studies have reported the results of auditing operational spreadsheets. Panko [24,25] cited seven reports on field audits. Most of these disclose no details on the procedure for identifying errors. One exception is Hicks (cited in Panko [25]), who reported on the audit of a large-scale capital budgeting spreadsheet at NYNEX. In this audit, each of the six main
9 136 S.G. Powell et al. / Decision Support Systems 46 (2008) modules of the spreadsheet was audited by a three-person team. The audit began with an explanation from the developer covering the module and its relation to the model as a whole. The team then verified formulas and checked cell references. One cell in each column was studied in detail and the others in the same row were checked for consistency. Test data were used to audit some portions of the module. Finally, Excel's formula auditing tool was used. The auditing procedure used by HM Customs and Excise was described by Butler [3]. This procedure involves the use of a software tool (SpACE) created for government auditing of small-business tax returns. This auditing procedure has been documented in [13], which is the most explicit published auditing protocol available. The Customs and Excise procedure works as follows. First, the auditor identifies the chain of cells from inputs to end result and uses the software to follow the chain of dependent cells so that the key formulas can be checked. Then the auditor checks the original formulas that were used to copy related formulas, and checks that the copies are correct. Again the software is used to speed up this task. Finally, fourteen types of high-risk cells are checked for arithmetic and logical correctness. These include, for example, cells that contain constants, have no dependents, or involve complex functions such as NPV. Clermont et al. [6] used specially-developed auditing software in a field audit of three large spreadsheets. The auditor first discussed each workbook with its developer and collected summary data about the workbook (number of cells, number of formulas, and so on). The second step was to spotcheck the spreadsheet for errors. Finally, the software tool was run and highlighted cells were investigated. Ideally, spreadsheets would be created free of errors, but it seems more practical to develop efficient methods and tools for detecting errors. Unfortunately, the literature provides little guidance for this effort. Few laboratory experiments have been performed that test the effectiveness of different approaches to error detection. Although several field audits have been done to identify errors, most have not reported on the methods used, and none has compared different approaches to the task. Finally, we should keep in mind the complementary roles of prevention and detection. In terms of the end product, it may not matter whether a developer creates an errorfree result by making a lot of mistakes and correcting them or by doing careful planning and making relatively few mistakes. Nevertheless, there is an unavoidable economic trade-off between prevention efforts and detection efforts. Perhaps further research into these topics might help identify the kinds of initiatives that could reduce errors at least cost. Research that establishes effective methods for detecting errors obviously requires a well-tested taxonomy of errors. Laboratory experiments that compare the performance of subjects using different auditing procedures would contribute significantly to our understanding. Ideally, such experiments would use business analysts (not students) as subjects and operational spreadsheets as the objects of study. Field experiments would be more challenging but could offer deeper results. A natural experiment in which an organization or workgroup introduces new procedures for error detection would be an ideal research platform. 7. Research directions Research on spreadsheet errors can be conducted either in the laboratory or in the field. Each type of research offers its own insights and has its own limitations. In some ways, laboratory research is easier to conduct than field research, but its limitations are significant. In particular, error rates in laboratory experiments should not be used uncritically to infer error rates in operational spreadsheets because the underlying conditions differ. We don't yet know the impact of those differences on error rates. More research needs to be carried out in both domains before we can judge how the two error rates compare. Laboratory experiments can be particularly useful in identifying the types of errors that occur, the stage of the modeling process at which they occur, and their causes. Laboratory experiments would also be helpful in comparing auditing procedures and auditing software. It is perhaps remarkable that more laboratory experiments have not been performed on subjects creating spreadsheets. The ideal experiment would give a subject a written description of a problem to solve using a spreadsheet and then track the process through from start to finish. The method of verbal protocols [41] would seem to be ideal for this purpose, especially when used in conjunction with software that tracks keystrokes. A series of such experiments could determine the stage at which errors of different types are made (and discovered). They could also allow us to identify specific approaches to spreadsheet design, building, and testing that work best. The great advantage of field research on operational spreadsheets is that we are dealing with the real thing, not an experimental substitute. Unfortunately, field research is generally difficult and it has its own limitations. Certainly, more work needs to be done in identifying the types and frequency of errors that occur in the field. To be most useful, such research must precisely specify error types and detection methods. It must also describe the spreadsheets that were audited: how they were selected, their size, complexity, and so on. Field research is also needed on methods for detecting errors. Here are some of the questions research in this area could answer: Which procedures for error detection work best? How do teams and individuals compare at detecting errors? Can we identify high-risk cells before auditing them? Does auditing software improve error detection rates? To what extent do the results of laboratory experiments on error detection apply to error detection in the field? Field research can also uncover other useful information about spreadsheet use. For example, what best practices are used by developers whose spreadsheets are relatively errorfree? What organizational policies and norms exist in organizations with good spreadsheets? And what can we learn from the sociology of spreadsheet use; that is, how do individuals and groups come together to solve problems using spreadsheet technology? In summary, errors in spreadsheets appear to be frequent and are potentially costly. We have reviewed the research
10 S.G. Powell et al. / Decision Support Systems 46 (2008) literature on spreadsheet errors, focusing on five topics, and drawn the following conclusions: Classification: Taxonomies exist, but in most cases they are context-dependent, imprecisely defined, and untested. Impact: Little is known about the quantitative (economic) consequences of errors in spreadsheets. Frequency: If we want to estimate cell error rates accurately, we must standardize the procedures for doing so. Creation and prevention: Laboratory experiments have been inconclusive or suggestive, but broad progress has been limited. Detection: Serious research, detailing the methods used and comparing alternative approaches, is lacking on this subject. The state of the art allows us to outline several promising directions for future research. Laboratory research can better identify the types of errors that occur, when and how. It can also be used to compare auditing procedures and auditing software. Two kinds of field research are needed: narrow research into errors and their frequency, and broad research into best practices and the sociology of spreadsheet use. References [1] Y. Ayalew, M. Clermont, R. Mittermier, Detecting errors in spreadsheets, Proceedings of the European Spreadsheet Risks Interest Group Annual Conference University of Greenwich, London, 2000, pp [2] P. Brown, J. Gould, An experimental study of people creating spreadsheets, ACM Transactions on Office Information Systems 5 (1987) [3] R. Butler, Is this spreadsheet a tax evader? Proceedings of the 33rd Hawaii International Conference on System Sciences, 2000, pp [4] R. Butler, Risk assessment for spreadsheet developments: choosing which models to audit, H. M. Customs and Excise UK (2000). [5] J. Caulkins, E. Morrison, T. Weidemann, Spreadsheet errors and decision making: evidence from field interviews, Journal of End User Computing (to appear) (2006). [6] M. Clermont, A spreadsheet auditing tool evaluated in an industrial context, Proceedings of the European Spreadsheet Risks Interest Group Conference Cardiff Wales, 2002, pp [7] P. Cragg, M. King, Spreadsheet modelling abuse: an opportunity for OR? Journal of Operational Research Society 44 (8) (1993) [8] N. Davies, C. Ikin, Auditing spreadsheets, Australian Accountant 57 (11) (1987) [9] S. Ditlea, Spreadsheets can be hazardous to your health, Personal Computing (1987) [10] D. Freeman, How to make spreadsheets error-proof, Journal of Accountancy 181 (5) (1996) [11] F. Galletta, D. Abraham, M. El Louadi, W. Leske, Y. Pollalis, J. Sampler, An empirical study of spreadsheet error-finding performance, Accounting, Management & Information Technology 3 (2) (1993) [12] F. Galletta, K. Hartzel, S. Johnson, J. Joseph, S. Rustagi, Spreadsheet presentation and error detection: an experimental study, Journal of Management Information Systems 13 (3) (1997) [13] H. M., Customs and excise computer audit service, Methodology for the audit of spreadsheet models, [14] H. Howe, M.G. Simkin, Factors affecting the ability to detect spreadsheet errors, Decision Sciences Journal of Innovative Education 4 (1) (2006) [15] D. Janvrin, J. Morrison, Factors influencing risks and outcomes in enduser development, Proceedings of the 29th Annual Hawaii International Conference on System Sciences, 1996, pp [16] J. Kreie, T. Cronin, J. Pendley, J. Renwick, Applications development by end-users: can quality be improved? Decision Support Systems, 29 (2000) [17] S. Kruck, Testing spreadsheet accuracy theory, Information and Software Technology 48 (2006) [18] S. Kruck, S. Sheetz, Spreadsheet accuracy theory, Journal of Information Systems Education 12 (2001) [19] R. Lawrence, J. Lee, Financial modelling of project financing transactions, Institute of Actuaries of Australia Financial Services 19 (2004). [20] F. Lerch, Computerized financial planning: discovering cognitive difficulties in knowledge building, Unpublished Ph.D. dissertation, University of Michigan, Ann Arbor, MI (1988). [21] T. McGill, J. Klobas, The role of spreadsheet knowledge in userdeveloped application success, Decision Support Systems 39 (2005) [22] B. Nardi, J. Miller, Twinkling lights and nested loops: distributed problem solving and spreadsheet development, International Journal of Man-Machine Studies 34 (1991) [23] J. Olson, J.E. Nilsen, Analysis of the cognition involved in spreadsheet software interaction, Human-Computer Interaction 3 (4) ( ) [24] R. Panko, What we know about spreadsheet errors, Journal of End-User Computing 10 (1998) [25] R. Panko, What we know about spreadsheet errors, cba.hawaii.edu/ssr/mypapers/whatknow.htm, accessed September 2. [26] R. Panko, R. Halverson, Spreadsheets on trial: a survey of research on spreadsheet risks, Proceedings of the 29th Annual Hawaii International Conference on Systems Sciences, 1996, pp [27] R. Panko, R. Halverson, Are two heads better than one (at reducing errors in spreadsheet modeling?), Office Systems Research Journal 15 (1) (1997) [28] R. Panko, R. Halverson, Two corpuses of spreadsheet errors, Proceedings of the 33rd Annual Hawaii International Conference on Systems Sciences, 2000, pp [29] R. Panko, R. Sprague, Hitting the wall: errors in developing and code inspecting a simple spreadsheet model, Decision Support Systems 22 (1998) [30] S. Powell, K. Baker, B. Lawson, Errors in operational spreadsheets, Spreadsheet Engineering Research Project working paper, [31] S. Powell, K. Baker, B. Lawson, An auditing protocol for spreadsheet models, Spreadsheet Engineering Research Project working paper, [32] M. Purser, D. Chadwick, Does an awareness of differing types of spreadsheet errors aid end-users in identifying spreadsheet errors? Proceedings of the European Spreadsheet Risk Interest Group Annual Conference, Cambridge, UK, 2006, pp [33] K. Rajalingham, D. Chadwick, B. Knight, Classification of spreadsheet errors, Proceedings of the European Spreadsheet Risks Interest Group Annual Conference, Greenwich, England, 2000, pp [34] J. Reason, Human Error, Cambridge University Press, Cambridge, UK, [35] T. Reinhardt, N. Pillay, Analysis of spreadsheet errors made by computer literacy students, Proceedings of the IEEE International Conference on Advanced Learning Technologies, 2004, pp [36] B. Ronen, M. Palley, H. Lucas, Spreadsheet analysis and design, Communications of the ACM 32 (1) (1989) [37] P. Saariluoma, J. Sajaniemi, Transforming verbal descriptions into mathematical formulas in spreadsheet calculation, International Journal of Human-Computer Studies 41 (1994) [38] K. Sommerville, Software Engineering, 7th EditionAddison Wesley, [39] T. Teo, J. Lee-Partridge, Effects of error factors and prior incremental practice on spreadsheet error detection: an experimental study, Omega The International Journal of Management Science 29 (2001) [40] T. Teo, M. Tan, Quantitative and qualitative errors in spreadsheet development, Proceedings of the 30th Hawaii International Conference on Systems Sciences, 1997, pp [41] M. Van Someren, Y. Barnard, J. Sandberg, The Think Aloud Method: A Practical Guide to Modeling Cognitive Processes, Academic Press, New York, Steve Powell is a Professor at the Tuck School of Business at Dartmouth. His primary research interest lies in modeling production and services processes, but he has also been active in research in energy economics, marketing, and operations. At Tuck, he has developed a variety of courses in management science, including the core Decision Science course and electives in the Art of Modeling, Business Process Redesign, and Applications of Simulation. He originated the Teacher's Forum column in Interfaces, and has written a number of articles on teaching modeling to practitioners. He is the academic director of the INFORMS Annual Teaching of Management Science Workshop. In 2001 he was awarded the INFORMS Prize for the Teaching of Operations Research/Management Science Practice. He is the co-author with Kenneth Baker of The Art of Modeling with Spreadsheets (Wiley, 2004).
11 138 S.G. Powell et al. / Decision Support Systems 46 (2008) Ken Baker is a faculty member at Dartmouth College. He is currently Nathaniel Leverone Professor of Management at the Tuck School of Business and also adjunct professor at the Thayer School of Engineering. At Dartmouth, he has taught courses relating to decision science, manufacturing management, and environmental management. Over the years, much of his teaching and research has dealt with production planning and control, and he is widely known for his textbook Elements of Sequencing and Scheduling, in addition to a variety of technical articles. He has served as the Tuck School's associate dean and directed the Tuck School's management development programs in the manufacturing area. In 2001 he was named a Fellow of INFORMS' Manufacturing and Service Operations Management (MSOM) Society, and in 2004 a Fellow of INFORMS. He is the co-author with Stephen Powell of The Art of Modeling with Spreadsheets (Wiley, 2004). Barry Lawson is a research associate at the Tuck School of Business at Dartmouth and is also a visiting scholar in the geography department of the college. He founded and has served as president of Barry Lawson Associates, a consulting firm, since As visiting scholar, he coordinates the development of an atlas of the upper Connecticut River Watershed in New HampshireandVermont.AsresearchassociateatTuckheserves as the program manager for the Tuck Spreadsheet Engineering Research Project. Lawson has taught in graduate programs at Boston University and Wayne State University as well as in short courses at Bentley College. He has moderated a host of public hearings for local, state and federal governments on controversial environmental and energy-and waste-related projects, and has considerable experience in group facilitation, conflict resolution and simulation design.
Errors in Operational Spreadsheets: A Review of the State of the Art
Errors in Operational Spreadsheets: A Review of the State of the Art Stephen G. Powell Tuck School of Business Dartmouth College [email protected] Kenneth R. Baker Tuck School of Business Dartmouth College
Quality Control in Spreadsheets: A Software Engineering-Based Approach to Spreadsheet Development
Quality Control in Spreadsheets: A Software Engineering-Based Approach to Spreadsheet Development Kamalasen Rajalingham, David Chadwick, Brian Knight, Dilwyn Edwards Information Integrity Research Centre
Risk Assessment For Spreadsheet Developments: Choosing Which Models to Audit
Risk Assessment For Spreadsheet Developments: Choosing Which Models to Audit Raymond J. Butler, CISA H. M. Customs and Excise, UK Email [email protected] Crown Copyright reserved, published by permission
The Use of Spreadsheets: Considerations for Section 404 of the Sarbanes-Oxley Act*
The Use of Spreadsheets: Considerations for Section 404 of the Sarbanes-Oxley Act* July 2004 *connectedthinking The Use of Spreadsheets: Considerations for Section 404 of the Sarbanes-Oxley Act Introduction
CFSD 21 ST CENTURY SKILL RUBRIC CRITICAL & CREATIVE THINKING
Critical and creative thinking (higher order thinking) refer to a set of cognitive skills or strategies that increases the probability of a desired outcome. In an information- rich society, the quality
An Activity-Based Costing Assessment Task: Using an Excel Spreadsheet
e-journal of Business Education & Scholarship of Teaching Vol. 3, No. 1, 2009, pp:25-35. http://www.ejbest.org Instructional Note An Activity-Based Costing Assessment Task: Using an Excel Spreadsheet Damian
SPSS: Getting Started. For Windows
For Windows Updated: August 2012 Table of Contents Section 1: Overview... 3 1.1 Introduction to SPSS Tutorials... 3 1.2 Introduction to SPSS... 3 1.3 Overview of SPSS for Windows... 3 Section 2: Entering
White Paper from Global Process Innovation. Fourteen Metrics for a BPM Program
White Paper from Global Process Innovation by Jim Boots Fourteen Metrics for a BPM Program This white paper presents 14 metrics which may be useful for monitoring progress on a BPM program or initiative.
Spreadsheet Auditing Software
David Nixon, Mike O'Hara Information Systems Institute, University of Salford, United Kingdom Nixon & O Hara ABSTRACT It is now widely accepted that errors in spreadsheets are both common and potentially
Almost all spreadsheet programs are based on a simple concept: the malleable matrix.
MS EXCEL 2000 Spreadsheet Use, Formulas, Functions, References More than any other type of personal computer software, the spreadsheet has changed the way people do business. Spreadsheet software allows
Performance Assessment Task Which Shape? Grade 3. Common Core State Standards Math - Content Standards
Performance Assessment Task Which Shape? Grade 3 This task challenges a student to use knowledge of geometrical attributes (such as angle size, number of angles, number of sides, and parallel sides) to
Spreadsheets have become the principal software application for teaching decision models in most business
Vol. 8, No. 2, January 2008, pp. 89 95 issn 1532-0545 08 0802 0089 informs I N F O R M S Transactions on Education Teaching Note Some Practical Issues with Excel Solver: Lessons for Students and Instructors
Comparing Methods to Identify Defect Reports in a Change Management Database
Comparing Methods to Identify Defect Reports in a Change Management Database Elaine J. Weyuker, Thomas J. Ostrand AT&T Labs - Research 180 Park Avenue Florham Park, NJ 07932 (weyuker,ostrand)@research.att.com
Data Quality Assurance
CHAPTER 4 Data Quality Assurance The previous chapters define accurate data. They talk about the importance of data and in particular the importance of accurate data. They describe how complex the topic
ACCESS 2007. Importing and Exporting Data Files. Information Technology. MS Access 2007 Users Guide. IT Training & Development (818) 677-1700
Information Technology MS Access 2007 Users Guide ACCESS 2007 Importing and Exporting Data Files IT Training & Development (818) 677-1700 [email protected] TABLE OF CONTENTS Introduction... 1 Import Excel
Populating a Data Quality Scorecard with Relevant Metrics WHITE PAPER
Populating a Data Quality Scorecard with Relevant Metrics WHITE PAPER SAS White Paper Table of Contents Introduction.... 1 Useful vs. So-What Metrics... 2 The So-What Metric.... 2 Defining Relevant Metrics...
Unemployment Insurance Data Validation Operations Guide
Unemployment Insurance Data Validation Operations Guide ETA Operations Guide 411 U.S. Department of Labor Employment and Training Administration Office of Unemployment Insurance TABLE OF CONTENTS Chapter
CRITICAL PATH ANALYSIS AND GANTT CHARTS
CRITICAL PATH ANALYSIS AND GANTT CHARTS 1. An engineering project is modelled by the activity network shown in the figure above. The activities are represented by the arcs. The number in brackets on each
Q&As: Microsoft Excel 2013: Chapter 2
Q&As: Microsoft Excel 2013: Chapter 2 In Step 5, why did the date that was entered change from 4/5/10 to 4/5/2010? When Excel recognizes that you entered a date in mm/dd/yy format, it automatically formats
Preventing Duplicate Payments for Lawson Customers
White Paper Preventing Duplicate Payments for Lawson Customers By Brian Rosenberg, Senior Partner, RPI Consultants Duplicate Payments are far more common than most companies realize or are prepared to
Data Management for Multi-Environment Trials in Excel
Data Management for Multi-Environment Trials in Excel Cathy Garlick ([email protected]), Statistical Services Centre, University of Reading, UK 27 May 2010 1. Introduction Excel is widely available
Automating Spreadsheet Discovery & Risk Assessment
Abstract Keywords Automating Spreadsheet Discovery & Risk Assessment Automating Spreadsheet Discovery & Risk Assessment Prodiance Corporation 5000 Executive Parkway, Suite 270 San Ramon, CA 94583 USA [email protected]
CREATING LEARNING OUTCOMES
CREATING LEARNING OUTCOMES What Are Student Learning Outcomes? Learning outcomes are statements of the knowledge, skills and abilities individual students should possess and can demonstrate upon completion
Software Process for QA
Software Process for QA Basic approaches & alternatives CIS 610, W98 / M Young 1/7/98 1 This introduction and overview is intended to provide some basic background on software process (sometimes called
For example, estimate the population of the United States as 3 times 10⁸ and the
CCSS: Mathematics The Number System CCSS: Grade 8 8.NS.A. Know that there are numbers that are not rational, and approximate them by rational numbers. 8.NS.A.1. Understand informally that every number
Exemplars. Research Report
Exemplars Research Report Every effort has been made to trace and contact the authors of these works, but this has not been possible in every case. If notified, we will undertake to rectify any errors
AUDIT REPORT. Cybersecurity Controls Over a Major National Nuclear Security Administration Information System
U.S. Department of Energy Office of Inspector General Office of Audits and Inspections AUDIT REPORT Cybersecurity Controls Over a Major National Nuclear Security Administration Information System DOE/IG-0938
AUDIT PROCEDURES RECEIVABLE AND SALES
184 AUDIT PROCEDURES RECEIVABLE AND SALES Ștefan Zuca Abstract The overall objective of the audit of accounts receivable and sales is to determine if they are fairly presented in the context of the financial
Advanced Excel 10/20/2011 1
Advanced Excel Data Validation Excel has a feature called Data Validation, which will allow you to control what kind of information is typed into cells. 1. Select the cell(s) you wish to control. 2. Click
Checklist for Operational Risk Management
Checklist for Operational Risk Management I. Development and Establishment of Comprehensive Operational Risk Management System by Management Checkpoints - Operational risk is the risk of loss resulting
Position Classification Standard for Management and Program Clerical and Assistance Series, GS-0344
Position Classification Standard for Management and Program Clerical and Assistance Series, GS-0344 Table of Contents SERIES DEFINITION... 2 EXCLUSIONS... 2 OCCUPATIONAL INFORMATION... 3 TITLES... 6 EVALUATING
In-Depth Guide Advanced Spreadsheet Techniques
In-Depth Guide Advanced Spreadsheet Techniques Learning Objectives By reading and completing the activities in this chapter, you will be able to: Create PivotTables using Microsoft Excel Create scenarios
[Refer Slide Time: 05:10]
Principles of Programming Languages Prof: S. Arun Kumar Department of Computer Science and Engineering Indian Institute of Technology Delhi Lecture no 7 Lecture Title: Syntactic Classes Welcome to lecture
AN INTEGRATED APPROACH TO TEACHING SPREADSHEET SKILLS. Mindell Reiss Nitkin, Simmons College. Abstract
AN INTEGRATED APPROACH TO TEACHING SPREADSHEET SKILLS Mindell Reiss Nitkin, Simmons College Abstract As teachers in management courses we face the question of how to insure that our students acquire the
IBM SPSS Direct Marketing 23
IBM SPSS Direct Marketing 23 Note Before using this information and the product it supports, read the information in Notices on page 25. Product Information This edition applies to version 23, release
Peer Review Process Description
Peer Review Process Description Version 1.0 draft1 Table of Contents 1. Overview... 1 2. Work Aids... 1 3. Risk Assessment Guidance... 1 4. Participants... 2 5. Inspection
Introduction to QuickBooks Online Edition Course Manual
Introduction to QuickBooks Online Edition Course Manual Module 8 End of Period Activities and Financial Statements Copyright Notice. Each module of the Introduction To QuickBooks Course Manual may be viewed
Data Analysis Tools. Tools for Summarizing Data
Data Analysis Tools This section of the notes is meant to introduce you to many of the tools that are provided by Excel under the Tools/Data Analysis menu item. If your computer does not have that tool
IBM SPSS Direct Marketing 22
IBM SPSS Direct Marketing 22 Note Before using this information and the product it supports, read the information in Notices on page 25. Product Information This edition applies to version 22, release
Performance Assessment Task Bikes and Trikes Grade 4. Common Core State Standards Math - Content Standards
Performance Assessment Task Bikes and Trikes Grade 4 The task challenges a student to demonstrate understanding of concepts involved in multiplication. A student must make sense of equal sized groups of
Microsoft Excel 2007 Mini Skills Overview of Tables
To make managing and analyzing a group of related data easier, you can turn a range of cells into a Microsoft Office Excel table (previously known as an Excel list). A table typically contains related
5/19/2014. 1 Professor Lili Saghafi
5/19/2014 1 Professor Lili Saghafi MANAGING INFORMATION TECHNOLOGY Lecture 9 METHODOLOGIES FOR CUSTOM SOFTWARE DEVELOPMENT By : Prof. Lili Saghafi 1-2 METHODOLOGIES FOR CUSTOM SOFTWARE DEVELOPMENT Large
Part II Management Accounting Decision-Making Tools
Part II Management Accounting Decision-Making Tools Chapter 7 Chapter 8 Chapter 9 Cost-Volume-Profit Analysis Comprehensive Business Budgeting Incremental Analysis and Decision-making Costs Chapter 10
Quantitative and qualitative methods in process improvement and product quality assessment.
Quantitative and qualitative methods in process improvement and product quality assessment. Anna Bobkowska Abstract Successful improvement of the development process and product quality assurance should
Audit Risk and Materiality in Conducting an Audit
Audit Risk and Materiality in Conducting an Audit 1647 AU Section 312 Audit Risk and Materiality in Conducting an Audit (Supersedes SAS No. 47.) Source: SAS No. 107. See section 9312 for interpretations
INTERNATIONAL STANDARD ON AUDITING (UK AND IRELAND) 240 THE AUDITOR S RESPONSIBILITY TO CONSIDER FRAUD IN AN AUDIT OF FINANCIAL STATEMENTS CONTENTS
INTERNATIONAL STANDARD ON AUDITING (UK AND IRELAND) 240 THE AUDITOR S RESPONSIBILITY TO CONSIDER FRAUD IN AN AUDIT OF FINANCIAL STATEMENTS CONTENTS Paragraphs Introduction... 1-3 Characteristics of Fraud...
MICROSOFT EXCEL STEP BY STEP GUIDE
IGCSE ICT SECTION 14 DATA ANALYSIS MICROSOFT EXCEL STEP BY STEP GUIDE Mark Nicholls ICT Lounge Data Analysis Self Study Guide Contents Learning Outcomes Page 3 What is a Data Model?... Page 4 Spreadsheet
Microsoft Excel Tips & Tricks
Microsoft Excel Tips & Tricks Collaborative Programs Research & Evaluation TABLE OF CONTENTS Introduction page 2 Useful Functions page 2 Getting Started with Formulas page 2 Nested Formulas page 3 Copying
DATA QUALITY DATA BASE QUALITY INFORMATION SYSTEM QUALITY
DATA QUALITY DATA BASE QUALITY INFORMATION SYSTEM QUALITY The content of those documents are the exclusive property of REVER. The aim of those documents is to provide information and should, in no case,
Peer Review Process Description
Peer Review Process Description Version 1.0 draft1 Table of Contents 1.Overview...1 2.Work Aids...1 3.Risk Assessment Guidance...1 4.Participants...2 5.Inspection Procedure...4
D A Cowen Snowden Mining Industry Consultants
` VALUATION MODELLING STANDARDS: Guidelines for spreadsheet modelling in the mining industry D A Cowen Snowden Mining Industry Consultants 1. Introduction Spreadsheets have developed in concert with personal
The Discussion Paper. Conceptual Framework of Financial Accounting
The Discussion Paper Conceptual Framework of Financial Accounting Accounting Standards Board of Japan December 2006 (Tentative translation: 16 Mar. 2007) Contents Preface 1 Chapter 1 Objectives of Financial
University of Macau Department of History. Guidelines for Writing a BA Senior Thesis
University of Macau Department of History Guidelines for Writing a BA Senior Thesis In order to complete the requirements of the BA programme in History at the University of Macau, each senior student
How To Develop Software
Software Engineering Prof. N.L. Sarda Computer Science & Engineering Indian Institute of Technology, Bombay Lecture-4 Overview of Phases (Part - II) We studied the problem definition phase, with which
Using Excel for Statistics Tips and Warnings
Using Excel for Statistics Tips and Warnings November 2000 University of Reading Statistical Services Centre Biometrics Advisory and Support Service to DFID Contents 1. Introduction 3 1.1 Data Entry and
Managing Multi-Valued Attributes in Spreadsheet Applications
Managing Multi-Valued Attributes in Spreadsheet Applications Clare Churcher, Theresa McLennan and Wendy Spray Lincoln University New Zealand churcher/mclennan/[email protected] Abstract End-users frequently
What is the Purpose of College Algebra? Sheldon P. Gordon Farmingdale State College of New York
What is the Purpose of College Algebra? Sheldon P. Gordon Farmingdale State College of New York Each year, over 1,000,000 students [8] take College Algebra and related courses. Most of these courses were
Appendix B Data Quality Dimensions
Appendix B Data Quality Dimensions Purpose Dimensions of data quality are fundamental to understanding how to improve data. This appendix summarizes, in chronological order of publication, three foundational
The Analysis of Quality Escapes in the Aerospace & Defense Industry
The Analysis of Quality Escapes in the Aerospace & Defense Industry White Paper November 1, 2012 1825 Commerce Center Blvd Fairborn, Ohio 45324 937-322-3227 www.ren-rervices.com The Analysis of Quality
WVU STUDENT EVALUATION OF INSTRUCTION REPORT OF RESULTS INTERPRETIVE GUIDE
WVU STUDENT EVALUATION OF INSTRUCTION REPORT OF RESULTS INTERPRETIVE GUIDE The Interpretive Guide is intended to assist instructors in reading and understanding the Student Evaluation of Instruction Report
Macros allow you to integrate existing Excel reports with a new information system
Macro Magic Macros allow you to integrate existing Excel reports with a new information system By Rick Collard Many water and wastewater professionals use Microsoft Excel extensively, producing reports
Trends in Interdisciplinary Dissertation Research: An Analysis of the Survey of Earned Doctorates
Trends in Interdisciplinary Dissertation Research: An Analysis of the Survey of Earned Doctorates Working Paper NCSES 12-200 April 2012 by Morgan M. Millar and Don A. Dillman 1 Disclaimer and Acknowledgments
Creating CSV Files. Recording Data in Spreadsheets
Statistical Modeling: A Fresh Approach Creating CSV Files Recording Data in Spreadsheets Use spreadsheet software to record and organize your data as tables. Do data entry using the spreadsheet; do analysis
Protecting Business Information With A SharePoint Data Governance Model. TITUS White Paper
Protecting Business Information With A SharePoint Data Governance Model TITUS White Paper Information in this document is subject to change without notice. Complying with all applicable copyright laws
How to Review a Technical Paper
How to Review a Technical Paper Alan Meier, Berkeley Lab, University of California, Berkeley, CA 94720, USA First published in Energy and Buildings 19:75 78 (1992) Available online at: http://eetd.lbl.gov/ea/buildings/alan/publications/how.to.review.html
Commonly Used Excel Formulas
Microsoft Excel 2007 Advanced Formulas Windows XP Look Up Values in a List of Data: Commonly Used Excel Formulas Let's say you want to look up an employee's phone extension by using their badge number
How To Use Excel With A Calculator
Functions & Data Analysis Tools Academic Computing Services www.ku.edu/acs Abstract: This workshop focuses on the functions and data analysis tools of Microsoft Excel. Topics included are the function
Business & Marketing. Customised Training program (CTP)
Business & Marketing Customised Training Programs (CTPs) In 2011/12 TafeSA trained over 80,000 students across 48 campuses in over 400 qualifications ranging from Certificate levels through to Advanced
Creating A Grade Sheet With Microsoft Excel
Creating A Grade Sheet With Microsoft Excel Microsoft Excel serves as an excellent tool for tracking grades in your course. But its power is not limited to its ability to organize information in rows and
Experiences with Online Programming Examinations
Experiences with Online Programming Examinations Monica Farrow and Peter King School of Mathematical and Computer Sciences, Heriot-Watt University, Edinburgh EH14 4AS Abstract An online programming examination
Do you know? "7 Practices" for a Reliable Requirements Management. by Software Process Engineering Inc. translated by Sparx Systems Japan Co., Ltd.
Do you know? "7 Practices" for a Reliable Requirements Management by Software Process Engineering Inc. translated by Sparx Systems Japan Co., Ltd. In this white paper, we focus on the "Requirements Management,"
Foundations of Business Intelligence: Databases and Information Management
Foundations of Business Intelligence: Databases and Information Management Problem: HP s numerous systems unable to deliver the information needed for a complete picture of business operations, lack of
1. I have 4 sides. My opposite sides are equal. I have 4 right angles. Which shape am I?
Which Shape? This problem gives you the chance to: identify and describe shapes use clues to solve riddles Use shapes A, B, or C to solve the riddles. A B C 1. I have 4 sides. My opposite sides are equal.
Microsoft Excel 2010 Part 3: Advanced Excel
CALIFORNIA STATE UNIVERSITY, LOS ANGELES INFORMATION TECHNOLOGY SERVICES Microsoft Excel 2010 Part 3: Advanced Excel Winter 2015, Version 1.0 Table of Contents Introduction...2 Sorting Data...2 Sorting
META DATA QUALITY CONTROL ARCHITECTURE IN DATA WAREHOUSING
META DATA QUALITY CONTROL ARCHITECTURE IN DATA WAREHOUSING Ramesh Babu Palepu 1, Dr K V Sambasiva Rao 2 Dept of IT, Amrita Sai Institute of Science & Technology 1 MVR College of Engineering 2 [email protected]
A Performance Audit of the State s Purchasing Card Program
REPORT TO THE UTAH LEGISLATURE Number 2006-09 A Performance Audit of the State s Purchasing Card Program September 2006 Audit Performed By: Audit Manager Audit Supervisor Audit Staff Rick Coleman Susan
STOP THAT SUBVERSIVE SPREADSHEET!
STOP THAT SUBVERSIVE SPREADSHEET! David Chadwick School of Computing and Mathematical Sciences, University of Greenwich, London SE10 9LS, UK; [email protected] Abstract: This paper documents the formation
Calculating Cash Flow Using Excel Data Lists
Calculating Cash Flow Using Excel Data Lists By: Curtis D. Frye 7/26/2006 Bank and credit card statements tell you where your money has gone, but they don t show you how much money you ll have on hand
Sage Intelligence Reporting. Microsoft FRx to Sage Intelligence Report Designer Add-In Conversion Guide. Sage 100 ERP
Sage Intelligence Reporting Microsoft FRx to Sage Intelligence Report Designer Add-In Conversion Guide Sage 100 ERP The software described in this document is protected by copyright, and may not be copied
Evaluating a new programming language
In G. Kadoda (Ed). Proc. PPIG 13 Pages 275-289 Evaluating a new programming language Steven Clarke Microsoft Corporation 1 Microsoft Way Redmond, WA 98052 USA +1 425 705 5978 [email protected] Keywords:
TDWI strives to provide course books that are content-rich and that serve as useful reference documents after a class has ended.
Previews of TDWI course books are provided as an opportunity to see the quality of our material and help you to select the courses that best fit your needs. The previews can not be printed. TDWI strives
Working Paper No. 51/00 Current Materiality Guidance for Auditors. Thomas E. McKee. Aasmund Eilifsen
Working Paper No. 51/00 Current Materiality Guidance for Auditors By Thomas E. McKee Aasmund Eilifsen SNF-project No. 7844: «Finansregnskap, konkurs og revisjon The project is financed by the Norwegian
SOAL-SOAL MICROSOFT EXCEL 1. The box on the chart that contains the name of each individual record is called the. A. cell B. title C. axis D.
SOAL-SOAL MICROSOFT EXCEL 1. The box on the chart that contains the name of each individual record is called the. A. cell B. title C. axis D. legend 2. If you want all of the white cats grouped together
Checks and Controls in spreadsheets
Checks and Controls in spreadsheets Patrick O'Beirne Systems Modelling Ltd Tara Hill, Gorey, Co. Wexford, Ireland Tel +353-53-942-2294 Email [email protected] 4 June 2009 Abstract Spreadsheets that are informally
INTOSAI. Performance Audit Subcommittee - PAS. Designing performance audits: setting the audit questions and criteria
INTOSAI Performance Audit Subcommittee - PAS Designing performance audits: setting the audit questions and criteria 1 Introduction A difficult phase in performance auditing is planning and designing. In
Best Practices in Hiring
Best Practices in Hiring by Mike Poskey ZERORISK HR If a job candidate looks good on paper and looks good in the interview, one would naturally assume this candidate would be a good fit for the job. While
NHA. User Guide, Version 1.0. Production Tool
NHA User Guide, Version 1.0 Production Tool Welcome to the National Health Accounts Production Tool National Health Accounts (NHA) is an internationally standardized methodology that tracks public and
Data Coding and Entry Lessons Learned
Chapter 7 Data Coding and Entry Lessons Learned Pércsich Richárd Introduction In this chapter we give an overview of the process of coding and entry of the 1999 pilot test data for the English examination
The Series of Discussion Papers. Conceptual Framework of Financial Accounting
The Series of Discussion Papers Conceptual Framework of Financial Accounting Working Group on Fundamental Concepts September 2004 (Tentative translation: 28 Feb. 2005) Contents Issuance of the Series of
Modeling Guidelines Manual
Modeling Guidelines Manual [Insert company name here] July 2014 Author: John Doe [email protected] Page 1 of 22 Table of Contents 1. Introduction... 3 2. Business Process Management (BPM)... 4 2.1.
TDWI strives to provide course books that are content-rich and that serve as useful reference documents after a class has ended.
Previews of TDWI course books offer an opportunity to see the quality of our material and help you to select the courses that best fit your needs. The previews cannot be printed. TDWI strives to provide
White Paper May 2009. Spreadsheet-based planning in today s economy: Why you re up against the wall
White Paper May 2009 Spreadsheet-based planning in today s economy: Why you re up against the wall Spreadsheet-based planning in today s economy: Why you re up against the wall 2 Contents 3 The error-riddled
3 Guidance for Successful Evaluations
3 Guidance for Successful Evaluations In developing STEP, project leads identified several key challenges in conducting technology evaluations. The following subsections address the four challenges identified
Tutorial 2: Using Excel in Data Analysis
Tutorial 2: Using Excel in Data Analysis This tutorial guide addresses several issues particularly relevant in the context of the level 1 Physics lab sessions at Durham: organising your work sheet neatly,
