1 MANAGNG TH DVLOPMNT OF LARG SOFTWAR SYSTMS Dr. Winston W. Rovce NTRODUCTON l am going to describe my pe,-.~onal views about managing large software developments. have had various assignments during the past nit,.: years, mostly concerned with the development of software packages for spacecraft mission planning, commanding and post-flight analysis. n these assignments have experienced different degrees of successwith respect to arriving at an operational state, on-time, and within costs. have become prejudiced by my experiences and am going to relate some of these prejudices in this presentation. COMPUTR PROGRAM DVLOPMNT FUNCTONS There are two essential steps common to all computer program developments, regardless of size or complexity. There is first an analysis step, followed second by a coding step as depicted in Figure 1. This sort of very simple implementation concept is in fact all that is required if the effort is sufficiently small and if the final product is to be operated by those who built it - as is typically done with computer programs for internal use. t is also the kind of development effort for which most customers are happy to pay, since both steps involve genuinely creative work which directly contributes to the usefulness of the final product. An implementation plan to manufacture 13rger software systems, and keyed only to these steps, however, is doomed tofailure. Many additional development steps are required, none contribute as directly to the final product as analysis and coding, and all drive up the development costs. Customer personnel typically would rather not pay for them, and development personnel would rather not implement them. The prime function of management is to sell these concepts to both groups and then enforce compliance on the part of development personnel. ANALYSS CODNG Figure 1. mplementation steps to deliver a small computer program for internal operations. A more grandiose approach to software development is illustrated in Figure 2. The analysis and coding steps are still in the picture, but they are preceded by two levels of requirements analysis, are separated by a program design step, and followed by a testing step. These additions are treated separately from analysis and coding because they are distinctly different in the way they are executed. They must be planned and staffed differently for best utilization of program resources. Figure 3 portrays the iterative relationship between successive development phases for this scheme. The ordering of steps is based on the following concept: that as each step progresses and the design is further detailed, there is an iteration with the preceding and succeeding steps but rarely with the more remote steps in the sequence. The virtue of all of this is that as the design proceeds the change process is scoped down to manageable limits. At any point in the design process after the requirements analysis is completed there exists a firm and c~seup~ moving baseline to whi(:h to ~turn in the event of unforeseen design difficulties. What we have is an effective fallback position that tends to maximize the extent of early work that is salvageable and preserved. Reprinted from Proceedings, WSCON, August 1970, pages 1-9. Co_pyright 1_9_70 by The nstitute of lectrical and lectronics t)gineers,,.328 nc. Originally published by TRW.
2 SYST M ANALYSS PROGRAM DSGN coo,.o TSTNG OPRATONS Figure 2. mplementation steps to develop a large computer program for delivery to a customer. believe in this concept, but the implementation described above is risky and invites failure. The problem is illustrated in Figure 4. The testing phase which occurs at the end of the development cycle is the first event for which timing, storage, input/output transfers, etc., are experienced as distinguished from analyzed. These phenomena are not precisely analyzable. They are not the solutions to the standard partial differential equations of mathematical physics for instance. Yet if these phenomena fail to satisfy the various external constraints, then invariably a major redesign is required. A simple octal patch or redo of some isolated code will not fix these kinds of difficulties. The required design changes are likely to be so disruptive that the software requirements upon which the design is based and which provides the rationale for everything are violated. ither the requirements must be modified, or a substantial change in the design is required. n effect the development process has returned to the origin and one can expect up to a lo0-percent overrun in schedule and/or costs. One might note that there has been a skipping-over of the analysis and code phases. One cannot, of course, produce software without these steps, but generally these phases are managed with relative ease and have little impact on requirements, design, and testing. n my experience there are whole departments consumed with the analysis of orbit mechanics, spacecraft attitude determination, mathematical optimization of payload activity and so forth, but when these departments have completed their difficult and complex work, the resultant program steps involvea few lines of serial arithmetic code. f in the execution of their difficult and complex work the analysts have made a mistake, the correction is invariably implemented by a minor change in the code with no disruptive feedback into the other development bases. However, believe the illustrated approach to be fundamentally sound. The remainder of this discussion presents five additional features that must be added to this basic approach to eliminate most of the development risks. 329
3 SYSTM! RQURMNTSB~ ~"'i so,w.,~ ANALYSS ~1111~ pr~ogram ~lll CODNG i TSTNG OPRATONS Figure 3. Hopefully, the ~terat=ve interact=on between the various phases is confined to successive steps. SYSTM "1.~oo,.~-,..Sl.,~ so,w..~!. ANALYSS PROGRAM DSGN coo,.g,~! TSTNG O.ATO.S! Figure 4. Unfortunately, for the process illustrated, the design iterations are never confined to the successive steps. 330
4 STP 1: PROGRAM DSGN COMS FRST The first step towards a fix is illustrated in Figure 5. A preliminary program design phase has been inserted between the software requirements generation phase and the analysis phase. This procedure can be criticized on the basis that the program designer is forced to design in the relative vacuum of initial software requirements without any existing analysis..as a result, his preliminary design may be substantially in error as compared to his design if he were to wait until the analysis was complete. This criticism is correct but it misses the point. By this technique the program designer assures that the software will not fail because of storage, timing, and data flux reasons. As the analysis proceeds in the succeeding phase the program designer must impose on the analyst the storage, timing, and operational constraints in such a way that he senses the consequences. When he justifiably requires more of this kind of resource in order to implement his equations it must be simultaneously snatched from his analyst compatriots. n this way all the analysts and all the program designers will contribute to a meaningful design process which will culminate in the proper allocation of execution time and storage resources. f the total resources to be applied are insufficient or if the embryo operational design is wrong it will be recognized at this earlier stage and the iteration with requhements and preliminary design can be redone before final design, coding and test commences. How is this procedure implemented? The following steps are required. 1) Begin the design process with program designers, not analysts or programmers. 2) Design, define and allocate the data processing modes even at the risk of being wrong. Allocate processing, functions, design the data base, define data base processing, allocate execution time, define interfaces and processing modes with the operating system, describe input and output processing, and define preliminary operating procedures. 3) Write an overview document that is understandable, informative and current. ach and every worker must have an elemental understanding of the system. At least one person must have a deep understanding of the system which comes partially from having had to write an overview document. ALLOCAT ~ A DSCRB / so..ooo,,. / c % Figure 5. Step 1 : nsure that a preliminary program design is complete before analysis begins. 331
5 STP2: DOCUMNT TH DSGN At this point it is appropriate to raise the issue of - "how much documentation?" My own view is "quite a lot;" certainly more than most programmers, analysts, or program designers are willing to do if left to their own devices. The first rule of managing software development is ruthless enforcement of documentation requirements. Occasionally am called upon to review the progress of other software design efforts. My first step is to investigate the state of the documentation, f the documentation is in serious default my first recommendation is simple. Replace project management. Stop all activities not related to documentation. Bring the documentation up to acceptable standards. Management of software is simply impossible withouta very high degree of documentation. As an example, let me offer the following estimates for comparison. n order to procure a 5 million dollar hardware device, would expect that a 30 page specification would provide adequate detail to control the procurement. n order to procure5 million dollars of software would estimate ~ 1[,00 pa~e specification is about right in order to achieve comparable control, Why so much documentation? 1) ach designer must communicate with interfacing designers, with his management and possibly with thecustorner. A verbal record is too intangible to provide an adequate basis for an interface or manage- mentdecision. An acceptable written description forces the designer to take an unequivocal position and provide tangible evidence of completion. t prevents the designer from hiding behind the-"l finished" - syndrome month after month. am90-percent 2) During the early phase of software development the documentation i.sthe specification and i._~.s the design. Until coding begins these three nouns (documentation, specification, design) denoteasingtething. f the documentation is bad the design is bad. f the documentation does not yet exist there is as yet no design, only people thinking and talking about the design which is of some value, but not much. 3) The real monetary value of good documentation begins downstream in the development process during the testing phase and continues through operations and redesign. The value of documentation can be described in terms of three concrete, tangible situations that every program manager faces. a) During the testing phase, with good documentation the manager can concentrate personnel on the mistakes in the program. Without good documentation every mistake, large or small, is analyzed by one man who probably made the mistake in the first place because he is the only man who understands the program area. b) During the operational phase, with good documentation the manager can use operation-oriented personnel to operate the program and to do a better job, cheaper. Without good documentation the software must be operated by those who built it. Generally these people are relatively disinterested in operations and do not do as effective a job as operations-oriented personnel. t should be pointed out in this connection that in an operational situation, if there is some hangup the software is always blamed first. n order either to absolve the software or to fix the blame, the software documentation must speak clearly. c) Following initial operations, when system improvements are in order, good documentation permits effective redesign, updating, and retrofitting in the field. f documentation does not exist, generally the entire existing framework of operating software must be junked, even for relatively modest changes. Figure 6 shows a documentation plan which is keyed to the steps previously shown. Note that six documents are produced, and at the time of delivery of the final product, Documents No, 1, No. 3, No. 4, No. 5, and No. 6 are updated and current. 332
6 rl. o. i0. i ~,-,,*,1 =. ~ illl ~$~ z~_~ m u, X /, 0: wz g /oo ~ i,~ ll N / " ~,.~r" z_,,,. ~ ~ ~OLU a..~ N 8 "0.2 N t- i-,<~,~,- Z w LL 333
7 STP 3: DO T TWC After documentation, the second most important criterion for success revolves around whether the product is totally original. f the computer program in question is being developed for the first time, arrange matters so that the version finally delivered to the customer for operational deployment is actually the second version insofar as critical design/operations areas are concerned. Figure 7 iltustrates how this might be carried out by means of a simulation. Note that it is simply the entire process done in miniature, toatime scale that is relatively small with respect to the overall effort. The nature of this effort can vary widely depending primarily on the overall time scale and the nature of the critical problem areas to be modeled. f the effort runs 30 months then this early development ofapilot model might be scheduled for 10 months. For this schedule, fairly formal controls, documentation procedures, etc., can be utilized. f, however, the overall effort were reduced to 12 months, then the pilot effort could be compressed to three months perhaps, in order to gain sufficient leverage on the mainline development. n this case a very special kind of broad competence is required on the part of the personnel involved. They must have an intuitive feel for analysis, coding, and program design. They must quickly sense the trouble spots in the design, model them, model their alternatives, forget the straightforward aspects of the design which aren't worth studying at this early point, and finally arrive at an error-free program. n either case the point of all this, as with a simulation, is that questions of timing, storage, etc. which are otherwise matters of judgment, can now be studied with precision. Without this simulation the project manager is at the mercy of human judgment. With the simulation he can at least perform experimental tests of some key hypotheses and scope down what remains for human judgment, which in the area of computer program design (as in the estimation of takeoff gross weight, costs to complete, or the daily double) is invariably and seriously optimistic.,,, PRLMNARY % 1 PROGRAM DSGN ANALYSS ANALYSS! PROGRAM DS,GN -U coo,.o L.,s.,.o USAG i PROGRAM DSGN TSTNG [ OPRATONS Figure 7. Step 3: Attempt to do the job twice - the first result provides an early simulation of the final product. 334
8 STP 4: PLAN, CONTROL AND MONTOR TSTNG Without question the biggest user of project resources, whether it be manpower, computer time, or management judgment, is the test phase. t is the phase of greatest risk in terms of dollars and schedule. t occurs at the latest point in the schedule when backup alternatives are least available, if at all. The previous three recommendations to design the program before beginning analysis and coding, to document it completely, and to build a pilot model are all aimed at uncovering and solving problems before entering the test phase. However, even after doing these things there is stillatest phase and there are still important things to be done. Figure 81ists some additional aspects to testing. n planning for testing, would suggest the following for consideration. 1) Many parts of the test process are best handled by test specialists who did not necessarily contribute to the original design. f it is argued that only the designer can perform a thorough test because only he understands the area he built, this is a sure sign of a failure to document properly. With good documentation it is feasible to use specialists in software product assurance who will, in my judgment, do a better job of testing than the designer. 2) Most errors are of an obvious nature that carl be easily spotted by visual inspection. very bit of an analysis and every bit of code should be subjected to a simple visual scan by a second party who did not do the original analysis or code but who would spot things like dropped minus signs, missing factors of two, jumps to wrong addresses, etc., which are in the nature of proofrea0ing the analysis and code. Do not use the computer to detect this kind of thing - it is too expensive. 3) Test every logic path in the computer program at least once with some kind of numerical check. f wereacustomer, would not accept delivery until this procedure was completed and certified. This step will uncover the majority of coding errors. While this test procedure sounds simple, for a large, complex computer program it is relatively difficult to plow through every logic path with controlled values of input. n fact there are those who will argue that it is very nearly impossible. n spite of this would persist in my recommendation that every logic path be subjected to at least one authentic check. 4) After the simple errors (which are in the majority, and which obscure the big mistakes) are removed, then it is time to turn over the software to the test area for checkout purposes. At the proper time during the course of development and in the hands of the proper person the computer itself is the best device for checkout. Key management decisions are: when is the time and who is the person to do final checkout? STP 5: NVOLV TH CUSTOMR For some reason what a software design is going to do is subject to wide interpretation even after previous agreement. t is important to involve the customer ina formal way so that he has committed himself at earlier points before final delivery. To give the contractor free rein between requirement definition and operation is inviting trouble. Figure g indicates three points following requirementsdefinition where the insight, judgment, and commitment of the customer carl bolster the development effort. SUMMARY Figure 10 summarizes the five steps that feel necessary to transform a risky development process into one that will provide the desired product. would emphasize that each item costs some additional sum of money. f the relatively simpler process without the five complexities described here would work successfully, then of course the additional money is not well spent. i, my experience, however, the simpler method has never worked on large software development efforts and the costs to recover far exceeded those required to finance the five-step process listed. 335
9 l ~ ~m~ T _~ L_ ~.L wsig ~o_~, _ " o ~.~ O..va W r- 2 o '1 8 t- O "O o 8 t- e,. Q.. r,m LL /, 336
10 Z o_.- < r,- ill Q. o /'L J (.- r- 0 (J f- t.- 0. ~n >o,< z < +.~ o 4.., C o r- / l Q. Q. ii c~ (/) -- z~ i,,. u a,~l r.n rr n n. J ii1 337
11 :i]. ~ ' l l e ~$ ~ ~ i n ~ ~ u 8( '.. s"" O0 0 O ~ d.l. w R 338
General Principles of Software Validation; Final Guidance for Industry and FDA Staff Document issued on: January 11, 2002 This document supersedes the draft document, "General Principles of Software Validation,
Production of Large Computer Programs HERBERT D. BENINGTON The paper is adapted from a presentation at a symposium on advanced programming methods for digital computers sponsored by the Navy Mathematical
USE-CASE 2.0 The Guide to Succeeding with Use Cases Ivar Jacobson Ian Spence Kurt Bittner December 2011 USE-CASE 2.0 The Definitive Guide About this Guide 3 How to read this Guide 3 What is Use-Case 2.0?
IP ASSETS MANAGEMENT SERIES 1 Successful Technology Licensing 2 Successful Technology Licensing I. INTRODUCTION II. III. IV. PREPARATION FOR NEGOTIATION KEY TERMS CONDUCTING THE NEGOTIATION V. USING THE
Making Smart IT Choices Understanding Value and Risk in Government IT Investments Sharon S. Dawes Theresa A. Pardo Stephanie Simon Anthony M. Cresswell Mark F. LaVigne David F. Andersen Peter A. Bloniarz
The Scrum Guide The Definitive Guide to Scrum: The Rules of the Game July 2013 Developed and sustained by Ken Schwaber and Jeff Sutherland Table of Contents Purpose of the Scrum Guide... 3 Definition of
SESSION 1 PAPER 1 SOME METHODS OF ARTIFICIAL INTELLIGENCE AND HEURISTIC PROGRAMMING by Dr. MARVIN L. MINSKY (94009) 3 BIOGRAPHICAL NOTE Marvin Lee Minsky was born in New York on 9th August, 1927. He received
Experience with Processes and Monitors in Mesa 1 Abstract Butler W. Lampson Xerox Palo Alto Research Center David D. Redell Xerox Business Systems The use of monitors for describing concurrency has been
Eight Things Your Business Analysts Need to Know A Practical Approach to Recognizing and Improving Competencies An ESI International White Paper (877) 766-3337 www.esi-intl.com Table of Contents Abstract...3
Why Johnny Can t Encrypt: A Usability Evaluation of PGP 5.0 Alma Whitten School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 firstname.lastname@example.org J. D. Tygar 1 EECS and SIMS University
WHEN YOU CONSULT A STATISTICIAN... WHAT TO EXPECT SECTION ON STATISTICAL CONSULTING AMERICAN STATISTICAL ASSOCIATION 2003 When you consult a statistician, you enlist the help of a professional who is particularly
Programming Techniques R. Morris Editor On the Criteria To Be Used in Decomposing Systems into Modules D.L. Parnas Carnegie-Mellon University This paper discusses modularization as a mechanism for improving
Basic Marketing Research: Volume 1 Handbook for Research Professionals Official Training Guide from Qualtrics Scott M. Smith Gerald S. Albaum Copyright 2012, Qualtrics Labs, Inc. ISBN: 978-0-9849328-1-8
Financial Conduct Authority How the Financial Conduct Authority will investigate and report on regulatory failure April 2013 Contents Introduction 3 1. Context 4 2. What is a regulatory failure? 5 3 Event
Focus Groups as Qualitative Research PLANNING AND RESEARCH DESIGN FOR FOCUS GROUPS Contributors: David L. Morgan Print Pub. Date: 1997 Online Pub. Date: Print ISBN: 9780761903437 Online ISBN: 9781412984287
UML and Patterns.book Page 61 Thursday, September 16, 2004 9:48 PM Chapter 6 6 USE CASES The indispensable first step to getting the things you want out of life: decide what you want. Ben Stein Objectives
OVERVIEW Brief description This toolkit deals with the nuts and bolts (the basics) of setting up and using a monitoring and evaluation system for a project or an organisation. It clarifies what monitoring
In Security and Usability: Designing Secure Systems that People Can Use, eds. L. Cranor and G. Simson. O'Reilly, 2005, pp. 679-702 CHAPTER THIRTY- FOUR Why Johnny Can t Encrypt A Usability Evaluation of
How to Choose a Leadership Pattern by Robert Tannenbaum and Warren H. Schmidt No. 73311 MAY JUNE 1973 How to Choose a Leadership Pattern Robert Tannenbaum and Warren H. Schmidt Since its publication in
TRANSFORM YOUR CITY THROUGH INNOVATION THE INNOVATION DELIVERY MODEL FOR MAKING IT HAPPEN JANUARY 2014 CONTENTS INTRODUCTION... 1 THE IMPERATIVE FOR INNOVATION... 2 WHAT IS THE INNOVATION DELIVERY MODEL?....
Defining and Testing EMR Usability: Principles and Proposed Methods of EMR Usability Evaluation and Rating HIMSS EHR Usability Task Force June 2009 CONTENTS EXECUTIVE SUMMARY... 1 INTRODUCTION... 2 WHAT
Rich Data, Poor Data Designing Dashboards to Inform by Stephen Few A note about the author Stephen Few has worked for 24 years as an IT innovator, consultant, and educator. Today, as Principal of the consultancy
SOFTWARE ENGINEERING Report on a conference sponsored by the NATO SCIENCE COMMITTEE Garmisch, Germany, 7th to 11th October 1968 Chairman: Professor Dr. F. L. Bauer Co-chairmen: Professor L. Bolliet, Dr.
Your consent to our cookies if you continue to use this website.