Position Paper for Cognition and Collaboration Workshop: Analyzing Distributed Community Practices for Design
|
|
|
- Aubrey Allen
- 10 years ago
- Views:
Transcription
1 Position Paper for Cognition and Collaboration Workshop: Analyzing Distributed Community Practices for Design Jean Scholtz, Michelle Steves, and Emile Morse National Institute of Standards and Technology 100 Bureau Drive, MS 8940 Gaithersburg, MD , , Evaluation of Collaboration Support for the Intelligence Community In the United State, the events of 9/11 have triggered analysis of the intelligence community. Among the recommendations from the National Commission on Terrorist Attacks Upon the United States are the recommendations on unity of effort on the challenge of counterterrorism, including: unifying the many participants in the counterterrorism effort and their knowledge in a networkbased information sharing system that transcends traditional government boundaries 1 Although the work presented in this paper, has been carried out specifically in the United States, there are almost certain to be identical requirements in other countries. In the paper, we focus primarily on US aspects and leave to international readers to draw any parallels in their home countries. The problem US Government agencies today face a number of tensions. First, each agency needs to carryout their particular mission, while interacting appropriately with other agencies. An agency may rely on other agencies for input and may produce output for other agencies. They need to collaborate to ensure a high quality of work, but they need to retain their identity and mission focus. To that extent, the various agencies within the United States that produce intelligence focus on different aspects of counter terrorism, ranging from different types of data to different types of questions for different purposes. Today we can collect much more data than can be analyzed. Even sorting massive amounts of data to locate relevant data is a huge task. To make the task even more difficult, some of the information is meant to be misleading and deceptive. To work more effectively, we need internal collaboration (within agencies) and external collaboration (between agencies). Collaborating with people who we know is difficult enough. Locating and collaborating with people we don t know is extraordinarily difficult. During the past several years, we have been working as consultants advising on methodologies and metrics for evaluation of tools for the analytic community. One area that we have been evaluating is the use of collaborative tools and the impact they have on various processes. During the course of this work, we have installed collaboration tools on an experimental network available to a number of agencies. This work is both exciting and challenging as we have been given the opportunity to evaluate collaborative software use in complex real-world environments. With this, however, come a number of restrictions. We have used both qualitative and quantitative data to assess the use of the collaboration software and the impact on the various processes where it has been used. There are several issues: first, we are trying to evaluate this software in an operational environment. There are security restrictions on what we can collect and what we can report. There are constraints on analyst time. At times we are able to collect baseline data about processes and at other times, the software is being used in a new process for which no baseline exists. We are opportunistic about the experiments that we conduct, and that sometimes implies that there is no opportunity to collect baseline information. We, however, need to be able to draw some conclusions about collaboration software based on experiments with different users, for different purposes, and at different times. Currently, we have only one 1 Page 21, Executive Summary, The 9/11 Commission Report, National Commission of Terrorist Attacks Upon the United States, 2004,
2 collaboration tool available, but we should be installing another soon. How can we compare collaboration tools used in different agencies by different people? To address these questions, we began by developing a metrics model. The Metrics Model Our model consists of five elements or levels [3,4]. The levels are: system goals, evaluation objectives, conceptual metrics, and measures, both conceptual and implementation-specific. Each successive element in the model is a refinement in scope. This refinement can also be thought of as a level of abstraction or sphere of concern. For example, a system goal may have one or more evaluation goals. Each evaluation goal will have one or more conceptual metrics, which when assessed with its associated measures, will contribute to an assessment of if, and potentially how well, a particular system goal was met. The system goal is the intended benefit or functionality the software system will provide. This goal may reflect a technical performance goal for the software system or an organizational goal, e.g., process or product improvement. The system goal element provides two critical ingredients of system evaluation design: determining which aspect(s) of the system are of primary importance in the work domain and providing high-level questions the evaluation must answer whether the stated goals were met. The next level is the evaluation objective level. Each objective drives a particular part of an evaluation, within the context its associated system goal. Evaluation objectives (EO) partition evaluation concerns in the overall assessment of a particular goal. Clear articulation of these objectives provides further refinement in the top-down design of the evaluation. The lowest three levels in the model constitute metrics and measures. Often these terms are used inter changeably, however, we define the terms in relation to each other and describe how to distinguish them. We define metric as the interpretation of one or more contributing elements, e.g., measures or other metrics, corresponding to the degree to which the set of attribute elements affects its quality. Interpretation can be a computation or human assessment of the contributing elements. The computation can be the summed, weighted values, of the contributing elements. When weights are used, they are assigned to elements according to their importance to the respective attribute with respect to the particular assessment scenario. We define a measure, a noun, as a performance indicator that can be observed singly or collectively, computed, calculated, and may, at times, be automatically collected. A simple way to distinguish between metrics and measures is by the following statement: a measure is an observable value, while a metric associates meaning to that value by applying human judgment, often through a formula based on weighted values using the contributing measures. The lowest two levels in the model constitute measures. These levels represent measures that are required to substantiate the formulation of answers to evaluation questions. There are two levels of abstraction for measures: conceptual and implementation-specific. Conceptual measures identify the type of data to be collected. Implementation-specific measures identify the specifics about a particular collection instance, e.g., data element(s), associated tool, collection method, and so on. The figure below represents the metrics model
3 System goals Evaluation Objectives Conceptual Metrics Conceptual measures Implementation Specific measures A portion of the metrics model for evaluating collaboration is shown below. For each of these conceptual metrics that is appropriate for a given experiment, we specify implementation specific ways of obtaining this. The model is also dynamic. As we work on new processes and new software tools, we will certainly uncover new measures to collect.
4 Case Study One We are currently analyzing the use of Groove software used in a process for the past 1 ½ years. While we will not be able to report the results at this workshop due to the nature of the data, we will be able to discuss the success of the evaluation. To analyze the use of the collaboration software we are currently collecting and analyzing: - collaboration software logs - questionnaires - spreadsheets containing process data The collaboration software logs will let us examine: The overall use of the software The different components used in the software Usage/ individuals involved in the process Usage/ roles of individuals involved in the process The spreadsheets of process data contain dates that input was requested, dates that products were produced, people formally assigned to different requests. This data will give us Time to perform tasks Efficiency over time Effectiveness over time as reflected in number of requests Questionnaires will help us determine usage and satisfaction by analysts, supervisors, and consumers of analytic products. They will also help us determine the characteristics of the collaboration groups and provide information about user satisfaction and product quality and timeliness. By combining the software logs with information from the spreadsheets, we hope to see collaborative activities in supporting different requests. There are, however, caveats. Because we are looking at real world data, we must consider what is happening in the real world at the various times frames we are examining. External situations may contribute to increases or decreases in requests, and we must plot this type of information against our trends in order to properly interpret our data. Case Study Two We have been analyzing data from four open source analysts, each working on different tasks, for the past year. Prior work [2] has been conducted on two senior open source analysts each doing a single monitoring task over a one year period. We have collected data both through the user of Glass Box software [1] and by observations. While there was some collaboration happening in these activities, it was not a main focus. However, in the experiments the first part of the year, we will be requiring the analysts to use collaborative software and perform analytic tasks that require collaboration. We will be able to collect both Glass Box data and software logs from the collaboration software. We have a number of measures of process including: o Time spent in various phases of analysis o Growth of report documents over time o Number of relevant documents/ task o Comparisons of above based on length of task/ type of report requested o Time spent using various software applications o Ratings of task complexity We will be able to compare these measures for a non-collaborative analysis with those for collaborative work. This will enable us to determine changes in process. Additionally, we collect the products that each analyst generates. As we have access to a number of senior analysts, we can also evaluate the quality of the products. While the actual logged data will most likely not be analyzed by the time of the CHI workshop, we will have a number of observational studies that we can report on for the workshop.
5 Summary We are conducting evaluations of collaborative software for intelligence analysts in operational environments. Because it is the real world, we are able to see how users want to use the software and the different types of processes that need collaboration support. We have been able to provide feedback to developers concerning the usability and utility of their products. We collect both qualitative and quantitative data. The intelligence analysts have extremely limited time so we use the quantitative data to refine our qualitative data collection. It is also essential that we understand the various processes used in the different agencies so we can focus on collecting information about impacts on those processes due to collaborative software tools. Because it is the real world, there are limitations to the data we can collect due to security, access, and time limitations. As we do experiments involving different agencies and different software solutions, we developed a metrics model to attempt to bring some systematic data collection to a sometimes chaotic process. We are currently using this metrics model to analyze 1 ½ years of a collaborative process. By the time of the workshop, we will be able to report on the success of the evaluation methodology and recommendations. Due to the nature of the work, we will only be able to report general results from the actual analysis of the collaboration tool. One of the methodological questions we are exploring is the utility of the metrics model for different stages of software development and deployment. References 1. Cowley, P., Nowell, L., and Scholtz, J Glass Box: An Instrumented Infrastructure for Supporting Human Interaction with Information. HICSS 38, Jan Hawaii. 2. Scholtz, J., Morse, E., and Hewett, T In Depth Observational Studies of Professional Intelligence Analysts. Presented at Human Performance, Situation Awareness, and Automation Conference, March Daytona Beach, Fl. 3. Steves, M. and Scholtz, J A Framework for Real-World Software System Evaluations. CSCW. 4. Steves, M. and Scholtz, J A Framework for Evaluating Collaborative Systems in the Real World. HICSS 38, Jan Hawaii.
Building a Data Quality Scorecard for Operational Data Governance
Building a Data Quality Scorecard for Operational Data Governance A White Paper by David Loshin WHITE PAPER Table of Contents Introduction.... 1 Establishing Business Objectives.... 1 Business Drivers...
Practice Overview. REQUIREMENTS DEFINITION Issue Date: <mm/dd/yyyy> Revision Date: <mm/dd/yyyy>
DEPARTMENT OF HEALTH AND HUMAN SERVICES ENTERPRISE PERFORMANCE LIFE CYCLE FRAMEWORK PRACTIICES GUIIDE REQUIREMENTS DEFINITION Issue Date: Revision Date: Document
CDC UNIFIED PROCESS PRACTICES GUIDE
Document Purpose The purpose of this document is to provide guidance on the practice of Requirements Definition and to describe the practice overview, requirements, best practices, activities, and key
SECURITY METRICS: MEASUREMENTS TO SUPPORT THE CONTINUED DEVELOPMENT OF INFORMATION SECURITY TECHNOLOGY
SECURITY METRICS: MEASUREMENTS TO SUPPORT THE CONTINUED DEVELOPMENT OF INFORMATION SECURITY TECHNOLOGY Shirley Radack, Editor Computer Security Division Information Technology Laboratory National Institute
Paper 064-2014. Robert Bonham, Gregory A. Smith, SAS Institute Inc., Cary NC
Paper 064-2014 Log entries, Events, Performance Measures, and SLAs: Understanding and Managing your SAS Deployment by Leveraging the SAS Environment Manager Data Mart ABSTRACT Robert Bonham, Gregory A.
White Paper Operations Research Applications to Support Performance Improvement in Healthcare
White Paper Operations Research Applications to Support Performance Improvement in Healthcare Date: April, 2011 Provided by: Concurrent Technologies Corporation (CTC) 100 CTC Drive Johnstown, PA 15904-1935
Requirements Engineering: Elicitation Techniques
2008:PR003 Requirements Engineering: Elicitation Techniques Sai Ganesh. Gunda Source:http://www.marcocioffi.com/archives/2005/04/requirements-engineering/ MASTER S THESIS Software Engineering, 2008 Department
ANALYZING YOUR RESULTS
ANALYZING YOUR RESULTS Kelley Jarrett, Market Manager Design Billy Stephens, Account Manager Design Measuring performance and learning from your audiences behaviors and feedback. REVISIT BENCHMARKS
A Case Study in Software Enhancements as Six Sigma Process Improvements: Simulating Productivity Savings
A Case Study in Software Enhancements as Six Sigma Process Improvements: Simulating Productivity Savings Dan Houston, Ph.D. Automation and Control Solutions Honeywell, Inc. [email protected] Abstract
Business Intelligence Not a simple software development project
Business Intelligence Not a simple software development project By V. Ramanathan Executive Director, Saksoft Date: Mar 2006 India Phone: +91 44 2461 4501 Email: [email protected] USA Phone: +1 212 286 1083
GQM + Strategies in a Nutshell
GQM + trategies in a Nutshell 2 Data is like garbage. You had better know what you are going to do with it before you collect it. Unknown author This chapter introduces the GQM + trategies approach for
Software Quality Management
Software Lecture 9 Software Engineering CUGS Spring 2011 Kristian Sandahl Department of Computer and Information Science Linköping University, Sweden A Software Life-cycle Model Which part will we talk
How To Develop Software
Software Engineering Prof. N.L. Sarda Computer Science & Engineering Indian Institute of Technology, Bombay Lecture-4 Overview of Phases (Part - II) We studied the problem definition phase, with which
Shew-Fang Shieh, RN. MS. DBA. Cardinal Tien Hospital, Taiwan August-03-2012
Shew-Fang Shieh, RN. MS. DBA. Cardinal Tien Hospital, Taiwan August-03-2012 1 New Taipei City 2 Organizational performance is a key factor to influence the development of health care organization. Thus,
White Paper from Global Process Innovation. Fourteen Metrics for a BPM Program
White Paper from Global Process Innovation by Jim Boots Fourteen Metrics for a BPM Program This white paper presents 14 metrics which may be useful for monitoring progress on a BPM program or initiative.
Automation Sales/Marketing Professional
Automation Sales/Marketing Professional Alternate Titles: Inside sales: Customer Service Representative, Inside Sales Engineer, Technical Support Specialist Outside sales: Field Sales, Automation Product
WHITE PAPER. The Five Fundamentals of a Successful FCR Program
The Five Fundamentals of a Successful FCR Program April 2012 Executive Summary Industry analysts agree that First Contact Resolution (FCR) is the best way to measure the effectiveness of your contact center.
Measurement Information Model
mcgarry02.qxd 9/7/01 1:27 PM Page 13 2 Information Model This chapter describes one of the fundamental measurement concepts of Practical Software, the Information Model. The Information Model provides
Requirements engineering
Learning Unit 2 Requirements engineering Contents Introduction............................................... 21 2.1 Important concepts........................................ 21 2.1.1 Stakeholders and
Performance Management Systems: Conceptual Modeling
2011 International Conference on Economics and Business Information IPEDR vol.9 (2011) (2011) IACSIT Press, Bangkok, Thailand Performance Management Systems: Conceptual Modeling Dmitry Isaev Business Analytics
Fusion Center Technology Resources Road Map: Elements of an Enterprise Architecture for State and Major Urban Area Fusion Centers
Fusion Center Technology Resources Road Map: Elements of an Enterprise for State and Major Urban Area Fusion Centers General Overview The function of a fusion center is to provide a collection, analysis,
Situational Awareness Through Network Visualization
CYBER SECURITY DIVISION 2014 R&D SHOWCASE AND TECHNICAL WORKSHOP Situational Awareness Through Network Visualization Pacific Northwest National Laboratory Daniel M. Best Bryan Olsen 11/25/2014 Introduction
The SPES Methodology Modeling- and Analysis Techniques
The SPES Methodology Modeling- and Analysis Techniques Dr. Wolfgang Böhm Technische Universität München [email protected] Agenda SPES_XT Project Overview Some Basic Notions The SPES Methodology SPES_XT
White Paper Combining Attitudinal Data and Behavioral Data for Meaningful Analysis
MAASSMEDIA, LLC WEB ANALYTICS SERVICES White Paper Combining Attitudinal Data and Behavioral Data for Meaningful Analysis By Abigail Lefkowitz, MaassMedia Executive Summary: In the fast-growing digital
BI Dashboards the Agile Way
BI Dashboards the Agile Way Paul DeSarra Paul DeSarra is Inergex practice director for business intelligence and data warehousing. He has 15 years of BI strategy, development, and management experience
Background: Business Value of Enterprise Architecture TOGAF Architectures and the Business Services Architecture
Business Business Services Services and Enterprise and Enterprise This Workshop Two parts Background: Business Value of Enterprise TOGAF s and the Business Services We will use the key steps, methods and
Chapter 1 DECISION SUPPORT SYSTEMS AND BUSINESS INTELLIGENCE
Chapter 1 DECISION SUPPORT SYSTEMS AND BUSINESS INTELLIGENCE Learning Objectives Understand today s turbulent business environment and describe how organizations survive and even excel in such an environment
THE PRESIDENT S NATIONAL SECURITY TELECOMMUNICATIONS ADVISORY COMMITTEE
THE PRESIDENT S NATIONAL SECURITY TELECOMMUNICATIONS ADVISORY COMMITTEE BIG DATA ANALYTICS SCOPING REPORT July 20, 2015 DRAFT THE PRESIDENT S NATIONAL SECURITY TELECOMMUNICATIONS ADVISORY COMMITTEE BIG
WHITEPAPER. Creating and Deploying Predictive Strategies that Drive Customer Value in Marketing, Sales and Risk
WHITEPAPER Creating and Deploying Predictive Strategies that Drive Customer Value in Marketing, Sales and Risk Overview Angoss is helping its clients achieve significant revenue growth and measurable return
The Role of Information Technology Studies in Software Product Quality Improvement
The Role of Information Technology Studies in Software Product Quality Improvement RUDITE CEVERE, Dr.sc.comp., Professor Faculty of Information Technologies SANDRA SPROGE, Dr.sc.ing., Head of Department
SLDS Workshop Summary: Data Use
SLDS Workshop Summary: Data Use Developing a Data Use Strategy This publication aims to help states detail the current status of their State Longitudinal Data System (SLDS) data use strategy and identify
Customer Experience Strategy and Implementation
Customer Experience Strategy and Implementation Enterprise Customer Experience Transformation 2014 Andrew Reise, LLC. All Rights Reserved. Enterprise Customer Experience Transformation Executive Summary
Better decision making under uncertain conditions using Monte Carlo Simulation
IBM Software Business Analytics IBM SPSS Statistics Better decision making under uncertain conditions using Monte Carlo Simulation Monte Carlo simulation and risk analysis techniques in IBM SPSS Statistics
Assuming the Role of Systems Analyst & Analysis Alternatives
Assuming the Role of Systems Analyst & Analysis Alternatives Nature of Analysis Systems analysis and design is a systematic approach to identifying problems, opportunities, and objectives; analyzing the
Session-1: Business Enterprise Applications- Overview
Session-1: Business Enterprise Applications- Overview Important Points to Note All Participating colleges are requested to mute your telephone lines during the webinar session. Participants are requested
INTERMEDIATE QUALIFICATION
PROFESSIONAL QUALIFICATION SCHEME INTERMEDIATE QUALIFICATION SERVICE LIFECYCLE CONTINUAL SERVICE IMPROVEMENT CERTIFICATE SYLLABUS Page 2 of 18 Document owner The Official ITIL Accreditor Contents CONTINUAL
Enterprise Risk Management
Cayman Islands Society of Professional Accountants Enterprise Risk Management March 19, 2015 Dr. Sandra B. Richtermeyer, CPA, CMA What is Risk Management? Risk management is a process, effected by an entity's
Strategic Outcome- Based Metrics for the Federal Government
RG Perspective Strategic Outcome- Based Metrics for the Federal Government 11 Canal Center Plaza Alexandria, VA 22314 HQ 703-548-7006 Fax 703-684-5189 www.robbinsgioia.com 2013 Robbins Gioia, Inc. 1. Introduction
How To Write A Book On Risk Management
National Center for Risk and Economic Analysis of Terrorism Events CREATE FY2015 (Year 11) Call for White Papers CREATE, the DHS-sponsored Center of Excellence at the University of Southern California,
Creating An Excel-Based Balanced Scorecard To Measure the Performance of Colleges of Agriculture
Creating An Excel-Based Balanced Scorecard To Measure the Performance of Colleges of Agriculture Paper Presented For American Agricultural Economics Association (AAEA) Annual Meeting July 23-26, 2006 Long
Computing & Communications Services
2010 Computing & Communications Services 2010 / 10 / 04 Final Kent Percival, M.Sc., P.Eng. Defining the Value of the Business Analyst In achieving its vision, key CCS partnerships involve working directly
The Integrated Planning and Performance Framework
The Integrated Planning and Performance Framework Issue This briefing note provides an overview of the City's Integrated Planning Multi-Year Budgeting and Performance Framework and addresses four major
Fan Fu. Usability Testing of Cloud File Storage Systems. A Master s Paper for the M.S. in I.S. degree. April, 2013. 70 pages. Advisor: Robert Capra
Fan Fu. Usability Testing of Cloud File Storage Systems. A Master s Paper for the M.S. in I.S. degree. April, 2013. 70 pages. Advisor: Robert Capra This paper presents the results of a usability test involving
Five best practices for deploying a successful service-oriented architecture
IBM Global Services April 2008 Five best practices for deploying a successful service-oriented architecture Leveraging lessons learned from the IBM Academy of Technology Executive Summary Today s innovative
Key performance indicators
Key performance indicators Winning tips and common challenges Having an effective key performance indicator (KPI) selection and monitoring process is becoming increasingly critical in today s competitive
Develop Project Charter. Develop Project Management Plan
Develop Charter Develop Charter is the process of developing documentation that formally authorizes a project or a phase. The documentation includes initial requirements that satisfy stakeholder needs
Chapter 10 Practical Database Design Methodology and Use of UML Diagrams
Chapter 10 Practical Database Design Methodology and Use of UML Diagrams Copyright 2011 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 10 Outline The Role of Information Systems in
Framework for Improving Critical Infrastructure Cybersecurity
Framework for Improving Critical Infrastructure Cybersecurity Implementation of Executive Order 13636 8 April 2015 [email protected] Agenda Mission of NIST Cybersecurity at NIST Cybersecurity Framework
Business Intelligence and Decision Support Systems
Chapter 12 Business Intelligence and Decision Support Systems Information Technology For Management 7 th Edition Turban & Volonino Based on lecture slides by L. Beaubien, Providence College John Wiley
Proposal for the Assessment of Distance Education (DE) Programs in Nursing
Assessment of Distance Programs in Nursing Lois Magnussen and Joanne Itano, Department of Nursing This project focused on using the Flashlight Program to evaluate educational uses of the WWW in nursing.
ISTQB Certified Tester. Foundation Level. Sample Exam 1
ISTQB Certified Tester Foundation Level Version 2015 American Copyright Notice This document may be copied in its entirety, or extracts made, if the source is acknowledged. #1 When test cases are designed
PANORATIO. Big data : Benefits of a strategic vision. White Paper June 2013. Executive Summary
PANORATIO Big data : Benefits of a strategic vision White Paper June 2013 Executive Summary Following the massive deployment of new mobile technologies and social media, sources of data regarding organizations
A CASE STUDY ON SOFTWARE PROJECT MANAGEMENT IN INDUSTRY EXPERIENCES AND CONCLUSIONS
A CASE STUDY ON SOFTWARE PROJECT MANAGEMENT IN INDUSTRY EXPERIENCES AND CONCLUSIONS P. Mandl-Striegnitz 1, H. Lichter 2 1 Software Engineering Group, University of Stuttgart 2 Department of Computer Science,
Senior Executive Service (SES) Performance Appraisal System Training
Senior Executive Service (SES) Performance Appraisal System Training 1 Objectives After this briefing, you will Understand the basis and benefits of the new SES performance system. Understand how to develop
Telecommunication (120 ЕCTS)
Study program Faculty Cycle Software Engineering and Telecommunication (120 ЕCTS) Contemporary Sciences and Technologies Postgraduate ECTS 120 Offered in Tetovo Description of the program This master study
Qualitative data acquisition methods (e.g. Interviews and observations) -.
Qualitative data acquisition methods (e.g. Interviews and observations) -. Qualitative data acquisition methods (e.g. Interviews and observations) ( version 0.9, 1/4/05 ) Code: data-quali Daniel K. Schneider,
Levels of Software Testing. Functional Testing
Levels of Software Testing There are different levels during the process of Testing. In this chapter a brief description is provided about these levels. Levels of testing include the different methodologies
Instructional systems development
A Pebble-in-the-Pond Model For Instructional Design by M. David Merrill Instructional systems development (ISD) has recently come under attack to suggestions that it may not be an appropriate methodology
ISO/IEC 9126-1 Software Product Quality Model
Why do current systems fail? Standish Group found that 51% of projects failed 31% were partially successful Main causes were poor user requirements: 13.1% Incomplete requirements 12.4% Lack of user involvement
This alignment chart was designed specifically for the use of Red River College. These alignments have not been verified or endorsed by the IIBA.
Red River College Course Learning Outcome Alignment with BABOK Version 2 This alignment chart was designed specifically for the use of Red River College. These alignments have not been verified or endorsed
1.0 Introduction and Report Overview
1.0 Introduction and Report Overview Customer relationship management (CRM) represents the key business strategy that will determine successful enterprises in the 21st century. Enterprises engaged in CRM
The following is intended to outline our general product direction. It is intended for information purposes only, and may not be incorporated into
The following is intended to outline our general product direction. It is intended for information purposes only, and may not be incorporated into any contract. It is not a commitment to deliver any material,
ENERGY SECTOR CYBERSECURITY FRAMEWORK IMPLEMENTATION GUIDANCE
ENERGY SECTOR CYBERSECURITY FRAMEWORK IMPLEMENTATION GUIDANCE JANUARY 2015 U.S. DEPARTMENT OF ENERGY OFFICE OF ELECTRICITY DELIVERY AND ENERGY RELIABILITY Energy Sector Cybersecurity Framework Implementation
CHAPTER - 5 CONCLUSIONS / IMP. FINDINGS
CHAPTER - 5 CONCLUSIONS / IMP. FINDINGS In today's scenario data warehouse plays a crucial role in order to perform important operations. Different indexing techniques has been used and analyzed using
POLAR IT SERVICES. Business Intelligence Project Methodology
POLAR IT SERVICES Business Intelligence Project Methodology Table of Contents 1. Overview... 2 2. Visualize... 3 3. Planning and Architecture... 4 3.1 Define Requirements... 4 3.1.1 Define Attributes...
Data Analytics: Exploiting the Data Warehouse
Data Analytics: Exploiting the Data Warehouse Helena Galhardas DEI/IST References A. Vaisman and E. Zimányi, Data Warehouse Systems: Design and Implementation, Springer, 2014 (chpt 9) 2 1 Outline Data
OFFICE OF HOMELAND SECURITY AND PREPAREDNESS
First Name: Middle Initial: Last Name: Email Address: Home Phone Number: Cell Phone Number: Current Address: Street: City: State: Zip Code: County (if NJ) College/University currently enrolled in: Highest
LINKING INDIVIDUAL PERFORMANCE TO ORGANIZATIONAL GOALS. By Joseph Taracha [email protected]
LINKING INDIVIDUAL PERFORMANCE TO ORGANIZATIONAL GOALS By Joseph Taracha [email protected] INTRODUCTION Goal management can help organizations manage employee performance in a strategic, effective way.
Assessing Schedule Performance with Earned Value
FEATURED PAPER OCTOBER 2007 Time Is Up: Assessing Schedule Performance with Earned Value By Robert Van De Velde Time drives projects. Quick and accurate evaluation of schedule performance is crucial. Quantitative
Building a Digital. Create Value by Integrating Analytical Processes, Technology, and People into Business Operations.
Building a Digital Analytics Organization: Create Value by Integrating Analytical Processes, Technology, and People into Business Operations Judah Phillips Table of Contents Chapter 1 Using Digital Analytics
2.1 Initiation Phase Overview
2.1 Initiation Phase Overview The is the conceptualization of the project. This section describes the basic processes that must be performed to get a project started. Accordingly, the purpose of the is
Organizational Requirements Engineering
Chapter 9, Non-functional Requirements Organizational Requirements Engineering Prof. Dr. Armin B. Cremers Sascha Alda Armin B. Cremers, Sascha Alda Organizational Requirements Engineering 1 Overview of
Industrial Rapid Implementation Methodology (InRIM)
Industrial Rapid Implementation Methodology (InRIM) Introduction The path to improving asset effectiveness and productivity must be based on a firm foundation created from the interaction of three essential
Metadata-Based Project Management System. A Case Study at M-Files Corporation. Iulia Adomnita
Metadata-Based Project Management System. A Case Study at M-Files Corporation Iulia Adomnita University of Tampere School of Information Sciences Computer Science M.Sc. Thesis Supervisors: Timo Poranen,
ENTERPRISE RISK MANAGEMENT FRAMEWORK
ENTERPRISE RISK MANAGEMENT FRAMEWORK COVENANT HEALTH LEGAL & RISK MANAGEMENT CONTENTS 1.0 PURPOSE OF THE DOCUMENT... 3 2.0 INTRODUCTION AND OVERVIEW... 4 3.0 GOVERNANCE STRUCTURE AND ACCOUNTABILITY...
How To Create A Process Measurement System
Set Up and Operation of a Design Process Measurement System Included below is guidance for the selection and implementation of design and development process measurements. Specific measures can be found
GAININSIGHT. Custom Report Development Methodology. Requirements Scoping and Project Planning Process
GAININSIGHT Custom Report Development Methodology Requirements Scoping and Project Planning Process Symmetrics Custom Report Development Methodology A process to ensure maximum insight is gained from your
U.S. Defense Priorities OSD PA&E
1 U.S. Defense Priorities Deter potential adversaries and defend America and American interests Counter asymmetric threats including terrorism, cyber attacks and ballistic and cruise missiles Fight and
Forward Thinking for Tomorrow s Projects Requirements for Business Analytics
Seilevel Whitepaper Forward Thinking for Tomorrow s Projects Requirements for Business Analytics By: Joy Beatty, VP of Research & Development & Karl Wiegers, Founder Process Impact We are seeing a change
Position Classification Standard for Management and Program Clerical and Assistance Series, GS-0344
Position Classification Standard for Management and Program Clerical and Assistance Series, GS-0344 Table of Contents SERIES DEFINITION... 2 EXCLUSIONS... 2 OCCUPATIONAL INFORMATION... 3 TITLES... 6 EVALUATING
7 Best Practices for Speech Analytics. Autonomy White Paper
7 Best Practices for Speech Analytics Autonomy White Paper Index Executive Summary 1 Best Practice #1: Prioritize Efforts 1 Best Practice #2: Think Contextually to Get to the Root Cause 1 Best Practice
NASCIO EA Development Tool-Kit Solution Architecture. Version 3.0
NASCIO EA Development Tool-Kit Solution Architecture Version 3.0 October 2004 TABLE OF CONTENTS SOLUTION ARCHITECTURE...1 Introduction...1 Benefits...3 Link to Implementation Planning...4 Definitions...5
Holistic Development of Knowledge Management with KMMM
1 Karsten Ehms, Dr. Manfred Langen Holistic Development of Knowledge Management with KMMM Siemens AG / Corporate Technology Knowledge Management & Business Transformation If knowledge management is to
