Data Governance at Chevron GOM
|
|
|
- Ashlee Copeland
- 10 years ago
- Views:
Transcription
1 Data Governance at Chevron GOM A Case Study Authors: Dave Blosser Chevron North America Exploration and Production Paul Haines Noah Consulting PNEC 17 th International Conference On Petroleum Data Integration, Data and Information Management May 2013
2 Page 2 Table of Contents 1.0 Abstract: Problem Statement: Approach to Data Governance: Implementation of Data Governance: Building Sustainment Through a DG Organization: Beyond Data Governance as a Project: Summary: About the authors: About the companies: 34
3 Page Abstract: Data governance (DG) is one of the pillars of data management (DM) and is often considered hand-in-hand with data quality efforts and master data management programs. In our exploration and production (E&P) community, initiatives to implement data governance are often seen as either organizational efforts or technology efforts and are often spun up from the corporate office with a top-down emphasis. That approach, although well intended and with valuable recommendations, is often met with internal resistance and suspicion and falls short when it comes to implementing and sustaining data governance to make a truly significant and lasting impact within the business. The co-authors present this case study to demonstrate how data governance can be implemented in a unique and pragmatic federated manner. Our approach emphasizes a repeatable and sustainable methodology focused on supporting key business processes. We will review the methodology, components, and stages developed to implement data governance for specific data types through several proof of concepts. We will present lessons learned, challenges encountered, and business benefits realized through our efforts to date.
4 Page Problem Statement: Many papers have been presented in this conference that relate to data issues, data quality challenges, data inconsistencies, untrusted data, etc. At Chevron s Gulf of Mexico Business Unit (GOM or GOMBU), end-user surveys showed general dissatisfaction with data quality. There were known issues related to data access and data delivery to the business user. Initial assessments that led to this project showed inconsistent systems of record (SORs) and/or too many sources of data. Additionally, historical data tended to get lost or disorganized during post merger or acquisition activities. Essentially, Chevron GOM had problems similar to many other E&P companies. This section will not be the focus of this paper, but rather a short introduction so that the audience can relate to the historical situation. Chevron GOM had well sponsored data projects and data clean-up projects to address the challenges stated above. Data projects started but often lacked specific goals that ensured sustainment. Data projects were performed to make it clean and the data quality was improved, but it didn t stay clean. There was not a good understanding of the root cause; there was not a good understanding of how to keep it clean. Data cleanup can happen by brute force and short term improvements will be seen, especially if coupled with improved business processes, but process improvement is not enough. The proper implementation of DG makes it work; makes it sustainable. That is the conclusion that Chevron GOM came to. Over time and through evangelical meetings, the business came to agree that data cleanup efforts (remediation) were not enough. You can improve Data Quality for a short period of time, but without data governance, it will not last. There was a point in time where IT and the business said, We ve been talking about data governance for a long time let s do something. Through initial efforts and as a DG project was launched, it became clear that a majority of the business did not understand what DG really entailed. Business leaders knew what they wanted (they wanted trusted, clean data), but they had a misguided concept on how to get there. Business, management, and IT did not understand the fundamentals of how to do DG. Also, the project sponsors and senior business managers needed to see something real; something that would help them realize the value DG can provide. That realization is what led to the decision that formal training in the fundamentals, concepts, and framework was required a DG Academy was the first step.
5 Page Approach to Data Governance: This section will describe the overall DG process, the thinking that went into the DG Academy, and the project structure. Data governance along with data stewardship and data quality together provide the foundation for properly managing and sustaining information as a key asset in an E&P company. DG provides a framework of principles, policies, standards, roles and responsibilities, and processes to enable effective information management. DG lets the business take proper and responsible ownership of their information where everyone clearly understands their role and the data they are responsible for. Data governance includes clear descriptions of the tasks and activities that the business must perform to achieve and maintain known data quality. As will be shown, data governance refers to people and organizational capability, processes and controls, and technology and architecture. A proper balance between people, process, and technology is what enables data governance to be successful. There are clear and specific DG tasks and activities that can be described and documented and DG roles that need to be defined. Data governance activities include: Apply data policies Resolve cross functional or cross regional data issues Define data and business rules Maintain end-to-end responsibility for at least one subject area Monitor and sustain quality through data governance policies Define standardized data-related business performance measures Work with data quality (DQ) groups to ensure compliance with established standards and definitions Provide input into DQ program prioritization
6 Page 6 Additionally, in each client organization specific DG roles need to be defined and responsibilities documented. This is essential to establish a DG organizational capability. An example of a listing of roles might include: Data content owner Data steward Data custodian Data architect Data quality analyst Data technical analyst Data business analyst Data producer Data consumer Additionally, there are roles that provide sponsorship, steering, guidance, and/or pace and priority to the DG program. What does it mean for data to be governed? Since no one seemed to have a good definition, we developed our DG criteria to include those elements which, if true, give us confidence that our data is well governed. Thinking of this in terms of a check list, the four key areas to describe include: Standardization o Are SORs identified and in place? o Have standards been established and documented? o Are data models architected and documented? o Is secure access managed? Roles and responsibilities o Are data producers and data consumers identified? o Are data stewards identified, trained, and their role communicated? o Are data analysts and data custodians engaged? Process o Is master data managed? o Have business processes, DG processes, and business/data rules been applied? Data quality o Are data rules tied to business processes? o Are data rules tied to, mapped, and documented to appropriate repositories? o Is data monitored to a standard?
7 Page 7 These criteria have been established during this project to determine when data is governed to an agreed upon quality standard. The establishment of a DG Academy began as a way to bring the business up to a proper level of understanding of DG, what value it could provide, and how Chevron GOM would deliver DG and its associated standards. The DG project team wanted to create a reality based understanding of DG among IT project teams that were delivering data-related IT projects. The DG Academy was thought of as a way to effectively communicate what DG is, why it is important, and how to effectively implement it at a high level. Another question contemplated by the project team was how to administer the DG Academy. The strategy relating to the administration of the DG Academy included: Employ a light touch methodology. Create a classroom setting to establish DG concepts. Engage the participants in workshops to provide hands-on practice in the concepts. Through the DG Academy, present the DG concepts in a consistent manner. Use specific collateral relating to the data in the DG project. Use the data project team to bring out existing DQ experiences, anecdotes, and usage needs that would boost the value of the resultant standards. Leverage industry standards such as standard definitions created and accepted by the PPDM Business Rules Workgroup; of which both Chevron and Noah were contributing members. Create 1 st draft collateral during the workshops to accelerate the development of the standards and engage the participants from the data projects.
8 Page 8 Another strategic decision made for the development of DG within Chevron GOM was related to the overall approach. The overall approach focuses on key business processes as the primary dimension for developing data governance to sustain data quality. Key business processes are important because they can be used to reflect real business workflow problems and inefficiencies and allow us a pragmatic method to tie low level data issues to high level business concepts. It has been seen that poor quality data, missing data, etc. cause key business processes to break, fail, or be sub-optimal. The methodology includes: Describe key business processes that can be related to data flow. Map out these key business processes in standard business process workflow diagrams. Identify business rules that support the business data and/or business decisions where: o gaps in process are caused by a lack of standards or lack of business rules; o data quality issues are causing the process to fail or not be optimal; o business rules affect the design of the workflow; o the business rules are at a level that is data model and application independent. Identify specific data rules that support the business rules; these data rules will be at a granularity that can be acted upon to identify, understand, and rectify the data quality issues impeding the business process. The data rules are specific to a data type or data attribute and can be mapped to the SOR or application. The data rules should be such that a test can be created and administered that supports the data rule and can be executed manually or automatically. The business processes, business rules, and data rules are fully defined and attributed so they can be managed within the DG environment and DQ framework. Leverage industry standards such as the business rule and data rule attribution that came from the PPDM Business Rules Workgroup.
9 Page 9 The delivery of the DG Academy was designed to be: flexible, but comprehensive; taught in compact, segmented timeframe so concepts would build upon each other; split into tutorials and workshops (about 60% / 40% ratio). The workshops: o helped to solidify concepts into the participants working knowledge o helped to begin developing the variety of standards needed for creating the sustainable DG environment and DQ framework Once some of these (and other) fundamental concepts had been established within the DG project team, decisions needed to be made related to the details of the content and how it would be delivered within Chevron GOM. For that, the DG project team decided to bring in information management consultants who had experience in successfully delivering DG and DQ in the E&P industry and would be able to provide resources, thought leadership, and focus. These resources were added to the DG project team. There are various ways to begin a DG effort; there were active projects working to establish new SORs and an associated data cleanup effort in progress. The projects did not include an element of DG in their scope. We knew these projects were doomed to ultimate failure. It simply made sense to partner with these projects to enable their success and to provide fertile ground for the DG project team to test their ideas. There were several advantages to this: The data group worked closely with the business on their highest priority issues. The initiatives were business focused and business led. The initiatives were business and IT enabled both groups had skin in the game There was strong business acumen on the project. Some of these issues were directly related to data and data quality. A deliverable to provide a sustained data quality improvement was put into the group s project charters so the DG project was seen as an enabler to meet the group goals. The projects contemplated were data and/or workflow centric. It was anticipated that working with and training the data group in the fundamentals of DG and helping them establish a sustainable framework for their data types was a clear win for both sides. As the methodology was being developed and refined, both the structure of the project and the format of the DG Academy were maturing. Following after the Chevron project execution process structure, the project was divided into 6 stages corresponding to four of the Chevron process phases represented in the graphic as yellow, green, blue, and purple. The data type for the project is to be prioritized and identified as a precursor to the DG Academy methodology.
10 Page 10 Stage 1 Harvest o In this stage, the DG team worked with the data project to harvest artifacts that represented the current state of that data. These artifacts included business processes, data standards, data flows, data models, data rules, etc. These artifacts included the official or de facto standards relating to the data type. During this stage all the collateral for the DG Academy, including the content for the tutorials and the workshops are prepared. The harvested artifacts provide the team the information needed to build the workshop materials so they are specific to the project. The artifacts become an understanding of the starting point for the DG Academy efforts. Stage 2 Draft o In this stage, the DG team actually presents/teaches the school. Following the concepts of flexible but comprehensive, the school can be administered in two days to two weeks, depending on the availability of the participants. The intent is for the first draft of the standard to come out of the workshops. Stage 3 Refine o In this stage, the DG team works with the project team to take the results of the workshops and fully develop, expand, and refine them. The inputs for this stage are the business processes, business rules, data rules, data models, data quality metrics, etc. The output from this stage is the standards pack. The standards pack represents the final draft of a set of DG and DQ standards for a particular data type with elements required to create and sustain quality data. Stage 4 Approve o In this stage, the standards are presented for approval to the Standards Committee. As seen in a later section, this committee is comprised of SME s, Stewards, and the appropriate representatives who understand the data, understand the processes, and can make informed decisions with regards to the validity of the DG standards and the value proposition they will provide once implemented and executed. Stage 5 Apply the Approved Standards o In this stage, the standards are used as requirements and specifications to be put into the DQ framework. The standards are identified in the DG Compendium as approved and sanctioned DG standards. As seen in a later section, the Data Custodians are responsible for applying the data rules to the DQ engine. The DG Compendium will be more fully described in a later section.
11 Page 11 Stage 6 Monitor and Evergreen o In this stage, the DG standards have been applied into the DQ framework and are running against the data type. The DG processes to create, edit, delete, and manage all of the DG standards are in place and have been implemented so that the DG environment is properly maintained, sustained, and is evergreen. Thresholds, metrics, monitoring, and exception reporting are in place and have been implemented so that DQ framework reflects the DG environment. At this point, the DG criteria have been fulfilled and data is being fully governed. The development of the methodology of the DG Academy and the DG project was a work in progress. To accelerate the learning another strategic decision was to test with a series of proof of concepts. Serendipitously, there was plenty of food for the fodder: A variety of data projects already in-flight and at different phases in their project Projects representing differing functional groups within Chevron GOM A variety of data types A variety of data format structures; structured, unstructured, and temporal The development of the DG methodology and actual preparation of the DG Academy material leveraged consultants experience and a proven methodology, along with Chevron GOM s thoughts and experience. The resultant collateral was a blend of structure from academia, thoughts from Chevron GOM, and conceptual collateral from Noah Consulting.
12 Page Implementation of Data Governance: Once the DG standards pack has been created and approved, those standards must be properly leveraged in a data governance implementation for a given data type. The components of a DG implementation for a data type are building blocks for sustained data quality and include: Business lifecycle Data lifecycle Business processes Business rules Data rules Standards repository Data repository DG environment DQ framework and DQ engine For any data type there is a need to map how the data type interacts and influences decisions made throughout the business lifecycle or business value chain. Additionally, the data lifecycle should be mapped out to understand how the maturity of the data through its lifecycle affects business processes, business rules, and data rules. Additionally, the responsibility, content ownership, stewardship, etc. can/will change depending on both the business lifecycle and data lifecycle. The business and data lifecycles are independent but inter-related. As data matures and improves through its data lifecycle, the business decisions made based upon that data mature and improve as well. With proper data management and data governance the data quality, confidence, and trustworthiness improve. By the nature our upstream industry, data acquired early in the business and data lifecycles is used throughout the life of the asset. The creation, approval, socialization, management, and sustainment of the DG standards are keys to success as defined in the DG methodology. The development of a standards repository was seen as critical to the success of the DG project. Historically, in other DQ or DG efforts, the resultant collateral (e.g. the standards) are often drafted and finalized as unstructured documents with no indexing, no metadata, and no technology to systematically tie them
13 Page 13 together and keep them together as managed data. There are processes that need to be created to determine how the DG standards are stored, edited, added to, deleted, etc. There is technology required to sustain DG standards and deliver the material as needed. A data governance compendium was developed to satisfy that need. The DG Compendium addresses these failings and provides many benefits. The DG Compendium is the knowledge base for data asset teams and business users to find what they need. It is comprised of four key components: The Organization Registry o Provides the names and contact information for stewards, custodians, architects, analysts and DG/DQ support personnel The Processes Library o Describes the processes for remediation and data creation including required training The Rules Repository o Identifies both the business rules for data asset and the data rules for monitoring and maintaining data quality The Data Catalog o Contains the data definitions and logical data models for all data types under governance and SOR links and metadata for data content Besides a managed and maintained repository for the DG standards, the DG Compendium is envisioned to fulfill other important needs: It is used as a collaboration point to work with the DQ Support and Development team in IT responsible for taking the data rules and imbedding the rules into the DQ engine. It is used to on-board and train new business and IT resources so that data governance concepts become part of the fabric of the organization and help them understand how data-related processes affect their role in Chevron GOM. It is used to provide a platform to query, view, to better understand DG standards, request new rules and standards, request changes to existing DG standards, etc. The DG Compendium acts as the working standards repository as well as the final, approved standards repository. The templates, definitions, and workflows associated with the DG standards and loading up the DG Compendium are also accessed from the DG Compendium.
14 Page 14 After the DG Academy is completed the work begins to improve and detail the business processes, expand and validate the business rules, and create and map the data rules to the SOR or primary repository for data remediation, and monitor the entire system. During the development of the DG methodology and the DG project we learned that adding properly implemented DG standards and components to an already launched project often resulted in resistance from stakeholders. This was rectified by adding resources on to the data project from the DG project team to support the DG effort. The benefits of this approach included: Experienced DG consultants were able to work alongside the data project team to hone the DG standards deliverables. The project team met their established project time line. The project team gained knowledge about data governance and share/provide knowledge about the data and/or workflows in question. Creating and properly managing the DG standards through the DG Compendium is not enough, though. A DQ engine (Data Quality Engine) is required to take the specific data rules and act upon the data. Within Chevron GOM, two different technologies are already being used to act as that DQ engine and apply data rules to specific data types and data attributes. Since the DG methodology and standards within the DG Compendium are software agnostic, the DG project team leverages these technologies to act upon data. The collaboration and working relationship between the DG project team and the DQ support and development team (data custodians) is a key organizational aspect to effectively enable data governance and data quality to work together and for the data quality to be sustained. As depicted in the graphic below, the DG Compendium stores the DG standards required for the DQ engine. The data custodian implements the rules in the DQ engine and relates them specifically to the repository (SOR or application) as it runs through the DQ engine. The DQ engine is also synchronized with the DG Compendium for the data quality dimensions and data rule types: Data Quality Dimensions o Consistency o Accuracy o Validity o Uniqueness o Content o Trustworthiness Data Rule Type o Corrective o Detective o Preventative
15 Page 15 The output of the DQ engine includes results of the execution of the DQ rules and information related to whether the data passed or failed and metadata related to the results. The metadata includes run logs, error logs, exception reports, data quality level, etc. The metadata, metrics, results, updates, ranges, thresholds, etc. can then drive reporting and visualization related to the state of the quality of the data and/or the state of the data. Additionally, the data, having been tested against the DQ engine and with known quality can be passed to the appropriate applications, data hubs, etc. depending on the end-user needs and requirements. The DQ engine is a component of the overall DG/DQ Framework and environment. This Framework is comprised of several inter-related components. They are: Governance/Stewardship Model Information Architecture IQM Components IQM Processes Implementation and Sustainment The generic framework depicted below, coupled with the DG methodology not only enables the DG strategy to be successfully implemented, but also achieves and sustains improved data quality. Some of the most important of these components will be conceptually described.
16 Page 16 Data Governance and Stewardship Data governance is the practice of organizing and implementing policies, procedures and standards to effectively manage the quality and usage of data assets. Data governance specifies accountability and responsibility for data quality and data stewardship. Data Stewardship is the implementation of data management processes in accordance with the governance policies, principles, and standards. Data governance and data stewardship are required in order to attain a sustainable data quality capability. Data governance is about making data management better. Availability of data right people, right data and right time Limiting the redundancy of data stores Increasing the reuse of data integration and publishing services Standard business performance measures - definitions and calculations Determination of appropriate use of data, (e.g., business sensitivity, confidentiality, retention) Documentation of what data is added to which data repository and when Documentation of data sources used for data processing and publishing Commonality of business facts - rules, data names, definitions and structures for data integration Data governance serves many purposes within the DG Framework such as establishing roles and responsibilities for the players within the data quality and data governance community data stewards, data quality analysts, members of the data governance and/or stewardship communities, and the group that helps to facilitate the DG Framework. But additionally and
17 Page 17 perhaps most importantly, data governance establishes and ensures accountability between data producers and data consumers. A data producer is a person, process, or application that creates or derives data that is stored, accessed, and used by one or more data consumers. A data consumer is a person, process, or application that relies on data to perform its required functions. A lynchpin to achieving the benefit of sustained data quality is in the understanding of how these two players (often business people) relate to key business processes, business rules, and the underlying data rules that establish the level of data quality that satisfies the business needs. The data steward is the person who holds data producers accountable to the terms of their data quality metrics to fulfill the needs of the business process. When conflicts between a producer and its consumers arise, the data steward is responsible for mediating the situation by working with both parties to identify an appropriate solution. Additionally, data stewards use the historical quality measures and metrics produced by data quality controls to perform reviews and ongoing analysis to identify the root cause of data quality impairments and identify corrections or improvements that can be made to the producing person, process, or application to eliminate or mitigate the possibility of future impairments being generated. In many cases, data produced in one business function is consumed in one or more other business functions. This is certainly the case with master data and, thusly, shows the importance of the role of a BU or corporate data steward. Governance Organizational Model The following model illustrates a generic data governance hierarchy example that supports a DG framework implementation and sustainment program. It is important to note that this is not an organization chart per se, but shows the relationships of the roles to each other within the DG framework.
18 Page 18 One of the key benefits of this approach is that it supports an organic implementation. Additional DG roles and/or layers of DG responsibility are added as data types are added, as more key business functions are mapped and require support, and/or as the organization matures its implementation of a DG framework. The seed of data governance is planted by identifying subject matter experts, anointing them de facto business representatives and identifying resources that are doing data quality as a major aspect of their daily job and anointing them data quality analysts and/or data stewards. Information Architecture Much of the information architecture of the DG framework is developed in conjunction with, but not as a central part, of the DG efforts. The information architecture requires a business focus both on a logical data model for structured data and a primary taxonomy for unstructured data. For proper data governance, the focus is on (a) defining and documenting standards within both the structured and unstructured environment; (b) defining and documenting the data objects and attributes that support the key business processes and data rules; (c) creating and maintaining a data dictionary or data catalog of the standards; and (d) creating and maintaining data governance and data quality metadata. An understanding of proper capture and management of metadata related to data, data quality, and data governance enables successful data governance implementation. A key work product of the development of a DG framework and DG implementation is metadata metadata architecture, repository, and management. The success of implementing the IQM framework processes define, assess, remediate, and monitor to enable data governance, improve data quality, and sustain data quality is highly dependent on a robust metadata management strategy and system. Metadata is found in the DG Compendium and the DQ engine. Ultimately, there could/should be an effort to look at metadata across systems and functions on its own so that metadata can be modeled, architected, managed, and leveraged to
19 Page 19 its best efficiency and highest value for Chevron GOM. But, as previously suggested, in the meantime caring for metadata associated with data governance, data quality, and stewardship requires dedicated effort to understand how metadata will help to manage the standards, rules, exceptions, alerts, etc. that are associated with the DG Framework. The ability to add metadata in a straightforward and simple manner and access it in a timely manner is important to sustaining the data governance initiative and, ultimately, its success. This was accomplished by designing forms, processes, and worksheets to collect the metadata that directly relate to the DG standards, business processes, business and data rules, etc. The fully vetted and approved the data collected by these forms (etc.) become the DG standards and the metadata that supports the standards. While a normalized metadata model will make porting to a more comprehensive solution easier, it will make creating reports somewhat more difficult. In most cases, ad hoc reporting will be beyond the average business user. Thus, the DG Compendium also needs a library of reports to meet the business needs. At a minimum, reports should be created and maintained that: allow users to view, browse, and query business processes, business rules, and data rules and their relationship to each other; allow users to query and browse for any and all data standards, including the organizational perspective of who are the stewards, SME s, custodians, data quality analysts, business analysts, etc. for any given type of data and their associated attributes; allow consumers to view quality measures and metrics for their critical data elements; both current and historical; allow producers to view their current responsibilities and adherence to their business processes and data rules that govern their data quality; allow managers to monitor their goals and objectives related to data quality thresholds, data KPIs, etc. Ultimately, business will be best served with some type of dashboard technology to present quality measures, metrics, and statistics. IQM Framework Components The information quality management (IQM) framework components support the development, vetting, and approval of standards.
20 Page 20 DQ Engine The DQ engine was previously described. Stewardship Stewardship has already been discussed as it relates to governance and organization. Stewardship as a component here relates more to the function of stewardship, how the steward relates to the consumer and producer, how the steward relates to the Business Processes, Business Rules, and Data Rules, how stewards function to help govern data, etc. Business Rules and Data Rules Governance helps provide consistent and logical business and data rules. A business rule is a statement that defines or constrains some aspect of the business. A data rule is a statement that provides the opportunity to validate compliance to expected conditions or to identify and review exceptions. The goal of a data governance initiative is not the proliferation of a multitude of rules. Rather, business rules and data rules should be used to validate that a business process is working correctly, prevent new errors from being introduced, and, in limited cases, automatically correct errors. Standards Standards allow data rules and other controls to be more automated and allow different functions or groups to compare assets. Standards include definitions and terminology as well as processes and conventions. For the purposes of DG Framework standards are drafted, vetted, and approved by the appropriate body with the recognized authority. There are no de facto standards when data governance is being implemented; the standards need to be documented, managed, able to be modified, and communicated in the DG Compendium. Data Naming Convention Data naming conventions as components of the IQM Framework are described as businessfocused, consistent as possible, human-readable, machine generated, and informative. Data naming conventions can cause a lot of human churn because many people have angst and passion about how things are named.
21 Page 21 Data Model The Data Model as a component of the IQM Framework is described as a business-focused, logical data model to describe data, data attributes, and its appropriate metadata. As part of the DG Framework efforts a data model should be defined that supports the business needs, the business processes, the data rules, and the underlying data elements and attributes that the data rules will be run against. Data Taxonomy Data Taxonomy as a component of the IQM Framework is described as a business-focused, logical, fit-for-purpose hierarchical taxonomy to describe folder structures for properly putting away log files and seismic files for future needs. As part of the DG Framework efforts a taxonomy should be defined that supports the business needs, the business processes, the data rules, and the underlying data elements and attributes that the data rules will be run against. IQM Framework Processes The IQM framework processes as defined as part of the overall DG Framework relate to the overall conceptual process of how data quality issues are identified, made clean, and kept clean. That is the goal of the process. Additionally, it is a pragmatic process that can begin as a single data element and grow. The process can be thought of as an upwardly growing spiral. Define The first step in the process is to prioritize and define what is to be governed or quality managed. This prioritization can be made as part of the overall process, run as a separate project, or as part of another initiative. For our DG project, the data types are associated with the POC s selected. The define step includes the pace as well as the priority. The pace can be thought of as how big a bite is to be taken or how many parallel threads of effort are to be undertaken. As suggested by the title, the define step is a discovery step. Artifacts are gathered and analyzed; key repositories for the data type in scope are identified; the appropriate producers and consumers are identified. The geographic scope and/or areas to work are identified. During this step, it is also appropriate to talk with business users, data workers, producers, and consumers to determine their perspective on the state of the
22 Page 22 data. In the define step the artifacts need to be analyzed to see if they are of sufficient granularity for our DG efforts. If gaps and/or missing key artifacts are discovered then some efforts should be taken during the Define step to shore them up. Assess The Assess step assesses both the data and the artifacts. If there are data quality rules, then during the assess step the rules will be run, warnings and failures will be captured and analyzed, and a more quantitative state of the data can be determined. However best determined, an initial assessment of the data quality must be made. This needs to include overall data quality, specific areas for key problems and issues, and an understanding of how the quality of the data is affecting business decisions. The more effort put to establish as baseline then the better the future monitoring will help demonstrate improvement. In addition to assessing the data and data quality, the standards and artifacts also need to be assessed for completeness, accuracy, business applicability, and capability for automation. In each step, data rules will be documented, improved, and/or edited but in this assess step the business processes need to be complete, vetted, and ready for approval. This will be important when planning remediation. Capturing metadata related to this assessment is also important. Remediate Remediation occurs when the data gets cleaned, and data quality is improved. The assessment shows what needs improvement and modification and remediation performs that function based on priority and capability. During remediation, automation can play a significant role in lowering costs and improving efficiency of remediation. This is a main reason why the define and assess steps are so important. Without these steps and establishing certain DQ rules, remediation will be manual. Remediation is best performed by running automated error detection / correction procedures. Ultimately, users will validate the remediated data to ensure it is accurate and suitable for actual use. Monitor The Monitoring of data is an important step in the process. Monitoring data after it has been remediated is the only way to effectively keep it clean and of high quality (or at least known quality). Monitoring data allows for the business users to trust the data and have that trust remain constant and/or grow over time. Monitoring as depicted here includes monitoring of the data and (also) monitoring of the state of the data governance criteria. Monitoring includes establishing thresholds, ranges, and target KPIs for data quality. Over time, to improve data quality, the thresholds and ranges need to be tightened based on consumer needs and producer abilities. Also, the monitoring is based on the key business
23 Page 23 processes continuing to meet the business needs. Monitoring includes historical and current runs of the data quality values, understanding changes that have been made, referring to the metadata related to exceptions and alerts, etc. One key to successful monitoring is good reporting and visualization.
24 Page Building Sustainment Through a DG Organization: A goal of the DG project is to transform the project into a sustained practice a functional capability that is part of the core Chevron business. There are certain activities identified to assist in building this sustainment. These include: maintaining and improving the DG Academy as a part of the standard data enhancement process; establishing a proper organizational capability and building it into a service line; maintaining upper management and function level support and sponsorship; maintaining flexibility and an outward-focused awareness; looking for opportunities to prioritize efforts to drive towards success; documenting, measuring, and reporting on key success factors and objectives frequently, as well as celebrating successes. An important aspect of the DG project is to understand how to establish a DG organizational capability -- a capability that can begin relatively small and grow organically as the data types under governance grow and as the business benefit of sustained data quality is realized. Properly building the capability into the DG program structure requires that there is a clear understanding of overall process and methodology that are being developed. Organizationally, there are groups that will be directly responsible for certain aspects of data governance in Chevron GOM. These groups need to be clearly identified and named, and their roles and responsibilities clearly defined. DG organizational considerations include: the need to define and describe the various Data Governance bodies; the desire to leverage current functional organizations; the need to decide what roles are to be within the DG/DQ Framework; the need to give each role a label or name that can be agreed on; the need to define/describe roles and responsibilities as they relate to the specific data types to govern. The various data governance bodies will include the following groups and attributes. DG standards committee (currently, domain-oriented business leadership teams) o A hybrid group from the DG practice, the business, and (possibly) IT o Performs portfolio, champion, and standards guidance and approval o Approves the DG standards pack o Domain oriented o Accountable for the data types and information objects within a domain o Decisions and work effort facilitated by DG practice Data steward o Part of business; in the function by discipline o Manages a portfolio o Performs oversight on standards development
25 Page 25 o Monitors data quality metrics and results to agreed standard o Provides oversight on data remediation and clean-up o Resolves hard problems; the authoritative source Data custodian(s) o Part of IT s data quality team o Manages the implementation of the data rule base in the DQ engine o Manages or facilitates coordination of data remediation and clean-up Data analyst o Part of IT s data quality team o Performs analysis on data issues o Performs heavy lifting for data remediation and clean-up o Resolves most data problems Data governance practice o More details below; includes key DG responsibilities, key DG roles o DG practice lead o DG Compendium librarian o DG Academy support The graphic below shows the hierarchical nature of various DG roles as envisioned by the Chevron GOM DG project team. Additionally, as see in the graphic, these roles will be repeated for the major data domains or functions within Chevron GOM. As described above, the standards committee fulfills an important role in the approval of the DG standards that come out of the DG Academy efforts and the data project teams. The graphic below shows the structure of this committee.
26 Page 26 Setting up the data governance practice is a crucial activity of the DG project. Building the DG organizational capability begins with establishing the DG practice as a work team and the DG practice lead as a role. The DG practice: is dedicated to sustaining data quality through proper data governance; ensures roles are filled; ensures metrics are maintained and communicated; ensures alignment and proper leadership support; captures and communicates value; monitors budget for DG efforts; maintains and enhances the DG Academy, DG processes, and DG collateral; maintains and manages the DG Compendium; maintains and enhances DG and DQ roles and responsibilities; guides and/or takes guidance from the sponsors and business leadership for future changes in strategy and direction; on-boards new DG resources, business groups, and project teams as DG expands; and keeps the vision alive and communicates it effectively. Effective communication includes: describing what data governance entails in our vision; describing the current thinking related to DG and DQ roles and responsibilities; describing the overall DG and DQ organization and how the DG practice fits into it; describing the challenges with communicating the concepts related to DG and leveraging the DG Academy to address those challenges; being consistent in the naming of and communication of role names there has been inconsistency and ambiguity in the past; and
27 Page 27 describing the DG practice as it relates organizationally to both the business and IT. Much of the organizational capability described in the text and graphics above are designed to show the future state of DG in Chevron GOM. During 2013, the DG project and the project team will be very active in continuing to work with data project teams, perform additional POCs and progress through the Chevron project execution process. Related specifically to organizational capabilities, 2013 will be a ramp up and transition year. The DG project team will continue to work under the Chevron GOM data management (DM) program. The project will continue developing business and IT partnerships to determine how data content owners in the business and DQ support in IT can work together with the DG project team to prioritize, recommend, fund, and implement new projects to bring new data types under data governance. In 2013, the DG project team will continue to utilize consultants but the DG practice will bring on Chevron GOM staff to ramp up internal capabilities and begin to prepare for DG implementations beyond the end of the DG project. The DG project will support the DG practice and continue to develop components that will be consumed by the DG practice. As new data types come under data governance, whether initiated by the DG project or the DG practice, the results will be handed off to the DQ support team to integrate into the system. 6.0 Beyond Data Governance as a Project:
28 Page 28 For Chevron GOM, Operational Excellence (OE) drives everything the employees do There are certain foundational capabilities that are used to help achieve OE such as the standard method used to execute projects, the standard method used to map and change business process, the standard method used to set and measure performance, etc. As the data governance methodology and capability matures and grows beyond a project and into an operational practice, the DG project team sees data governance becoming one of those foundational capabilities: a standardized method used to define, assess, remediate, and monitor data quality into a sustained, high quality state. Data is an asset and must be managed accordingly. The DG project team sees the proper management, stewardship, and governance of these data assets in Chevron GOM as a differentiator; a competitive advantage. This is the path laid out. Some of the strategies that will help to pave this path include: engage full time data stewards by function; create and grow a data governance practice; remediate data only when necessary and consistent with business priorities; minimize change. As stated in the previous section, building a proper organizational data management capability with Chevron GOM staff is on the critical path. A viable service or capability cannot be operated effectively over the long term using only consultants. Since data is an asset and the decisions made based on governed data are central to the health and future of Chevron GOM, then data governance must be considered a core competency of the business. The shaping curve in the graphic above shows, at a very high level, some of the milestones envisioned for the DG project as it moves into an operational mode.
29 Page Establish A year of continued learning, continued development of various facets of the methodology, and continued work towards getting some specific data types under governance. Additionally this year, a focus will be on establishing the ground work for the organizational capability. Working with the business and senior management who are part of the organizational capability, the prioritization of data governance deployment will be ordered likely the top 10 data types will provide the highest value. o Foundational work to establish a sustainable DG environment will be performed in 2013 and includes: an on-boarding process for various functions; a data-type prioritization model; a value/benefit model and investment model; a scalability model; alignment with other Chevron initiatives and corporate groups; alignment and collaboration with other Chevron BU s. o Train IT projects so that any and all data-intensive projects, application upgrades or change-outs, and/or system migrations that involve data migrations will follow data management best practices and DG guidelines Deploy Data governance will be deployed in Chevron GOM. The project will be completed early 2014 and the DG practice will be in place and operating. Two functions will have been on-boarded. The DG practice will grow and be matured and a core team will be established to assist specific data teams on implementing the DG methodology. Data governance will be expanded based on functional need and business priority Operate The DG components, methodology, and technology will be mature. Data being governed will be expanded in several dimensions. All of the DG components will be reviewed for scalability, maturity, quality, etc. and updated as appropriate. Monitoring and reporting on the quality of specific data being governed will greatly increase the confidence that data consumers have in the data. Additionally, the ability to monitor and report on what data is being governed in a more holistic sense will provide the business with one aspect of the success of the DG project. It is envisioned that a series of charts, reports, and graphs will be available via a dashboard at varying levels of granularity to show DQ metrics, DG metrics, and the state of overall efforts
30 Page 30 The dashboard will show the current state and plans for data that are currently being governed and that data which is next up to be governed. The DG team and DQ Support team will work with the business to understand requirements and design appropriate dashboards, reports, and presentations for monitoring data quality and the governance criteria status. They will work with IT to create processes and technology to generate the data and metadata needed for the dashboards. The data project teams will test the dashboards ability to accurately represent the progress of their projects and to visually represent the state of the data being governed. Finally, working with the leadership teams, the DG project team will present reports and graphs via the dashboard that will represent measured business value. Project pace, data priority, and DG spend must be balanced with business needs. A refresh rate will be established that is in line with business needs and timing. These efforts from the DG project through the DG transition and into the DG operated state need to be somewhat nimble, but not haphazard. The refresh or re-prioritization should not be so frequent that efforts cannot get to a final governed state, but neither should effort be expended to govern data that is not a priority. Whatever is created as the ongoing DG practice needs to be consistent with the business imperative, IT planning, and Chevron Operational Excellence.
31 Page Summary: Data governance is a difficult concept to master but sustaining data quality requires good data governance. The data asset in a major oil company like Chevron is huge. The investment to improve this asset should be very large as well. In order to sustain this investment, the data must be treated as the asset it is. That implies there must be recognition of the asset value, an organizational and governance structure to support it, and funding to maintain it. Chevron GOM embarked on a journey to create and implement a program for attaining and sustaining known quality data through a data governance development process. The process and methodology developed is a fresh approach to managing the quality of information and this paper has described the thoughts and actions behind the process, the methodology, and the implementation. The methodology has been broken into six stages (Harvest, Draft, Refine, Approve, Apply, and Monitor) that correspond to four of Chevron GOM s execution process phases. As described, this approach is pragmatic and generic and has been designed to present a light touch to most of the people directly impacted. It includes a repeatable series of clearly defined activities and tasks to achieve desired goals and consistent results. Achieving a state of data being governed is accomplished with a balance of processes, people, and technology. The paper has described the methodology with significant detail, but it can be summarized as: a data governance and stewardship model that is taught through a DG Academy and is based on placing specific data types under governance, meeting the criteria of what it means to be governed; a framework for data quality controls that is based on key business processes that are impacted by data with associated standards built around the business process and supporting the business process through rules for both business and data. The data rules are imposed on the data through a data quality engine; a data governance standards repository called the DG Compendium stores and manages the standards and a variety of metadata from the results is monitored and reported; an organizational capability to set priority and pace, govern the standards, and mature the data governance practice. The DG project has so far focused on four data types through POCs and can be thought of as an agent of change. The project: realigns the concept of data ownership; removes data and the discussions centered on data issues as an impediment to change and innovation;
32 Page 32 breaks down the data fortresses between different groups in the business by clearly establishing relationships such as data producer, data consumer, data content owner, data steward, and data custodian; and creates a reality for the sponsors, senior business management, IT, and data consumer so they can better understand, appreciate, and value data governance and the important part it plays in their decisions. The other critical aspect of the DG project is establishing the DG organizational structure and capability to effectively create, approve, manage, and administer the DG standards. The transition from a DG project to a DG operational capability will transform over the next 18 to 24 months. The establishment of a DG practice and the role of DG practice lead is the first step. Through the DG project, it has been seen how data governance and data stewardship contribute to information management (IM). Data governance specifies accountability and responsibility for IM. Data stewardship enables IM to be more effective across the company. Although the DG project is not over yet, there are clear indications of key success factors and lessons learned: Key success factors: o The sponsors have requested that the project be accelerated so it can go into operational mode more quickly. That is seen as a clear validation of the project s success to date with the DG Academy, the DG Compendium, and the communication with the project sponsors and the business. o Data project groups are coming to the DG project team requesting assistance in establishing a DG capability as part of their project o The DG project and the DQ support team (data custodians) are getting traction in moving data rules from the DG Compendium into the DQ engine and making them part of the DQ tool set. Lessons learned o Be focused on why you are implementing standards. o Use a phased approach. o Leverage standards or de facto standards. o Determine who is best positioned to discover/document standards to improve data quality; use a team approach. o Involve SMEs to ensure accuracy. o Standards are key products (measureable results) of data stewards and DG teams. o Remediation (clean up) is difficult. Ensure it is focused to bring business value. o Find ways to measure business value.
33 Page About the authors: Dave Blosser Chevron North America Exploration and Production - Data Management, Program Manager Dave has been in his current position for three years. His purview for this project was, We are having a real problem with our data. Go fix it. Prior to this position, Dave was program manager for GOM s Upstream Foundation Program. During that time several leading edge applications were developed resulting in over $200 million of realized value and leading to the team s selection for Chevron s prestigious Chairmanship Award. Dave has led or been a part of many IT development efforts over the other 18 years of his career with Chevron. Prior to joining Chevron in 1988, Dave was a working petroleum engineer for a small independent where he gained experience in all aspects of the oil industry from exploration to abandonment. Too often exploration and abandonment happened in the same well, which led him to a new career path with IT. Paul Haines Noah Consulting Partner and Upstream and IM Subject Matter Expert (SME) As a subject matter expert, Paul has been consulting and advising about information management for E&P companies, both large and small, for the past seven years. Before his entry into consulting, he worked for Kerr-McGee Oil & Gas for five years. There, he was manager of Data Management and Integration. This role included responsibilities over all E&P technical subsurface and operations data. Prior to mid-2001, Paul spent 23 years with Schlumberger, beginning as a wireline field engineer. During these years, he gained a broad understanding of the industry and the data associated with the E&P industry. He held a variety of positions in areas of operations, technique, sales, marketing, and management. Paul served on the board of directors of the Professional Petroleum Data Management Association (PPDM) from 2007 to In 2011, Paul was honored with the Cornerstone Award for his contributions to the field of E&P Data Management at this conference. Paul has been a presenter and contributor at PNEC conferences frequently over the years.
34 Page About the companies: Chevron Corporation Chevron is one of the world's leading integrated energy companies and conducts business worldwide. Our success is driven by our people and their commitment to get results the right way by operating responsibly, executing with excellence, applying innovative technologies and capturing new opportunities for profitable growth. We are involved in virtually every facet of the energy industry. We explore for, produce and transport crude oil and natural gas; refine, market and distribute transportation fuels and lubricants; manufacture and sell petrochemical products; generate power and produce geothermal energy; provide energy efficiency solutions; and develop the energy resources of the future, including research for advanced biofuels. Chevron s business momentum is driven by three enterprise strategies people, execution and growth. We invest in people to maintain world-class workforce. We follow systems and processes that allow us to execute with excellence. And we use our competitive advantages to grow our company profitably. Based in Covington, LA, Chevron s Gulf of Mexico business unit has approximately 1,800 employees and approximately 2,000 contractors daily. Chevron is one of the largest producers and second largest leaseholder on the Gulf of Mexico shelf; and one of the top leaseholders in the deepwater Gulf of Mexico. Noah Consulting Noah Consulting makes a significant difference in the lives of Business and Data Management professionals in the companies they serve through energy industry specific information management services that transform data into valuable assets. We believe your data should work for you 24 hours, 7 days a week. Re-creating or re-purchasing data can be costly, error prone and time consuming. As such, excellent data management, optimal extraction and aggregation of data, and the proper analysis of that data are investments in a company's future. Noah Consulting strives to improve the wellness of the industry by helping clients navigate their information. This mission is clear by our alignment with Best-in-Class organizations concerned with industry standards, such as PPDM. Founded in 2008 by Information Services professionals, Noah Consulting is exclusively focused on the data and information domain within the energy sector. By creating a home for focused experts with a common passion, Noah Consulting has put together a team of professionals who average over 18 years of experience and who have both deep industry knowledge and technical expertise in Strategy and Planning, Data Governance, Program Management, Architecture and Business Intelligence.
Enterprise Data Governance
DATA GOVERNANCE Enterprise Data Governance Strategies and Approaches for Implementing a Multi-Domain Data Governance Model Mark Allen Sr. Consultant, Enterprise Data Governance WellPoint, Inc. 1 Introduction:
Data Governance Primer. A PPDM Workshop. March 2015
Data Governance Primer A PPDM Workshop March 2015 Agenda - SETTING THE STAGE - DATA GOVERNANCE BASICS - METHODOLOGY - KEYS TO SUCCESS Copyright 2015 Noah Consulting LLC. All Rights Reserved. Industry Drivers
Data Governance on Well Header. Not Only is it Possible, Where Else Would you Start!
Data Governance on Well Header Not Only is it Possible, Where Else Would you Start! Agenda Intro (Noah) A (not so) Brief History Methodology Prioritization Why Well Header Attribute versus Process Oriented
The following is intended to outline our general product direction. It is intended for informational purposes only, and may not be incorporated into
The following is intended to outline our general product direction. It is intended for informational purposes only, and may not be incorporated into any contract. It is not a commitment to deliver any
Better Data is Everyone s Job! Using Data Governance to Accelerate the Data Driven Organization
Better Data is Everyone s Job! Using Data Governance to Accelerate the Data Driven Organization Intros - Name - Interest / Challenge - Role Data Governance is a Business Function Data governance should
DISCIPLINE DATA GOVERNANCE GOVERN PLAN IMPLEMENT
DATA GOVERNANCE DISCIPLINE Whenever the people are well-informed, they can be trusted with their own government. Thomas Jefferson PLAN GOVERN IMPLEMENT 1 DATA GOVERNANCE Plan Strategy & Approach Data Ownership
Enabling Data Quality
Enabling Data Quality Establishing Master Data Management (MDM) using Business Architecture supported by Information Architecture & Application Architecture (SOA) to enable Data Quality. 1 Background &
EXPLORING THE CAVERN OF DATA GOVERNANCE
EXPLORING THE CAVERN OF DATA GOVERNANCE AUGUST 2013 Darren Dadley Business Intelligence, Program Director Planning and Information Office SIBI Overview SIBI Program Methodology 2 Definitions: & Governance
Agile Master Data Management TM : Data Governance in Action. A whitepaper by First San Francisco Partners
Agile Master Data Management TM : Data Governance in Action A whitepaper by First San Francisco Partners First San Francisco Partners Whitepaper Executive Summary What do data management, master data management,
17 th Petroleum Network Education Conferences
1.1.1. 17 th Petroleum Network Education Conferences Making Data Governance Work At All Levels of the Organization Prepared by: Joseph Seila Devon Energy and James Soos Noah Consulting March 21, 2013 Table
Analytics Strategy Information Architecture Data Management Analytics Value and Governance Realization
1/22 As a part of Qlik Consulting, works with Customers to assist in shaping strategic elements related to analytics to ensure adoption and success throughout their analytics journey. Qlik Advisory 2/22
NASCIO EA Development Tool-Kit Solution Architecture. Version 3.0
NASCIO EA Development Tool-Kit Solution Architecture Version 3.0 October 2004 TABLE OF CONTENTS SOLUTION ARCHITECTURE...1 Introduction...1 Benefits...3 Link to Implementation Planning...4 Definitions...5
Data Governance Maturity Model Guiding Questions for each Component-Dimension
Data Governance Maturity Model Guiding Questions for each Component-Dimension Foundational Awareness What awareness do people have about the their role within the data governance program? What awareness
POLAR IT SERVICES. Business Intelligence Project Methodology
POLAR IT SERVICES Business Intelligence Project Methodology Table of Contents 1. Overview... 2 2. Visualize... 3 3. Planning and Architecture... 4 3.1 Define Requirements... 4 3.1.1 Define Attributes...
Whitepaper Data Governance Roadmap for IT Executives Valeh Nazemoff
Whitepaper Data Governance Roadmap for IT Executives Valeh Nazemoff The Challenge IT Executives are challenged with issues around data, compliancy, regulation and making confident decisions on their business
The following is intended to outline our general product direction. It is intended for information purposes only, and may not be incorporated into
The following is intended to outline our general product direction. It is intended for information purposes only, and may not be incorporated into any contract. It is not a commitment to deliver any material,
A discussion of information integration solutions November 2005. Deploying a Center of Excellence for data integration.
A discussion of information integration solutions November 2005 Deploying a Center of Excellence for data integration. Page 1 Contents Summary This paper describes: 1 Summary 1 Introduction 2 Mastering
Building a Data Quality Scorecard for Operational Data Governance
Building a Data Quality Scorecard for Operational Data Governance A White Paper by David Loshin WHITE PAPER Table of Contents Introduction.... 1 Establishing Business Objectives.... 1 Business Drivers...
WHY DO I NEED A PROGRAM MANAGEMENT OFFICE (AND HOW DO I GET ONE)?
WHY DO I NEED A PROGRAM MANAGEMENT OFFICE (AND HOW DO I GET ONE)? Due to the often complex and risky nature of projects, many organizations experience pressure for consistency in strategy, communication,
Enterprise Data Governance
Enterprise Aligning Quality With Your Program Presented by: Mark Allen Sr. Consultant, Enterprise WellPoint, Inc. ([email protected]) 1 Introduction: Mark Allen is a senior consultant and enterprise
Using SAP Master Data Technologies to Enable Key Business Capabilities in Johnson & Johnson Consumer
Using SAP Master Data Technologies to Enable Key Business Capabilities in Johnson & Johnson Consumer Terry Bouziotis: Director, IT Enterprise Master Data Management JJHCS Bob Delp: Sr. MDM Program Manager
Business Architecture Scenarios
The OMG, Business Architecture Special Interest Group Business Architecture Scenarios Principal Authors William Ulrich, President, TSG, Inc. Co chair, OMG BASIG [email protected] Neal McWhorter, Principal,
Vermont Enterprise Architecture Framework (VEAF) Master Data Management (MDM) Abridged Strategy Level 0
Vermont Enterprise Architecture Framework (VEAF) Master Data Management (MDM) Abridged Strategy Level 0 EA APPROVALS EA Approving Authority: Revision
IRMAC SAS INFORMATION MANAGEMENT, TRANSFORMING AN ANALYTICS CULTURE. Copyright 2012, SAS Institute Inc. All rights reserved.
IRMAC SAS INFORMATION MANAGEMENT, TRANSFORMING AN ANALYTICS CULTURE ABOUT THE PRESENTER Marc has been with SAS for 10 years and leads the information management practice for canada. Marc s area of specialty
Master Data Management
Master Data Management Managing Data as an Asset By Bandish Gupta Consultant CIBER Global Enterprise Integration Practice Abstract: Organizations used to depend on business practices to differentiate them
Five best practices for deploying a successful service-oriented architecture
IBM Global Services April 2008 Five best practices for deploying a successful service-oriented architecture Leveraging lessons learned from the IBM Academy of Technology Executive Summary Today s innovative
Data Governance Overview
Data Governance Overview Anthony Chalker Managing Director August 12, 2014 2:05 2:55 Session What is Data Governance? Data Governance is the specification of decision rights and an accountability framework
Data Migration through an Information Development Approach An Executive Overview
Data Migration through an Approach An Executive Overview Introducing MIKE2.0 An Open Source Methodology for http://www.openmethodology.org Management and Technology Consultants Data Migration through an
OPTIMUS SBR. Optimizing Results with Business Intelligence Governance CHOICE TOOLS. PRECISION AIM. BOLD ATTITUDE.
OPTIMUS SBR CHOICE TOOLS. PRECISION AIM. BOLD ATTITUDE. Optimizing Results with Business Intelligence Governance This paper investigates the importance of establishing a robust Business Intelligence (BI)
Enterprise Business Service Management
Technical white paper Enterprise Business Service Management Key steps and components of a successful solution Table of contents Executive Summary... 2 Setting the goal establishing an IT initiative...
Enterprise Architecture (Re)Charter Template
Enterprise Architecture (Re)Charter Template To learn more about this full research or to inquire about membership, contact us: +1-866-913-8101 IT.Support@ executiveboard.com www.cebglobal.com/it CEB Enterprise
October 8, 2014. User Conference. Ronald Layne Manager, Data Quality and Data Governance [email protected]
Ensuring the highest quality data is delivered throughout the university providing valuable information serving individual and organizational need October 8, 2014 Ronald Layne Manager, Data Quality and
Agile Master Data Management A Better Approach than Trial and Error
Agile Master Data Management A Better Approach than Trial and Error A whitepaper by First San Francisco Partners First San Francisco Partners Whitepaper Executive Summary Market leading corporations are
Explore the Possibilities
Explore the Possibilities 2013 HR Service Delivery Forum Best Practices in Data Management: Creating a Sustainable and Robust Repository for Reporting and Insights 2013 Towers Watson. All rights reserved.
Data Governance. Unlocking Value and Controlling Risk. Data Governance. www.mindyourprivacy.com
Data Governance Unlocking Value and Controlling Risk 1 White Paper Data Governance Table of contents Introduction... 3 Data Governance Program Goals in light of Privacy... 4 Data Governance Program Pillars...
US Department of Education Federal Student Aid Integration Leadership Support Contractor January 25, 2007
US Department of Education Federal Student Aid Integration Leadership Support Contractor January 25, 2007 Task 18 - Enterprise Data Management 18.002 Enterprise Data Management Concept of Operations i
White Paper. An Overview of the Kalido Data Governance Director Operationalizing Data Governance Programs Through Data Policy Management
White Paper An Overview of the Kalido Data Governance Director Operationalizing Data Governance Programs Through Data Policy Management Managing Data as an Enterprise Asset By setting up a structure of
Information Governance Workshop. David Zanotta, Ph.D. Vice President, Global Data Management & Governance - PMO
Information Governance Workshop David Zanotta, Ph.D. Vice President, Global Data Management & Governance - PMO Recognition of Information Governance in Industry Research firms have begun to recognize the
Big Data Services From Hitachi Data Systems
SOLUTION PROFILE Big Data Services From Hitachi Data Systems Create Strategy, Implement and Manage a Solution for Big Data for Your Organization Big Data Consulting Services and Big Data Transition Services
Assessing and implementing a Data Governance program in an organization
Assessing and implementing a Data Governance program in an organization Executive Summary As companies realize the importance of data and the challenges they face in integrating the data from various sources,
An RCG White Paper The Data Governance Maturity Model
The Dataa Governance Maturity Model This document is the copyrighted and intellectual property of RCG Global Services (RCG). All rights of use and reproduction are reserved by RCG and any use in full requires
BI Dashboards the Agile Way
BI Dashboards the Agile Way Paul DeSarra Paul DeSarra is Inergex practice director for business intelligence and data warehousing. He has 15 years of BI strategy, development, and management experience
State of California Department of Transportation. Transportation System Data Business Plan
DRAFT Page i State of California Department of Transportation Transportation System Data Business Plan RFO# TSI DPA-0003 September 29, 2011 DRAFT Page ii Table of Contents Executive Summary... 4 Chapter
JOURNAL OF OBJECT TECHNOLOGY
JOURNAL OF OBJECT TECHNOLOGY Online at www.jot.fm. Published by ETH Zurich, Chair of Software Engineering JOT, 2008 Vol. 7, No. 8, November-December 2008 What s Your Information Agenda? Mahesh H. Dodani,
Fortune 500 Medical Devices Company Addresses Unique Device Identification
Fortune 500 Medical Devices Company Addresses Unique Device Identification New FDA regulation was driver for new data governance and technology strategies that could be leveraged for enterprise-wide benefit
TDWI strives to provide course books that are content-rich and that serve as useful reference documents after a class has ended.
Previews of TDWI course books offer an opportunity to see the quality of our material and help you to select the courses that best fit your needs. The previews cannot be printed. TDWI strives to provide
EIM Strategy & Data Governance
EIM Strategy & Data Governance August 2008 Any Information management program must utilize a framework and guiding principles to leverage the Enterprise BI Environment Mission: Provide reliable, timely,
The Future of Census Bureau Operations
The Future of Census Bureau Operations Version 1.0 April 25, 2013 The Future of Census Bureau Operations Page ii [This page intentionally left blank] The Future of Census Bureau Operations Page iii Document
An Oracle White Paper. December 2011. Cloud Computing Maturity Model Guiding Success with Cloud Capabilities
An Oracle White Paper December 2011 Cloud Computing Maturity Model Guiding Success with Cloud Capabilities Executive Overview... 3 Introduction... 4 Cloud Maturity Model... 4 Capabilities and Domains...
Data Governance A Big Step for your Big Data Initiatives
Data Governance A Big Step for your Big Data Initiatives RoundWorld Solutions, Inc. 191 Calle Magdalena Suite 290 Encinitas, CA 92024 Phone: 866-868-5130 12.29.2014 CONTENTS 1.0 Introduction...... 3 1.1
Solutions Master Data Governance Model and Mechanism
www.pwc.com Solutions Master Data Governance Model and Mechanism Executive summary Organizations worldwide are rapidly adopting various Master Data Management (MDM) solutions to address and overcome business
DATA QUALITY MATURITY
3 DATA QUALITY MATURITY CHAPTER OUTLINE 3.1 The Data Quality Strategy 35 3.2 A Data Quality Framework 38 3.3 A Data Quality Capability/Maturity Model 42 3.4 Mapping Framework Components to the Maturity
Evaluating Data Warehousing Methodologies: Objectives and Criteria
Evaluating Data Warehousing Methodologies: Objectives and Criteria by Dr. James Thomann and David L. Wells With each new technical discipline, Information Technology (IT) practitioners seek guidance for
Marathon Information Management Program
Case Study: Implementing Enterprise Content Management at Marathon Oil Reid G. Smith ECM Director & IT Upstream Services Manager Marathon Oil Corporation March 12, 2012 Who we are Global independent exploration
Welcome to the Data Analytics Toolkit PowerPoint presentation on data governance. The complexity of healthcare delivery, the exploding demand for
Welcome to the Data Analytics Toolkit PowerPoint presentation on data governance. The complexity of healthcare delivery, the exploding demand for actionable information, pressure for greater public accountability,
GEOSPATIAL LINE OF BUSINESS PROGRAM MANAGEMENT OFFICE CONCEPT OF OPERATIONS
GEOSPATIAL LINE OF BUSINESS PROGRAM MANAGEMENT OFFICE CONCEPT OF OPERATIONS March 2007 Adjudicated Draft TABLE OF CONTENTS 1 INTRODUCTION 3 2 VALUE PROPOSITION 3 3 ORGANIZING FRAMEWORK 3 31 PMO Organizational
Business Analysis Standardization & Maturity
Business Analysis Standardization & Maturity Contact Us: 210.399.4240 [email protected] Copyright 2014 Enfocus Solutions Inc. Enfocus Requirements Suite is a trademark of Enfocus Solutions Inc.
Solutions. Master Data Governance Model and the Mechanism
Solutions Master Data Governance Model and the Mechanism Executive summary Organizations worldwide are rapidly adopting various Master Data Management (MDM) solutions to address and overcome business issues
HP SOA Systinet software
HP SOA Systinet software Govern the Lifecycle of SOA-based Applications Complete Lifecycle Governance: Accelerate application modernization and gain IT agility through more rapid and consistent SOA adoption
Kalido Data Governance Maturity Model
White Paper Kalido Data Governance Maturity Model September 2010 Winston Chen Vice President, Strategy and Business Development Kalido Introduction Data management has gone through significant changes
Implementing a Data Governance Initiative
Implementing a Data Governance Initiative Presented by: Linda A. Montemayor, Technical Director AT&T Agenda AT&T Business Alliance Data Governance Framework Data Governance Solutions: o Metadata Management
How To Develop An Enterprise Architecture
OSI Solution Architecture Framework Enterprise Service Center April 2008 California Health and Human Services Agency Revision History REVISION HISTORY REVISION/WORKSITE # DATE OF RELEASE OWNER SUMMARY
The Information Management Center of Excellence: A Pragmatic Approach
1 The Information Management Center of Excellence: A Pragmatic Approach Peter LePine & Tom Lovell Table of Contents TABLE OF CONTENTS... 2 Executive Summary... 3 Business case for an information management
Using Organizational Change Management Principles to Create a Scalable OCM Methodology
Using Organizational Change Management Principles to Create a Scalable OCM Methodology Cynthia Onstott John Spurrell May 16, 2016 2 Today s Learning Objectives How to develop a new Organizational Change
Business Intelligence and Analytics: Leveraging Information for Value Creation and Competitive Advantage
PRACTICES REPORT BEST PRACTICES SURVEY: AGGREGATE FINDINGS REPORT Business Intelligence and Analytics: Leveraging Information for Value Creation and Competitive Advantage April 2007 Table of Contents Program
TOGAF TOGAF & Major IT Frameworks, Architecting the Family
Fall 08 TOGAF TOGAF & Major IT Frameworks, Architecting the Family Date: February 2013 Prepared by: Danny Greefhorst, MSc., Director of ArchiXL TOGAF is a registered trademark of The Open Group. TOGAF
NCOE whitepaper Master Data Deployment and Management in a Global ERP Implementation
NCOE whitepaper Master Data Deployment and Management in a Global ERP Implementation Market Offering: Package(s): Oracle Authors: Rick Olson, Luke Tay Date: January 13, 2012 Contents Executive summary
The Business in Business Intelligence. Bryan Eargle Database Development and Administration IT Services Division
The Business in Business Intelligence Bryan Eargle Database Development and Administration IT Services Division Defining Business Intelligence (BI) Agenda Goals Identify data assets Transform data and
White Paper. Software Development Best Practices: Enterprise Code Portal
White Paper Software Development Best Practices: Enterprise Code Portal An Enterprise Code Portal is an inside the firewall software solution that enables enterprise software development organizations
GOING BEYOND TRADITIONAL DATA GOVERNANCE: SIX STEPS FOR BUILDING AN INFLUENTIAL DATA-DRIVEN ORGANIZATION
GOING BEYOND TRADITIONAL DATA GOVERNANCE: SIX STEPS FOR BUILDING AN INFLUENTIAL DATA-DRIVEN ORGANIZATION In an environment of rapidly evolving technology, customer behavior, and competition, data has become
04 Executive Summary. 08 What is a BI Strategy. 10 BI Strategy Overview. 24 Getting Started. 28 How SAP Can Help. 33 More Information
1 BI STRATEGY 3 04 Executive Summary 08 What is a BI Strategy 10 BI Strategy Overview 24 Getting Started 28 How SAP Can Help 33 More Information 5 EXECUTIVE SUMMARY EXECUTIVE SUMMARY TOP 10 BUSINESS PRIORITIES
Jump Starting Data Governance: a Program Manager's Story
Jump Starting Data Governance: a Program Manager's Story ABSTRACT One of the many challenges facing a federal organization endeavoring to implement an enterprise Data Governance framework involves the
Data Governance Baseline Deployment
Service Offering Data Governance Baseline Deployment Overview Benefits Increase the value of data by enabling top business imperatives. Reduce IT costs of maintaining data. Transform Informatica Platform
Cisco Unified Communications and Collaboration technology is changing the way we go about the business of the University.
Data Sheet Cisco Optimization s Optimize Your Solution using Cisco Expertise and Leading Practices Optimizing Your Business Architecture Today, enabling business innovation and agility is about being able
Wilhelmenia Ravenell IT Manager Eli Lilly and Company
Wilhelmenia Ravenell IT Manager Eli Lilly and Company Agenda Introductions The Service Management Framework Keys of a successful Service management transformation Why transform? ROI and the customer experience
White Paper. An Introduction to Informatica s Approach to Enterprise Architecture and the Business Transformation Toolkit
White Paper An Introduction to Informatica s Approach to Enterprise Architecture and the Business Transformation Toolkit This document contains Confidential, Proprietary and Trade Secret Information (
Whitepaper: Creating an ECM Advisory Board and Program Charter
Whitepaper: Creating an ECM Advisory Board and Program Charter By: Ronda Ringo November 2013 The Problem A year ago, Todd, an IT Director, and Melanie, a Director of Business Process Improvement, had been
HOW TO USE THE DGI DATA GOVERNANCE FRAMEWORK TO CONFIGURE YOUR PROGRAM
HOW TO USE THE DGI DATA GOVERNANCE FRAMEWORK TO CONFIGURE YOUR PROGRAM Prepared by Gwen Thomas of the Data Governance Institute Contents Why Data Governance?... 3 Why the DGI Data Governance Framework
Role and Skill Descriptions. For An ITIL Implementation Project
Role and Skill Descriptions For An ITIL Implementation Project The following skill traits were identified as fairly typical of those needed to execute many of the key activities identified: Customer Relationship
Data Governance Best Practices
Data Governance Best Practices Rebecca Bolnick Chief Data Officer Maya Vidhyadharan Data Governance Manager Arizona Department of Education Key Issues 1. What is Data Governance and why is it important?
Realizing business flexibility through integrated SOA policy management.
SOA policy management White paper April 2009 Realizing business flexibility through integrated How integrated management supports business flexibility, consistency and accountability John Falkl, distinguished
Minnesota Health Insurance Exchange (MNHIX)
Minnesota Health Insurance Exchange (MNHIX) Project Status Report Week Ending: 09-19-2012 Page - 1 Executive Summary The Executive Summary provides an executive level review of general project activities,
Ten Steps to Quality Data and Trusted Information
Ten Steps to Quality Data and Trusted Information ABSTRACT Do these situations sound familiar? Your company is involved in a data integration project such as building a data warehouse or migrating several
Adopting a Continuous Integration / Continuous Delivery Model to Improve Software Delivery
Customer Success Stories TEKsystems Global Services Adopting a Continuous Integration / Continuous Delivery Model to Improve Software Delivery COMMUNICATIONS AGILE TRANSFORMATION SERVICES Executive Summary
HP Service Manager software
HP Service Manager software The HP next generation IT Service Management solution is the industry leading consolidated IT service desk. Brochure HP Service Manager: Setting the standard for IT Service
Master Data Management Architecture
Master Data Management Architecture Version Draft 1.0 TRIM file number - Short description Relevant to Authority Responsible officer Responsible office Date introduced April 2012 Date(s) modified Describes
Concept of Operations for Line of Business Initiatives
Concept of Operations for Line of Business Initiatives Version 1.0 Office of E-Gov and IT, OMB March 2006 Table of Contents FOREWORD...2 1 OBJECTIVES OF THE LINES OF BUSINESS CONCEPT OF OPERATIONS...3
Effective Data Governance
perspective Effective Data Governance Abstract Data governance is no more just another item that is good to talk about and nice to have, for global data management organizations. This PoV looks into why
How To Change A Business Model
SOA governance and organizational change strategy White paper November 2007 Enabling SOA through organizational change Sandy Poi, Global SOA Offerings Governance lead, associate partner, Financial Services
California Enterprise Architecture Framework
Version 2.0 August 01, 2013 This Page is Intentionally Left Blank Version 2.0 ii August 01, 2013 TABLE OF CONTENTS 1 Executive Summary... 1 1.1 What is Enterprise Architecture?... 1 1.2 Why do we need
Project, Program & Portfolio Management Help Leading Firms Deliver Value
in collaboration with Project, Program & Portfolio Help Leading Firms Deliver Value Managing Effectively & Efficiently Through an Enterprise PMO Program & Portfolio : Aligning IT Capabilities with Business
Framework for SOA services
Advisory Services Business Systems Integration Framework for SOA services Service-oriented architecture can transform the IT landscape by increasing efficiencies and decreasing costs. But the architecture
The ROI of Data Governance: Seven Ways Your Data Governance Program Can Help You Save Money
A DataFlux White Paper Prepared by: Gwen Thomas The ROI of Data Governance: Seven Ways Your Data Governance Program Can Help You Save Money Leader in Data Quality and Data Integration www.dataflux.com
TOGAF. TOGAF & Major IT Frameworks, Architecting the Family. by Danny Greefhorst, MSc., Director of ArchiXL. IT Governance and Strategy
TOGAF TOGAF & Major IT Frameworks, Architecting the Family by Danny Greefhorst, MSc., Director of ArchiXL TOGAF is a registered trademark of The Open Group. Copyright 2013 ITpreneurs. All rights reserved.
