Data Integration Alternatives Managing Value and Quality
|
|
|
- Hugo Neal
- 10 years ago
- Views:
Transcription
1 Solutions for Enabling Lifetime Customer Relationships Data Integration Alternatives Managing Value and Quality Using a Governed Approach to Incorporating Data Quality Services Within the Data Integration Process WHITE PAPER: DATA QUALITY & DATA INTEGRATION David Loshin President, Knowledge Integrity, Inc.
2 WHITE PAPER: DATA QUALITY & DATA INTEGRATION Data Integration Alternatives Managing Value and Quality Using a Governed Approach to Incorporating Data Quality Services Within the Data Integration Process 2 ABSTRACT WITH THE RATE OF DATA VOLUME GROWTH INCREASING AT A BREAKNECK PACE, SMART ORGANIZATIONS ARE INCREASINGLY RELYING ON REPORTING AND ANALYTICS TO NOT JUST RUN, BUT IMPROVE THE WAY THAT BUSINESS IS DONE. YET WHILE DECADES OF INVESTMENTS IN TRANSACTIONAL AND OPERATIONAL BUSINESS APPLICATIONS HAVE LED TO VIRTUAL ISLANDS OF DATA, THERE IS A GROWING LIST OF ENTERPRISE APPLICATIONS SUCH AS BUSINESS INTELLIGENCE AND DATA WAREHOUSING, CUSTOMER RELATIONSHIP MANAGEMENT, ENTERPRISE RESOURCE PLANNING, AND EVEN MORE COMPLEX APPROACHES TO BUSINESS ANALYTICS THAT REQUIRE ACCESS TO DATA SETS FROM A VARIETY OF SOURCES.
3 KEY DATA GOVERNANCE PROCESSES FACILITATE THE COLLECTION AND STANDARDIZATION OF ENTERPRISE DATA QUALITY REQUIREMENTS Data Integration is Everywhere With the rate of data volume growth increasing at a breakneck pace, smart organizations are increasingly relying on reporting and analytics to not just run, but improve the way that business is done. Data centralization becomes key to deploying strategic enterprise applications. Operational data stores, data warehouses, data marts, mash ups, federated operational systems, self-service reporting, data exchanges, and other analytic and operational applications require a greater degree of data sharing than ever before. Satisfying the information demands of these secondary-use business applications becomes a primary objective, and that means moving data from the original sources to the target business data systems. In other words, those needs must be satisfied using data integration. Data Integration Alternatives As the needs of downstream consumers have become more sophisticated, different approaches to data integration have evolved. The more traditional extract, transform, load (ETL) approach takes the data from its sources to a staging area in which data sets are manipulated and transformed into a target representation. An alternate approach is data virtualization, in which the data remains stored at the source and a conceptual view is materialized on demand. Traditional ETL The most common approach to data integration relies on some variation of the ETL paradigm. Because data sources that are used to populate downstream and secondaryuse business applications often live in many different formats, file and/or table structures, and sometimes even using different underlying character encodings, there is a predisposition to normalize data set representations before attempting to merge them into a target downstream system. The extraction component implies specially-engineered routines employed to fetch data from the sources, which will also require specially-designed transformations that apply a series of functions to normalize, cleanse, standardize, derive, translate, and other functions necessary to massage the data into a format that is suitably consistent with other transformed data sources in preparation for the target data systems. At that point, the data is ready to be propagated and loaded into the target destination, either overwriting the existing data or periodically augmenting the existing target data set. Data Virtualization As opposed to the traditional approach of extracting data from multiple sources and temporarily storing those data sets at a staging area, a different approach, called data federation or data virtualization, allows the source data sets to remain in their original locations. Data virtualization introduces abstraction layers over a variety of native data sources and, as a byproduct, provides relational views without requiring that data be extracted from its source. This approach to abstraction enables the creation of reusable data services, and the data abstraction layers typically deployed within a data virtualization environment allow for the presentation of a standardized a logical representation of enterprise data concepts, thereby allowing many different downstream data consumers to see a view of the data that is both structurally and semantically consistent. Data Challenges: Completeness, Consistency, Reasonableness With an increased interest in developing business applications that repurpose primary data sources for secondary uses, downstream consumers may have widely different expectations of the data, especially in terms of data quality. But when data sets are used for purposes for which they were not originally intended, those data users are often forced to redefine and reinterpret the meaning of the original data sets. 3
4 WHITE PAPER: DATA QUALITY & DATA INTEGRATION Data Integration Alternatives Managing Value and Quality Using a Governed Approach to Incorporating Data Quality Services Within the Data Integration Process 4 Extracting and transforming the data multiple times in different ways for different applications may lead to variant results and continual need for reconciliations, leading to mistrust, wasted time, rework, and questionable results. Some typical issues include: Missing data element values that skew counts and other aggregations Variance in use of commonly-accepted reference data introduce inconsistencies and inaccuracies Differences in formats, structures, and semantics presumed by downstream business applications may lead to drawing different conclusions from similar analyses Inconsistency in reporting leads to an ongoing need for reconciliation of generated reports Different implied semantics leads to misinterpretations and missed reasonableness expectations Essentially, the absence of standards for structure, formats, and definitions of repurposed data leads to issues that emerge as a result of ungoverned data integration processes incomplete data, inconsistent data, and data that does not meet reasonableness expectations. Adding Value Using Governed Data Quality Services If the challenges are introduced as a byproduct of the need for data integration, one approach to solving it is by recognizing that the issues exist, figuring out what the issues are, and then retooling the data integration process so that it incorporates ways to remediate data flaws before they have material impact. Fortunately, a governed approach to incorporating data quality services within consolidation is possible. The first step is to institute some key data governance processes to facilitate the collection and standardization of enterprise data quality requirements. The second step is embedding data quality management techniques within the data integration strategy. Data Governance Practices While an enterprise data governance program will encompass a wide variety of data management processes, the practical demands of data integration suggest focusing on a subset of those practices that directly support data integration, such as: Data requirements analysis Typical business application development considers the collection of data requirements as subsidiary to the functional requirements analysis process. But because enterprise projects such as data warehousing and customer relationship management cross line-of-business boundaries, there is a need for a well-defined process for soliciting, documenting, and synthesizing the collected information expectations that all downstream users will expect, and then translate those expectations into data requirements to be imposed on all candidate data sources. Not only does this impose a radical change in requirements gathering, it also requires the kind of oversight provided by a data governance infrastructure. Data standards review Defining data standards can address the challenge of inconsistency, especially aligning data element definitions and semantics. When key stakeholders from across the enterprise participate in a review and approval process for proposed data standards, there is a degree of confidence that the standards will be defined so that the collected downstream data consumer requirements will be observed.
5 HAVING DATA GOVERNANCE PRACTICES IN PLACE SIMPLIFIES THE INCORPORATION OF DATA QUALITY TECHNIQUES Metadata management These include processes for documenting the approved standard structures and definitions for reference data domains and data exchange and providing a means for communicating those standards. Data Quality Services Having data governance practices in place simplifies the incorporation of data quality techniques such as these: Parsing and Standardization Parsing is a process that relies on defined formats, patterns, and structures to determine when data values conform to a common set of expectations. Parsing is used in concert with a set of standardization rules triggered to transform the input data into a form that can be more effectively used, either to standardize the representation (presuming a valid representation) or to correct the values (should known errors be identified). Parsing and standardization can employ a library of data domains and rules to split data values into multiple components and rearrange the components into a normalized format. Standardization can also change full words to abbreviations, or abbreviations to full words, transform nicknames into a standard name form, translate across languages (e.g., Spanish to English), correct common misspellings, and reduce value variance to improve record linkage for deduplication and data cleansing. Data cleansing When data values are recognized as being inconsistent or incorrect, and the data flaws cannot be corrected at the point of origin, an alternative is to apply transformation rules to impute data values, correct names or addresses, eliminate extraneous and/ or meaningless data, and even merge duplicate records. Cleansing the data directly ensures that the data that meets some level of suitability; incorporating the same approach to parsing, standardization, and cleansing as part of the data integration process standardizes the transformations so that a consistent view of the data is provided to all downstream data consumers. Data validation When there is no coordination among the data consumers they might not necessarily apply the same data validations in the same way. Even if their rules are the same or similar, when validations are applied at the point of use, it is unlikely that the rules would be executed in the same order, or that the thresholds for acceptability would be the same. As a result, even though the same sources are being used, the results of the validation may vary as well. By incorporating a standard set of data validations within the data integration process, the constraints can be tested at specific points in the information flow process, thereby reducing the risk of inconsistency. Soliciting data quality requirements from the collection of downstream data consumers allows you to define data quality rules; implementing validation of compliance to these rules early in the process can help ensure that the quality of the data is sufficient to meet the business needs and allow any potential issues to be identified and remediated early and consistently. Considerations The approach used for implementing data integration should not interfere with the desire to improve the quality of the data in a coherent and consistent manner. Most traditional ETL tools vendors have recognized the need for incorporating data quality techniques, and many have either forged partner relationships with, or have completely acquired data quality tools vendors. Today, it is rare to find an end-to-end extraction, transformation, and loading tool that does not encourage the definition and implementation of embedded parsing, standardization, cleansing, and validation. 5
6 WHITE PAPER: DATA QUALITY & DATA INTEGRATION Data Integration Alternatives Managing Value and Quality Using a Governed Approach to Incorporating Data Quality Services Within the Data Integration Process 6 Alternatively, data virtualization tools are also increasingly tethered to data quality tools and technology. The abstraction provided via data virtualization provides an opportunity for the data management team to engage the downstream consumers, solicit their data quality requirements, and directly embed data attribute-based validations within one of the layers of abstraction. Consolidating data quality requirements and implementing data validation constraints at specific points in the information production flow reduces the risk of inconsistency, helps to ensure that the data quality is sufficient to meet the downstream data consumer needs, and alerts the data stewards to any potential issues that can be identified and remediated early in the data integration process. Here is the bottom line: Data integration is becoming pervasive across the organization. By introducing governed process that simplify information reuse in a consistent manner, trust in reporting and analytics will increase, benefitting all stakeholders across the organization! Spectrum TM technology from Pitney Bowes helps organizations provide trusted customer data and associated insights, in context, to such business applications as business intelligence, analytics, data warehousing and CRM. This allows organizations to personalize the customer experience across channels, accelerate compliance initiatives, better manage risk associated with customers and make business operations more efficient. FOR MORE INFORMATION ON DATA QUALITY AND DATA INTEGRATION SOLUTIONS, CALL PITNEY BOWES SOFTWARE AT OR VISIT PB.COM/SOFTWARE TODAY OR VISIT OUR WEBSITES.
7 7 David Loshin is the President of Knowledge Integrity, Inc., a consulting and development company focusing on customized information management solutions including information quality solutions consulting, information quality training and business rules solutions. Loshin is the author of Master Data Management, Enterprise Knowledge Management The Data Quality Approach and Business Intelligence The Savvy Manager s Guide and is a frequent speaker on maximizing the value of information. He can be reached at [email protected] or at (301)
8 Every connection is a new opportunity UNITED STATES One Global View Troy, NY [email protected] CANADA 26 Wellington Street East Suite 500 Toronto, ON M5E 1S [email protected] EUROPE/UNITED KINGDOM The Smith Centre The Fairmile Henley on Thames Oxon RG9 6AB UK [email protected] ASIA PACIFIC/AUSTRALIA Level 7, 1 Elizabeth Plaza North Sydney NSW [email protected] [email protected] Pitney Bowes Software Inc. is a wholly-owned subsidiary of Pitney Bowes Inc. Pitney Bowes, the Corporate logo, Spectrum and Portrait are trademarks of Pitney Bowes Inc. or a subsidiary. All other trademarks are the property of their respective owners Pitney Bowes Software Inc. All rights reserved AMER 1401
Data Integration Alternatives Managing Value and Quality
Solutions for Customer Intelligence, Communications and Care. Data Integration Alternatives Managing Value and Quality Using a Governed Approach to Incorporating Data Quality Services Within the Data Integration
Effecting Data Quality Improvement through Data Virtualization
Effecting Data Quality Improvement through Data Virtualization Prepared for Composite Software by: David Loshin Knowledge Integrity, Inc. June, 2010 2010 Knowledge Integrity, Inc. Page 1 Introduction The
Three Fundamental Techniques To Maximize the Value of Your Enterprise Data
Three Fundamental Techniques To Maximize the Value of Your Enterprise Data Prepared for Talend by: David Loshin Knowledge Integrity, Inc. October, 2010 2010 Knowledge Integrity, Inc. 1 Introduction Organizations
Five Fundamental Data Quality Practices
Five Fundamental Data Quality Practices W H I T E PA P E R : DATA QUALITY & DATA INTEGRATION David Loshin WHITE PAPER: DATA QUALITY & DATA INTEGRATION Five Fundamental Data Quality Practices 2 INTRODUCTION
Business Drivers for Data Quality in the Utilities Industry
Solutions for Enabling Lifetime Customer Relationships. Business Drivers for Data Quality in the Utilities Industry Xxxxx W HITE PAPER: UTILITIES WHITE PAPER: UTILITIES Business Drivers for Data Quality
Enterprise Location Intelligence
Solutions for Customer Intelligence, Communications and Care. Enterprise Location Intelligence Bringing Location-related Business Insight to Support Better Decision Making and More Profitable Operations
Enterprise Location Intelligence
Solutions for Enabling Lifetime Customer Relationships Enterprise Location Intelligence Bringing Location-related Business Insight to Support Better Decision Making and More Profitable Operations W HITE
Data Governance, Data Architecture, and Metadata Essentials
WHITE PAPER Data Governance, Data Architecture, and Metadata Essentials www.sybase.com TABLE OF CONTENTS 1 The Absence of Data Governance Threatens Business Success 1 Data Repurposing and Data Integration
The Geocoding Advantage: Best Practices for Managing Customer and Location-based Data in Telecommunications
Solutions for Enabling Lifetime Customer Relationships The Geocoding Advantage: Best Practices for Managing Customer and Location-based Data in Telecommunications WHITE PAPER: LOCATION INTELLIGENCE Scott
Master Data Management Drivers: Fantasy, Reality and Quality
Solutions for Customer Intelligence, Communications and Care. Master Data Management Drivers: Fantasy, Reality and Quality A Review and Classification of Potential Benefits of Implementing Master Data
Data Governance, Data Architecture, and Metadata Essentials Enabling Data Reuse Across the Enterprise
Data Governance Data Governance, Data Architecture, and Metadata Essentials Enabling Data Reuse Across the Enterprise 2 Table of Contents 4 Why Business Success Requires Data Governance Data Repurposing
Busting 7 Myths about Master Data Management
Knowledge Integrity Incorporated Busting 7 Myths about Master Data Management Prepared by: David Loshin Knowledge Integrity, Inc. August, 2011 Sponsored by: 2011 Knowledge Integrity, Inc. 1 (301) 754-6350
5 Best Practices for SAP Master Data Governance
5 Best Practices for SAP Master Data Governance By David Loshin President, Knowledge Integrity, Inc. Sponsored by Winshuttle, LLC 2012 Winshuttle, LLC. All rights reserved. 4/12 www.winshuttle.com Introduction
The Actionable Barcode
Solutions for Enabling Lifetime Customer Relationships The Actionable Barcode How organizations have turned a new postal mandate into much-needed customer insight WHITE PAPER: CUSTOMER EXPERIENCE MANAGEMENT
Data Governance. Data Governance, Data Architecture, and Metadata Essentials Enabling Data Reuse Across the Enterprise
Data Governance Data Governance, Data Architecture, and Metadata Essentials Enabling Data Reuse Across the Enterprise 2 Table of Contents 4 Why Business Success Requires Data Governance Data Repurposing
Building a Data Quality Scorecard for Operational Data Governance
Building a Data Quality Scorecard for Operational Data Governance A White Paper by David Loshin WHITE PAPER Table of Contents Introduction.... 1 Establishing Business Objectives.... 1 Business Drivers...
Supporting Your Data Management Strategy with a Phased Approach to Master Data Management WHITE PAPER
Supporting Your Data Strategy with a Phased Approach to Master Data WHITE PAPER SAS White Paper Table of Contents Changing the Way We Think About Master Data.... 1 Master Data Consumers, the Information
Considerations: Mastering Data Modeling for Master Data Domains
Considerations: Mastering Data Modeling for Master Data Domains David Loshin President of Knowledge Integrity, Inc. June 2010 Americas Headquarters EMEA Headquarters Asia-Pacific Headquarters 100 California
The secret to reducing churn
Solutions for Enabling Lifetime Customer Relationships The secret to reducing churn Leveraging analytics for better customer insight WHITE PAPER: COMMUNICATIONS Dr. Patrick Surry Global Solution Owner
Sphere of influence. The Importance of Social Network Analysis. Solutions for Enabling Lifetime Customer Relationships SOCIAL NETWORK ANALYSIS
Solutions for Enabling Lifetime Customer Relationships Sphere of influence The Importance of Social Network Analysis WHITE PAPER: SOCIAL NETWORK ANALYSIS Navin Sharma VP, Product Management, Customer Information
Challenges in the Effective Use of Master Data Management Techniques WHITE PAPER
Challenges in the Effective Use of Master Management Techniques WHITE PAPER SAS White Paper Table of Contents Introduction.... 1 Consolidation: The Typical Approach to Master Management. 2 Why Consolidation
CHAPTER SIX DATA. Business Intelligence. 2011 The McGraw-Hill Companies, All Rights Reserved
CHAPTER SIX DATA Business Intelligence 2011 The McGraw-Hill Companies, All Rights Reserved 2 CHAPTER OVERVIEW SECTION 6.1 Data, Information, Databases The Business Benefits of High-Quality Information
Solutions for Customer Intelligence, Communications, and Care. Public Sector. DMV Capabilities. Every connection is a new opportunity TM
Solutions for Customer Intelligence, Communications, and Care. Public Sector DMV Capabilities Every connection is a new opportunity TM Pitney Bowes Business Insight provides State Governments with a combination
Portrait Customer Analytic Solutions
Solutions for Enabling Lifetime Customer Relationships Portrait Customer Analytic Solutions Explore, understand and predict customer behavior for optimal ROI Every connection is a new opportunity TM Pitney
Effective Strategies for Unifying Customer Communications
Enabling Lifetime Customer Relationships. Effective Strategies for Unifying Customer Communications Leveraging Your Unique Customer Exchanges to Stay Ahead of the Competition W HITE PAPER: CUSTOMER COMMUNICATION
Master Data Management
Master Data Management David Loshin AMSTERDAM BOSTON HEIDELBERG LONDON NEW YORK OXFORD PARIS SAN DIEGO Ик^И V^ SAN FRANCISCO SINGAPORE SYDNEY TOKYO W*m k^ MORGAN KAUFMANN PUBLISHERS IS AN IMPRINT OF ELSEVIER
Operationalizing Data Governance through Data Policy Management
Operationalizing Data Governance through Data Policy Management Prepared for alido by: David Loshin nowledge Integrity, Inc. June, 2010 2010 nowledge Integrity, Inc. Page 1 Introduction The increasing
Data Quality and Cost Reduction
Data Quality and Cost Reduction A White Paper by David Loshin WHITE PAPER SAS White Paper Table of Contents Introduction Data Quality as a Cost-Reduction Technique... 1 Understanding Expenses.... 1 Establishing
JOURNAL OF OBJECT TECHNOLOGY
JOURNAL OF OBJECT TECHNOLOGY Online at www.jot.fm. Published by ETH Zurich, Chair of Software Engineering JOT, 2008 Vol. 7, No. 8, November-December 2008 What s Your Information Agenda? Mahesh H. Dodani,
Data Governance for Master Data Management and Beyond
Data Governance for Master Data Management and Beyond A White Paper by David Loshin WHITE PAPER Table of Contents Aligning Information Objectives with the Business Strategy.... 1 Clarifying the Information
How To Manage Your Parcel Management
Shipping & Mailing The new parcel management How automated services help schools attract students and lower costs. New challenges call for new solutions. 77% of college students made an online purchase
Enterprise Data Quality
Enterprise Data Quality An Approach to Improve the Trust Factor of Operational Data Sivaprakasam S.R. Given the poor quality of data, Communication Service Providers (CSPs) face challenges of order fallout,
Predictive Analytics. Helping you uncover valuable insights and develop effective business strategies
Predictive Analytics Helping you uncover valuable insights and develop effective business strategies Predictive Analytics Helping you uncover valuable insights and develop effective business strategies.
5 Best Practices for SAP Master Data Governance
5 Best Practices for SAP Master Data Governance By David Loshin President, Knowledge Integrity, Inc. Sponsored by Winshuttle, LLC Executive Summary Successful deployment of ERP solutions can revolutionize
Branch Sales Opportunity Based Myths
Branch Sales Opportunity Based Myths Common Misconceptions of Branch Sales Performance W H I T E PA P E R : FINANCIAL SERVICES Brian Diepold PhD, Senior Client Services Manager, Pitney Bowes Business Insight
White Paper. Thirsting for Insight? Quench It With 5 Data Management for Analytics Best Practices.
White Paper Thirsting for Insight? Quench It With 5 Data Management for Analytics Best Practices. Contents Data Management: Why It s So Essential... 1 The Basics of Data Preparation... 1 1: Simplify Access
Principal MDM Components and Capabilities
Principal MDM Components and Capabilities David Loshin Knowledge Integrity, Inc. 1 Agenda Introduction to master data management The MDM Component Layer Model MDM Maturity MDM Functional Services Summary
Measure Your Data and Achieve Information Governance Excellence
SAP Brief SAP s for Enterprise Information Management SAP Information Steward Objectives Measure Your Data and Achieve Information Governance Excellence A single solution for managing enterprise data quality
Knowledgent White Paper Series. Developing an MDM Strategy WHITE PAPER. Key Components for Success
Developing an MDM Strategy Key Components for Success WHITE PAPER Table of Contents Introduction... 2 Process Considerations... 3 Architecture Considerations... 5 Conclusion... 9 About Knowledgent... 10
Real-time customer information data quality and location based service determination implementation best practices.
White paper Location Intelligence Location and Business Data Real-time customer information data quality and location based service determination implementation best practices. Page 2 Real-time customer
Using analytics to drive better communications
Solutions for Enabling Lifetime Customer Relationships Link customer preferences and behavior with business solutions and document requirements to deliver WHITE PAPER: COMMUNICATION MANAGEMENT Lisa Sutrick
Customer Segmentation and Profitability
Customer Segmentation and Profitability Building Loyal, Profitable Customers S O L U T I O N P A P E R : FINANCIAL SERVICES SOLUTION PAPER: FINANCIAL SERVICES Customer Segmentation and Profitability Building
EAI vs. ETL: Drawing Boundaries for Data Integration
A P P L I C A T I O N S A W h i t e P a p e r S e r i e s EAI and ETL technology have strengths and weaknesses alike. There are clear boundaries around the types of application integration projects most
MDM Components and the Maturity Model
A DataFlux White Paper Prepared by: David Loshin MDM Components and the Maturity Model Leader in Data Quality and Data Integration www.dataflux.com 877 846 FLUX International +44 (0) 1753 272 020 One common
DATA QUALITY MATURITY
3 DATA QUALITY MATURITY CHAPTER OUTLINE 3.1 The Data Quality Strategy 35 3.2 A Data Quality Framework 38 3.3 A Data Quality Capability/Maturity Model 42 3.4 Mapping Framework Components to the Maturity
Business Intelligence Enabling Transparency across the Enterprise
White Paper Business Intelligence Enabling Transparency across the Enterprise Business solutions through information technology Entire contents 2004 by CGI Group Inc. All rights reserved. Reproduction
dxhub Denologix MDM Solution Page 1
Most successful large organizations are organized by lines of business (LOB). This has been a very successful way to organize for the accountability of profit and loss. It gives LOB leaders autonomy to
Portrait Interaction Optimizer TM. Convert inbound interactions into revenue and retention opportunities
Portrait Interaction Optimizer TM Convert inbound interactions into revenue and retention opportunities Deliver real-time best-next-action recommendations to each customer, wherever and whenever they choose
Conversing with Your Customers
Solutions for Enabling Lifetime Customer Relationships Conversing with Your Customers Turning call centers into profit centers while building customer lifetime value WHITE W PAPER: COMMUNICATIONS Jeff
Take control of inbound packages
Shipping & Mailing Take control of inbound packages Track, monitor and deliver with confidence Say goodbye to lost or misplaced Information is everything. packages. Every carrier knows that. They barcode,
Business Performance & Data Quality Metrics. David Loshin Knowledge Integrity, Inc. [email protected] (301) 754-6350
Business Performance & Data Quality Metrics David Loshin Knowledge Integrity, Inc. [email protected] (301) 754-6350 1 Does Data Integrity Imply Business Value? Assumption: improved data quality,
Master Your Data and Your Business Using Informatica MDM. Ravi Shankar Sr. Director, MDM Product Marketing
Master Your and Your Business Using Informatica MDM Ravi Shankar Sr. Director, MDM Product Marketing 1 Driven Enterprise Timely Trusted Relevant 2 Agenda Critical Business Imperatives Addressed by MDM
Next Generation Business Performance Management Solution
Next Generation Business Performance Management Solution Why Existing Business Intelligence (BI) Products are Inadequate Changing Business Environment In the face of increased competition, complex customer
Making Business Intelligence Easy. Whitepaper Measuring data quality for successful Master Data Management
Making Business Intelligence Easy Whitepaper Measuring data quality for successful Master Data Management Contents Overview... 3 What is Master Data Management?... 3 Master Data Modeling Approaches...
Data Governance. David Loshin Knowledge Integrity, inc. www.knowledge-integrity.com (301) 754-6350
Data Governance David Loshin Knowledge Integrity, inc. www.knowledge-integrity.com (301) 754-6350 Risk and Governance Objectives of Governance: Identify explicit and hidden risks associated with data expectations
BUSINESSOBJECTS DATA INTEGRATOR
PRODUCTS BUSINESSOBJECTS DATA INTEGRATOR IT Benefits Correlate and integrate data from any source Efficiently design a bulletproof data integration process Improve data quality Move data in real time and
Harness the value of information throughout the enterprise. IBM InfoSphere Master Data Management Server. Overview
IBM InfoSphere Master Data Management Server Overview Master data management (MDM) allows organizations to generate business value from their most important information. Managing master data, or key business
The Infrastructure for Information Management: A Brave New World for the CIO WHITE PAPER
The Infrastructure for Information Management: A Brave New World for the CIO WHITE PAPER SAS White Paper Table of Contents Trends and Drivers for Information Infrastructure.... 1 Objectives for Organizational
Populating a Data Quality Scorecard with Relevant Metrics WHITE PAPER
Populating a Data Quality Scorecard with Relevant Metrics WHITE PAPER SAS White Paper Table of Contents Introduction.... 1 Useful vs. So-What Metrics... 2 The So-What Metric.... 2 Defining Relevant Metrics...
Informatica Solutions for Healthcare Providers. Unlock the Potential of Data Driven Healthcare
S O L U T I O N S B R O C H U R E Informatica Solutions for Healthcare Providers Unlock the Potential of Data Driven Healthcare Informatica Solutions For Healthcare Providers Fundamental change in the
The ROI of Data Quality
Solutions for Enabling Lifetime Customer Relationships The ROI of Data Quality WHITE PAPER: DATA MANAGEMENT WHITE PAPER: INDUSTRY The ROI of Data Quality 2 ABSTRACT FOR AT LEAST A DECADE NOW, ANALYSTS
Integrating Data Governance into Your Operational Processes
TDWI rese a rch TDWI Checklist Report Integrating Data Governance into Your Operational Processes By David Loshin Sponsored by tdwi.org August 2011 TDWI Checklist Report Integrating Data Governance into
From the Document to the Discourse
From the Document to the Discourse Why Interactive, Multi-Channel Customer Communications are Key for Insurance Carriers Growth & Success W H I T E PA P E R : HEALTHCARE/INSURANCE Pitney Bowes Business
Digital Self Service. Reduce costs and maximize customer relationships by creating meaningful personalized self-service experiences.
Digital Self Service Reduce costs and maximize customer relationships by creating meaningful personalized self-service experiences. Give customers the answers they re looking for and you ll get the results
A Design Technique: Data Integration Modeling
C H A P T E R 3 A Design Technique: Integration ing This chapter focuses on a new design technique for the analysis and design of data integration processes. This technique uses a graphical process modeling
Confirm. Infrastructure asset management solutions for improved efficiency, lower costs, and simplified compliance.
Confirm Infrastructure asset management solutions for improved efficiency, lower costs, and simplified compliance. Confirm Confirm is the recognised leader in infrastructure asset management and maintenance
Big Data Analytics with IBM Cognos BI Dynamic Query IBM Redbooks Solution Guide
Big Data Analytics with IBM Cognos BI Dynamic Query IBM Redbooks Solution Guide IBM Cognos Business Intelligence (BI) helps you make better and smarter business decisions faster. Advanced visualization
Successful Outsourcing of Data Warehouse Support
Experience the commitment viewpoint Successful Outsourcing of Data Warehouse Support Focus IT management on the big picture, improve business value and reduce the cost of data Data warehouses can help
ORACLE ENTERPRISE DATA QUALITY PRODUCT FAMILY
ORACLE ENTERPRISE DATA QUALITY PRODUCT FAMILY The Oracle Enterprise Data Quality family of products helps organizations achieve maximum value from their business critical applications by delivering fit
The Informatica Solution for Improper Payments
The Informatica Solution for Improper Payments Reducing Improper Payments and Improving Fiscal Accountability for Government Agencies WHITE PAPER This document contains Confidential, Proprietary and Trade
BUSINESSOBJECTS DATA INTEGRATOR
PRODUCTS BUSINESSOBJECTS DATA INTEGRATOR IT Benefits Correlate and integrate data from any source Efficiently design a bulletproof data integration process Accelerate time to market Move data in real time
Data virtualization: Delivering on-demand access to information throughout the enterprise
IBM Software Thought Leadership White Paper April 2013 Data virtualization: Delivering on-demand access to information throughout the enterprise 2 Data virtualization: Delivering on-demand access to information
Data warehouse and Business Intelligence Collateral
Data warehouse and Business Intelligence Collateral Page 1 of 12 DATA WAREHOUSE AND BUSINESS INTELLIGENCE COLLATERAL Brains for the corporate brawn: In the current scenario of the business world, the competition
Practical Fundamentals for Master Data Management
Practical Fundamentals for Master Data Management How to build an effective master data capability as the cornerstone of an enterprise information management program WHITE PAPER SAS White Paper Table of
Next Generation Customer Onboarding
Next Generation Customer Onboarding Driving an Automated, Integrated Multi-Channel Onboarding Strategy S O L U T I O N P A P E R : FINANCIAL SERVICES SOLUTION PAPER: FINANCIAL SERVICES Next Generation
Chapter 6 Basics of Data Integration. Fundamentals of Business Analytics RN Prasad and Seema Acharya
Chapter 6 Basics of Data Integration Fundamentals of Business Analytics Learning Objectives and Learning Outcomes Learning Objectives 1. Concepts of data integration 2. Needs and advantages of using data
Understanding the Financial Value of Data Quality Improvement
Understanding the Financial Value of Data Quality Improvement Prepared by: David Loshin Knowledge Integrity, Inc. January, 2011 Sponsored by: 2011 Knowledge Integrity, Inc. 1 Introduction Despite the many
