SATISFYING NEW REQUIREMENTS FOR DATA INTEGRATION
|
|
- Augustus Clark
- 8 years ago
- Views:
Transcription
1 TDWI RESEARCH TDWI CHECKLIST REPORT SATISFYING NEW REQUIREMENTS FOR DATA INTEGRATION By David Loshin Sponsored by tdwi.org
2 JUNE 2012 TDWI CHECKLIST REPORT SATISFYING NEW REQUIREMENTS FOR DATA INTEGRATION By David Loshin TABLE OF CONTENTS 2 FOREWORD 2 NUMBER ONE Increase performance and efficiency. 3 NUMBER TWO Integrate the cloud. 3 NUMBER THREE Protect information in the integration layer. 4 NUMBER FOUR Embed master data services. 4 NUMBER FIVE Process big data and enterprise data. 5 NUMBER SIX Satisfy real-time demands. 5 NUMBER SEVEN Develop data quality and data governance policies and practices. 6 ABOUT OUR SPONSOR 6 ABOUT THE TDWI CHECKLIST REPORT SERIES 6 ABOUT THE AUTHOR 6 ABOUT TDWI RESEARCH 1201 Monster Road SW, Suite 250 Renton, WA T F E info@tdwi.org tdwi.org 2012 by TDWI (The Data Warehousing Institute TM ), a division of 1105 Media, Inc. All rights reserved. Reproductions in whole or in part are prohibited except by written permission. requests or feedback to info@tdwi.org. Product and company names mentioned herein may be trademarks and/or registered trademarks of their respective companies.
3 FOREWORD NUMBER ONE INCREASE PERFORMANCE AND EFFICIENCY. New trends in the industry are making data integration more paramount than ever. Essentially, the concept of data integration is being redefined; it is no longer limited to extracting data sets from internal sources and loading them into a data warehouse, but instead focuses on effectively facilitating the delivery of information to the right places within the appropriate time. Data integration goes beyond extract, transform, and load (ETL); data replication; and changed data capture, although these remain key components of the integration fabric. It is also hard to ignore the big data revolution as organizations seek to adapt their information management environments to accommodate massive data volumes coming from a large variety of sources (especially internal ones!) at accelerating speeds. Some of the challenges include absorbing numerous data feeds (both internal and external), moving data to analytical appliances designed for big data analytics, coupling the results with existing data warehouse and business intelligence (BI) architectures, and delivering results to a variety of downstream knowledge workers and information consumers. When coupled with the migration of storage, infrastructure, and business applications to the cloud, we see how these trends require moving large amounts of data to many different places in real time without allowing data replicas to become out of sync with each other. Mounds of structured data, unstructured data, big data, and advancements in cloud technology are imposing new requirements for data integration. This TDWI Checklist Report will explore some of the key drivers these new requirements are intended to address. Whether you are looking to support the performance needs of big data applications, filter concepts from unstructured data, monitor hundreds of data feeds for unexpected behavior, export data across enterprise boundaries, or provide real-time reporting and analysis, there is a rapidly expanding need for data integration competency that extends well beyond traditional ETL. People are increasingly recognizing that repurposing transactional data for analytical purposes yields significant value for improving many aspects of the business. Part of this epiphany is attributable to data volume growth, improved data visualization methods, and the lowered barrier to entry for business intelligence and analytics. Yet, some BI initiatives are at risk of succumbing to the perils of their own success: a growing user community, increased system demand, the need to make optimal use of high-performance platforms and programming models, and the need for simultaneous and rapid access to lots of data. Failed expectations for increased performance are linked to the familiar bottlenecks of data-access latency, coupled with the general need for increased performance in information delivery for both traditional and emerging techniques: 1. Platforms: Analytical appliances provide in-memory computation, yet are throttled by the need to stream data in and flow results out. 2. Demands: Today s environments must satisfy a mixture of workloads, requiring sophisticated queries with many-way joins that require numerous data exchanges to satisfy evaluation of the join conditions. 3. Execution: There is a need to optimize data access for query processing on appliances that support online transaction processing (OLTP) and data warehousing, especially when it comes to intermediate data transformations. 4. Big data analytics: Hadoop requires data integration and exchange support for the phase transitions between the calculation and reduction phases of the MapReduce programming model, especially as the sizes of the inputs grow and the complexity of the algorithms increases. Each of these demands requires innovation in optimizing the use of both new and existing data integration techniques to reduce the data-access bottleneck. Some approaches include using knowledge of canned reports to optimize data distribution and layouts, employing alternate data alignments (such as column orientation, or realignment of column order to improve hardware cache utilization), aggressive use of data compression, use of high-speed network technology (such as Infiniband), data replication, changed data capture, and data federation techniques, all for improved performance. 2 TDWI RESEARCH tdwi.org
4 NUMBER TWO INTEGRATE THE CLOUD. NUMBER THREE PROTECT INFORMATION IN THE INTEGRATION LAYER. The evolution toward simultaneously exploiting both on-premises and cloud-based environments implies a need for greater agility in data acquisition, integration, and absorption. Moving data beyond the corporate firewall goes beyond solely extracting data from existing internal legacy data sources. The challenge becomes more acute as you look at the different scenarios where data instances and data sets cross the enterprise boundary. Some examples: Incorporating analytical results with cloud-based systems (such as SaaS-based sales, marketing, and contact management applications) Incorporating external data into an analytical platform (such as social network graph analysis that examines continuous data streams) When exposing reporting services to customers (such as account statements and analysis in the financial industry) Collaborative data analysis in which data sources are pooled from different enterprises (such as comparative effectiveness research in the healthcare industry) These examples suggest a need to smooth out the differences in structure and semantics in a bidirectional manner without introducing additional latencies and access bottlenecks. Sharing data across enterprise boundaries has implications for data exchange and integration. Here are some key characteristics for data exchange and sharing capabilities in the big data world: The ability to seamlessly access a variety of large data sets inside the organization (such as transaction data from different operational systems, or data in an enterprise data warehouse) The ability to seamlessly access a variety of large data sets outside the organization (including licensed data sets, public domain data sets, data in the cloud, data feeds, and streamed data) The ability to harmonize your data to unify meaning and structure of your data elements for all consumers The ability to share data with applications and environments hosted outside the organization (again, with a focus on interfacing with cloud-based systems and applications) Extra-enterprise data integration must enable these types of exchanges by providing data standards, canonical models, exchange schemas, and high-performance methods for cloud-based data access. It must also provide integration as services, which allows for information availability with different levels of transparency and reduces the effort required for application development. The need for extra-enterprise and cloud-oriented data integration is clear. But any scenario in which sensitive data is destined to flow outside the corporate boundary is bound to raise eyebrows from the compliance department. In spite of organizational motivations to migrate their applications and data to cloud-based systems, ensuring the security of protected data (especially in the financial and healthcare environments) is not only a good practice, but it is also the law. Most regulated industries have legal requirements for protecting personally identifiable data. And although many organizations feel secure that their IT security program effectively protects private data, the number of security breaches and the scale of the data that is exposed tell a different story. Whether that is a byproduct of deliberate hacking or social engineering, exposure of private data is on the rise. In other words, even in the face of existing IT security frameworks, there are situations in which those barriers are breached, exposing the supposedly secure data. Data integration is a particular point of potential security weakness. Anytime data is exchanged (either within or outside the enterprise), there is risk of exposure. This means that another new requirement for data integration is incorporating different aspects of information protection (especially in the cloud), including identity management, authentication, and authorization for any data exchange, whether it is delivering information to individuals or even in machine-tomachine transfers. Because sharing clear data poses an exposure risk, another mitigation strategy is the use of encryption for any data exchange as part of the integration layer. This addresses concerns about a breach of the typical IT security layer in the event of a security failure, the encrypted data is still unusable. 3 TDWI RESEARCH tdwi.org
5 NUMBER FOUR EMBED MASTER DATA SERVICES. NUMBER FIVE PROCESS BIG DATA AND ENTERPRISE DATA. Many organizations are putting significant emphasis on the use of master data repositories and master data management (MDM), which is intended to provide universal access to a unified presentation of information about uniquely identifiable entities. Typically, these entities are represented in different ways in multiple data sets across the environment. The typical approaches to MDM are simplistic: build a relatively static master data hub with simple data extraction and transformations prior to loading. But the data integration challenge becomes more complex for a number of reasons, such as: The increased number and granularity of type for master domains The greater emphasis on master domain usability Broad differences in size, format, structure, and meaning among the different (both static and dynamic) data sources Different requirements for use of the shared master information A growing number of domain-specific applications to align The need to resolve references to individual entities from a variety of semi-structured and unstructured data streams Entity identities are embedded within unstructured data in different ways, and the variety of unstructure makes it difficult to create a single set of standardization and transformation rules that can be universally applied. Plus, applications using master data will need coherent views of master domains (that is, their data instance values are consistent, timely, and up to date) to ensure consistency of business process operation. Master data management will require increasingly sophisticated methods for seamless integration of data into master data sets, indicating that the data integration framework directly embed master data services such as: Text analysis, automated tokenization, and semantic resolution Automated data standardization Entity extraction Data validation Integrated hierarchical mapping Big-data analytics applications are often intended to absorb many large structured and unstructured data sets and then calculate results, providing better predictive models and enhancing customer profiles. The analytical results add value when they are recombined with information persisted in existing data warehouses. This presents a data dependency between the two analytics styles: The big data analytics applications must be adept at parsing out and resolving entity concepts from unstructured data and linking those entities with recognized entities (such as customers or products) that are accessed from the data warehouse. The important results (such as customer profile enhancements) from the big data analytics applications must be appended to the persistent data warehouse models to support reporting, queries, dimensional analysis, etc. For example, a big data analytics program might scan many simultaneous social media streams, parse out individual identities and corresponding product preferences, access the individuals profiles from the data warehouse, recalculate product affinity scores, then update the profiles in the warehouse. Leveraging this integrated analysis implies two data integration requirements: 1. Unified data integration tooling: The data integration tools and runtimes must be complete and provide unified support for enterprise data (including both structured and unstructured data), along with a variety of data access methods, such as text files, traditional SQL, Hadoop, and other NoSQL frameworks. 2. Integrated analytics: The data integration layer should embed capabilities for parsing the relevant content in unstructured data. This combines newer techniques such as text analytics and entity extraction/identification with techniques associated with ETL and data cleansing, such as parsing, standardization, and identity resolution. More to the point: emerging big data technologies will need data integration tooling that will enable integration of business analytics and data warehouses. Searching, matching, and linking entity information in real time Real-time synchronization and coherence among the consumers of master data 4 TDWI RESEARCH tdwi.org
6 NUMBER SIX SATISFY REAL-TIME DEMANDS. NUMBER SEVEN DEVELOP DATA QUALITY AND DATA GOVERNANCE POLICIES AND PRACTICES. The need for real-time data integration emanates from two directions: the first is to reduce the latency between transaction events and the time those events can be incorporated into reports and analyses, and the second is to use real-time technology for continuous availability solutions (i.e., active-active) or live standby systems. Broader adoption of business intelligence in a pervasive manner across the organization enables a wide community of decision makers to be informed with actionable knowledge. A side effect of this broad adoption is heightened expectations for availability of that actionable knowledge, especially in relation to embedded business intelligence within operational processes. The results of BI and analytics need to be integrated directly into production operational applications to address real-time process needs, and transaction information must be fed in real time to the analytical applications. In other words, there is a need for real-time data acquisition as well as real-time delivery of actionable information to a variety of different types of channels and devices across different locations and networks. The concept of real-time data integration seems like a natural fit for all aspects of pervasive business intelligence, and offers the opportunity for an amazing upside. One might even hypothesize that everything should be real-time enabled for all BI and data warehouses. Yet the logistics of real-time data synchronization can impose significant performance criteria, especially in environments with many simultaneously executing production transaction and operational applications that are informed by different BI applications. Satisfying real-time data demands puts focus on adapting timeproven mechanisms for continuous availability to the world of information, including high-availability methods for systemic faulttolerance, live standby, synchronous replication, asynchronous replication, and changed data capture to maintain coherence across replicas. And let s not forget the big-data integration scenario either. The desire to analyze high-volume/high-velocity data streams and push the results to the right consumers creates another intersection point that benefits from real-time data integration. Data reuse and repurposing creates a dilemma in attempting to verify the suitability, usability, and quality of data. Data governance is a set of policies and practices intended to institute control over data quality and usability, as well as standardize business terms, data element definitions, and conformance to defined business rules. Requiring data governance policies and protocols for data integration within and across different organizations is not strictly a product requirement; it is a systemic one in which observation of data policies must be supported within the data integration fabric. The need to demonstrate how data complies with business policies can be satisfied through integrating data quality services within the data integration fabric. For example, there are various complexities of customer and party data when it comes to compliance with laws, especially when it comes to protection of private information, customer awareness (such as Know Your Customer in the financial sector), or identification of individuals on government watch lists. The preferred method for implementation is using data quality services that can be tightly coupled with the data integration layer instead of relying on jury-rigged solutions made of different tools cobbled together. With increased data exchanges, sharing, and repurposing, introducing policies and practices will guide the definition of standards and controls. This will also provide a means for monitoring compliance with agreed-to standards that can help reduce the uncertainty associated with data reuse and sharing among collaborative partners. The first step involves recognizing the need for data governance within organizational boundaries, and establishing the policies and procedures for ensuring data quality within the firewall. The more mature phase involves data governance that extends to collaborative communities (extending beyond the enterprise boundaries), which requires sharing common metadata and documenting, tracking, and managing data lineage, policy management, and data quality servicelevel agreements. In addition, improved alignment with metadata management capabilities helps trace the introduction of data errors and enables impact analysis and scoping of changes to eliminate their root causes. 5 TDWI RESEARCH tdwi.org
7 ABOUT OUR SPONSOR ABOUT THE AUTHOR Oracle Data Integration provides a fully unified set of products for building, deploying, and managing data-centric architectures for operational and analytical data integration across the enterprise. Oracle s products combine to provide all the elements of data integration real-time data movement, transformation, big data processing, data synchronization, data quality, data management, and data services to ensure that information is timely, accurate, and consistent across complex systems. Oracle (NASDAQ: ORCL) is the world s most complete, open, and integrated business software and hardware systems company. Visit David Loshin president of Knowledge Integrity, Inc. ( is a recognized thought leader, TDWI instructor, and expert consultant in the areas of data management and business intelligence. David is a prolific author regarding business intelligence best practices, as the author of numerous books and papers on data management, including The Practitioner s Guide to Data Quality Improvement, with additional content provided at David is a frequent invited speaker at conferences, Web seminars, and sponsored Web sites and channels, including His bestselling book, Master Data Management, has been endorsed by data management industry leaders, and his valuable MDM insights can be reviewed at David can be reached at loshin@knowledge-integrity.com. ABOUT THE TDWI CHECKLIST REPORT SERIES ABOUT TDWI RESEARCH TDWI Checklist Reports provide an overview of success factors for a specific project in business intelligence, data warehousing, or a related data management discipline. Companies may use this overview to get organized before beginning a project or to identify goals and areas of improvement for current projects. TDWI Research provides research and advice for business intelligence and data warehousing professionals worldwide. TDWI Research focuses exclusively on BI/DW issues and teams up with industry thought leaders and practitioners to deliver both broad and deep understanding of the business and technical challenges surrounding the deployment and use of business intelligence and data warehousing solutions. TDWI Research offers in-depth research reports, commentary, inquiry services, and topical conferences as well as strategic planning services to user and vendor organizations. 6 TDWI RESEARCH tdwi.org
Effecting Data Quality Improvement through Data Virtualization
Effecting Data Quality Improvement through Data Virtualization Prepared for Composite Software by: David Loshin Knowledge Integrity, Inc. June, 2010 2010 Knowledge Integrity, Inc. Page 1 Introduction The
More informationIntegrating Data Governance into Your Operational Processes
TDWI rese a rch TDWI Checklist Report Integrating Data Governance into Your Operational Processes By David Loshin Sponsored by tdwi.org August 2011 TDWI Checklist Report Integrating Data Governance into
More informationData Warehousing in the Cloud
TDWI RESEARCH TDWI CHECKLIST REPORT Data Warehousing in the Cloud By David Loshin Sponsored by: tdwi.org JULY 2015 TDWI CHECKLIST REPORT Data Warehousing in the Cloud By David Loshin TABLE OF CONTENTS
More informationThree Fundamental Techniques To Maximize the Value of Your Enterprise Data
Three Fundamental Techniques To Maximize the Value of Your Enterprise Data Prepared for Talend by: David Loshin Knowledge Integrity, Inc. October, 2010 2010 Knowledge Integrity, Inc. 1 Introduction Organizations
More informationUsing and Choosing a Cloud Solution for Data Warehousing
TDWI RESEARCH TDWI CHECKLIST REPORT Using and Choosing a Cloud Solution for Data Warehousing By Colin White Sponsored by: tdwi.org JULY 2015 TDWI CHECKLIST REPORT Using and Choosing a Cloud Solution for
More informationData Governance, Data Architecture, and Metadata Essentials
WHITE PAPER Data Governance, Data Architecture, and Metadata Essentials www.sybase.com TABLE OF CONTENTS 1 The Absence of Data Governance Threatens Business Success 1 Data Repurposing and Data Integration
More informationDATA REPLICATION FOR REAL-TIME DATA WAREHOUSING AND ANALYTICS
TDWI RESE A RCH TDWI CHECKLIST REPORT DATA REPLICATION FOR REAL-TIME DATA WAREHOUSING AND ANALYTICS By Philip Russom Sponsored by tdwi.org APRIL 2012 TDWI CHECKLIST REPORT DATA REPLICATION FOR REAL-TIME
More informationBuilding a Data Quality Scorecard for Operational Data Governance
Building a Data Quality Scorecard for Operational Data Governance A White Paper by David Loshin WHITE PAPER Table of Contents Introduction.... 1 Establishing Business Objectives.... 1 Business Drivers...
More informationIBM AND NEXT GENERATION ARCHITECTURE FOR BIG DATA & ANALYTICS!
The Bloor Group IBM AND NEXT GENERATION ARCHITECTURE FOR BIG DATA & ANALYTICS VENDOR PROFILE The IBM Big Data Landscape IBM can legitimately claim to have been involved in Big Data and to have a much broader
More informationBusting 7 Myths about Master Data Management
Knowledge Integrity Incorporated Busting 7 Myths about Master Data Management Prepared by: David Loshin Knowledge Integrity, Inc. August, 2011 Sponsored by: 2011 Knowledge Integrity, Inc. 1 (301) 754-6350
More informationDATA VISUALIZATION AND DISCOVERY FOR BETTER BUSINESS DECISIONS
TDWI research TDWI BEST PRACTICES REPORT THIRD QUARTER 2013 EXECUTIVE SUMMARY DATA VISUALIZATION AND DISCOVERY FOR BETTER BUSINESS DECISIONS By David Stodder tdwi.org EXECUTIVE SUMMARY Data Visualization
More informationHow to Enhance Traditional BI Architecture to Leverage Big Data
B I G D ATA How to Enhance Traditional BI Architecture to Leverage Big Data Contents Executive Summary... 1 Traditional BI - DataStack 2.0 Architecture... 2 Benefits of Traditional BI - DataStack 2.0...
More informationMDM and Data Warehousing Complement Each Other
Master Management MDM and Warehousing Complement Each Other Greater business value from both 2011 IBM Corporation Executive Summary Master Management (MDM) and Warehousing (DW) complement each other There
More informationPrincipal MDM Components and Capabilities
Principal MDM Components and Capabilities David Loshin Knowledge Integrity, Inc. 1 Agenda Introduction to master data management The MDM Component Layer Model MDM Maturity MDM Functional Services Summary
More informationBig Data Integration: A Buyer's Guide
SEPTEMBER 2013 Buyer s Guide to Big Data Integration Sponsored by Contents Introduction 1 Challenges of Big Data Integration: New and Old 1 What You Need for Big Data Integration 3 Preferred Technology
More informationA Next-Generation Analytics Ecosystem for Big Data. Colin White, BI Research September 2012 Sponsored by ParAccel
A Next-Generation Analytics Ecosystem for Big Data Colin White, BI Research September 2012 Sponsored by ParAccel BIG DATA IS BIG NEWS The value of big data lies in the business analytics that can be generated
More informationKlarna Tech Talk: Mind the Data! Jeff Pollock InfoSphere Information Integration & Governance
Klarna Tech Talk: Mind the Data! Jeff Pollock InfoSphere Information Integration & Governance IBM s statements regarding its plans, directions, and intent are subject to change or withdrawal without notice
More informationTen Mistakes to Avoid
EXCLUSIVELY FOR TDWI PREMIUM MEMBERS TDWI RESEARCH SECOND QUARTER 2014 Ten Mistakes to Avoid In Big Data Analytics Projects By Fern Halper tdwi.org Ten Mistakes to Avoid In Big Data Analytics Projects
More informationTraditional BI vs. Business Data Lake A comparison
Traditional BI vs. Business Data Lake A comparison The need for new thinking around data storage and analysis Traditional Business Intelligence (BI) systems provide various levels and kinds of analyses
More informationKey Issues for Data Management and Integration, 2006
Research Publication Date: 30 March 2006 ID Number: G00138812 Key Issues for Data Management and Integration, 2006 Ted Friedman The effective management and leverage of data represent the greatest opportunity
More informationTDWI research. TDWI Checklist report. Data Federation. By Wayne Eckerson. Sponsored by. www.tdwi.org
TDWI research TDWI Checklist report Data Federation By Wayne Eckerson Sponsored by www.tdwi.org NOVEMBER 2009 TDWI Checklist report Data Federation By Wayne Eckerson TABLE OF CONTENTS 2 FOREWORD 2 NUMBER
More informationThe Liaison ALLOY Platform
PRODUCT OVERVIEW The Liaison ALLOY Platform WELCOME TO YOUR DATA-INSPIRED FUTURE Data is a core enterprise asset. Extracting insights from data is a fundamental business need. As the volume, velocity,
More informationOracle Data Integration: CON7926 Oracle Data Integration: A Crucial Ingredient for Cloud Integration
Oracle Data Integration: CON7926 Oracle Data Integration: A Crucial Ingredient for Cloud Integration Julien Testut Principal Product Manager, Oracle Data Integration Sumit Sarkar Principal Systems Engineer,
More informationTRANSITIONING TO BIG DATA:
TRANSITIONING TO BIG DATA: A Checklist for Operational Readiness Moving to a Big Data platform: Key recommendations to ensure operational readiness Overview Many factors can drive the decision to augment
More informationBIG DATA AND THE ENTERPRISE DATA WAREHOUSE WORKSHOP
BIG DATA AND THE ENTERPRISE DATA WAREHOUSE WORKSHOP Business Analytics for All Amsterdam - 2015 Value of Big Data is Being Recognized Executives beginning to see the path from data insights to revenue
More informationAn Architecture for Integrated Operational Business Intelligence
An Architecture for Integrated Operational Business Intelligence Dr. Ulrich Christ SAP AG Dietmar-Hopp-Allee 16 69190 Walldorf ulrich.christ@sap.com Abstract: In recent years, Operational Business Intelligence
More informationData virtualization: Delivering on-demand access to information throughout the enterprise
IBM Software Thought Leadership White Paper April 2013 Data virtualization: Delivering on-demand access to information throughout the enterprise 2 Data virtualization: Delivering on-demand access to information
More informationData Governance, Data Architecture, and Metadata Essentials Enabling Data Reuse Across the Enterprise
Data Governance Data Governance, Data Architecture, and Metadata Essentials Enabling Data Reuse Across the Enterprise 2 Table of Contents 4 Why Business Success Requires Data Governance Data Repurposing
More informationExperience studies data management How to generate valuable analytics with improved data processes
www.pwc.com/us/insurance Experience studies data management How to generate valuable analytics with improved data processes An approach to managing data for experience studies October 2015 Table of contents
More informationData Virtualization and ETL. Denodo Technologies Architecture Brief
Data Virtualization and ETL Denodo Technologies Architecture Brief Contents Data Virtualization and ETL... 3 Summary... 3 Data Virtualization... 7 What is Data Virtualization good for?... 8 Applications
More informationOperationalizing Data Governance through Data Policy Management
Operationalizing Data Governance through Data Policy Management Prepared for alido by: David Loshin nowledge Integrity, Inc. June, 2010 2010 nowledge Integrity, Inc. Page 1 Introduction The increasing
More informationHow To Turn Big Data Into An Insight
mwd a d v i s o r s Turning Big Data into Big Insights Helena Schwenk A special report prepared for Actuate May 2013 This report is the fourth in a series and focuses principally on explaining what s needed
More informationten mistakes to avoid
second quarter 2010 ten mistakes to avoid In Predictive Analytics By Thomas A. Rathburn ten mistakes to avoid In Predictive Analytics By Thomas A. Rathburn Foreword Predictive analytics is the goal-driven
More informationEvolving Data Warehouse Architectures
Evolving Data Warehouse Architectures In the Age of Big Data Philip Russom April 15, 2014 TDWI would like to thank the following companies for sponsoring the 2014 TDWI Best Practices research report: Evolving
More informationNews and trends in Data Warehouse Automation, Big Data and BI. Johan Hendrickx & Dirk Vermeiren
News and trends in Data Warehouse Automation, Big Data and BI Johan Hendrickx & Dirk Vermeiren Extreme Agility from Source to Analysis DWH Appliances & DWH Automation Typical Architecture 3 What Business
More informationEnabling Business Transformation with a Modern Approach to Data Management
SAP Overview Brochure SAP Technology Enabling Business Transformation with a Modern Approach to Data Management Table of Contents 4 Reenvisioning the Data Management Landscape Deliver Real-Time Insight
More informationMDM Components and the Maturity Model
A DataFlux White Paper Prepared by: David Loshin MDM Components and the Maturity Model Leader in Data Quality and Data Integration www.dataflux.com 877 846 FLUX International +44 (0) 1753 272 020 One common
More informationDeploying an Operational Data Store Designed for Big Data
Deploying an Operational Data Store Designed for Big Data A fast, secure, and scalable data staging environment with no data volume or variety constraints Sponsored by: Version: 102 Table of Contents Introduction
More informationThe Role of Data Integration in Public, Private, and Hybrid Clouds
The Role of Data Integration in Public, Private, and Hybrid Clouds In today s information-driven economy, data is a fundamental asset to most businesses. As more and more of that data moves to the cloud,
More informationManaging Data in Motion
Managing Data in Motion Data Integration Best Practice Techniques and Technologies April Reeve ELSEVIER AMSTERDAM BOSTON HEIDELBERG LONDON NEW YORK OXFORD PARIS SAN DIEGO SAN FRANCISCO SINGAPORE SYDNEY
More informationBeyond the Single View with IBM InfoSphere
Ian Bowring MDM & Information Integration Sales Leader, NE Europe Beyond the Single View with IBM InfoSphere We are at a pivotal point with our information intensive projects 10-40% of each initiative
More informationPRIME DIMENSIONS. Revealing insights. Shaping the future.
PRIME DIMENSIONS Revealing insights. Shaping the future. Service Offering Prime Dimensions offers expertise in the processes, tools, and techniques associated with: Data Management Business Intelligence
More informationData Integration Checklist
The need for data integration tools exists in every company, small to large. Whether it is extracting data that exists in spreadsheets, packaged applications, databases, sensor networks or social media
More informationORACLE DATA INTEGRATOR ENTERPRISE EDITION
ORACLE DATA INTEGRATOR ENTERPRISE EDITION Oracle Data Integrator Enterprise Edition 12c delivers high-performance data movement and transformation among enterprise platforms with its open and integrated
More informationData Modeling for Big Data
Data Modeling for Big Data by Jinbao Zhu, Principal Software Engineer, and Allen Wang, Manager, Software Engineering, CA Technologies In the Internet era, the volume of data we deal with has grown to terabytes
More information5 Keys to Unlocking the Big Data Analytics Puzzle. Anurag Tandon Director, Product Marketing March 26, 2014
5 Keys to Unlocking the Big Data Analytics Puzzle Anurag Tandon Director, Product Marketing March 26, 2014 1 A Little About Us A global footprint. A proven innovator. A leader in enterprise analytics for
More informationORACLE BUSINESS INTELLIGENCE, ORACLE DATABASE, AND EXADATA INTEGRATION
ORACLE BUSINESS INTELLIGENCE, ORACLE DATABASE, AND EXADATA INTEGRATION EXECUTIVE SUMMARY Oracle business intelligence solutions are complete, open, and integrated. Key components of Oracle business intelligence
More informationChapter 6 Basics of Data Integration. Fundamentals of Business Analytics RN Prasad and Seema Acharya
Chapter 6 Basics of Data Integration Fundamentals of Business Analytics Learning Objectives and Learning Outcomes Learning Objectives 1. Concepts of data integration 2. Needs and advantages of using data
More informationSubmitted to: Service Definition Document for BI / MI Data Services
Submitted to: Service Definition Document for BI / MI Data Services Table of Contents 1. Introduction... 3 2. Data Quality Management... 4 3. Master Data Management... 4 3.1 MDM Implementation Methodology...
More informationSQL Server 2012 Performance White Paper
Published: April 2012 Applies to: SQL Server 2012 Copyright The information contained in this document represents the current view of Microsoft Corporation on the issues discussed as of the date of publication.
More informationData Governance. Data Governance, Data Architecture, and Metadata Essentials Enabling Data Reuse Across the Enterprise
Data Governance Data Governance, Data Architecture, and Metadata Essentials Enabling Data Reuse Across the Enterprise 2 Table of Contents 4 Why Business Success Requires Data Governance Data Repurposing
More information5 Best Practices for SAP Master Data Governance
5 Best Practices for SAP Master Data Governance By David Loshin President, Knowledge Integrity, Inc. Sponsored by Winshuttle, LLC 2012 Winshuttle, LLC. All rights reserved. 4/12 www.winshuttle.com Introduction
More informationAn Oracle White Paper November 2010. Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics
An Oracle White Paper November 2010 Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics 1 Introduction New applications such as web searches, recommendation engines,
More informationData Virtualization Usage Patterns for Business Intelligence/ Data Warehouse Architectures
DATA VIRTUALIZATION Whitepaper Data Virtualization Usage Patterns for / Data Warehouse Architectures www.denodo.com Incidences Address Customer Name Inc_ID Specific_Field Time New Jersey Chevron Corporation
More informationWhy Big Data in the Cloud?
Have 40 Why Big Data in the Cloud? Colin White, BI Research January 2014 Sponsored by Treasure Data TABLE OF CONTENTS Introduction The Importance of Big Data The Role of Cloud Computing Using Big Data
More informationLuncheon Webinar Series May 13, 2013
Luncheon Webinar Series May 13, 2013 InfoSphere DataStage is Big Data Integration Sponsored By: Presented by : Tony Curcio, InfoSphere Product Management 0 InfoSphere DataStage is Big Data Integration
More informationHow Transactional Analytics is Changing the Future of Business A look at the options, use cases, and anti-patterns
How Transactional Analytics is Changing the Future of Business A look at the options, use cases, and anti-patterns Table of Contents Abstract... 3 Introduction... 3 Definition... 3 The Expanding Digitization
More informationInformatica PowerCenter Data Virtualization Edition
Data Sheet Informatica PowerCenter Data Virtualization Edition Benefits Rapidly deliver new critical data and reports across applications and warehouses Access, merge, profile, transform, cleanse data
More informationMaster Data Management. Zahra Mansoori
Master Data Management Zahra Mansoori 1 1. Preference 2 A critical question arises How do you get from a thousand points of data entry to a single view of the business? We are going to answer this question
More informationOracle Big Data Management System
Oracle Big Data Management System A Statement of Direction for Big Data and Data Warehousing Platforms O R A C L E S T A T E M E N T O F D I R E C T I O N A P R I L 2 0 1 5 Disclaimer The following is
More informationExtend your analytic capabilities with SAP Predictive Analysis
September 9 11, 2013 Anaheim, California Extend your analytic capabilities with SAP Predictive Analysis Charles Gadalla Learning Points Advanced analytics strategy at SAP Simplifying predictive analytics
More informationData Integration Alternatives Managing Value and Quality
Solutions for Enabling Lifetime Customer Relationships Data Integration Alternatives Managing Value and Quality Using a Governed Approach to Incorporating Data Quality Services Within the Data Integration
More informationData Integration for Real-Time Data Warehousing and Data Virtualization
TDWI RESEARCH TDWI CHECKLIST REPORT Data Integration for Real-Time Data Warehousing and Data Virtualization By Philip Russom Sponsored by tdwi.org O C T OBER 2 010 TDWI CHECKLIST REPORT Data Integration
More informationIntegrating Netezza into your existing IT landscape
Marco Lehmann Technical Sales Professional Integrating Netezza into your existing IT landscape 2011 IBM Corporation Agenda How to integrate your existing data into Netezza appliance? 4 Steps for creating
More informationThe Growing Practice of Operational Data Integration. Philip Russom Senior Manager, TDWI Research April 14, 2010
The Growing Practice of Operational Data Integration Philip Russom Senior Manager, TDWI Research April 14, 2010 Sponsor: 2 Speakers: Philip Russom Senior Manager, TDWI Research Gavin Day VP of Operations
More informationFast, Low-Overhead Encryption for Apache Hadoop*
Fast, Low-Overhead Encryption for Apache Hadoop* Solution Brief Intel Xeon Processors Intel Advanced Encryption Standard New Instructions (Intel AES-NI) The Intel Distribution for Apache Hadoop* software
More informationEstablish and maintain Center of Excellence (CoE) around Data Architecture
Senior BI Data Architect - Bensenville, IL The Company s Information Management Team is comprised of highly technical resources with diverse backgrounds in data warehouse development & support, business
More informationEnabling Real-Time Sharing and Synchronization over the WAN
Solace message routers have been optimized to very efficiently distribute large amounts of data over wide area networks, enabling truly game-changing performance by eliminating many of the constraints
More informationOracle s Big Data solutions. Roger Wullschleger. <Insert Picture Here>
s Big Data solutions Roger Wullschleger DBTA Workshop on Big Data, Cloud Data Management and NoSQL 10. October 2012, Stade de Suisse, Berne 1 The following is intended to outline
More informationEnterprise Data Management
TDWI research TDWI Checklist report Enterprise Data Management By Philip Russom Sponsored by www.tdwi.org OCTOBER 2009 TDWI Checklist report Enterprise Data Management By Philip Russom TABLE OF CONTENTS
More informationConverged, Real-time Analytics Enabling Faster Decision Making and New Business Opportunities
Technology Insight Paper Converged, Real-time Analytics Enabling Faster Decision Making and New Business Opportunities By John Webster February 2015 Enabling you to make the best technology decisions Enabling
More informationData Integration: Using ETL, EAI, and EII Tools to Create an Integrated Enterprise. Colin White Founder, BI Research TDWI Webcast October 2005
Data Integration: Using ETL, EAI, and EII Tools to Create an Integrated Enterprise Colin White Founder, BI Research TDWI Webcast October 2005 TDWI Data Integration Study Copyright BI Research 2005 2 Data
More informationBig Data Comes of Age: Shifting to a Real-time Data Platform
An ENTERPRISE MANAGEMENT ASSOCIATES (EMA ) White Paper Prepared for SAP April 2013 IT & DATA MANAGEMENT RESEARCH, INDUSTRY ANALYSIS & CONSULTING Table of Contents Introduction... 1 Drivers of Change...
More informationIBM Software Integrating and governing big data
IBM Software big data Does big data spell big trouble for integration? Not if you follow these best practices 1 2 3 4 5 Introduction Integration and governance requirements Best practices: Integrating
More informationWhitepaper Data Governance Roadmap for IT Executives Valeh Nazemoff
Whitepaper Data Governance Roadmap for IT Executives Valeh Nazemoff The Challenge IT Executives are challenged with issues around data, compliancy, regulation and making confident decisions on their business
More informationHow to Choose Between Hadoop, NoSQL and RDBMS
How to Choose Between Hadoop, NoSQL and RDBMS Keywords: Jean-Pierre Dijcks Oracle Redwood City, CA, USA Big Data, Hadoop, NoSQL Database, Relational Database, SQL, Security, Performance Introduction A
More informationredesigning the data landscape to deliver true business intelligence Your business technologists. Powering progress
redesigning the data landscape to deliver true business intelligence Your business technologists. Powering progress The changing face of data complexity The storage, retrieval and management of data has
More informationData Integration for the Real Time Enterprise
Executive Brief Data Integration for the Real Time Enterprise Business Agility in a Constantly Changing World Overcoming the Challenges of Global Uncertainty Informatica gives Zyme the ability to maintain
More informationSupporting Your Data Management Strategy with a Phased Approach to Master Data Management WHITE PAPER
Supporting Your Data Strategy with a Phased Approach to Master Data WHITE PAPER SAS White Paper Table of Contents Changing the Way We Think About Master Data.... 1 Master Data Consumers, the Information
More informationHow To Make Data Streaming A Real Time Intelligence
REAL-TIME OPERATIONAL INTELLIGENCE Competitive advantage from unstructured, high-velocity log and machine Big Data 2 SQLstream: Our s-streaming products unlock the value of high-velocity unstructured log
More informationA Whole New World. Big Data Technologies Big Discovery Big Insights Endless Possibilities
A Whole New World Big Data Technologies Big Discovery Big Insights Endless Possibilities Dr. Phil Shelley Query Execution Time Why Big Data Technology? Days EDW Hours Hadoop Minutes Presto Seconds Milliseconds
More informationBig Data and Your Data Warehouse Philip Russom
Big Data and Your Data Warehouse Philip Russom TDWI Research Director for Data Management April 5, 2012 Sponsor Speakers Philip Russom Research Director, Data Management, TDWI Peter Jeffcock Director,
More informationWhat's New in SAS Data Management
Paper SAS034-2014 What's New in SAS Data Management Nancy Rausch, SAS Institute Inc., Cary, NC; Mike Frost, SAS Institute Inc., Cary, NC, Mike Ames, SAS Institute Inc., Cary ABSTRACT The latest releases
More informationORACLE DATA INTEGRATOR ENTERPRISE EDITION
ORACLE DATA INTEGRATOR ENTERPRISE EDITION ORACLE DATA INTEGRATOR ENTERPRISE EDITION KEY FEATURES Out-of-box integration with databases, ERPs, CRMs, B2B systems, flat files, XML data, LDAP, JDBC, ODBC Knowledge
More informationVIEWPOINT. High Performance Analytics. Industry Context and Trends
VIEWPOINT High Performance Analytics Industry Context and Trends In the digital age of social media and connected devices, enterprises have a plethora of data that they can mine, to discover hidden correlations
More informationIntroduction to TIBCO MDM
Introduction to TIBCO MDM 1 Introduction to TIBCO MDM A COMPREHENSIVE AND UNIFIED SINGLE VERSION OF THE TRUTH TIBCO MDM provides the data governance process required to build and maintain a comprehensive
More informationIBM Analytics The fluid data layer: The future of data management
IBM Analytics The fluid data layer: The future of data management Why flexibility and adaptability are crucial in the hybrid cloud world 1 2 3 4 5 6 The new world vision for data architects Why the fluid
More informationEII - ETL - EAI What, Why, and How!
IBM Software Group EII - ETL - EAI What, Why, and How! Tom Wu 巫 介 唐, wuct@tw.ibm.com Information Integrator Advocate Software Group IBM Taiwan 2005 IBM Corporation Agenda Data Integration Challenges and
More informationI D C A N A L Y S T C O N N E C T I O N. T h e C r i t i cal Role of I/O in Public Cloud S e r vi c e P r o vi d e r E n vi r o n m e n t s
($B) I D C A N A L Y S T C O N N E C T I O N Rick Villars Vice President, Information and Cloud T h e C r i t i cal Role of I/O in Public Cloud S e r vi c e P r o vi d e r E n vi r o n m e n t s August
More informationSee the Big Picture. Make Better Decisions. The Armanta Technology Advantage. Technology Whitepaper
See the Big Picture. Make Better Decisions. The Armanta Technology Advantage Technology Whitepaper The Armanta Technology Advantage Executive Overview Enterprises have accumulated vast volumes of structured
More informationData Virtualization A Potential Antidote for Big Data Growing Pains
perspective Data Virtualization A Potential Antidote for Big Data Growing Pains Atul Shrivastava Abstract Enterprises are already facing challenges around data consolidation, heterogeneity, quality, and
More informationInformatica PowerCenter The Foundation of Enterprise Data Integration
Informatica PowerCenter The Foundation of Enterprise Data Integration The Right Information, at the Right Time Powerful market forces globalization, new regulations, mergers and acquisitions, and business
More informationBangkok, Thailand 22 May 2008, Thursday
Bangkok, Thailand 22 May 2008, Thursday Proudly Sponsored By: BI for Customers Noam Berda May 2008 Agenda Next Generation Business Intelligence BI Platform Road Map BI Accelerator Q&A 2008 / 3 NetWeaver
More informationDetecting Anomalous Behavior with the Business Data Lake. Reference Architecture and Enterprise Approaches.
Detecting Anomalous Behavior with the Business Data Lake Reference Architecture and Enterprise Approaches. 2 Detecting Anomalous Behavior with the Business Data Lake Pivotal the way we see it Reference
More informationHow To Handle Big Data With A Data Scientist
III Big Data Technologies Today, new technologies make it possible to realize value from Big Data. Big data technologies can replace highly customized, expensive legacy systems with a standard solution
More informationBUSINESSOBJECTS DATA INTEGRATOR
PRODUCTS BUSINESSOBJECTS DATA INTEGRATOR IT Benefits Correlate and integrate data from any source Efficiently design a bulletproof data integration process Accelerate time to market Move data in real time
More informationData Virtualization for Agile Business Intelligence Systems and Virtual MDM. To View This Presentation as a Video Click Here
Data Virtualization for Agile Business Intelligence Systems and Virtual MDM To View This Presentation as a Video Click Here Agenda Data Virtualization New Capabilities New Challenges in Data Integration
More information