VelociData Solving the Need for Speed in DataOps. Inside Analysis / Bloor Group Briefing June 13, 2014

Size: px
Start display at page:

Download "VelociData Solving the Need for Speed in DataOps. Inside Analysis / Bloor Group Briefing June 13, 2014"

Transcription

1 VelociData Solving the Need for Speed in DataOps Inside Analysis / Bloor Group Briefing June 13,

2 Transforming Speed and Economics of Data Operations to Achieve Time-Bound Service Levels, Gain Wire-Speed Analytics Advantage and Reduce Cost and Complexity of BI/Analytics Architectures VelociData is a purpose built, data operations micro-supercomputing appliance that is orders of magnitude faster, more scalable, and more cost effective than conventional approaches to data transformation, data quality, data encryption, and data sorting 2

3 VelociData Hyper-Acceleration Data Operations Hubs Solving Data Integrations Challenges for the Enterprise vfusion = ETL/ELT function offload Addressing performance challenges Improving data quality on ingest Cost avoidance of ETL and analytics platforms build-outs zfusion = vfusion + Mainframe data and data operations offload Reducing mainframe MSU related charges and deferring upgrades Improving data operations performance Migrating and converting mainframe data for new analytics architectures (e.g., Big Data) 3

4 VelociData Solution Palette VelociData Suites VelociData Solutions Examples Conventional (records/second) VelociData (records/second) Lookup and Replace Data enrichment by populating fields from a master file, dictionary translations, etc. (e.g. CP à Cardiopulmonologist) ,000 Type Conversions XML à Fixed; Binary à Char; Date/Time Formats ,000 Format Conversions Rearrange, add, drop, merge, split, and resize fields to change layouts , ,000 Key Generation Hash multiple field values into a unique key, (e.g. SHA-2) ,000 > 1,000,000 zfusion vfusion Data Masking USPS Address Processing Domain Set Validation Obfuscate data for non-production uses: Persistent or Dynamic; Format preserving; AES-256 Standardization, verification, and cleansing (CASS certification in process) Validate a value based on a list of acceptable values (e.g., all product codes at a retailer; all countries in the world) ,000 > 1,000, , ,000 Field Content Validation Validates based on patterns such as s, dates, and phone numbers > 1,000,000 Data type validation and bounds checking > 1,000,000 Accelerated Data Sort Mainframe Data Conversion Sort data using complex sort keys from multiple fields within records Copybook parsing & data layout discovery; EBCDIC, COMP, COMP-3, à ASCII, Integer, Float, Results are system dependent but data intended to provide magnitude comparison ,000 1,000, > 200,000 9

5 Common ETL Bottlenecks Extract Transform Load CSV Mainframe XML RDBMS Social Media ETL Server Task #1 Task #2 Task #3 Task #4 Task #5 Task #6 Task #7 Task #8 Candidates for Acceleration Hadoop ETL Server Data Warehouse Database Appliances BI Tools Cloud Sensor Hadoop Staging DB 13

6 Example ETL Processes Offloaded to VelociData Extract Transform Load CSV Keep Existing Input Interfaces Remove Bottlenecks Reduce ETL Server Workload Faster Total Processing Time Mainframe Hadoop ETL Server ETL Server Data Warehouse XML Task #1 Task #2 Task #6 Task #7 Database Appliances BI Tools Cloud RDBMS Task #3 Task #8 Task #4 Social Media Task #5 Staging DB Sensor Hadoop 6

7 vfusion Data Operations Acceleration Hub Wire-rate transformations enable fast data access between systems ETL Server Preprocess data for fast movement into and out of Data Integration tools Hadoop Convert data to ASCII and improve quality in flight VelociData feeds Hadoop pre-processed, quality data for real-time BI efforts MPP Platforms (e.g., Netezza) Format and improve data for ready insertion into Data Analytics architectures VelociData enables real-time data access by Netezza for real-time analytics 7

8 zfusion Data Operations Acceleration Hub Wire-rate transformations enable fast data access between systems ETL Server Preprocess data for fast movement into and out of Data Integration tools Hadoop Convert data to ASCII and improve quality in flight VelociData feeds Hadoop pre-processed, quality data for real-time BI efforts MPP Platforms (e.g., Netezza) Format and improve data for ready insertion into Data Analytics architectures Mainframe Conversion into and out of EBCDIC and packed decimal formats VelociData enables real-time data access by Netezza for real-time analytics 8

9 Live Demonstration And now for something completely different

10 Examples of New World Data Challenges Being Solved Property casualty company shortens by 5x a daily task of processing 540 million records to enable more accurate real-time quoting Retailer now integrates full customer data from instore, on-line, and mobile sources in realtime processing 40,000 records/s (up from 100/s) Pharmaceutical discovery query is reduced from 8 days to 20 minutes Health benefits provider shortens a data integration process from 16 hours to 45 seconds to enable better customer support Logistics firm standardizes USPS address at 10 billion / hour for data cleansing on ingest Manufacturer sorts billions of records with multifields keys at nearly M/s for analytics and data quality Credit card company reduces mainframe costs and improves analytics performance by integrating historical and fresh data into Hadoop at line rates Financial processing network masks 5 million fields/s of production data to sell opportunity information to retailers 10

11 For More Information Please Visit: velocidata.com 11

12 Thank You!

13 Questions?

14 Additional Slides

15 Today s Update from VelociData Fast access to real time data Addressing Total Flow bottlenecks VelociData s solutions New capabilities Live Demo Questions 15

16 Enabling Three Layers of Data Access Wire-rate transformations and convergence of fresh and historical data VelociData enables real-time data access for immediate analytics and visualization Sensors Weblogs Transactions Mainframe Hadoop Social Media RDBMS VelociData feeds databases and warehouses pre-analytic, aggregated data for operational analytics VelociData delivers Hadoop pre-processed, quality data to keep the lake clean Hadoop 5

17 Accessing Realtime and Historical Data Business Excellence Real-time Operational Analytics Realtime Analysis for Competitive Advantage Enabling the speed of business to match business opportunities Integrating Historical Data for Operational Excellence Informing traditional BI with real-time inputs Conventional Batch-oriented BI Iterative Modeling 17

18 How We Achieve Orders of Magnitude Improvement in Cost-Performance VelociData Data Operations Appliance Micro-supercomputing appliance in a 4U rack form factor Heterogeneous system architecture that includes FPGAs, GPUs, and CPUs with internal parallelism that dramatically outperforms general purpose computers Purpose-built solutions that combine software, firmware, and massively parallel hardware to achieve acceleration approaching wire-rates Pricing and terms - Low-risk subscription terms without upfront license fees - Subscription includes high availability production system, Q/A system, disaster recovery system, unlimited usage, maintenance, support, and updates - Fixed fee for initial installation and services 18

19 Architecture Price/Performance/Functionality Criteria Transformation Complexity Format conversions Intense Lookups SID Generations Latency SLAs Limited batch window Low latency/ R-T Data Volumes Large data sets (e.g., 10Ms records) High txn volumes (e.g., 100,000s txns/sec) BI/Analytics Architecture Choices ILM Feature Rich Integration Data Quality MDM, PIM, IR Governance Outcome Use the right tool for the job 10

20 Parallelism in IT Processing is Compelling Amdahl s Law High Performance Computing history Systems were expensive Unique tools and training required Scaling performance is often sub-linear Issues with timing and thread synchronization HPC has struggled for 40 years to deliver widespread accessibility mostly due to cost and poor abstraction, development tools, and design environment If we could just deliver accessibility at an affordable cost Hardware is now becoming inexpensive Application development improvements still needed to enable productivity ü Abstract through implementation of streaming as the paradigm 20

21 Complementary Approach: Heterogeneous System Architecture Leverage a variety of compute resources Not just parallel threads on identical resources Right resources at the right times Functional elements use appropriate processing components where needed Accommodate stream processing Source à processing à target Streaming data model enables pipelining, data flow acceleration Embrace fine-grained pipeline / functional parallelism Especially data / direct parallelism Separate latency and throughput Engineered system Manage thread, memory, and resource timing and contention 21

22 Offloading ELT Work from DW/Analytics Systems Offload from Teradata, Netezza, SAP, and other distributed platforms or appliances ETL tasks often migrate into these systems due to available capacity and performance problems (as opposed to true infrastructure design) Sometimes as much as 80% of the target is used to perform T Performing cleansing, transformation, and sort in the load step can offload a tremendous amount of push-down ETL work Land clean, properly formatted data in the initial load Misplaced workload can be right-placed to a purpose-built system Improves overall workflow performance Future-proofs architecture for increasing data volumes and variety Recovers target resources freeing them up to do what they were built to do Often lowers total costs 22

23 zfusion: Mainframe Function and Data Offload Use Cases BI applications such as operational analytics and dashboards Reduce Mainframe MSU/MIPS costs associated with data processing Joins / aggregations / deduplication outside of DB Data quality / filtering / masking Data movement enabling access to mainframe data (especially to Big Data) Basel II, SOX and other regulatory reporting Mobile applications Unique Capabilities Automatic COBOL copybook processing and layout generation Data conversion between mainframe and x86 formats at mainframe speed Mainframe processing offload (transform, sort, mask, data quality, ) at line rate Some Customer Characteristics Volatile/valuable mainframe data usage that creates unpredictable demands (e.g., financial services) Competitive advantage based on time to insight (e.g., retail, CPG) Highly mainframe TCO conscious with misplaced data integration workloads (e.g., insurance) 23

24 Example Mainframe-to-Hadoop Workflow Mainframe Input Validation Key Generation Formatter Lookup Address Standardization CSV Out Simple, configuration-driven workflow Sample shows Mainframe à HDFS Along the way data are validated, cleansed, reformatted, enriched, Land analytics-ready data as fast as it can move across the wire Workflow can also work in reverse to return processed data to the mainframe 24

25 VelociData: Continuous Innovation 1Q14 Accelerated Sorting Operation FTP-Driven Workflows Data Routing 2Q14 Custom Mainframe Type Support Aggregation / Dedupe Transformation Enhancements 2H14 Pre-Analytics Statistics Calculations Expression Transformation Data JOIN Enhanced User Interface Platform Update Customer Quote: What is unique about VelociData is you can prove the story quickly 25

26 Heterogeneous System Architecture Standard CPUs General purpose not bad at everything - Good branch prediction, fast access to large memory Graphics Boards (GPUs) Thousands of cores performing very specific tasks - Excellent matrix and floating point FPGA Coprocessors Fully customizable with extreme opportunities for parallelism - Excels at bit manipulation for regex, cryptography, searching, 26

27 Stream Processing AND Hadoop Leveraging stream processing with batch-oriented Hadoop Access to more data for analytics Process data on ingest (also land raw data if desired) Transformation Cleansing Security Never read a COBOL copybook again Stream sort for integrating data, aggregation, and dedupe 27

28 Integration with ETL Vendors Scripting In Tasks Simplest and quickest integration Call VelociData command-line from scheduling or ETL tools Often used when offloading entire stages of ETL processing Custom Connector Utilize a custom VelociData-built interface into ETL components for tighter mid-job integration Examples include a Buildop in DataStage and a Custom Transform in Informatica GUI Level Integration Tighter integration allowing GUI developers to directly configure and call into VelociData For DataStage this is a Custom Stage; VelociData is working closely with IBM to develop Metadata Incorporation Communicates with existing metadata environment for robust compliance Cooperates with existing data lineage and data governance tools Developing independent metadata strategy 28

29 Source / Target Connectors Cloudera Impala DB2 Greenplum HDFS Hive Informix Microsoft SQL Server / Azure 2012 MySQL 5.x Oracle 11g PostgreSQL Salesforce.com Sqoop Sybase Sybase IQ Teradata Text Files XML 29

30 VelociData vs. Database Appliances A single 4U system VelociData Appliance Hardware accelerated data transformation Highly optimized for data flow Custom accelerated operators for ETL tasks Tools and syntax designed for transformations and filtering Real-time conversion of data between disparate systems Database Appliances Typically entire cabinets of equipment Hardware accelerated SQL Highly optimized for storing and retrieving tables Accelerated general SQL tasks Set-oriented syntax for database operations Fast processing of tables that are already resident on the appliance Use the right tool for the job 30

31 Assisting Workflows by Offloading ELT Data transformation and formatting Transform / convert non-database types (like XML and mainframe) into formatted rows and columns Sort and aggregate data on ingest Compute surrogate keys Join reference data through lookup operations Pivoting or de-pivot to shape data for warehousing Data quality and cleansing Filtering and selecting various data Filter out empty values, bad values, or improperly formatted elements Standardize and regularize data Security Masking data to remove PII / PCI data on the way to the warehouse 31

Example Use Cases. Solving the Need for Speed in Data Ops. Doc Version 2.1

Example Use Cases. Solving the Need for Speed in Data Ops. Doc Version 2.1 Example Use Cases Solving the Need for Speed in Data Ops Doc Version 2.1 Table of Contents 1 Introduction to VelociData...3 1.1 Solution Templates for Accelerating Data Ops... 3 1.2 Extending the Life

More information

End to End Solution to Accelerate Data Warehouse Optimization. Franco Flore Alliance Sales Director - APJ

End to End Solution to Accelerate Data Warehouse Optimization. Franco Flore Alliance Sales Director - APJ End to End Solution to Accelerate Data Warehouse Optimization Franco Flore Alliance Sales Director - APJ Big Data Is Driving Key Business Initiatives Increase profitability, innovation, customer satisfaction,

More information

Luncheon Webinar Series May 13, 2013

Luncheon Webinar Series May 13, 2013 Luncheon Webinar Series May 13, 2013 InfoSphere DataStage is Big Data Integration Sponsored By: Presented by : Tony Curcio, InfoSphere Product Management 0 InfoSphere DataStage is Big Data Integration

More information

Data processing goes big

Data processing goes big Test report: Integration Big Data Edition Data processing goes big Dr. Götz Güttich Integration is a powerful set of tools to access, transform, move and synchronize data. With more than 450 connectors,

More information

IBM AND NEXT GENERATION ARCHITECTURE FOR BIG DATA & ANALYTICS!

IBM AND NEXT GENERATION ARCHITECTURE FOR BIG DATA & ANALYTICS! The Bloor Group IBM AND NEXT GENERATION ARCHITECTURE FOR BIG DATA & ANALYTICS VENDOR PROFILE The IBM Big Data Landscape IBM can legitimately claim to have been involved in Big Data and to have a much broader

More information

SAS Enterprise Data Integration Server - A Complete Solution Designed To Meet the Full Spectrum of Enterprise Data Integration Needs

SAS Enterprise Data Integration Server - A Complete Solution Designed To Meet the Full Spectrum of Enterprise Data Integration Needs Database Systems Journal vol. III, no. 1/2012 41 SAS Enterprise Data Integration Server - A Complete Solution Designed To Meet the Full Spectrum of Enterprise Data Integration Needs 1 Silvia BOLOHAN, 2

More information

What's New in SAS Data Management

What's New in SAS Data Management Paper SAS034-2014 What's New in SAS Data Management Nancy Rausch, SAS Institute Inc., Cary, NC; Mike Frost, SAS Institute Inc., Cary, NC, Mike Ames, SAS Institute Inc., Cary ABSTRACT The latest releases

More information

Data Integration Checklist

Data Integration Checklist The need for data integration tools exists in every company, small to large. Whether it is extracting data that exists in spreadsheets, packaged applications, databases, sensor networks or social media

More information

Information Architecture

Information Architecture The Bloor Group Actian and The Big Data Information Architecture WHITE PAPER The Actian Big Data Information Architecture Actian and The Big Data Information Architecture Originally founded in 2005 to

More information

Data Governance in the Hadoop Data Lake. Michael Lang May 2015

Data Governance in the Hadoop Data Lake. Michael Lang May 2015 Data Governance in the Hadoop Data Lake Michael Lang May 2015 Introduction Product Manager for Teradata Loom Joined Teradata as part of acquisition of Revelytix, original developer of Loom VP of Sales

More information

Big Data Analytics - Accelerated. stream-horizon.com

Big Data Analytics - Accelerated. stream-horizon.com Big Data Analytics - Accelerated stream-horizon.com Legacy ETL platforms & conventional Data Integration approach Unable to meet latency & data throughput demands of Big Data integration challenges Based

More information

Integrating data in the Information System An Open Source approach

Integrating data in the Information System An Open Source approach WHITE PAPER Integrating data in the Information System An Open Source approach Table of Contents Most IT Deployments Require Integration... 3 Scenario 1: Data Migration... 4 Scenario 2: e-business Application

More information

ORACLE DATA INTEGRATOR ENTERPRISE EDITION

ORACLE DATA INTEGRATOR ENTERPRISE EDITION ORACLE DATA INTEGRATOR ENTERPRISE EDITION ORACLE DATA INTEGRATOR ENTERPRISE EDITION KEY FEATURES Out-of-box integration with databases, ERPs, CRMs, B2B systems, flat files, XML data, LDAP, JDBC, ODBC Knowledge

More information

Real-time Data Replication

Real-time Data Replication Real-time Data Replication from Oracle to other databases using DataCurrents WHITEPAPER Contents Data Replication Concepts... 2 Real time Data Replication... 3 Heterogeneous Data Replication... 4 Different

More information

The Future of Data Management

The Future of Data Management The Future of Data Management with Hadoop and the Enterprise Data Hub Amr Awadallah (@awadallah) Cofounder and CTO Cloudera Snapshot Founded 2008, by former employees of Employees Today ~ 800 World Class

More information

Ganzheitliches Datenmanagement

Ganzheitliches Datenmanagement Ganzheitliches Datenmanagement für Hadoop Michael Kohs, Senior Sales Consultant @mikchaos The Problem with Big Data Projects in 2016 Relational, Mainframe Documents and Emails Data Modeler Data Scientist

More information

Deploy. Friction-free self-service BI solutions for everyone Scalable analytics on a modern architecture

Deploy. Friction-free self-service BI solutions for everyone Scalable analytics on a modern architecture Friction-free self-service BI solutions for everyone Scalable analytics on a modern architecture Apps and data source extensions with APIs Future white label, embed or integrate Power BI Deploy Intelligent

More information

Big Data Success Step 1: Get the Technology Right

Big Data Success Step 1: Get the Technology Right Big Data Success Step 1: Get the Technology Right TOM MATIJEVIC Director, Business Development ANDY MCNALIS Director, Data Management & Integration MetaScale is a subsidiary of Sears Holdings Corporation

More information

SQL Server 2012 Business Intelligence Boot Camp

SQL Server 2012 Business Intelligence Boot Camp SQL Server 2012 Business Intelligence Boot Camp Length: 5 Days Technology: Microsoft SQL Server 2012 Delivery Method: Instructor-led (classroom) About this Course Data warehousing is a solution organizations

More information

ENTERPRISE EDITION ORACLE DATA SHEET KEY FEATURES AND BENEFITS ORACLE DATA INTEGRATOR

ENTERPRISE EDITION ORACLE DATA SHEET KEY FEATURES AND BENEFITS ORACLE DATA INTEGRATOR ORACLE DATA INTEGRATOR ENTERPRISE EDITION KEY FEATURES AND BENEFITS ORACLE DATA INTEGRATOR ENTERPRISE EDITION OFFERS LEADING PERFORMANCE, IMPROVED PRODUCTIVITY, FLEXIBILITY AND LOWEST TOTAL COST OF OWNERSHIP

More information

Providing Secure Representative Data Sets

Providing Secure Representative Data Sets Test Data Protection Providing Secure Representative Data Sets By Dr. Ron Indeck VelociData Inc. - www.velocidata.com World Headquarters 321 North Clark Street, Suite 740 Chicago, IL 60654 Telephone: 312-600-4422

More information

Next-Generation Cloud Analytics with Amazon Redshift

Next-Generation Cloud Analytics with Amazon Redshift Next-Generation Cloud Analytics with Amazon Redshift What s inside Introduction Why Amazon Redshift is Great for Analytics Cloud Data Warehousing Strategies for Relational Databases Analyzing Fast, Transactional

More information

dbspeak DBs peak when we speak

dbspeak DBs peak when we speak Data Profiling: A Practitioner s approach using Dataflux [Data profiling] employs analytic methods for looking at data for the purpose of developing a thorough understanding of the content, structure,

More information

ORACLE DATA INTEGRATOR ENTEPRISE EDITION FOR BUSINESS INTELLIGENCE

ORACLE DATA INTEGRATOR ENTEPRISE EDITION FOR BUSINESS INTELLIGENCE ORACLE DATA INTEGRATOR ENTEPRISE EDITION FOR BUSINESS INTELLIGENCE KEY FEATURES AND BENEFITS (E-LT architecture delivers highest performance. Integrated metadata for alignment between Business Intelligence

More information

Dell Cloudera Syncsort Data Warehouse Optimization ETL Offload

Dell Cloudera Syncsort Data Warehouse Optimization ETL Offload Dell Cloudera Syncsort Data Warehouse Optimization ETL Offload Drive operational efficiency and lower data transformation costs with a Reference Architecture for an end-to-end optimization and offload

More information

SOLUTION BRIEF. JUST THE FAQs: Moving Big Data with Bulk Load. www.datadirect.com

SOLUTION BRIEF. JUST THE FAQs: Moving Big Data with Bulk Load. www.datadirect.com SOLUTION BRIEF JUST THE FAQs: Moving Big Data with Bulk Load 2 INTRODUCTION As the data and information used by businesses grow exponentially, IT organizations face a daunting challenge moving what is

More information

Next Generation Data Warehousing Appliances 23.10.2014

Next Generation Data Warehousing Appliances 23.10.2014 Next Generation Data Warehousing Appliances 23.10.2014 Presentert av: Espen Jorde, Executive Advisor Bjørn Runar Nes, CTO/Chief Architect Bjørn Runar Nes Espen Jorde 2 3.12.2014 Agenda Affecto s new Data

More information

Testing Big data is one of the biggest

Testing Big data is one of the biggest Infosys Labs Briefings VOL 11 NO 1 2013 Big Data: Testing Approach to Overcome Quality Challenges By Mahesh Gudipati, Shanthi Rao, Naju D. Mohan and Naveen Kumar Gajja Validate data quality by employing

More information

Datenverwaltung im Wandel - Building an Enterprise Data Hub with

Datenverwaltung im Wandel - Building an Enterprise Data Hub with Datenverwaltung im Wandel - Building an Enterprise Data Hub with Cloudera Bernard Doering Regional Director, Central EMEA, Cloudera Cloudera Your Hadoop Experts Founded 2008, by former employees of Employees

More information

MDM and Data Warehousing Complement Each Other

MDM and Data Warehousing Complement Each Other Master Management MDM and Warehousing Complement Each Other Greater business value from both 2011 IBM Corporation Executive Summary Master Management (MDM) and Warehousing (DW) complement each other There

More information

Oracle Big Data SQL Technical Update

Oracle Big Data SQL Technical Update Oracle Big Data SQL Technical Update Jean-Pierre Dijcks Oracle Redwood City, CA, USA Keywords: Big Data, Hadoop, NoSQL Databases, Relational Databases, SQL, Security, Performance Introduction This technical

More information

The Data Warehouse ETL Toolkit

The Data Warehouse ETL Toolkit 2008 AGI-Information Management Consultants May be used for personal purporses only or by libraries associated to dandelon.com network. The Data Warehouse ETL Toolkit Practical Techniques for Extracting,

More information

High-Volume Data Warehousing in Centerprise. Product Datasheet

High-Volume Data Warehousing in Centerprise. Product Datasheet High-Volume Data Warehousing in Centerprise Product Datasheet Table of Contents Overview 3 Data Complexity 3 Data Quality 3 Speed and Scalability 3 Centerprise Data Warehouse Features 4 ETL in a Unified

More information

News and trends in Data Warehouse Automation, Big Data and BI. Johan Hendrickx & Dirk Vermeiren

News and trends in Data Warehouse Automation, Big Data and BI. Johan Hendrickx & Dirk Vermeiren News and trends in Data Warehouse Automation, Big Data and BI Johan Hendrickx & Dirk Vermeiren Extreme Agility from Source to Analysis DWH Appliances & DWH Automation Typical Architecture 3 What Business

More information

SAP Data Services 4.X. An Enterprise Information management Solution

SAP Data Services 4.X. An Enterprise Information management Solution SAP Data Services 4.X An Enterprise Information management Solution Table of Contents I. SAP Data Services 4.X... 3 Highlights Training Objectives Audience Pre Requisites Keys to Success Certification

More information

Capitalize on Big Data for Competitive Advantage with Bedrock TM, an integrated Management Platform for Hadoop Data Lakes

Capitalize on Big Data for Competitive Advantage with Bedrock TM, an integrated Management Platform for Hadoop Data Lakes Capitalize on Big Data for Competitive Advantage with Bedrock TM, an integrated Management Platform for Hadoop Data Lakes Highly competitive enterprises are increasingly finding ways to maximize and accelerate

More information

ORACLE DATA INTEGRATOR ENTERPRISE EDITION

ORACLE DATA INTEGRATOR ENTERPRISE EDITION ORACLE DATA INTEGRATOR ENTERPRISE EDITION Oracle Data Integrator Enterprise Edition 12c delivers high-performance data movement and transformation among enterprise platforms with its open and integrated

More information

Integrating Ingres in the Information System: An Open Source Approach

Integrating Ingres in the Information System: An Open Source Approach Integrating Ingres in the Information System: WHITE PAPER Table of Contents Ingres, a Business Open Source Database that needs Integration... 3 Scenario 1: Data Migration... 4 Scenario 2: e-business Application

More information

THE DATA WAREHOUSE ETL TOOLKIT CDT803 Three Days

THE DATA WAREHOUSE ETL TOOLKIT CDT803 Three Days Three Days Prerequisites Students should have at least some experience with any relational database management system. Who Should Attend This course is targeted at technical staff, team leaders and project

More information

Course Outline. Module 1: Introduction to Data Warehousing

Course Outline. Module 1: Introduction to Data Warehousing Course Outline Module 1: Introduction to Data Warehousing This module provides an introduction to the key components of a data warehousing solution and the highlevel considerations you must take into account

More information

IBM BigInsights for Apache Hadoop

IBM BigInsights for Apache Hadoop IBM BigInsights for Apache Hadoop Efficiently manage and mine big data for valuable insights Highlights: Enterprise-ready Apache Hadoop based platform for data processing, warehousing and analytics Advanced

More information

Cost-Effective Business Intelligence with Red Hat and Open Source

Cost-Effective Business Intelligence with Red Hat and Open Source Cost-Effective Business Intelligence with Red Hat and Open Source Sherman Wood Director, Business Intelligence, Jaspersoft September 3, 2009 1 Agenda Introductions Quick survey What is BI?: reporting,

More information

Converged, Real-time Analytics Enabling Faster Decision Making and New Business Opportunities

Converged, Real-time Analytics Enabling Faster Decision Making and New Business Opportunities Technology Insight Paper Converged, Real-time Analytics Enabling Faster Decision Making and New Business Opportunities By John Webster February 2015 Enabling you to make the best technology decisions Enabling

More information

Harnessing the Power of the Microsoft Cloud for Deep Data Analytics

Harnessing the Power of the Microsoft Cloud for Deep Data Analytics 1 Harnessing the Power of the Microsoft Cloud for Deep Data Analytics Today's Focus How you can operate your business more efficiently and effectively by tapping into Cloud based data analytics solutions

More information

SAP Real-time Data Platform. April 2013

SAP Real-time Data Platform. April 2013 SAP Real-time Data Platform April 2013 Agenda Introduction SAP Real Time Data Platform Overview SAP Sybase ASE SAP Sybase IQ SAP EIM Questions and Answers 2012 SAP AG. All rights reserved. 2 Introduction

More information

Copyr i g ht 2012, SAS Ins titut e Inc. All rights res er ve d. DATA MANAGEMENT FOR ANALYTICS

Copyr i g ht 2012, SAS Ins titut e Inc. All rights res er ve d. DATA MANAGEMENT FOR ANALYTICS DATA MANAGEMENT FOR ANALYTICS WHAT IS ANALYTICS? A VERY BROAD TERM OFTEN CONFUSED Descriptive What happened? When? Why? Advanced What will happen? When? Why? How do we benefit? What actions should I take?

More information

TRANSFORM BIG DATA INTO ACTIONABLE INFORMATION

TRANSFORM BIG DATA INTO ACTIONABLE INFORMATION TRANSFORM BIG DATA INTO ACTIONABLE INFORMATION Make Big Available for Everyone Syed Rasheed Solution Marketing Manager January 29 th, 2014 Agenda Demystifying Big Challenges Getting Bigger Red Hat Big

More information

CitusDB Architecture for Real-Time Big Data

CitusDB Architecture for Real-Time Big Data CitusDB Architecture for Real-Time Big Data CitusDB Highlights Empowers real-time Big Data using PostgreSQL Scales out PostgreSQL to support up to hundreds of terabytes of data Fast parallel processing

More information

Decision Ready Data: Power Your Analytics with Great Data. Murthy Mathiprakasam

Decision Ready Data: Power Your Analytics with Great Data. Murthy Mathiprakasam Decision Ready Data: Power Your Analytics with Great Data Murthy Mathiprakasam 2 Your Mission Repeatably deliver trusted and timely data for great analytics and great social impact 3 Great Data Powers

More information

SQL Maestro and the ELT Paradigm Shift

SQL Maestro and the ELT Paradigm Shift SQL Maestro and the ELT Paradigm Shift Abstract ELT extract, load, and transform is replacing ETL (extract, transform, load) as the usual method of populating data warehouses. Modern data warehouse appliances

More information

Big Data Technologies Compared June 2014

Big Data Technologies Compared June 2014 Big Data Technologies Compared June 2014 Agenda What is Big Data Big Data Technology Comparison Summary Other Big Data Technologies Questions 2 What is Big Data by Example The SKA Telescope is a new development

More information

Oracle s Big Data solutions. Roger Wullschleger. <Insert Picture Here>

Oracle s Big Data solutions. Roger Wullschleger. <Insert Picture Here> s Big Data solutions Roger Wullschleger DBTA Workshop on Big Data, Cloud Data Management and NoSQL 10. October 2012, Stade de Suisse, Berne 1 The following is intended to outline

More information

Best Practices for Hadoop Data Analysis with Tableau

Best Practices for Hadoop Data Analysis with Tableau Best Practices for Hadoop Data Analysis with Tableau September 2013 2013 Hortonworks Inc. http:// Tableau 6.1.4 introduced the ability to visualize large, complex data stored in Apache Hadoop with Hortonworks

More information

Managing Big Data with Hadoop & Vertica. A look at integration between the Cloudera distribution for Hadoop and the Vertica Analytic Database

Managing Big Data with Hadoop & Vertica. A look at integration between the Cloudera distribution for Hadoop and the Vertica Analytic Database Managing Big Data with Hadoop & Vertica A look at integration between the Cloudera distribution for Hadoop and the Vertica Analytic Database Copyright Vertica Systems, Inc. October 2009 Cloudera and Vertica

More information

White Paper. Unified Data Integration Across Big Data Platforms

White Paper. Unified Data Integration Across Big Data Platforms White Paper Unified Data Integration Across Big Data Platforms Contents Business Problem... 2 Unified Big Data Integration... 3 Diyotta Solution Overview... 4 Data Warehouse Project Implementation using

More information

Unified Data Integration Across Big Data Platforms

Unified Data Integration Across Big Data Platforms Unified Data Integration Across Big Data Platforms Contents Business Problem... 2 Unified Big Data Integration... 3 Diyotta Solution Overview... 4 Data Warehouse Project Implementation using ELT... 6 Diyotta

More information

Cloud Integration and the Big Data Journey - Common Use-Case Patterns

Cloud Integration and the Big Data Journey - Common Use-Case Patterns Cloud Integration and the Big Data Journey - Common Use-Case Patterns A White Paper August, 2014 Corporate Technologies Business Intelligence Group OVERVIEW The advent of cloud and hybrid architectures

More information

Presenters: Luke Dougherty & Steve Crabb

Presenters: Luke Dougherty & Steve Crabb Presenters: Luke Dougherty & Steve Crabb About Keylink Keylink Technology is Syncsort s partner for Australia & New Zealand. Our Customers: www.keylink.net.au 2 ETL is THE best use case for Hadoop. ShanH

More information

EII - ETL - EAI What, Why, and How!

EII - ETL - EAI What, Why, and How! IBM Software Group EII - ETL - EAI What, Why, and How! Tom Wu 巫 介 唐, wuct@tw.ibm.com Information Integrator Advocate Software Group IBM Taiwan 2005 IBM Corporation Agenda Data Integration Challenges and

More information

An Oracle White Paper June 2012. High Performance Connectors for Load and Access of Data from Hadoop to Oracle Database

An Oracle White Paper June 2012. High Performance Connectors for Load and Access of Data from Hadoop to Oracle Database An Oracle White Paper June 2012 High Performance Connectors for Load and Access of Data from Hadoop to Oracle Database Executive Overview... 1 Introduction... 1 Oracle Loader for Hadoop... 2 Oracle Direct

More information

Data Lake In Action: Real-time, Closed Looped Analytics On Hadoop

Data Lake In Action: Real-time, Closed Looped Analytics On Hadoop 1 Data Lake In Action: Real-time, Closed Looped Analytics On Hadoop 2 Pivotal s Full Approach It s More Than Just Hadoop Pivotal Data Labs 3 Why Pivotal Exists First Movers Solve the Big Data Utility Gap

More information

Offload Enterprise Data Warehouse (EDW) to Big Data Lake. Ample White Paper

Offload Enterprise Data Warehouse (EDW) to Big Data Lake. Ample White Paper Offload Enterprise Data Warehouse (EDW) to Big Data Lake Oracle Exadata, Teradata, Netezza and SQL Server Ample White Paper EDW (Enterprise Data Warehouse) Offloads The EDW (Enterprise Data Warehouse)

More information

Course 10777A: Implementing a Data Warehouse with Microsoft SQL Server 2012

Course 10777A: Implementing a Data Warehouse with Microsoft SQL Server 2012 Course 10777A: Implementing a Data Warehouse with Microsoft SQL Server 2012 OVERVIEW About this Course Data warehousing is a solution organizations use to centralize business data for reporting and analysis.

More information

Informatica Data Replication: Maximize Return on Data in Real Time Chai Pydimukkala Principal Product Manager Informatica

Informatica Data Replication: Maximize Return on Data in Real Time Chai Pydimukkala Principal Product Manager Informatica Informatica Data Replication: Maximize Return on Data in Real Time Chai Pydimukkala Principal Product Manager Informatica Terry Simonds Technical Evangelist Informatica 2 Agenda Replication Business Drivers

More information

Trusted, Enterprise QlikViewreporting. data Integration and data Quality (It s all about data)

Trusted, Enterprise QlikViewreporting. data Integration and data Quality (It s all about data) Trusted, Enterprise QlikViewreporting with Informatica data Integration and data Quality (It s all about data) Arjan Hijstek senior sales consultant Informatica Nederland bv ahijstek@informatica.com 06-22.454.327

More information

Implementing a Data Warehouse with Microsoft SQL Server 2012 MOC 10777

Implementing a Data Warehouse with Microsoft SQL Server 2012 MOC 10777 Implementing a Data Warehouse with Microsoft SQL Server 2012 MOC 10777 Course Outline Module 1: Introduction to Data Warehousing This module provides an introduction to the key components of a data warehousing

More information

OWB Users, Enter The New ODI World

OWB Users, Enter The New ODI World OWB Users, Enter The New ODI World Kulvinder Hari Oracle Introduction Oracle Data Integrator (ODI) is a best-of-breed data integration platform focused on fast bulk data movement and handling complex data

More information

Managing Third Party Databases and Building Your Data Warehouse

Managing Third Party Databases and Building Your Data Warehouse Managing Third Party Databases and Building Your Data Warehouse By Gary Smith Software Consultant Embarcadero Technologies Tech Note INTRODUCTION It s a recurring theme. Companies are continually faced

More information

Cisco Data Preparation

Cisco Data Preparation Data Sheet Cisco Data Preparation Unleash your business analysts to develop the insights that drive better business outcomes, sooner, from all your data. As self-service business intelligence (BI) and

More information

Safe Harbor Statement

Safe Harbor Statement Safe Harbor Statement The following is intended to outline our general product direction. It is intended for information purposes only, and may not be incorporated into any contract. It is not a commitment

More information

WHITE PAPER LOWER COSTS, INCREASE PRODUCTIVITY, AND ACCELERATE VALUE, WITH ENTERPRISE- READY HADOOP

WHITE PAPER LOWER COSTS, INCREASE PRODUCTIVITY, AND ACCELERATE VALUE, WITH ENTERPRISE- READY HADOOP WHITE PAPER LOWER COSTS, INCREASE PRODUCTIVITY, AND ACCELERATE VALUE, WITH ENTERPRISE- READY HADOOP CLOUDERA WHITE PAPER 2 Table of Contents Introduction 3 Hadoop's Role in the Big Data Challenge 3 Cloudera:

More information

Implementing a Data Warehouse with Microsoft SQL Server 2012

Implementing a Data Warehouse with Microsoft SQL Server 2012 Course 10777A: Implementing a Data Warehouse with Microsoft SQL Server 2012 Length: Audience(s): 5 Days Level: 200 IT Professionals Technology: Microsoft SQL Server 2012 Type: Delivery Method: Course Instructor-led

More information

IBM WebSphere DataStage Online training from Yes-M Systems

IBM WebSphere DataStage Online training from Yes-M Systems Yes-M Systems offers the unique opportunity to aspiring fresher s and experienced professionals to get real time experience in ETL Data warehouse tool IBM DataStage. Course Description With this training

More information

Green Migration from Oracle

Green Migration from Oracle Green Migration from Oracle Greenplum Migration Approach Strong Experiences on Oracle Migration Automate all tasks DDL Migration Data Migration PL-SQL and SQL Scripts Migration Data Quality Tests ETL and

More information

SAP Sybase Replication Server What s New in 15.7.1 SP100. Bill Zhang, Product Management, SAP HANA Lisa Spagnolie, Director of Product Marketing

SAP Sybase Replication Server What s New in 15.7.1 SP100. Bill Zhang, Product Management, SAP HANA Lisa Spagnolie, Director of Product Marketing SAP Sybase Replication Server What s New in 15.7.1 SP100 Bill Zhang, Product Management, SAP HANA Lisa Spagnolie, Director of Product Marketing Agenda SAP Sybase Replication Server Overview Replication

More information

Workday Big Data Analytics

Workday Big Data Analytics Workday Big Data Analytics Today s fast-paced business climate demands that decision-makers stay informed. Having access to key information gives them the best insight into their business. However, many

More information

Collaborative Big Data Analytics. Copyright 2012 EMC Corporation. All rights reserved.

Collaborative Big Data Analytics. Copyright 2012 EMC Corporation. All rights reserved. Collaborative Big Data Analytics 1 Big Data Is Less About Size, And More About Freedom TechCrunch!!!!!!!!! Total data: bigger than big data 451 Group Findings: Big Data Is More Extreme Than Volume Gartner!!!!!!!!!!!!!!!

More information

How To Handle Big Data With A Data Scientist

How To Handle Big Data With A Data Scientist III Big Data Technologies Today, new technologies make it possible to realize value from Big Data. Big data technologies can replace highly customized, expensive legacy systems with a standard solution

More information

Chapter 6 Basics of Data Integration. Fundamentals of Business Analytics RN Prasad and Seema Acharya

Chapter 6 Basics of Data Integration. Fundamentals of Business Analytics RN Prasad and Seema Acharya Chapter 6 Basics of Data Integration Fundamentals of Business Analytics Learning Objectives and Learning Outcomes Learning Objectives 1. Concepts of data integration 2. Needs and advantages of using data

More information

Data Doesn t Communicate Itself Using Visualization to Tell Better Stories

Data Doesn t Communicate Itself Using Visualization to Tell Better Stories SAP Brief Analytics SAP Lumira Objectives Data Doesn t Communicate Itself Using Visualization to Tell Better Stories Tap into your data big and small Tap into your data big and small In today s fast-paced

More information

NEWLY EMERGING BEST PRACTICES FOR BIG DATA

NEWLY EMERGING BEST PRACTICES FOR BIG DATA 2000-2012 Kimball Group. All rights reserved. Page 1 NEWLY EMERGING BEST PRACTICES FOR BIG DATA Ralph Kimball Informatica October 2012 Ralph Kimball Big is Being Monetized Big data is the second era of

More information

Il mondo dei DB Cambia : Tecnologie e opportunita`

Il mondo dei DB Cambia : Tecnologie e opportunita` Il mondo dei DB Cambia : Tecnologie e opportunita` Giorgio Raico Pre-Sales Consultant Hewlett-Packard Italiana 2011 Hewlett-Packard Development Company, L.P. The information contained herein is subject

More information

Oracle Data Integration: CON7926 Oracle Data Integration: A Crucial Ingredient for Cloud Integration

Oracle Data Integration: CON7926 Oracle Data Integration: A Crucial Ingredient for Cloud Integration Oracle Data Integration: CON7926 Oracle Data Integration: A Crucial Ingredient for Cloud Integration Julien Testut Principal Product Manager, Oracle Data Integration Sumit Sarkar Principal Systems Engineer,

More information

Implementing a Data Warehouse with Microsoft SQL Server 2012

Implementing a Data Warehouse with Microsoft SQL Server 2012 Course 10777 : Implementing a Data Warehouse with Microsoft SQL Server 2012 Page 1 of 8 Implementing a Data Warehouse with Microsoft SQL Server 2012 Course 10777: 4 days; Instructor-Led Introduction Data

More information

An Oracle White Paper November 2010. Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics

An Oracle White Paper November 2010. Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics An Oracle White Paper November 2010 Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics 1 Introduction New applications such as web searches, recommendation engines,

More information

Enabling Better Business Intelligence and Information Architecture With SAP Sybase PowerDesigner Software

Enabling Better Business Intelligence and Information Architecture With SAP Sybase PowerDesigner Software SAP Technology Enabling Better Business Intelligence and Information Architecture With SAP Sybase PowerDesigner Software Table of Contents 4 Seeing the Big Picture with a 360-Degree View Gaining Efficiencies

More information

EMC Federation Big Data Solutions. Copyright 2015 EMC Corporation. All rights reserved.

EMC Federation Big Data Solutions. Copyright 2015 EMC Corporation. All rights reserved. EMC Federation Big Data Solutions 1 Introduction to data analytics Federation offering 2 Traditional Analytics! Traditional type of data analysis, sometimes called Business Intelligence! Type of analytics

More information

Bringing Big Data into the Enterprise

Bringing Big Data into the Enterprise Bringing Big Data into the Enterprise Overview When evaluating Big Data applications in enterprise computing, one often-asked question is how does Big Data compare to the Enterprise Data Warehouse (EDW)?

More information

Constructing a Data Lake: Hadoop and Oracle Database United!

Constructing a Data Lake: Hadoop and Oracle Database United! Constructing a Data Lake: Hadoop and Oracle Database United! Sharon Sophia Stephen Big Data PreSales Consultant February 21, 2015 Safe Harbor The following is intended to outline our general product direction.

More information

Course Outline: Course: Implementing a Data Warehouse with Microsoft SQL Server 2012 Learning Method: Instructor-led Classroom Learning

Course Outline: Course: Implementing a Data Warehouse with Microsoft SQL Server 2012 Learning Method: Instructor-led Classroom Learning Course Outline: Course: Implementing a Data with Microsoft SQL Server 2012 Learning Method: Instructor-led Classroom Learning Duration: 5.00 Day(s)/ 40 hrs Overview: This 5-day instructor-led course describes

More information

Klarna Tech Talk: Mind the Data! Jeff Pollock InfoSphere Information Integration & Governance

Klarna Tech Talk: Mind the Data! Jeff Pollock InfoSphere Information Integration & Governance Klarna Tech Talk: Mind the Data! Jeff Pollock InfoSphere Information Integration & Governance IBM s statements regarding its plans, directions, and intent are subject to change or withdrawal without notice

More information

Addressing Risk Data Aggregation and Risk Reporting Ben Sharma, CEO. Big Data Everywhere Conference, NYC November 2015

Addressing Risk Data Aggregation and Risk Reporting Ben Sharma, CEO. Big Data Everywhere Conference, NYC November 2015 Addressing Risk Data Aggregation and Risk Reporting Ben Sharma, CEO Big Data Everywhere Conference, NYC November 2015 Agenda 1. Challenges with Risk Data Aggregation and Risk Reporting (RDARR) 2. How a

More information

Simplifying Big Data Analytics: Unifying Batch and Stream Processing. John Fanelli,! VP Product! In-Memory Compute Summit! June 30, 2015!!

Simplifying Big Data Analytics: Unifying Batch and Stream Processing. John Fanelli,! VP Product! In-Memory Compute Summit! June 30, 2015!! Simplifying Big Data Analytics: Unifying Batch and Stream Processing John Fanelli,! VP Product! In-Memory Compute Summit! June 30, 2015!! Streaming Analy.cs S S S Scale- up Database Data And Compute Grid

More information

The Evolution of ETL

The Evolution of ETL The Evolution of ETL -From Hand-coded ETL to Tool-based ETL By Madhu Zode Data Warehousing & Business Intelligence Practice Page 1 of 13 ABSTRACT To build a data warehouse various tools are used like modeling

More information

IBM Netezza 1000. High-performance business intelligence and advanced analytics for the enterprise. The analytics conundrum

IBM Netezza 1000. High-performance business intelligence and advanced analytics for the enterprise. The analytics conundrum IBM Netezza 1000 High-performance business intelligence and advanced analytics for the enterprise Our approach to data analysis is patented and proven. Minimize data movement, while processing it at physics

More information

An Overview of SAP BW Powered by HANA. Al Weedman

An Overview of SAP BW Powered by HANA. Al Weedman An Overview of SAP BW Powered by HANA Al Weedman About BICP SAP HANA, BOBJ, and BW Implementations The BICP is a focused SAP Business Intelligence consulting services organization focused specifically

More information

Lofan Abrams Data Services for Big Data Session # 2987

Lofan Abrams Data Services for Big Data Session # 2987 Lofan Abrams Data Services for Big Data Session # 2987 Big Data Are you ready for blast-off? Big Data, for better or worse: 90% of world s data generated over last two years. ScienceDaily, ScienceDaily

More information

A TECHNICAL WHITE PAPER ATTUNITY VISIBILITY

A TECHNICAL WHITE PAPER ATTUNITY VISIBILITY A TECHNICAL WHITE PAPER ATTUNITY VISIBILITY Analytics for Enterprise Data Warehouse Management and Optimization Executive Summary Successful enterprise data management is an important initiative for growing

More information

Integrating Netezza into your existing IT landscape

Integrating Netezza into your existing IT landscape Marco Lehmann Technical Sales Professional Integrating Netezza into your existing IT landscape 2011 IBM Corporation Agenda How to integrate your existing data into Netezza appliance? 4 Steps for creating

More information