BBBT Podcast Transcript



Similar documents
BBBT Podcast Transcript

BBBT Podcast Transcript

BBBT Podcast Transcript

BBBT Podcast Transcript

Modern Data Integration

BBBT Podcast Transcript

IBM AND NEXT GENERATION ARCHITECTURE FOR BIG DATA & ANALYTICS!

Data Integration Checklist

Louis Gudema: Founder and President of Revenue + Associates

Ten Things You Need to Know About Data Virtualization

Next Generation Tech-Talk. Cloud Based Business Collaboration with Cisco Spark

SQL Maestro and the ELT Paradigm Shift

Datenverwaltung im Wandel - Building an Enterprise Data Hub with

Are You Big Data Ready?

Cloud Integration and the Big Data Journey - Common Use-Case Patterns

SELLING PROJECTS ON THE MICROSOFT BUSINESS ANALYTICS PLATFORM

Transcription. Founder Interview - Panayotis Vryonis Talks About BigStash Cloud Storage. Media Duration: 28:45

Data Integration Hub

Information Architecture

How to Manage Your Data as a Strategic Information Asset

"Breakthrough New Software Automates The Optimization Process To Get You A #1 Ranking - All With The Single Click Of A Button!"

How to Enhance Traditional BI Architecture to Leverage Big Data

Whitepaper. An Rx for Enterprise Big Data Success NVP. NewVantage Partners

A Next-Generation Analytics Ecosystem for Big Data. Colin White, BI Research September 2012 Sponsored by ParAccel

Bringing agility to Business Intelligence Metadata as key to Agile Data Warehousing. 1 P a g e.

Decoding the Big Data Deluge a Virtual Approach. Dan Luongo, Global Lead, Field Solution Engineering Data Virtualization Business Unit, Cisco

Managed Service Provider for the IT Industry Using Digital Robotics with Machine Learning and Intelligence

<Insert Picture Here> Oracle and/or Hadoop And what you need to know

joel_maloff-phone.com-cloud_based_phone_service Page 1 of 5

The Enterprise Data Hub and The Modern Information Architecture

AWS Service Catalog. User Guide

Independent process platform

Trends and Research Opportunities in Spatial Big Data Analytics and Cloud Computing NCSU GeoSpatial Forum

Sustainable Development with Geospatial Information Leveraging the Data and Technology Revolution

Big Data Integration: A Buyer's Guide

Hadoop and Relational Database The Best of Both Worlds for Analytics Greg Battas Hewlett Packard

Data Virtualization for Agile Business Intelligence Systems and Virtual MDM. To View This Presentation as a Video Click Here

Integrating and Implementing Unified Communications and Messaging Into Your Real Estate Business With TeleCenter Real Estate Pro

A: We really embarrassed ourselves last night at that business function.

SAP BusinessObjects Business Intelligence 4.1 One Strategy for Enterprise BI. May 2013

A: I thought you hated business. What changed your mind? A: MBA's are a dime a dozen these days. Are you sure that is the best route to take?

AREVOLUTIONIN COMPETITIVEINTELLIGENCE

Leveraging Machine Data to Deliver New Insights for Business Analytics

Data Virtualization. Paul Moxon Denodo Technologies. Alberta Data Architecture Community January 22 nd, Denodo Technologies

Automated Data Ingestion. Bernhard Disselhoff Enterprise Sales Engineer

Agile Business Intelligence Data Lake Architecture

White Paper. Unified Data Integration Across Big Data Platforms

Unified Data Integration Across Big Data Platforms

Jenesis Software - Podcast Episode 3

Big Data Success Step 1: Get the Technology Right

ORACLE DATA INTEGRATOR ENTERPRISE EDITION

Data Virtualization A Potential Antidote for Big Data Growing Pains

CIO 24/7 Podcast No.6 Cloud Computing

Managing Data in Motion

How To Handle Big Data With A Data Scientist

Start Learning Joomla!

Introduction. What is RAID? The Array and RAID Controller Concept. Click here to print this article. Re-Printed From SLCentral

Transcription. Crashplan vs Backblaze. Which service should you pick the short version

Big Data Zurich, November 23. September 2011

Moving Large Data at a Blinding Speed for Critical Business Intelligence. A competitive advantage

Splunk Company Overview

CREATING PACKAGED IP FOR BUSINESS ANALYTICS PROJECTS

Build a Streamlined Data Refinery. An enterprise solution for blended data that is governed, analytics-ready, and on-demand

The Most Popular CRM Customization Equals Deeper Integration

YOU WILL NOT BE EFFECTIVE READING THIS.

IBM chooses AppZero as application virtualization partner for SmartCloud

Databricks. A Primer

Data virtualization: Delivering on-demand access to information throughout the enterprise

ENABLING OPERATIONAL BI

Databricks. A Primer

Integrating Netezza into your existing IT landscape

Let Big Data connect the dots in your business

Cisco IT Hadoop Journey

Affordable, Scalable, Reliable OLTP in a Cloud and Big Data World: IBM DB2 purescale

Bringing Big Data into the Enterprise

Data processing goes big

A TECHNICAL WHITE PAPER ATTUNITY VISIBILITY

Introduction to Open Atrium s workflow

Simplifying Big Data Analytics: Unifying Batch and Stream Processing. John Fanelli,! VP Product! In-Memory Compute Summit! June 30, 2015!!

Building Your Big Data Team

USEReady Introduction

Accelerate Data Loading for Big Data Analytics Attunity Click-2-Load for HP Vertica

Self-service BI: The Top 7 Selection Criteria. Claudia Imhoff Intelligent Solutions, Inc. Sept. 25, 2012

securing aws accounts with iam and

How Big Is Big Data Adoption? Survey Results. Survey Results Big Data Company Strategy... 6

Cisco Data Preparation

Top 10 Business Intelligence Trends

Introduction to Oracle Business Intelligence Standard Edition One. Mike Donohue Senior Manager, Product Management Oracle Business Intelligence

Hadoop Data Hubs and BI. Supporting the migration from siloed reporting and BI to centralized services with Hadoop

Big Data Architecture & Analytics A comprehensive approach to harness big data architecture and analytics for growth

HOW THE DATA LAKE WORKS

WHITE PAPER. Data Migration and Access in a Cloud Computing Environment INTELLIGENT BUSINESS STRATEGIES

THE ENSIGHTEN PROMISE. The Power to Collect, Own and Activate Omni-Channel Data

INNOVATOR S PLAYBOOK A TALK WITH GLOBAL LEADERS BRITISH TELECOM (BT)

Deploy. Friction-free self-service BI solutions for everyone Scalable analytics on a modern architecture

Replicating to everything

Migrating Within the Cloud, SaaS to SaaS

The Predictive Marketer Episode 1 Published: October 8, 2015

The Future of Data Management

Capitalize on Big Data for Competitive Advantage with Bedrock TM, an integrated Management Platform for Hadoop Data Lakes

Transcription:

BBBT Podcast Transcript About the BBBT Vendor: The Boulder Brain Trust, or BBBT, was founded in 2006 by Claudia Imhoff. Its mission is to leverage business intelligence for industry vendors, for its members, who are independent analysts and experts, and for its subscribers, who are practitioners. To accomplish this mission, the BBBT provides a variety of services, centered around vendor presentations. For more, see: www.bbbt.us. Diyotta Date recorded: August 7, 2015 Host: Guest(s): Run time: 00:14:02 Audio link: Transcript: Claudia Imhoff, President, BBBT Sanjay Vyas, President and CEO John Santaferraro, Chief Marketing Officer Podcast [See next page] Copyright 2015 BBBT - All Rights Reserved Page 1 of 8

Claudia Imhoff: Hello, and welcome to this edition of the Boulder BI Brain Trust, or the BBBT. We're a gathering of international consultants, analysts, and experts in business intelligence, who meet with interesting and innovative BI companies here in beautiful Boulder, Colorado. We not only get briefed on the latest news and releases, but we share our ideas with the vendor on where the BI industry is going, and help them with their technological directions and marketing messages. I m Claudia Imhoff and the BBBT podcasts are produced by my company, Intelligent Solutions. I'm pleased to introduce my guests today. They are Sanjay Vyas and John Santaferraro. Sanjay is the president and CEO, and John is the chief marketing officer for Diyotta. Welcome to you both. Sanjay Vyas: John Santaferraro: Thank you, Claudia. Nice to be here. Nice to be here, Claudia. Let's start with you, Sanjay. A little bit of an overview of Diyotta. What were the drivers in forming the company? Let me start by talking about our background. Diyotta was founded by three individuals, myself being one. Ravi Punuru was the chief strategy officer. Sripathi Tumati was the chief technology officer. All three of us had seen in our past life large enterprise data warehouses being built. We were right in the center of everything. We have seen and experienced the bottlenecks, which challenged us to address them in some different ways. That's where we started to formulize the architecture of Diyotta. I would say, as we have seen the evolution between the different data architectures and as we are moving towards the modern data architecture, we are in a very early stage of what's going to come next. We consider ourselves fortunate enough to be part of that evolution and revolution. We are building a solution, which can stand up to today's challenges and also look beyond that and look at the future what is coming as a Yottabyte generation. Copyright 2015 BBBT - All Rights Reserved Page 2 of 8

Very interesting paradigm shift that you certainly showed us today. I was actually quite fascinated by it. John, let me bring you into the conversation to explain a little bit about that paradigm shift and also the fact that Diyotta is based on five principles. I'd like to hear what those are as well. Let's talk about the paradigm shift and then the five principles. Claudia, if you think back 20 years when the first ETL tools were designed, the world looked very different. There were a few operational data sources. People were creating a data warehouse and a couple of data marts. Everything flowed in a really nice left-to-right diagram. Today the world looks completely different with data living on premise, in the cloud or near SaaS applications. Users accessing information either from their laptop or a mobile device, accessing data that is either on premise or in the cloud. Data flowing in from mobile devices, the Internet of Things, trillions of timestamped events pouring into the system. The nice left-to-right diagram doesn't exist anymore. It's a matrix of data flowing in all directions with hundreds of thousands of users now that are actually accessing the data. The old paradigm of ETL doesn't work anymore. We've created our product based on five principles of modern data integration. The first principle, we take the processing to where the data lives. You can't bring everything through a server anymore. You don't want everything to have to go through some kind of an operational data store or common storage area. Second, we fully leverage all platforms based on what they were designed to do well. If there are functions that live inside of a database, we call those functions and use them in the database. If there are functions like text analytics that are done well in Hadoop, we leverage those. The third principle of modern data integration is move data point-to-point to avoid server bottlenecks. You only move the amount of data when you want to move it and when it makes sense to move it. You're not moving everything. Copyright 2015 BBBT - All Rights Reserved Page 3 of 8

The fourth principle of modern data integration is manage all of the business rules and data logic centrally. Creating a design studio that captures metadata from a source system automatically stores that metadata. When you move that data into a platform like Hadoop, where that metadata doesn't exist or can't be stored, you still have it stored centrally. You can go back and track things like lineage and better understand the context of the data and the eventual analytics you're going to do. The fifth principle of modern data integration is to make changes quickly using existing logic and rules. By having this robust storage of metadata and templating everything, if you want to make changes in your data processing or your data blending and enrichment, or you want to bring in a new set of data or you want to migrate once and migrate again, most of that work is already done. We built Diyotta to basically model these five principles of modern data integration. Excellent. I really like that. It seems like it's also an evolution from the traditional ETL, extract, transform and load, to more of the ELT or ELTL type of model, where you extract, you may load it into the platform. That's basically what you're talking about, right? Yeah. I have to give credit to Yves de Montcheuil, who was on the BBBT event today, at least tweeting. He described it as ELTLTLTLTL. I think he threw in a couple of Hs as well for Hadoop. In the paradigm shift, in the world of modern data, data is flowing everywhere. You may have data that actually lives today in Netezza. You want to bring it into Hadoop. You're going to combine it with some data that you've scraped off of the Internet or that's coming off of a sensor. You may actually process it back out to Netezza to do some kind of analytic, based on a function that runs in database back in Netezza. It's a very flexible kind of environment that supports dataflows in all directions and of all kinds. Copyright 2015 BBBT - All Rights Reserved Page 4 of 8

Is that what makes it unique? It makes Diyotta unique. The other thing that makes it unique is that we automate a lot of that process. We automatically capture data from some of the source systems. We automatically define structures and the target systems where nothing has yet been defined. Because we capture all of that metadata and separate the movement and transformation of data from the actual logic, we're able to move from one platform to another at just a click of a button. You can today move from Netezza to Hadoop. Tomorrow you might want to move from Hadoop to Spark. It's easy. In a drag-down menu, you literally just choose the next platform and it goes. What I like about it is the incredible reusability of the product in that case, right? Absolutely. Sanjay, let me go back to you. If you don't mind, a brief overview of the Diyotta architecture. Before I jump in to answer that question, I will take the liberty to quote one of the members again, who said that, "Diyotta is taking the integration to the data, not data to the integration." If I have to explain that in one sentence, Diyotta architecture allows you to leverage your investment, which you already put in, by creating all the business rules in an abstract layer and then taking the processing to the engines, where that processing can be done in a much faster and efficient manner. We have agent-based processing as well for globally distributed configurations. At the same time, while we are seeing the modern data integration within Hadoop, we have Hadoop architecture, where you don't have to invest in a separate ETL or separate server bottleneck to process data on Hadoop. Copyright 2015 BBBT - All Rights Reserved Page 5 of 8

Diyotta collocates with Hadoop in physical architecture manner. It gives you the capability to run all the processing within Hadoop, whether it is on MapReduce, Tez or Spark execution engine. Let's stay with you, Sanjay, and talk a little about the magic, the secret sauce, if you will, of Diyotta and that is its abstraction layer. I was actually quite taken by your explanation of the abstraction layer and what that means to someone who is perhaps not as knowledgeable in all of these various targets that we're talking about, Netezza, Hadoop. It could be Oracle, it could be DB2, whatever it is. If you don't mind, talk a little bit about the abstraction layer. What does that mean to someone trying to do all this data integration? Claudia, that's a great question. We had seen in our experience as well that while we were devising these architectures, the business focus was still addressing the problems which we had been finding in the data. As the complexities are increasing, the problems still remain for the most part the same, like loading the data into your modern data architecture, performing the processing there. What we saw is that instead of reengineering while you are changing your underlying data platforms why not have an abstraction business layer, which allows you to define your business rules once and reuse them all the time, whenever any technology change happened. That should not impact your data engineering platform. That's where the abstraction layer helps a lot. It has all bells and whistles to perform those different processing on the platform of your choice. It actually hides the complexity of whatever technology it is. Anyone can use the tool because they don't have to know all that, right? Right. That's where the self-service BI, self-service ETL or ELT, whatever you want to call it. That's where we are heading towards. I'll be going to cover in the next question. Copyright 2015 BBBT - All Rights Reserved Page 6 of 8

Excellent. Very briefly, we got a couple of minutes left. If you don't mind, give me a few examples of how your customers are actually using your technology. We have customers who are in cloud using the technology. They have Hadoop cluster in cloud. They have sensitive information originating on premise, which we have to encrypt and then load into the Hadoop cluster, in Amazon for example. We have customers who are doing telecom type use-cases, where they are loading data in a micro-batch environment every five minutes. On a period of a day, they are loading almost a terabyte of information every day. We have customers who are using as enterprise data link initiative, where they are loading different types of data, including Twitter feeds, including mainframe, including flat file, DB2 traditional. A gamut of customers having different types of use-cases, financial as well as telecom as well as health care. That's where we are seeing a lot of traction with the modern data architecture. Let's wrap up with a little bit of a view toward the future. What's going to be happening with Diyotta? We are in a very early stage of Hadoop becoming a mainstream data architecture. We are seeing customers leveraging that to the benefit of evolving their data architecture into multitudes of intelligent systems. You talk about businesses using Hadoop for deriving intelligence for predictive analytics and things like that. Where we see ourselves is we are in the middle of a modern data evolution. That's why we say that we are a preferred choice. We want to be the preferred choice for modern data architecture and integration in that ecosystem. We are going from a paradigm shift, where we started this paradigm shift of modern data integration and want to continue with that. As we go forwards, we will see a lot of IoT kind of use cases. We will see social Copyright 2015 BBBT - All Rights Reserved Page 7 of 8

media, pure cloud, Software as a Service, as well as data in motion. That's where we are heading towards with Diyotta. Excellent. I would say that it's a remarkable paradigm and one that I think is, first of all, very innovative and certainly very needed today. That's it for this edition of the BBBT podcasts. Again, I'm Claudia Imhoff. It has been a great pleasure to speak with Sanjay Vyas and John Santaferraro of Diyotta today. Thanks to both of you. Thanks, Claudia. Thank you, Claudia. I hope you enjoyed today's podcast. You'll find more podcasts from other vendors at our web site www.bbbt.us. If you want to read more about today s session, please search for our hash tag on Twitter. That's #BBBT. And please join me again for another interview. Good bye and good business! Copyright 2015 BBBT - All Rights Reserved Page 8 of 8