TERADATA QUERY GRID. Teradata User Group September 2014



Similar documents
Teradata s Big Data Technology Strategy & Roadmap

BIG DATA: FROM HYPE TO REALITY. Leandro Ruiz Presales Partner for C&LA Teradata

Artur Borycki. Director International Solutions Marketing

Data Warehouse Hadoop. Shimpei Kodama 2015/9/29

Bringing Intergalactic Data Speak (a.k.a.: SQL) to Hadoop Martin Willcox Director Big Data Centre of Excellence (Teradata

CERULIUM TERADATA COURSE CATALOG

UNIFY YOUR (BIG) DATA

Architecting for the Internet of Things & Big Data

Welcome. Host: Eric Kavanagh. The Briefing Room. Twitter Tag: #briefr

HIGH PERFORMANCE ANALYTICS FOR TERADATA

Lofan Abrams Data Services for Big Data Session # 2987

INVESTOR PRESENTATION. First Quarter 2014

Oracle Big Data SQL Technical Update

Performance Tuning for the Teradata Database

Teradata Unified Big Data Architecture

SAS and Teradata Partnership

Up Your R Game. James Taylor, Decision Management Solutions Bill Franks, Teradata

INVESTOR PRESENTATION. Third Quarter 2014

Oracle Big Data Strategy Simplified Infrastrcuture

Data processing goes big

Oracle BI EE Implementation on Netezza. Prepared by SureShot Strategies, Inc.

Oracle Database 12c Plug In. Switch On. Get SMART.

Hadoop Job Oriented Training Agenda

Ganzheitliches Datenmanagement

Exploring the Synergistic Relationships Between BPC, BW and HANA

Integrated Big Data: Hadoop + DBMS + Discovery for SAS High Performance Analytics

Inge Os Sales Consulting Manager Oracle Norway

IBM BigInsights Has Potential If It Lives Up To Its Promise. InfoSphere BigInsights A Closer Look

Constructing a Data Lake: Hadoop and Oracle Database United!

Investor Presentation. Second Quarter 2015

Yu Xu Pekka Kostamaa Like Gao. Presented By: Sushma Ajjampur Jagadeesh

Programming Hadoop 5-day, instructor-led BD-106. MapReduce Overview. Hadoop Overview

An Oracle White Paper November Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics

GO BIG WITH DATA PLATFORMS: HADOOP AND TERADATA 1700

Using MySQL for Big Data Advantage Integrate for Insight Sastry Vedantam

2009 Oracle Corporation 1

Big Data SQL and Query Franchising

SAP Business Objects Business Intelligence platform Document Version: 4.1 Support Package Data Federation Administration Tool Guide

Luncheon Webinar Series May 13, 2013

Capitalize on Big Data for Competitive Advantage with Bedrock TM, an integrated Management Platform for Hadoop Data Lakes

Integrating VoltDB with Hadoop

Data Integration Checklist

Teradata Connector for Hadoop Tutorial

MyOra 3.0. User Guide. SQL Tool for Oracle. Jayam Systems, LLC

WHAT S NEW IN SAS 9.4

Using RDBMS, NoSQL or Hadoop?

Introduction to Big data. Why Big data? Case Studies. Introduction to Hadoop. Understanding Features of Hadoop. Hadoop Architecture.

Cost-Effective Business Intelligence with Red Hat and Open Source

Using Tableau Software with Hortonworks Data Platform

The Inside Scoop on Hadoop

Bringing Big Data to People

Apache Sqoop. A Data Transfer Tool for Hadoop

Sharding with postgres_fdw

SQL Server 2016 New Features!

Oracle Big Data Handbook

An Oracle White Paper June High Performance Connectors for Load and Access of Data from Hadoop to Oracle Database

Teradata Utilities Class Outline

Data Warehouse as a Service. Lot 2 - Platform as a Service. Version: 1.1, Issue Date: 05/02/2014. Classification: Open

Big Data Course Highlights

Safe Harbor Statement

HDP Hadoop From concept to deployment.

Comparing SQL and NOSQL databases

The Future of Data Management

Oracle s Big Data solutions. Roger Wullschleger. <Insert Picture Here>

CitusDB Architecture for Real-Time Big Data

Qsoft Inc

GO BIG WITH DATA PLATFORMS: HADOOP AND TERADATA

OLH: Oracle Loader for Hadoop OSCH: Oracle SQL Connector for Hadoop Distributed File System (HDFS)

Oracle Database 10g: New Features for Administrators

HDP Enabling the Modern Data Architecture

SAP Database Strategy Overview. Uwe Grigoleit September 2013

The Sierra Clustered Database Engine, the technology at the heart of

Hadoop and Map-Reduce. Swati Gore

BIG DATA ANALYTICS REFERENCE ARCHITECTURES AND CASE STUDIES

Oracle Big Data Building A Big Data Management System

Real World Big Data Architecture - Splunk, Hadoop, RDBMS

SQL Server Administrator Introduction - 3 Days Objectives

Customized Report- Big Data

Integrate Master Data with Big Data using Oracle Table Access for Hadoop

Big Data Analytics in LinkedIn. Danielle Aring & William Merritt

Well packaged sets of preinstalled, integrated, and optimized software on select hardware in the form of engineered systems and appliances

X4-2 Exadata announced (well actually around Jan 1) OEM/Grid control 12c R4 just released

Executive Summary... 2 Introduction Defining Big Data The Importance of Big Data... 4 Building a Big Data Platform...

Jeffrey D. Ullman slides. MapReduce for data intensive computing

Best Practices for Hadoop Data Analysis with Tableau

FIFTH EDITION. Oracle Essentials. Rick Greenwald, Robert Stackowiak, and. Jonathan Stern O'REILLY" Tokyo. Koln Sebastopol. Cambridge Farnham.

Polybase for SQL Server 2016

Using distributed technologies to analyze Big Data

Big Data Architecture & Analytics A comprehensive approach to harness big data architecture and analytics for growth

CA Big Data Management: It s here, but what can it do for your business?

CA Database Performance

brief contents PART 1 BACKGROUND AND FUNDAMENTALS...1 PART 2 PART 3 BIG DATA PATTERNS PART 4 BEYOND MAPREDUCE...385

Big Data Management. Big Data Management. (BDM) Autumn Povl Koch September 16,

THE ATLAS DISTRIBUTED DATA MANAGEMENT SYSTEM & DATABASES

TUT NoSQL Seminar (Oracle) Big Data

Hortonworks & SAS. Analytics everywhere. Page 1. Hortonworks Inc All Rights Reserved

ESS event: Big Data in Official Statistics. Antonino Virgillito, Istat

SAS 9.4 Intelligence Platform: Migration Guide, Second Edition

Cloudera Certified Developer for Apache Hadoop

Transcription:

TERADATA QUERY GRID Teradata User Group September 2014

2 9/15/2014 Teradata Confidential

Teradata s View Big Data and Data in General DATA enables INSIGHTS which drive ACTIONS to provide BUSINESS ADVANTAGE thru BETTER DECISIONS Big Data is DATA. It adds to the depth and types of INSIGHTS... But the rest of the story is the SAME. However, the economics have changed, increasing the amount of data you can capture the tools have been enhanced, expanding the data types you can analyze The architecture framework has evolved so you can use the right tool for the right job 3 9/15/2014 Teradata Confidential

Analysts Recommend: Shift from a Single Platform to an Ecosystem "Logical" Data Warehouse Hybrid Ecosystem We will abandon the old models based on the desire to implement for high-value analytic applications. Big Data requirements are solved by a range of platforms including analytical databases, discovery platforms, and NoSQL solutions beyond Hadoop. 4 9/15/2014 Teradata Confidential Source: Big Data Comes of Age. EMA and 9sight Consulting. Nov 2012.

UNIFIED DATA ARCHITECTURE ERP MANAGE MOVE ACCESS Marketing Marketing Executives SCM CRM Images DATA PLATFORM INTEGRATED DATA WAREHOUSE Business Intelligence Predictive Analytics Applications Business Intelligence Operational Systems Frontline Workers Audio and Video Fast Loading Operational Intelligence Data Mining Customers Partners Machine Logs Text Filtering and Processing Online Archival DISCOVERY PLATFORM Data Discovery Math and Stats Engineers Data Scientists Web and Social Path, graph, time-series analysis Pattern Detection Languages Business Analysts SOURCES ANALYTIC TOOLS USERS

Fabric Based Computing Optimized for BI The backbone of UDA > High performance infrastructure > Aggregate Teradata IDW, Aster Discovery and Hadoop > Industry approach optimized for Big Analytics use Key Teradata Elements > BYNET V5 on InfiniBand interconnect > Teradata Managed Servers > System management across all of the FBC 6 9/15/2014 Teradata Confidential

Teradata QueryGrid Optimize, simplify, and orchestrate processing across and beyond the Teradata UDA Run the right analytic on the right platform > Take advantage of specialized processing engines while operating as a cohesive analytic environment Automated and optimized work distribution through pushdown processing across platforms > Minimize data movement, process data where it resides > Minimize data duplication > Transparently automate analytic processing and data movement between systems > Bi-directional data movement Integrated processing; within and outside the UDA Easy access to data and analytics through existing SQL skills and tools 7 9/15/2014 Teradata Confidential

Teradata QueryGrid Vision Business users IDW Discovery Data Scientists TERADATA DATABASE TERADATA ASTER DATABASE HADOOP TERADATA ASTER DATABASE TERADATA DATABASE RDBMS DATABASES MONGODB DATABASE COMPUTE CLUSTER Push-down to Hadoop System SQL, SQL-MR, SQL-GR Multiple Teradata Systems Push-down to Other Database Push-down to NoSQL Databases Run SAS, Perl, Ruby, Python, R When fully implemented, the Teradata Database or the Teradata Aster Database will be able to intelligently use the functionality and data of multiple heterogeneous processing engines 8 9/15/2014 Teradata Confidential

Quickly Building Out the QueryGrid Family Connector One-Way Bi-Directional Push-Down Teradata Hortonworks Hadoop Teradata Cloudera Hadoop Planned Planned Aster - Teradata Aster 6 - Hortonworks Hadoop Planned Planned Aster 6 - Cloudera Hadoop Planned Planned 9 9/15/2014 Teradata Confidential

Teradata Database 15 Teradata QueryGrid Leverage Hadoop resources, Reduce data movement Bi-directional to Hadoop Query push-down Easy configuration of server connections Query through Teradata Sent to Hadoop through Hive Results returned to Teradata Additional processing joins data in Teradata Final results sent back to application/user 10 9/15/2014 Teradata Confidential

Teradata QueryGrid Teradata-Hadoop QueryGrid > Server grammar Simplify via server name > Hadoop import operator Load_from_hcatalog Added server grammar > Hadoop export operator (new) Load_to_hcatalog Supports files: Delimited Text, JSON, RCFile Sequence File, ORCfile, Avro Query push-down Bi-directional data transfer Provide access rights 11 9/15/2014 Teradata Confidential

Foreign Server Syntax table CREATE FOREIGN SERVER Hadoop_sysd USING HOSTTYPE('hadoop') SERVER ( sysd.labs.teradata.com') PORT ('9083') HIVESERVER ('sysd.labs.teradata.com') HIVEPORT ('10000') USERNAME( Hive ) DEFAULT_STRING_SIZE( 2048 ) HADOOP_PROPERTIES( org.apache.hadoop.io.compress.gzipcodec ); DO IMPORT WITH syslib.load_from_hcatalog_hdp1_3_2, DO EXPORT WITH syslib.load_to_hcatalog_hdp1_3_2 Merge_hdfs_files( True ) Compression_codec( org.apache.hadoop.io.compress.gzipcodec ; SELECT source, session FROM clickstream@hadoop_sysd WHERE session_ts = 2013-12-01 ; Server name 12 9/15/2014 Teradata Confidential

QueryGrid Server Objects and Privileges TD_SERVER_DB contains all servers objects Servers are global objects Users have SELECT and INSERT granted to them > GRANT SELECT ON hdp132 TO tom; > GRANT INSERT ON hdp143 TO lisa; Being able to create and drop a server is a privilege > GRANT CREATE SERVER > GRANT DROP SERVER 13 9/15/2014 Teradata Confidential

Remote SQL Execution Push SQL to remote Hive system Hive filters data on non-partitioned columns Foreign table Select executed on remote system SELECT source, session FROM FOREIGN TABLE(select session, source from clickstream where source = Mozilla )@Hadoop_sysd WHERE session = current_date AS dt; 14 9/15/2014 Teradata Confidential

QueryGrid Data Transfer Import SELECT source, session FROM clickstream@hadoop_sysd WHERE session_ts = 2013-01-01 ; insert/select & create table as to instantiate data locally. Export INSERT INTO cust_loc@hadoop_sysd SELECT cust_id, cust_zip FROM cust_data WHERE last_update = current_date; 15 9/15/2014 Teradata Confidential

QueryGrid Insert Explained Part 1 EXPLAIN INSERT INTO cardata@hdp132 SELECT * FROM newcars; ***Success: Activity Count = 41 Explanation --------------------------------------------------------------------------- 1) First, we lock a distinct ut1."pseudo table" for read on a RowHash to prevent global deadlock for ut1.tab1. 2) Next, we lock ut1.tab1 for read. 3) We do an all-amps RETRIEVE step from ut1.newcars by way of an all-rows scan with no residual conditions executing table operator SYSLIB.load_to_hcatalog with a condition of ("(1=1)") into Spool 2 (used to materialize view, derived table, table function or table operator drvtab_inner) (all_amps), which is built locally on the AMPs. The size of Spool 2 is estimated with low confidence to be 8 rows (11,104 bytes). The estimated time for this step is 0.16 seconds. 16 9/15/2014 Teradata Confidential

QueryGrid Explained Part 2 4) We do an all-amps RETRIEVE step from Spool 2 (Last Use) by way of an all-rows scan into Spool 3 (used to materialize view, derived table, table function or table operator TblOpInputSpool) (all_amps), which is redistributed by hash code to all AMPs. The size of Spool 3 is estimated with low confidence to be 8 rows ( 11,104 bytes). The estimated time for this step is 0.16 seconds. 5) We do an all-amps RETRIEVE step from Spool 3 (Last Use) by way of an all-rows scan executing table operator SYSLIB.load_to_hcatalog with a condition of ("(1=1)") into Spool 4 (used to materialize view, derived table, table function or table operator h4) (all_amps), which is built locally on the AMPs. < BEGIN EXPLAIN FOR REMOTE QUERY --> TD: 3 column(s); Hadoop: 3 column(s), with 2 partition column(s); doors(integer) -> doors(string); make(varchar) -> make*(string); model(varchar) -> model*(string); * denotes partition column; <-- END EXPLAIN FOR REMOTE QUERY > The size of Spool 4 is estimated with low confidence to be 8 rows (200 bytes). The estimated time for this step is 0.16 seconds. 17 9/15/2014 Teradata Confidential

Create and Drop Hadoop Tables Stored procedures to create and drop Hadoop tables Allows SQL scripts to export data in stand alone fashion CALL SYSLIB.HDROP('t3', hdp132 ); CALL SYSLIB.HCTAS('t3','c2,c3','LOCATION "/user/hive/table_t12"', hdp132','default'); 18 9/15/2014 Teradata Confidential

Requirements for QueryGrid to Hadoop Teradata 15.0 > Node memory > 96GB Network > All Teradata nodes able to connect to all Hadoop data nodes Proxy user on Hadoop 19 9/15/2014 Teradata Confidential

Connection Flow The client connects to the system through the PE in node 1. The query is parsed in the PE. During the parsing phase, the table operator s contract function contacts the HCatalog component through the External Access Handler (EAH), which is a one-per-node Java Virtual Machine The HCatalog returns the metadata about the table, the number of columns, and the types for the columns. The parser uses this info and also uses this connection d to obtain the Hadoop splits of data that underlie the Hadoop table. The splits are assigned to the AMPs in a round-robin fashion so that each AMP gets a split. The parser phase completes and produces an AMP step containing the table operator. This is sent to all the AMPs in parallel. Each AMP then begins to execute the table operator s execute function providing a parallel import of Hadoop data. The execute function opens and reads the split data reading in Hadoop rows. These are converted to Teradata data types in each column, and the rows are written to spool. When all the data has been written, the spool file is redistributed as input into the next part of the query plan. 20 9/15/2014 Teradata Confidential

Quickly Building Out the QueryGrid Family Connector One-Way Bi-Directional Push-Down Teradata Hortonworks Hadoop Teradata Cloudera Hadoop August Planned Planned Teradata Oracle (11g) 3Q 2014 3Q 2014 Teradata - MongoDB 1Q2015 1Q2015 1Q2015 Teradata DB2 1H2015 1H2015 1H2015 Teradata SQLServer 2Q2015 2QH2015 2QH2015 Aster - Teradata Aster 6 - Hortonworks Hadoop Planned Planned Aster 6 - Cloudera Hadoop Planned Planned 21 9/15/2014 Teradata Confidential

Unity Source Link -- High Level Description USL enables dynamic SQL access to foreign database (currently Oracle only) from Teradata. The first phase 14.10 was released with Import capability only. > Import is selecting data from a foreign database > Export is inserting data into a foreign database This release includes > Export capability > Dynamic database links for both Import and Export via SQL-H Foreign Server grammar > Single sign-on across local and remote databases. 22 9/15/2014 Teradata Confidential 3

Quickly Building Out the QueryGrid Family Connector One-Way Bi-Directional Push-Down Teradata Hortonworks Hadoop Teradata Cloudera Hadoop August Planned Planned Teradata Oracle (11g) 3Q 2014 3Q 2014 Teradata - Teradata Oct Oct Oct Teradata - MongoDB 1Q2015 1Q2015 1Q2015 Teradata DB2 1H2015 1H2015 1H2015 Teradata SQLServer 2Q2015 2QH2015 2QH2015 Aster - Teradata Aster 6 - Hortonworks Hadoop Planned Planned Aster 6 - Cloudera Hadoop Planned Planned 23 9/15/2014 Teradata Confidential

Teradata-to-Teradata: Overview Provides SQLs to read from or write to a remote Teradata Database The feature will be released asynchronously Teradata-to-Teradata is supported when both local & remote boxes are at least on Teradata 15.00 When 15.10 is released, user can use the connector across releases 15.00 -> 15.10 or 15.10 -> 15.00, NOTE if there is data type mismatch it may lead to error Both the local and remote system should have the connector software installed 24 9/15/2014 Teradata Confidential

TERADATA -TO- TERADATA : GENERAL ARCHITECTURE Bi-directional Operate with either IB or Ethernet connection of any speeds Supports Teradata 15.0 on SUSE Linux 10 and 11 Parallel import & export Buffered mode Vs. Row mode Remote execution (Pushdown processing, Predicate pushdown) Rights to execute Select, Insert Trusted user Partial data type support. 15.0 supports basic data types. On roadmap is support for LOB and JSON 25 9/15/2014 Teradata Confidential

TERADATA TO TERADATA CONNECTOR: FEATURES AND CAPABILITIES Feature Ability to create links to remote Teradata system Ability to query on another Teradata system using Server object Push execution on to remote system Push or insert data to remote system Ability to see data definitions of objects on remote system Ability to execute SQL on the remote system from the local Ability to stich execution plan, from local and remote system SQL CREATE SERVER object SELECT statement SELECT FOREIGN TABLE statement INSERT statement HELP FOREIGN [SERVER DATABASE TABLE] Execute Foreign SQL EXPLAIN on INSERT / SELECT statements 26 9/15/2014 Teradata Confidential

Teradata-to-Teradata: Create Server Object Create Server Statement: CREATE AUTHORIZATION sdll7151proxy AS DEFINER TRUSTED USER 'proxyuser' PASSWORD 'secret'; CREATE FOREIGN SERVER sdll7151 EXTERNAL SECURITY DEFINER TRUSTED sdll7151proxy USING Hosttype('Teradata') remotehost ('sdll7151.labs.teradata.com') local_ips('sdlc0062.labs.teradata.com') port('5000') read_timeout(200) listen_timeout(60) concurrentstreams(1) DO IMPORT WITH syslib.load_from_td, DO EXPORT WITH syslib.load_to_td; TD_SERVER_DB contains all servers objects. > Servers are global objects. 27 9/15/2014 Teradata Confidential Teradata Confidential

Teradata-to-Teradata: Query Provides capability to access remote Teradata data easily using the server grammar Import examples: SELECT * FROM store.pricelist@sdll7151 WHERE part_price<2.00; SELECT * FROM store.pricelist@sdll7151 UNION SELECT * FROM newitems; Query can be part of a larger query User can limit the columns and rows imported in order to control the amount of data imported and speed the import process. Returned rows can be joined with local tables Import can be used in joins and any place a normal table is referenced including views, macros, and stored procedures. 28 9/15/2014 Teradata Confidential Teradata Confidential

Teradata-to-Teradata: Import using Foreign table syntax SELECT * FROM FOREIGN TABLE( SELECT p.partno, SUM(quantity * part_price) FROM store.pricelist p, store.inventory i GROUP BY p.partno WHERE p.partno=i.partno )@sdll7151 as dt; Select FOREIGN TABLE provides a grammar pass through mechanism. Grammar in the parenthesis is unchecked but tokenized and then passed to the remote for execution. Limitations: ON Clause sub-query Order By not supported in Foreign Table Select. Ordering would be lost during the data transfer. 29 9/15/2014 Teradata Confidential

Teradata-to-Teradata: Stitching Local and Remote explains During Import EXPLAIN SELECT * FROM store.pricelist@sdll7151 WHERE part_price<2.00; 1) First, we do an all-amps RETRIEVE step executing table ope syslib.load_from_td with a condition of ("pricelist.part_p 2.00"). < BEGIN EXPLAIN FOR REMOTE QUERY --> 1) First, we lock a distinct store."pseudo table" for read RowHash to prevent global deadlock for store.pricelist. 2) Next, we lock store.pricelist for read. 3) We do an all RETRIEVE step from store.pricelist by way of an all-rows s a condition of ("store.pricelist.part_price < 2.00") into (group_amps), which is built locally on the AMPs. The size Spool 1 is estimated with no confidence to be 2 rows (202 The estimated time for this step is0.07 seconds. -> The co of Spool 1 are flushed to disk and sent back to the user a result of statement 1. The total estimated time is 0.07 se <-- END EXPLAIN FOR REMOTE QUERY > The size of Spool 1 is estimated with low confidence to be (272 bytes). The estimated time for this step is 0.06 sec 30 9/15/2014 Teradata Confidential

Teradata-to-Teradata: Stitching Local and Remote explains During Import SELECT p.partno, SUM(quantity * part_price) FROM store.pricelists@sdll7151 p, inventory i WHERE p.partno=i.partno; < BEGIN EXPLAIN FOR REMOTE QUERY --> 2) Next, we lock store.pricelist for read. 3) We do an all-amps RETRIEVE step from store.pricelist by way of an all-rows scan with no residual conditions into Spool 1 (group_amps), which is built locally on the AMPs. The size of Spool 1 is estimated with low confidence to be 4 rows (132 bytes). The estimated time for this step is 0.07 seconds. -> The contents of Spool 1 are flushed to disk and sent back to the user as the result of statement 1. The total estimated time is 0.07 seconds. <-- END EXPLAIN FOR REMOTE QUERY > The size of Spool 1 is estimated with low confidence to be 4 rows (132 bytes). The estimated time for this step is 0.06 seconds. 4) We do an all-amps RETRIEVE step from Spool 1 (Last Use) by way of an all-rows scan with a condition of ("NOT (p.partno IS NULL)") into Spool 5 (all_amps), which is redistributed by hash code to all AMPs. Then we do a SORT to order Spool 5 by row hash. The size of Spool 5 is estimated with low confidence to be 4 rows ( 100 bytes). The estimated time for this step is 0.03 seconds. 5) We do an all-amps JOIN step from UT1.i by way of a RowHash match scan, which is joined to Spool 5 (Last Use) by way of a RowHash match scan. UT1.i and Spool 5 are joined using a merge join, with a join condition of ("PARTNO = UT1.i.partno"). The result goes into Spool 4 (all_amps), which is built locally on the AMPs. 31 9/15/2014 Teradata Confidential

Teradata-to-Teradata: Export Use INSERT statement to export data to remote system Supports Insert-Select from local table syntax: INSERT INTO store.pricelist@sdll7151 SELECT * FROM newitems; Also, supports inserting specific named columns into the remote table as well as constant data INSERT INTO store.pricelist@sdll7151 VALUES (1223315, 'UP Flat Car', 3.99); Table may be empty or an existing table with data. Data is appended to the existing table s data 32 9/15/2014 Teradata Confidential

Teradata-to-Teradata: Stitching Local and Remote explains During Export EXPLAIN INSERT INTO store.pricelist@sdll7151 SELECT * FROM newitems; 3) We do an all-amps RETRIEVE step from UT1.newitems by way of an all-rows scan with no residual conditions executing table operator syslib.load_to_td with a condition of ("(1=1)") into Spool 2 (used to materialize view, derived table, table function or table operator pricelist) (all_amps), which is built locally on the AMPs. < BEGIN EXPLAIN FOR REMOTE QUERY --> 1) First, we lock a distinct store."pseudo table" for write on a RowHash to prevent global deadlock for store.pricelist. 2) Next, we lock store.pricelist for write. 3) We do an all-amps RETRIEVE step executing table operator SYSLIB.load_from_tdRemote with a condition of ("(1=1)"). The size of Spool 1 is estimated with low confidence to be 4 rows (272 bytes). The estimated time for this step is 0.06 seconds. 4) We do an all-amps RETRIEVE step from Spool 1 (Last Use) by way of an all-rows scan into Spool 2 (all_amps), which is redistributed by hash code to all AMPs. Then we do a SORT to order Spool 2 by row hash. The size of Spool 2 is estimated with low confidence to be 4 rows (240 bytes). The estimated time for this step is 0.03 seconds. 5) We do an all-amps MERGE into store.pricelist from Spool 2 (Last Use). <-- END EXPLAIN FOR REMOTE QUERY > 33 9/15/2014 Teradata Confidential

Teradata-to-Teradata: Monitoring data movement Bytestransferred in and out are logged in DBQL Show size of data transfer Select Bytestransferred from dbc.dbqlsteptbl; Query Monitor portlet in Viewpoint can show data imported and exported per query 34 9/15/2014 Teradata Confidential

Quickly Building Out the QueryGrid Family Connector One-Way Bi-Directional Push-Down Teradata Hortonworks Hadoop Teradata Cloudera Hadoop August Planned Planned Teradata Oracle (11g) 3Q 2014 3Q 2014 Teradata - Teradata Oct Oct Oct Teradata - Aster 4Q2014 4Q2014 4Q2014 Teradata - MongoDB 1Q2015 1Q2015 1Q2015 Teradata DB2 1H2015 1H2015 1H2015 Teradata SQLServer 2Q2015 2QH2015 2QH2015 Aster - Teradata Aster 6 - Hortonworks Hadoop Planned Planned Aster 6 - Cloudera Hadoop Planned Planned Even more connectors being planned (Plans change though of course!) 35 9/15/2014 Teradata Confidential

BI Tools Run SQL-MR Through Teradata 36 9/15/2014 Teradata Confidential

Teradata Unity and Teradata QueryGrid Teradata Differentiators Work together to provide seamless user experience Simplify multi-system analytics Teradata Unity Multi-System Management Route queries Synchronize data UDA ecosystem mgmt Teradata QueryGrid Orchestrate processing across and beyond the Teradata UDA Right analytic, right platform Push-down processing Bi-directional data movement 37 9/15/2014 Teradata Confidential

38 9/15/2014 Teradata Confidential

QUESTIONS?