ORACLE GOLDENGATE BIG DATA ADAPTER FOR HIVE



Similar documents
ORACLE GOLDENGATE BIG DATA ADAPTER FOR FLUME

Oracle Retail MICROS Stores2 Functional Document Sales - Reasons Release September 2015

Oracle Retail MICROS Stores2 Functional Document Tax Free - Manual Receipt Number Assignment Release September 2015

Oracle Enterprise Manager

Oracle Fusion Middleware. 1 Oracle Identity Management Templates

NetFlow Collection and Processing Cartridge Pack User Guide Release 6.0

Oracle Enterprise Manager

Oracle Enterprise Manager Ops Center. Introduction. Tuning Monitoring Rules and Policies 12c Release 1 ( )

Oracle Retail MICROS Stores2 Functional Document General - License Control Release September 2015

Oracle Retail MICROS Stores2 Functional Document Customers - Online Centralized Customer Management Release September 2015

Oracle Cloud E

Note : It may be possible to run Test or Development instances on 32-bit systems with less memory.

StorageTek Library Attach for Window Servers

Review Employee Leave Balances

Siebel Application Deployment Manager Guide. Siebel Innovation Pack 2013 Version 8.1/8.2 September 2013

Oracle Cloud E

Oracle Communications Network Charging and Control. Release: 4.4

Oracle Retail Customer Engagement Cloud Service (Relate) Release Notes Release March 2015

Oracle WebLogic Server

Revenue/Expenses Balance by Fund

1 Changes in this release

Oracle Enterprise Manager

Oracle Fusion Middleware

JD Edwards EnterpriseOne Tools. 1 Understanding JD Edwards EnterpriseOne Business Intelligence Integration. 1.1 Oracle Business Intelligence

New Features in Primavera Contract Management 14.1

Oracle Enterprise Manager. Introduction to the Oracle Virtual Networking Plug-in. Requirements. Supported Versions

Oracle Cloud. Creating a Business Intelligence Cloud Extract E

Pre-Installation Instructions

Introduction to Virtual Datacenter

Oracle VM. Paravirtual Drivers Installation Guide for Microsoft Windows for Release E May 2012

MySQL and Virtualization Guide

Oracle Field Service Cloud SmartCollaboration Administration Panel. Release 4.5

Oracle FLEXCUBE Direct Banking Release Help Desk User Manual. Part No. E

Oracle WebCenter Sites. Backup and Recovery Guide 11g Release 1 (11.1.1)

Oracle Fusion Transactional Business Intelligence

Vendor Performance Summary Report

Oracle Fusion Middleware

About Contract Management

PeopleSoft Customer Relationship Management 9.1 Hardware and Software Requirements Guide

Oracle Audit Vault Administrator s Guide Oracle Audit Vault Auditor s Guide Oracle Enterprise Manager Cloud Control Administrator s Guide

Reviewing Employee History

Oracle Audit Vault Oracle FLEXCUBE Universal Banking Release [April] [2014]

Oracle Hospitality Payment Gateway Services Release Notes Release Part Number: E January 2016

Oracle Cloud. What s New for Oracle Compute Cloud Service (IaaS) Topics. July What's New for Oracle Compute Cloud Service (IaaS) Release 16.

MySQL Installer Guide

Oracle Banking Digital Experience

Oracle Fusion Middleware

Siebel Installation Guide for UNIX. Siebel Innovation Pack 2013 Version 8.1/8.2, Rev. A April 2014

IBM WebSphere Portal Reference Guide Release 9.2

Application Interface Services Server for Mobile Enterprise Applications Configuration Guide Tools Release 9.2


Oracle Fusion Middleware

Siebel Installation Guide for Microsoft Windows. Siebel Innovation Pack 2013 Version 8.1/8.2, Rev. A April 2014

Oracle Java Micro Edition Software Development Kit

Approve or Reject Purchase Requistions, Contracts, and Purchase Orders

Oracle Virtual Desktop Client for Android. Release Notes for Release 1.2

Oracle Retail MICROS Retail-J. Inventory Management: Stock Adjustments Release 13.0

1 Review Information About this Guide

Third Party System Management Integration Solution

Siebel Correspondence, Proposals, and Presentations Guide. Siebel Innovation Pack 2013 Version 8.1/8.2 September 2013

PeopleSoft Enterprise Campus Solutions 9.0 Enrollment Web Services

Required Ports and Protocols. Communication Direction Protocol and Port Purpose Enterprise Controller Port 443, then Port Port 8005

ORACLE 11g RDBMS Features: Oracle Total Recall Oracle FLEXCUBE Enterprise Limits and Collateral Management Release 12.1 [December] [2014]

Oracle Secure Payment Gateway for HIS Profit Series SQL Server Upgrade Procedures. May 2016

Oracle Cloud E

Oracle Virtual Desktop Infrastructure. VDI Demo (Microsoft Remote Desktop Services) for Version 3.2

Oracle Virtual Desktop Client for ipad. User Guide for Version 1.0

Oracle CRM On Demand Connected Mobile Sales Administration Guide. Version March 2012

Customer Order Portal Created on 9/24/ :45 PM

COPYRIGHT & TRADEMARKS

Oracle Virtual Desktop Client. Release Notes for Version 3.0

ORACLE USER PRODUCTIVITY KIT USAGE TRACKING ADMINISTRATION & REPORTING RELEASE 3.6 PART NO. E

Oracle Beehive. Using iphone or ipad with Oracle Beehive Release 2 ( )

PeopleSoft Enterprise FSCM 9.1 Mobile Inventory Management Hardware and Software Requirements

Oracle Transactional Business Intelligence Enterprise for Human Capital Management Cloud Service

How To Customize An Orgsync App On Anorus Mobile Security Suite On A Microsoft Ipad Oracle 2.5 (Ios) On A Pc Orca 2.2 (Iphone) On An Android Orca2 (Ip

Contents Legal Notices... 2 Preface... 5 Introduction... 7 Installation Instructions... 8

Collections Manager Release 7.5

Oracle Enterprise Manager

Oracle Order to Activate Integration Pack for Siebel CRM and Oracle Communications Order and Service Management

Oracle Banking Digital Experience

Oracle Hospitality OPERA Cloud Services Security Guide Release 1.20 Part Number: E April 2016

Integrating Oracle Sales Cloud, Release 9 with JD Edwards EnterpriseOne release 9.1 Implementation Guide

Oracle Virtual Desktop Client. Release Notes for Release 3.2

JD Edwards World. Database Audit Manager Release A9.3 E

Oracle Hospitality Cruise Shipboard Property Management System 3M AT9000MKII Installation Guide Release

Oracle Enterprise Manager Ops Center. Ports and Protocols. Ports and Protocols 12c Release 3 ( )

Oracle Sales Cloud Reporting and Analytics Overview. Release 13.2 Part Number E January 2014

P R O V I S I O N I N G O R A C L E H Y P E R I O N F I N A N C I A L M A N A G E M E N T

Oracle Fusion Middleware

JD Edwards EnterpriseOne Applications

Version Using Oracle Commerce Cloud Service Agent Console

Configuring Secure Network Communications for SAP

Functional Enhancements

Scheduler JAVA JOB Creation Oracle FLEXCUBE Investor Servicing Release [September] [2014]

Oracle WebLogic Server

2 Software requirements and installation

Oracle Business Intelligence Publisher. 1 Oracle Business Intelligence Publisher Certification. Certification Information 10g Release 3 (

Bank Account Numbers for ACH Payments

Oracle Enterprise Single Sign-on Logon Manager How-To: Configuring ESSO-LM Event Logging with Microsoft SQL Server 2005 Release

Transcription:

ORACLE GOLDENGATE BIG DATA ADAPTER FOR HIVE Version 1.0 Oracle Corporation i

Table of Contents TABLE OF CONTENTS... 2 1. INTRODUCTION... 3 1.1. FUNCTIONALITY... 3 1.2. SUPPORTED OPERATIONS... 4 1.3. UNSUPPORTED OPERATIONS... 4 2. BUILDING HIVE ADAPTER... 5 2.2 BUILDING WITH MAVEN... 5 2.3. BUILDING WITH BUILD SCRIPT... 6 3. CONFIGURATION... 7 3.1. HADOOP PREREQUISITE... 7 3.2. RUNTIME AND CLASSPATH... 7 3.3. ACCESSING THE PRECOMPILED HIVE ADAPTER... 8 3.4. HIVE ADAPTER PROPERTIES... 8 4. PERFORMANCE CONSIDERATIONS... 10 5. GUIDE TO CONFIGURE AND RUN HIVE ADAPTER... 11 5.1. HADOOP SETUP... 11 5.2. BUILDING HIVE ADAPTER USING MAVEN... 11 5.3. CONFIGURING HIVE ADAPTER... 11 5.4. STARTING HIVE ADAPTER... 11 5.5. CREATING TABLES IN HIVE... 12 5.6. QUERYING RECORDS FROM HDFS USING HIVE... 12 6. LIMITATION... 13 This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing. If this is software or related documentation that is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, then the following notice is applicable: U.S. GOVERNMENT END USERS: Oracle programs, including any operating system, integrated software, any programs installed on the hardware, and/or documentation, delivered to U.S. Government end users are "commercial computer software" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, use, duplication, disclosure, modification, and adaptation of the programs, including any operating system, integrated software, any programs installed on the hardware, and/or documentation, shall be subject to license terms and license restrictions applicable to the programs. No other rights are granted to the U.S. Government. This software or hardware is developed for general use in a variety of information management applications. It is not developed or intended for use in any inherently dangerous applications, including applications that may create a risk of personal injury. If you use this software or hardware in dangerous applications, then you shall be responsible to take all appropriate fail-safe, backup, redundancy, and other measures to ensure its safe use. Oracle Corporation and its affiliates disclaim any liability for any damages caused by use of this software or hardware in dangerous applications. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Intel and Intel Xeon are trademarks or registered trademarks of Intel Corporation. All SPARC trademarks are used under license and are trademarks or registered trademarks of SPARC International, Inc. AMD, Opteron, the AMD logo, and the AMD Opteron logo are trademarks or registered trademarks of Advanced Micro Devices. UNIX is a registered trademark of The Open Group. This software or hardware and documentation may provide access to or information about content, products, and services from third parties. Oracle Corporation and its affiliates are not responsible for and expressly disclaim all warranties of any kind with respect to third-party content, products, and services unless otherwise set forth in an applicable agreement between you and Oracle. Oracle Corporation and its affiliates will not be responsible for any loss, costs, or damages incurred due to your access to or use of thirdparty content, products, or services, except as set forth in an applicable agreement between you and Oracle. ii

1. INTRODUCTION The Oracle GoldenGate Big Data Adapter for Hive is designed to stream change capture data into the Hadoop Distributed File System (HDFS) and query the records using Hive. This example is developed to provide a ready-made functionality, while at time maintaining simplicity so that the example is easily understood. Oracle GoldenGate customers can use, modify, and extend this example to fulfill their specific needs. 1.1. Functionality The Hive Adapter takes the unfiltered operations from the source trail and writes them into multiple files in HDFS based on the file size configured. 1. Records for each table will be written into its own directory structure. With this approach tables can be created in Hive for each source table by pointing the LOCATION of the Hive table to the corresponding directory path in HDFS. Directory structure and format of records in the file are as noted below. Directory structure of files HDFS_ROOT_DIRECTORY/SCHEMA_NAME/TABLE_NAME/ HDFS_ROOT_DIRECTORY:- The root directory structure in HDFS as configured in the properties. SCHEMA_NAME:- The schema in which the source table exists. TABLE_NAME:- The name of the table for which the change data is being captured. File Naming <TABLE_NAME>_<SYSTEM_TIMESTAMP> TABLE_NAME:- The name of the table for which the change data is being captured. SYSTEM_TIMESTAMP:- The timestamp of the system when the file is being created. Timestamp is being used in the file name to maintain unique file names. Change data records <OPERATION_TYPE><COLUMN_DELIMITER><COLUMN_1_VALUE><COLUMN_DELIMITER>...<COLUM N_N_VALUE><COLUMN_DELIMITER><OPERATION_TIMESTAMP><LINE_DELIMITER> OPERATION_TYPE:- Indicated the type of operation i.e. Insert or Update or Delete. COLUMN_DELIMITER:- Column delimiter as specified in the properties. COLUMN_1_VALUE:- Value of the 1 st column.... :- COLUMN_DELIMITER and COLUMN_VALUE will be repeated for all column values between the first and last column. COLUMN_N_VALUE:- Value of the last column. OPERATION_TIMESTAMP:- Timestamp of the operation. LINE_DELIMITER:- Line delimiter as specified in properties. iii

2. All records written to the HDFS file will be in Delimiter Separated Values (DSV) format. Column and row delimiters can be configured in the properties. 3. Files for each table will be rolled based on size configured in properties. 4. On every restart of the adapter pump process, the current open files, if any, will be closed, resulting in the opening of a new file in HDFS. 1.2. Supported Operations Inserts Updates including change to primary key(s) Deletes 1.3. Unsupported Operations Truncate Table iv

2. BUILDING HIVE ADAPTER 2.1 Prerequisite 1. The Oracle GoldenGate Java Adapter 12 is built and runs with Java 7. It is required that the Hive Adapter also be built with the Java SE Development Kit 7. 2. The recommended method of building the Hive Adapter is using Apache Maven. Maven is an open source build tool available for download at http://maven.apache.org. One of the biggest benefits provided by Maven is that it will resolve and download the build and runtime dependencies of the Hive Adapter from the Maven central repository. Internet connectivity is required to access the Maven central repository. The Hive Adapter was built and tested with version 3.0.5 of Maven. 3. The precompiled Hive Adapter was built and tested using version 2.5.1 of the HDFS client and 0.13.1 version of Hive client. NOTE: - Step 2 is required only if maven is used to build the Hive adapter. 2.2 Building with Maven 1. Navigate to GoldenGate (Java Adapters) root directory GG_ROOT_DIRECTORY /AdapterExamples/big-data/hive 2. Type the following command to build mvn clean package 3. Maven will download all the required dependencies and build the Hive Adapter. The created jar and the downloaded dependent jars will be placed in the following directory GG_ROOT_DIRECTORY/AdapterExamples/big-data/hive/target/hive-lib 4. One of the dependencies that Maven does not download is the Oracle GoldenGate Java adapter (ggjava.jar). This dependency is located in the following directory GG_ROOT_DIRECTORY/ggjava/ggjava.jar This dependency is instead resolved using a relative path inside the pom.xml file. If the Hive Adapter project is copied to a different directory, the relative path to the ggjava.jar file will need to be fixed by editing the pom.xml file in: GG_ROOT_DIRECTORY/AdapterExamples/big-data/hive Note: - The pom.xml file in the GG_ROOT_DIRECTORY/AdapterExamples/big-data/hive directory instructs Maven what version of the HDFS client to download and to build against. It is important that the HDFS client version match the version of HDFS to which it is connecting. Using mismatched versions of the HDFS client and HDFS server can result in serious problems. The version of the HDFS client in the pom.xml file is 2.5.1. If using a version of HDFS other than 2.5.1 it is highly recommended that the user v

modify the pom.xml file so that HDFS Adapter is built against the correct version of the HDFS client, and the correct jar files are downloaded for use at runtime. WARNING: - Many companies employ proxy servers to help protect their network from the hazards of the Internet. Performing a Maven build while connected to the company network or VPN may require that dependency downloads from the Maven central repository go through a proxy server. This may require special configuration of the Maven settings.xml file which is generally located in the USER HOME/.m2 directory. Consult your network administrator for proxy settings. Consult the Maven documentation on how to correctly configure the settings.xml file for your required proxy settings. 2.3. Building with Build Script A build script, build.sh, is provided in the GG_ROOT_DIRECTORY/AdapterExamples/bigdata/hdfs directory. This script is only an alternative to the Maven build mentioned above. Note: - The build script requires HDFS_HOME environment variable to be set. vi

3. Configuration 3.1. Hadoop Prerequisite In order to successfully run the Hive Adapter, Hadoop single instance or Hadoop cluster must be installed, running, and network accessible to the machine running the Hive Adapter. 3.2. Runtime and Classpath 1. Hadoop Connectivity The Oracle GoldenGate Big Data Adapter for Hive obtains connectivity information to HDFS via the Hadoop core-site.xml file. This file is generally located in the HADOOP_HOME/etc/hadoop directory. The directory containing the properly configured core-site.xml file must be in the CLASSPATH. 2. Hive Dependencies Classpath The GoldenGate Hive Adapter requires the HDFS client, which in turn requires a number of dependent jars. The easiest way to fulfill the runtime dependencies of the Hive Adapter is by configuring the gg.classpath property in the properties file using the wildcard asterisk character (*).. The GoldenGate Java properties file is very specific as to how the wildcard character is used. The following works: gg.classpath= GG_ROOT_DIRECTORY/AdapterExamples/bigdata/hive/target/hive-lib/* The following does NOT work: gg.classpath= GG_ROOT_DIRECTORY/AdapterExamples/bigdata/hive/target/hive-lib/*.jar 3. GG Adapter Boot Options The following can be configured using the boot options property. Memory allocation for GG Java Adapter JVM (-Xmx and -Xms). HDFS core-site.xml GG Adapter dependencies (ggjava.jar) javawriter.bootoptions=-xmx512m -Xms32m -Djava.class.path=ggjava/ggjava.jar -Dlog4j.configuration=log4j.properties vii

PATH_TO_HADOOP_CONFIG_DIR: Directory containing Hadoop core-site.xml. CLASSPATH_DELIMITER: The Classpath delimiter for Windows is semicolon (;) and Linux is colon (:). 3.3. Accessing the Precompiled Hive Adapter The Oracle GoldenGate Big Data Adapter for Hive installation includes the pre-built jar file. The Hive Adapter binary is available in the following location GG_ROOT_DIRCTORY/AdapterExamples/big-data/hive/bin/ogg-hive-adapter-1.0.jar If the If the above jar is being used instead of building the Hive Adapter from source, then ogg-hiveadapter-1.0.jar should be added to the gg.classpath variable in the properties file. The gg.classpath variable must also include the HDFS dependency jars. This is generally resolved by adding HDFS_ROOT_DIRECTORY/share/hadoop/common/lib/*:HDFS_ROOT_DIRECTORY/share/hadoo p/common/* to the gg.classpath variable. 3.4. Hive Adapter Properties The below properties are specific to the Hive Adapter. The name attribute represents the handler name configured as part of gg.handlerlist. 1. gg.handler.name.type The value of this property should not be changed and always should be com.goldengate.delivery.handler.hdfs.hivehandler. This property is mandatory. 2. gg.handler.name.fielddelimiter The delimiter to be used for field/column value separation. Values like semicolon (;) or comma (,) or any other character can be set. Nonprintable characters like \u0001, \u0002 are also supported. Use the Unicode format in the configuration file when using nonprintable characters as delimiters. The default value is \u0001. This property is optional. 3. gg.handler.name.linedelimiter The delimiter to be used for line separation.this property is optional. The default value is \n. gg.handler.name.rootdirectorypath viii

The root directory in HDFS under which the directories and files will be created. This property is mandatory. 4. gg.handler.name.deleteopkey The actual key that should be used to represent the DELETE operation type in DSV. This property is optional. The default value is D. 5. gg.handler.name.updateopkey The actual key that should be used to represent the UPDATE operation type in DSV This property is optional. Default value is U. 6. gg.handler.name.insertopkey The actual key that should be used to represent the INSERT operation type in DSV. This property is optional. Default value is I. 7. gg.handler.name.pkupdateopkey The actual key that should be used to represent the PRIMARY KEY UPDATE operation type in DSV. This property is optional. Default value is P. 8. gg.handler.name.includeoptype A value of true indicates the operation type is to be included in DSV. A value of false indicates the operation type is not to be included in DSV. This property is optional. The default value is true. 9. gg.handler.name.includeoptimestamp A value of true indicates the operation timestamp is to be included in DSV. A value of false indicates the operation timestamp is not to be included in DSV. This property is optional. The default value is true. 10. gg.handler.name.maxfilesize The maximum size of the file in HDFS. Once the maximum size is reached, the current file will be closed and a new file will be opened. The value should be entered in bytes. This property is optional. The default value is 10485760 (i.e.10 MB). ix

4. Performance Considerations By default, data is flushed to the HDFS server at the end of each transaction. This behavior is the same if operating in op or tx mode, so performance differences are not likely to be significant. If transactions are small (contain few operations), performance may be significantly increased by employing the new transaction grouping functionality. The transaction grouping functionality allows operations which cross multiple transaction boundaries to be grouped together as a single transaction. Because performance is dependent upon many variables, the use of the transaction grouping functionality cannot guarantee increased performance. It is simply a mechanism to help customers tune the Oracle GoldenGate Java Adapter to their specific needs. 4.1 Grouping Configuration 1. gg.handler.name.mode To enable grouping the value of this property should be set to tx. 2. gg.handler.name.maxgroupsize Controls the maximum number of operations that can be held by an operation group irrespective of whether the operation group holds operations from a single transaction or multiple transactions. The operation group will send a transaction commit and end the group as soon as this number of operations is reache. This property leads to splitting of transactions across multiple operation groups 3. gg.handler.name.mingroupsize This is the minimum number of operations that must exist in a group before the group can end. This property helps to avoid groups that are too small by grouping multiple small transactions into one operation group so that it can be more efficiently processed. NOTE: maxgroupsize should always be greater than or equal to mingroupsize; i.e. maxgroupsize >= mingroupsize. Example:- Consider a scenario where the source trail has 100 transactions with 10 operations each and the handler is configured in tx mode. Without grouping functionality, transaction commit occurs for every 10 records ultimately flushing the records to HDFS for every 10 records. With grouping enabled by setting max and min group size to 1000, transaction commit occurs for 1000 records ultimately flushing 1000 records to HDFS at once. x

5. Guide to configure and run Hive Adapter The following steps assume GG_ROOT_DIRECTORY to be the GoldenGate installation directory. 5.1. Hadoop Setup Make sure Hadoop is started using the script available in $HADOOP_HOME/sbin. 1../start-dfs.sh 2../start-yarn.sh 5.2. Building Hive Adapter using Maven 1. Navigate to $GG_ROOT_DIRECTORY/AdapterExamples/big-data/hive 2. Execute command mvn clean package 5.3. Configuring Hive Adapter 1. Copy $GG_ROOT_DIRECTORY/AdapterExamples/big-data/hive/dirprm/hive.prm to $GG_ROOT_DIRECTORY/dirprm 2. Copy $GG_ROOT_DIRECTORY/AdapterExamples/big-data/hive/dirprm/hive.props to $GG_ROOT_DIRECTORY/dirprm 3. Edit {HDFS_INSTALL_DIRECTORY" in $GG_ROOT_DIRECTORY/dirprm/hive.props to point to a valid Hadoop installation path 4. At the GGSCI command prompt, execute the following command: GGSCI> ADD EXTRACT HIVE, EXTTRAILSOURCE./AdapterExamples/javadelivery/tc 5.4. Starting Hive Adapter 1. Make sure that the LD_LIBRARY_PATH environment variable set. For example: export LD_LIBRARY_PATH=$JAVA_HOME/jre/lib/amd64/server/ 2. At the GGSCI command prompt, execute the following commands: GGSCI 1> START MGR GGSCI 2> START HIVE GGSCI 3> INFO ALL Both the Manager and Hive processes should be in RUNNING status. xi

5.5. Creating tables in Hive 1. Navigate to $HIVE_HOME/bin and execute the following command:./hive 2. Execute the following create table scripts: hive> CREATE EXTERNAL TABLE TCUSTMER( OPERATION_TYPE CUST_CODE NAME CITY STATE OPERATION_TIMESTAMP STRING ) ROW FORMAT DELIMITED FIELDS TERMINATED BY '\u0001' LINES TERMINATED BY '\n' LOCATION '/gg/replication/hive/gg/tcustmer'; hive> CREATE EXTERNAL TABLE TCUSTORD( OPERATION_TYPE CUST_CODE ORDER_DATE PRODUCT_CODE ORDER_ID INT, PRODUCT_PRICE DOUBLE, PRODUCT_AMOUNT INT, TRANSACTION_ID INT, OPERATION_TIMESTAMP STRING ) ROW FORMAT DELIMITED FIELDS TERMINATED BY '\u0001' LINES TERMINATED BY '\n' LOCATION '/gg/replication/hive/gg/tcustord'; 5.6. Querying records from HDFS using Hive 1. Navigate to $HIVE_HOME/bin and execute the following command:./hive 2. Verify data replication into HDFS using the following queries: hive> select * from TCUSTMER; hive> select * from TCUSTORD xii

6. Limitation 1. Tables in Hive will not be created by the Hive Adapter. Tables should be manually created in Hive using the Hive create table syntax. 2. HDFS DataStream will be flushed using hflush on every transaction commit event, but records may not be written onto the physical disc until the DataStream is closed (i.e. once the max file size is reached or if the adapter pump process is stopped). Due to this, the Hive client cannot retrieve the latest records from HDFS until the data is written to disc. 3. The new unified update record is not yet supported by the Hive Adapter and will cause the Extract process to ABEND. 4. Hive adapter does not support the truncate table operation. A truncate table operation will cause the Extract process to ABEND. xiii