Cloudera Backup and Disaster Recovery



Similar documents
Cloudera Backup and Disaster Recovery

How To Use Cloudera Manager Backup And Disaster Recovery (Brd) On A Microsoft Hadoop (Clouderma) On An Ubuntu Or 5.3.5

Cloudera Navigator Installation and User Guide

Configuring TLS Security for Cloudera Manager

Cloudera Manager Introduction

Cloudera Enterprise Reference Architecture for Google Cloud Platform Deployments

Important Notice. (c) Cloudera, Inc. All rights reserved.

Cloudera Manager Health Checks

Cloudera Manager Health Checks

Cloudera Enterprise Reference Architecture for Google Cloud Platform Deployments

Cloudera Manager Training: Hands-On Exercises

CRM to Exchange Synchronization

Big Data Operations Guide for Cloudera Manager v5.x Hadoop

CDH 5 Quick Start Guide

Cloudera Manager Monitoring and Diagnostics Guide

Cloudera Manager Monitoring and Diagnostics Guide

Cloudera ODBC Driver for Apache Hive Version

Security Provider Integration Kerberos Server

Cloudera Administration

CRM to Exchange Synchronization

Data Domain Profiling and Data Masking for Hadoop

Contents Notice to Users

Web Remote Access. User Guide

How to Install and Configure EBF15328 for MapR or with MapReduce v1

DocAve 6 Service Pack 1 Job Monitor

Cloudera Manager Administration Guide

CRM to Exchange Synchronization

StarWind iscsi SAN Software: Tape Drives Using StarWind and Symantec Backup Exec

SolarWinds Migrating SolarWinds NPM Technical Reference

Cloudera Navigator Installation and User Guide

Cloudera ODBC Driver for Apache Hive Version 2.5.5

Single Sign-On Guide for Blackbaud NetCommunity and The Patron Edge Online

Group Management Server User Guide

Introduction to Hyper-V High- Availability with Failover Clustering

Cloud Services ADM. Agent Deployment Guide

Management Reporter Integration Guide for Microsoft Dynamics AX

Windows Server Update Services 3.0 SP2 Step By Step Guide

BDR for ShadowProtect Solution Guide and Best Practices

SafeNet Authentication Manager Express. Upgrade Instructions All versions

Active Directory Management. Agent Deployment Guide

ShadowControl ShadowStream

Configuration Guide. Remote Backups How-To Guide. Overview

CA Performance Center

Dell Enterprise Reporter 2.5. Configuration Manager User Guide

Xcalibur Global Version 1.2 Installation Guide Document Version 3.0

Red Hat Enterprise Linux OpenStack Platform 7 OpenStack Data Processing

Synchronization Agent Configuration Guide

DocAve 4.1 SharePoint Disaster Recovery High Availability (SPDR HA) User Guide

How To Set Up A Load Balancer With Windows 2010 Outlook 2010 On A Server With A Webmux On A Windows Vista V (Windows V2) On A Network With A Server (Windows) On

Veeam Backup Enterprise Manager. Version 7.0

Omniquad Exchange Archiving

User's Guide. Product Version: Publication Date: 7/25/2011

Virtual Web Appliance Setup Guide

File Management Utility User Guide

EventTracker: Configuring DLA Extension for AWStats Report AWStats Reports

Appointment Scheduler

Integrating VMware Horizon Workspace and VMware Horizon View TECHNICAL WHITE PAPER

Configuring Hadoop Security with Cloudera Manager

efolder BDR for Veeam Cloud Connection Guide

Virtual Managment Appliance Setup Guide

Microsoft FTP Configuration Guide for Helm 4

RoomWizard Synchronization Software Manual Installation Instructions

Lepide Software. LepideAuditor for File Server [CONFIGURATION GUIDE] This guide informs How to configure settings for first time usage of the software

Mobility Services Platform Software Installation Guide

Installation and Configuration Guide

MANUFACTURER RamSoft Incorporated 243 College St, Suite 100 Toronto, ON M5T 1R5 CANADA

Active Directory Provider User s Guide

USER GUIDE WEB-BASED SYSTEM CONTROL APPLICATION. August 2014 Phone: Publication: , Rev. C

Pipeliner CRM Phaenomena Guide Getting Started with Pipeliner Pipelinersales Inc.

Database Administration Guide

Pipeliner CRM Phaenomena Guide Administration & Setup Pipelinersales Inc.

WhatsUp Gold v16.2 Installation and Configuration Guide

ORACLE USER PRODUCTIVITY KIT USAGE TRACKING ADMINISTRATION & REPORTING RELEASE 3.6 PART NO. E

Getting Started with Database-as-a-Service

Configuring Failover

DocAve 4.1 Backup User Guide

Frequently Asked Questions

TASKE Call Center Management Tools

DIGIPASS KEY series and smart card series for Juniper SSL VPN Authentication

Pipeliner CRM Phaenomena Guide Add-In for MS Outlook Pipelinersales Inc.

Disaster Recovery. Websense Web Security Web Security Gateway. v7.6

VMware Identity Manager Connector Installation and Configuration

Contents Overview... 5 Configuring Project Management Bridge after Installation... 9 The Project Management Bridge Menu... 14

Archiving User Guide Outlook Plugin. Manual version 3.1

FileMaker Server 13. FileMaker Server Help

How To Login To The Mft Internet Server (Mft) On A Pc Or Macbook Or Macintosh (Macintosh) With A Password Protected (Macbook) Or Ipad (Macro) (For Macintosh) (Macros

Pipeliner CRM Phaenomena Guide Sales Target Tracking Pipelinersales Inc.

Hadoop Basics with InfoSphere BigInsights

SC-T35/SC-T45/SC-T46/SC-T47 ViewSonic Device Manager User Guide

Server Installation Guide ZENworks Patch Management 6.4 SP2

Dell Statistica Statistica Enterprise Installation Instructions

USER GUIDE CLOUDME FOR WD SENTINEL

Centrify Identity and Access Management for Cloudera

Pipeliner CRM Phaenomena Guide Sales Pipeline Management Pipelinersales Inc.

Device LinkUP + Desktop LP Guide RDP

MadCap Software. Upgrading Guide. Pulse

-lead Grabber Business 2010 User Guide

Sophos Mobile Control Startup guide. Product version: 3

Content Filtering Client Policy & Reporting Administrator s Guide

fåíéêåéí=péêîéê=^çãáåáëíê~íçêûë=dìáçé

Transcription:

Cloudera Backup and Disaster Recovery

Important Notice (c) 2010-2013 Cloudera, Inc. All rights reserved. Cloudera, the Cloudera logo, Cloudera Impala, and any other product or service names or slogans contained in this document are trademarks of Cloudera and its suppliers or licensors, and may not be copied, imitated or used, in whole or in part, without the prior written permission of Cloudera or the applicable trademark holder. Hadoop and the Hadoop elephant logo are trademarks of the Apache Software Foundation. All other trademarks, registered trademarks, product names and company names or logos mentioned in this document are the property of their respective owners. Reference to any products, services, processes or other information, by trade name, trademark, manufacturer, supplier or otherwise does not constitute or imply endorsement, sponsorship or recommendation thereof by us. Complying with all applicable copyright laws is the responsibility of the user. Without limiting the rights under copyright, no part of this document may be reproduced, stored in or introduced into a retrieval system, or transmitted in any form or by any means (electronic, mechanical, photocopying, recording, or otherwise), or for any purpose, without the express written permission of Cloudera. Cloudera may have patents, patent applications, trademarks, copyrights, or other intellectual property rights covering subject matter in this document. Except as expressly provided in any written license agreement from Cloudera, the furnishing of this document does not give you any license to these patents, trademarks copyrights, or other intellectual property. The information in this document is subject to change without notice. Cloudera shall not be liable for any damages resulting from technical errors or omissions which may be present in this document, or from use of this document. Cloudera, Inc. 1001 Page Mill Road, Building 2 Palo Alto, CA 94304-1008 info@cloudera.com US: 1-888-789-1488 Intl: 1-650-362-0488 www.cloudera.com Release Information Version: 4.7.0 Date: September 5, 2013

Table of Contents About this Guide...5 Designating a Replication Source...7 Modifying the Peer Configuration...7 HDFS Replication...9 Viewing Replication Job Status...10 Hive Replication...11 Viewing Replication Job Status...13 Enabling Replication Between Clusters in Different Kerberos Realms...15

About this Guide About this Guide This guide describes Cloudera Backup and Disaster Recovery (BDR), a separately-licensed Cloudera product that provides an integrated, easy-to-use management solution for enabling data protection in the Hadoop platform. The information in this guide is also available in the online Help included with Cloudera Manager. The following sections are covered in this Guide: Designating a Replication Source on page 7 HDFS Replication on page 9 Hive Replication on page 11 Enabling Replication Between Clusters in Different Kerberos Realms on page 15 Cloudera Backup and Disaster Recovery 5

Designating a Replication Source Designating a Replication Source From the Cloudera Manager Admin Console, you can designate a Cloudera Manager server as the source for data (files) to be replicated to a service managed by the Cloudera Manager server you are logged into. To set up a peer relationship as a replication source: 1. From the Service page for either Hive or HDFS, select the Replication tab. 2. Click the link Add Replication Source to go to the Administration page Peers tab. You can also go directly to the Peers page by pulling down the Administration tab and selecting Peers. If there are no existing peers, you will see only an Add Peer button in addition to a short message. If you have existing peers, they are listed here. 3. Click the Add Peer button. 4. In the Add Peer pop-up, provide a name, the URL (including the port) of the Cloudera Manager Server that will act as the source for the data to be replicated, and the login credentials for that server. Note that the Data Replication feature recommends that SSL be used, and a warning is shown if the URL uses http instead of https. However, you can ignore the warning and proceed if SSL is not available. 5. Click the Add Peer button in the pop-up to create the peer relationship. 6. To test the connectivity between your current Cloudera Manager server and the remote server select Test Connectivity from the Actions menu associated with the peer. Note that the current Cloudera Manager system is also available as a replication source. Modifying the Peer Configuration To modify the peer configuration (to change the login or password): 1. Pull down the Admin tab, and select Peers. 2. From the Actions menu for the peer, select Edit. 3. Make your changes. 4. Click Update Peer to save your changes. You can also delete a peer relationship: From the Actions menu for the peer, select Delete. Cloudera Backup and Disaster Recovery 7

HDFS Replication HDFS Replication HDFS Replication enables you to copy (replicate) your HDFS data from a remote (or local) Peer Cloudera Manager server to your local Cloudera Manager server (the server whose Admin console you are currently logged into) and keep the data sets synchronized. You can add Peers though the Administration > Peers tab (see Designating A Replication Source). You can also use the Add Replication Source link on the HDFS Replication page to go to the Peers page. Note: HDFS replication will not work between a source cluster that has encryption enabled and a target cluster running CDH 4.0. This is because the CDH 4.0 client is used for replication in this case, and it does not support encryption. Once you have a peer relationship set up with a Cloudera Manager server, you can configure replication of your HDFS data. 1. From the Services tab, go to the CDH4 HDFS service where you want to host the replicated data. 2. Click the Replication tab at the top of the page. 3. Select the HDFS service to be the source of the replicated data. If the peer Cloudera Manager Server has multiple CDH4 HDFS services (for example, if it is managing multiple CDH4 clusters) you will be able to select the HDFS service you want to use as the source. Note that the local CDH4 HDFS service (being managed by the Cloudera Manager server you are logged into) is also available as a replication source. If the peer whose HDFS service you want is not listed, click the Add Peer link to go to the Peers page to add a Cloudera Manager peer. When you select a replication source, the Create Replication pop-up opens. 4. Enter the path to the directory (or file) you want to replicate (the source). 5. Enter the path where the target files should be placed. 6. Select a schedule: You can have it run immediately, run once at a scheduled time in the future, or at regularly scheduled intervals. If you select "Once" or "Recurring" you are presented with fields that let you set the date and time and (if appropriate) the interval between runs. 7. If you want to modify the parameters of the MapReduce job, click More Options. Here you will be able to select a MapReduce service (if there is more than one in your cluster) and change the following parameters: The MapReduce service to use. The scheduler pool to use. The user that should run the MapReduce job. By default this is hdfs. If you want to run the MR job as a different user, you can enter that here. If you are using Kerberos, you MUST provide a user name here, and it must be one with an ID greater than 1000. An alternative path for the logs. Limits for the number of map slots and for bandwidth per mapper. The defaults are unlimited. Whether to abort the job on an error (default is not to do so). This means that files copied up to that point will remain on the destination, but no additional files will be copied. Whether to skip checksum checks (default is to perform them). If checked, checksum validation will not be performed. Whether to remove deleted files from the target directory if they have been removed on the source. When this option is enabled, files deleted from the target directory are sent to trash if HDFS trash is enabled, or are deleted permanently if trash is not enabled. Further, with this option enabled, if files unrelated to the source exist in the target location, then those files will also be deleted. Whether to preserve the block size, replication count, and permissions as they exist on the source file system, or to use the settings as configured on the target file system. The default is to preserve these settings as on the source. Cloudera Backup and Disaster Recovery 9

HDFS Replication Note: If you leave the setting to preserve permissions, then you must be running as a superuser. You can use the "Run as" option to ensure that is the case. Whether to generate alerts for various state changes in the replication workflow. You can alert on failure, on start, on success, or when the replication workflow is aborted. 8. Click Save Schedule to save the replication specs. When saved, the replication job appears in the Replication list, with relevant information about the source and target locations, and the timestamp of the last run and the next scheduled run (if there is a recurring schedule). A scheduled job will show a calendar icon to the left of the job specification. If it is scheduled to run once, the calendar icon will disappear after the job has run. To specify additional replication tasks, click the Create Replication button that appears once you have added the first replication task. Note that only one replication can occur at a time; if another replication job starts before the previous one has finished, the second one is canceled. You can test the replication task without actually transferring data using the "Dry Run" feature: From the Actions menu for the replication task you want to test, click Dry Run. From the Actions menu for a replication task, in addition to Dry Run you can also: Edit the job configuration Run the job (immediately) Delete the job Disable/Enable the job (if the job is on a recurring schedule) When a task is disabled, instead of the calendar icon you will see a Stopped icon, and the job entry will appear in gray. Disabling and enabling a job is only available if the job is on a recurring schedule. Viewing Replication Job Status While a run is in progress, the calendar icon turns into spinner, and each stage of the replication task is indicated in the message after the replication specification. If the replication is successful, the number of files copied is indicated. If there have been no changes to a file at the source since the previous replication, then that file will not be copied. As a result, after the initial replication run, only a subset of the files may actually be copied, and this will be indicated in the success message. If the replication fails, that will be indicated and the timestamp will appear in Red text. To view more information about completed replication runs, click anywhere in the replication job entry row in the replication list. This displays sub-entries for each past replication run. To view detailed information about a particular past run, click the entry for that replication run. This opens another sub-entry that shows: A result message The start and end time of the replication job. A link to the command details for that replication run. Details about the data that was replicated. When viewing a sub-entry, you can dismiss the sub-entry by clicking anywhere in its parent entry, or by clicking the return arrow icon!at the top left of the sub-entry area. 10 Cloudera Backup and Disaster Recovery

Hive Replication Hive Replication Hive Replication enables you to copy (backup) and keep in sync the Hive Metastore and data from clusters managed by a remote peer or local Cloudera Manager server, and keep the copy on a cluster managed by your local Cloudera Manager server (the server whose Admin console you are currently logged into). You can add Peers though the Administration > Peers tab (see Designating a Replication Source). You can use the Add Peer link on the Replication page to go to the Peers page to add a new peer Cloudera Manager server. Once you have a peer relationship set up with a Cloudera Manager server, you can configure replication of your Hive Metastore data. Note: Hive replication between CDH4.2 or later and CDH4.0 does not work if the Hive schema contains views. If data replication is desired, Hive replication does not work between a source cluster that has encryption enabled and a target cluster running CDH 4.0. This is because the CDH 4.0 client used for replication does not support encryption. Hive replication (even without data replication) does not work between a source cluster running CDH 4.0 and a target cluster that has encryption enabled. 1. From the Services tab, go to the CDH4 Hive service where you want to host the replicated data. 2. Click the Replication tab at the top of the page. 3. Select the Hive service to be the source of the replicated data. If the peer Cloudera Manager Server has multiple CDH4 Hive services (for example, if it is managing multiple CDH4 clusters) you will be able to select the service you want to use as the source. If the peer whose Hive service you want is not listed, click the Add Peer link to go to the Peers page to add a Cloudera Manager peer. When you select a replication source, the Create Replication pop-up opens. 4. Leave Replicate All checked to replicate all the Hive metastore databases from the source. To replicate only selected databases, uncheck this option and enter the Database name(s) and tables you want to replicate. You can specify multiple data bases and tables using the plus symbol to add more rows to the specification. You can specify multiple databases on a single line by separating their names with the "pipe" character. For example: mydbname1 mydbname2 mydbname3 Regex can be used in either Database or Table fields. For example: [\w_]+ (?!\\b(myname)\\b).* db1 db2 [\w_]+ db1 [\w_]+ Click the "+" button and then enter any database/table name any database/table except the one named "myname" To get all tables of the db1 and db2 databases Alternate way to get all tables of the db1 and db2 databases Cloudera Backup and Disaster Recovery 11

Hive Replication db2 [\w_]+ 5. Select the target destination. If there is only one Hive service managed by Cloudera Manager available as a target, then this will be specified as the target. If there are more than one Hive services managed by this Cloudera Manager, then you will be able to select among those. 6. Select a schedule: You can have it run immediately, run once at a scheduled time in the future, or at regularly scheduled intervals. If you select "Once" or "Recurring" you are presented with fields that let you set the date and time and (if appropriate) the interval between runs. 7. Uncheck the Replicate HDFS Files checkbox to skip replicating the associated data files if you uncheck this, only the Hive metadata will be replicated. These are replicated to a default location; to specify a different location, you can change the Export Path and Destination under the More Options section, described below. 8. Use the More Options section to specify an export location, modify the parameters of the MapReduce job that will perform the replication, and other options. Here you will be able to select a MapReduce service (if there is more than one in your cluster) and change the following parameters: By default, Cloudera Manager exports the Hive Metadata to a default HDFS location (/user/${user.name}/.cm/hive) and then imports from this HDFS file to the target Hive Metastore. The default HDFS location for this export file can be overridden by specifying a path in the Export Path field. The Force Overwrite option, if checked, forces overwriting data in the target metastore if there are incompatible changes detected. For example, if the target metastore was modified and a new partition was added to a table, this option would force deletion of that partition, overwriting the table with the version found on the source. Important: If the Force Overwrite option is not set and the Hive replication process detects incompatible changes on the source cluster, Hive replication will fail. By default, Cloudera Manager replicates Hive's HDFS data files to a default location (/). To override the default, enter a path in the Destination field. Select the MapReduce service to use for this replication (if there is more than one in your cluster). The user is set in the Run As option. To specify the user that should run the MapReduce job, use the Run As option. By default MapReduce jobs run as hdfs. If you want to run the MR job as a different user, you can enter that here. If you are using Kerberos, you MUST provide a user name here, and it must be one with an ID greater than 1000. An alternative path for the logs. Limits for the number of map slots and for bandwidth per mapper. The defaults are unlimited. Whether to abort the job on an error (default is not to abort the job). Check the checkbox to enable this. This means that files copied up to that point will remain on the destination, but no additional files will be copied. Whether to skip checksum checks (default is to perform them). Whether to remove deleted files from the target directory if they have been removed on the source. Whether to preserve the block size, replication count, and permissions as they exist on the source file system, or to use the settings as configured on the target file system. The default is to preserve these settings as on the source. Note: If you leave the setting to preserve permissions, then you must be running as a superuser. You can use the "Run as" option to ensure that is the case. Whether to generate alerts for various state changes in the replication workflow. You can alert on failure, on start, on success, or when the replication workflow is aborted. 9. Click Save Schedule to save the replication specs. 12 Cloudera Backup and Disaster Recovery

Hive Replication When saved, the replication job appears in the Replication list, with relevant information about the source and target locations, and the timestamp of the last run and the next scheduled run (if there is a recurring schedule). A scheduled job will show a calendar icon to the left of the job specification. If it is scheduled to run once, the calendar icon will disappear after the job has run. To specify additional replication tasks, click the Create Replication button that appears once you have added the first replication tasks. If the replication failed, the timestamp will appear in Red text. Note that only one replication can occur at a time; if another replication job starts before the previous one has finished, the second one is cancelled. You can test the replication task without actually transferring data using the "Dry Run" feature: From the Actions menu for the replication task you want to test, click Dry Run. From the Actions menu for a replication task, in addition to Dry Run you can also: Edit the task configuration Run the task (immediately) Delete the task Disable/Enable the job (if the job is on a recurring schedule). When a task is disabled, instead of the calendar icon you will see a Stopped icon, and the job entry will appear in grey. Disabling and enabling a job is only available if the job is on a recurring schedule. Viewing Replication Job Status While a run is in progress, the calendar icon turns into spinner, and each stage of the replication task is indicated in the message after the replication specification. If the replication is successful, the number of files copied is indicated. If there have been no changes to a file at the source since the previous replication, then that file will not be copied. As a result, after the initial replication run, only a subset of the files may actually be copied, and this will be indicated in the success message. The Replication task can be aborted. While replication is running (the spinner is running), click on Commands and it will show the Hive Replication with another spinner, and next to it an Abort button. Click the Abort button to terminate the task. If the remote export task is still running, the Abort will cause the remote task to be terminated also. If the replication fails, that will be indicated and the timestamp will appear in Red text. To view more information about completed replication runs, click anywhere in the replication job entry row in the replication list. This displays sub-entries for each past replication run. To view detailed information about a particular past run, click the entry for that replication run. This opens another sub-entry that shows: A result message The start and end time of the replication job. A link to the command details for that replication run. Details about the data that was replicated. When viewing a sub-entry, you can dismiss the sub-entry by clicking anywhere in its parent entry, or by clicking the return arrow icon!at the top left of the sub-entry area. Cloudera Backup and Disaster Recovery 13

Enabling Replication Between Clusters in Different Kerberos Realms Enabling Replication Between Clusters in Different Kerberos Realms If you want to enable replication between clusters that reside in different Kerberos Realms, there are some additional setup steps you need to perform to ensure that the source and target clusters can communicate. Note: If either the source or target cluster is running Cloudera Manager 4.6 or later, then both clusters (source and target) must be running 4.6 or later. Cross-realm authentication does not work if one cluster is running Cloudera Manager 4.5.x and one is running Cloudera Manager 4.6 or later. For HDFS replication: 1. On the hosts in the target cluster, ensure that the krb5.conf file on each host has the following information: The kdc information for the source cluster's Kerberos realm. Domain/host to realm mapping for the source cluster NameNode hosts. 2. On the target cluster, through Cloudera Manager, add the realm of the source cluster to the Trusted Kerberos Realms configuration property. a. Go the the HDFS service page and from the Configuration tab select View and Edit. b. In the search field type "Trusted Kerberos" to find the Trusted Kerberos Realms property. c. Enter the source cluster realm and save your changes. 3. It is recommended that you restart all the affected services. However, you MUST restart the JobTracker to enable it to pick up the new Trusted Kerberos Realm settings. Failure to restart the JobTracker prior to the first replication attempt may cause the JobTracker to fail. For Hive replication: 1. Perform the steps described above on the target cluster, including restarting the JobTracker. 2. On the hosts in the source cluster, ensure that the krb5.conf file on each host has the following information: The kdc information for the target cluster's Kerberos realm. Domain/host to realm mapping for the target cluster NameNode hosts. 3. On the source cluster, through Cloudera Manager, add the realm of the target cluster to the Trusted Kerberos Realms configuration property. a. Go the the HDFS service page and from the Configuration tab select View and Edit. b. In the search field type "Trusted Kerberos" to find the Trusted Kerberos Realms property. c. Enter the target cluster realm and save your changes. 4. It is not necessary to restart any services on the source cluster. Cloudera Backup and Disaster Recovery 15