NetApp Solutions for Hadoop Reference Architecture



Similar documents
Successfully Deploying Alternative Storage Architectures for Hadoop Gus Horn Iyer Venkatesan NetApp

Building & Optimizing Enterprise-class Hadoop with Open Architectures Prem Jain NetApp

NetApp Open Solution for Hadoop Solutions Guide

Hadoop Architecture. Part 1

How To Run Apa Hadoop 1.0 On Vsphere Tmt On A Hyperconverged Network On A Virtualized Cluster On A Vspplace Tmter (Vmware) Vspheon Tm (

Apache Hadoop Cluster Configuration Guide

Welcome to the unit of Hadoop Fundamentals on Hadoop architecture. I will begin with a terminology review and then cover the major components

Dell Reference Configuration for Hortonworks Data Platform

ADVANCED NETWORK CONFIGURATION GUIDE

Open source software framework designed for storage and processing of large scale data on clusters of commodity hardware

Redpaper. Big Data Networked Storage Solution for Hadoop. Front cover. ibm.com/redbooks. Learn about the IBM big analytics storage solutions

Platfora Big Data Analytics

HP reference configuration for entry-level SAS Grid Manager solutions

Red Hat Enterprise linux 5 Continuous Availability

Accelerating and Simplifying Apache

HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW

SAN Conceptual and Design Basics

Big Data Storage Options for Hadoop Sam Fineberg, HP Storage

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

HP Reference Architecture for Hortonworks Data Platform on HP ProLiant SL4540 Gen8 Server

UCS M-Series Modular Servers

IBM System Storage DS5020 Express

Brocade Solution for EMC VSPEX Server Virtualization

How To Write An Article On An Hp Appsystem For Spera Hana

FlexPod Big Data Solutions for Hadoop

MapR Enterprise Edition & Enterprise Database Edition

CDH AND BUSINESS CONTINUITY:

Scala Storage Scale-Out Clustered Storage White Paper

ENABLING GLOBAL HADOOP WITH EMC ELASTIC CLOUD STORAGE

High Availability with Windows Server 2012 Release Candidate

Increasing Hadoop Performance with SanDisk Solid State Drives (SSDs)

Elasticsearch on Cisco Unified Computing System: Optimizing your UCS infrastructure for Elasticsearch s analytics software stack

Intel RAID SSD Cache Controller RCS25ZB040

FOR SERVERS 2.2: FEATURE matrix

Video Surveillance Storage and Verint Nextiva NetApp Video Surveillance Storage Solution

WHITEPAPER: Understanding Pillar Axiom Data Protection Options

VTrak SATA RAID Storage System

Cisco for SAP HANA Scale-Out Solution on Cisco UCS with NetApp Storage

MESOS CB220. Cluster-in-a-Box. Network Storage Appliance. A Simple and Smart Way to Converged Storage with QCT MESOS CB220

RED HAT ENTERPRISE VIRTUALIZATION FOR SERVERS: COMPETITIVE FEATURES

IBM System x reference architecture for Hadoop: MapR

Pivot3 Desktop Virtualization Appliances. vstac VDI Technology Overview

VIDEO SURVEILLANCE WITH SURVEILLUS VMS AND EMC ISILON STORAGE ARRAYS

OPTIMIZING SERVER VIRTUALIZATION

The Benefits of Virtualizing

Cisco, Citrix, Microsoft, and NetApp Deliver Simplified High-Performance Infrastructure for Virtual Desktops

Parallels Cloud Storage

CSE-E5430 Scalable Cloud Computing Lecture 2

Big Data With Hadoop

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA

Storage Architectures for Big Data in the Cloud

Enterprise Disk Storage Subsystem Directions

Oracle Database Scalability in VMware ESX VMware ESX 3.5

The Future of Computing Cisco Unified Computing System. Markus Kunstmann Channels Systems Engineer

Best Practice of Server Virtualization Using Qsan SAN Storage System. F300Q / F400Q / F600Q Series P300Q / P400Q / P500Q / P600Q Series

HadoopTM Analytics DDN

Cisco Prime Home 5.0 Minimum System Requirements (Standalone and High Availability)

EMC Backup and Recovery for Microsoft SQL Server

Hortonworks Data Platform Reference Architecture

CloudSpeed SATA SSDs Support Faster Hadoop Performance and TCO Savings

Post-production Video Editing Solution Guide with Microsoft SMB 3 File Serving AssuredSAN 4000

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Enabling High performance Big Data platform with RDMA

Get More Scalability and Flexibility for Big Data

Best Practices for Data Sharing in a Grid Distributed SAS Environment. Updated July 2010

Can Storage Fix Hadoop

A Platform Built for Server Virtualization: Cisco Unified Computing System

Networking in the Hadoop Cluster

The Advantages of Multi-Port Network Adapters in an SWsoft Virtual Environment

Dell High Availability Solutions Guide for Microsoft Hyper-V

June Blade.org 2009 ALL RIGHTS RESERVED

Parallels. Clustering in Virtuozzo-Based Systems

Real-time Protection for Hyper-V

The functionality and advantages of a high-availability file server system

Block based, file-based, combination. Component based, solution based

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database

Apache Hadoop Storage Provisioning Using VMware vsphere Big Data Extensions TECHNICAL WHITE PAPER

IBM BladeCenter H with Cisco VFrame Software A Comparison with HP Virtual Connect

Processing of Hadoop using Highly Available NameNode

SUN HARDWARE FROM ORACLE: PRICING FOR EDUCATION

Windows Server 2008 R2 Hyper-V Live Migration

Nutanix Tech Note. Failure Analysis All Rights Reserved, Nutanix Corporation

Open-E Data Storage Software and Intel Modular Server a certified virtualization solution

Deploying Microsoft SQL Server 2008 R2 with Logical Partitioning on the Hitachi Virtual Storage Platform with Hitachi Dynamic Tiering

Deploying a 48,000-user Exchange Server 2010 Environment with Hitachi Compute Blade 2000 and Hitachi Adaptable Modular Storage 2500

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

SUN ORACLE DATABASE MACHINE

Dell PowerVault MD Family. Modular storage. The Dell PowerVault MD storage family

Server and Storage Virtualization with IP Storage. David Dale, NetApp

ORACLE BIG DATA APPLIANCE X3-2

Private cloud computing advances

August Transforming your Information Infrastructure with IBM s Storage Cloud Solution

Solution Brief Network Design Considerations to Enable the Benefits of Flash Storage

Introduction to Hadoop. New York Oracle User Group Vikas Sawhney

NetApp E-Series Storage Systems

Hadoop Distributed File System. Jordan Prosch, Matt Kipps

Protect SQL Server 2012 AlwaysOn Availability Group with Hitachi Application Protector

HP recommended configuration for Microsoft Exchange Server 2010: HP LeftHand P4000 SAN

Transcription:

White Paper NetApp Solutions for Hadoop Reference Architecture Gus Horn, Iyer Venkatesan, NetApp April 2014 WP-7196 Abstract Today s businesses need to store, control, and analyze the unprecedented complexity, speed, or volume of data they are capturing. Hadoop has gained popularity by handling these large and diverse data types. However, there are technical challenges that enterprises face in deploying Hadoop, specifically in the areas of cluster availability, operations, and scale. NetApp has developed reference architectures with leading Hadoop vendors to deliver a solution that overcomes some of these challenges so that business can ingest, store, and manage this big data with higher reliability and scalability, and with less time spent on operations and maintenance. This white paper discusses a flexible, validated, enterprise-class Hadoop architecture based on NetApp storage in a managed or external direct-attached storage (DAS) configuration with any compatible distribution of Apache Hadoop. Included are guidelines and best practices on how to build and configure modular Hadoop systems, including recommendations on servers, file systems, and networking topology. This enables groups or organizations to gain insights for competitive advantage, in-depth customer knowledge, and many other business benefits.

TABLE OF CONTENTS 1 Introduction... 4 1.1 Benefits...4 1.2 Brief Hadoop Overview...4 1.3 Basic Hadoop Architecture...5 2 Hadoop Challenges... 5 2.1 HDFS Replication Trade-Offs...6 2.2 When Drives Fail...6 3 NetApp Solutions for Hadoop Architecture... 7 3.1 High-Level Architecture...7 3.2 Technical Advantages of the NetApp Solutions for Hadoop...7 3.3 Validated Base Configuration...7 3.4 Other Supported Configurations...8 4 Solution Architecture Details... 8 4.1 Network Architecture...9 4.2 Storage Architecture... 10 4.3 NameNode Metadata Protection... 11 4.4 Rack Awareness Implementation... 13 5 Key Components of Validated Configuration... 13 6 Other Supported Features, Products, and Protocols... 14 7 Conclusion... 14 Additional Information... 15 Version History... 15 LIST OF TABLES Table 1) Recommended products for the validated reference architecture.... 13 Table 2) Components supported by NetApp and its partners.... 14 LIST OF FIGURES Figure 1) Basic Hadoop architecture....5 Figure 2) Architecture diagram for NetApp solutions for Hadoop....7 Figure 3) Network architecture.... 10 Figure 4) E-Series configured with four hosts per array.... 11 2 NetApp Solutions for Hadoop Reference Architecture

Figure 5) E-Series configured with eight hosts per array.... 11 Figure 6) NameNode backup configuration using NFS.... 12 Figure 7) NameNode HA configuration using NFS storage.... 12 3 NetApp Solutions for Hadoop Reference Architecture

1 Introduction At present, the term big data is well known and is having a significant impact on businesses that need to store, manage, and analyze the complex and large amounts of data that they are capturing. Hadoop and its growing ecosystem of products have enabled many businesses to handle this volume and these diverse types of data so they can start gaining valuable insights from this data. However, Hadoop poses some challenges in enterprise data centers, centering on operations, availability, and implementation. Validated reference architectures for Hadoop are an effective way to overcome some of these challenges in implementing Hadoop. NetApp Solutions for Hadoop validate NetApp storage with the leading Hadoop distributions in a flexible and open environment. Recommendations for servers and networking are also included so a fully configured cluster can be built. Asgroups begin to move Hadoop pilots or proofs of concept into production; they typically expect a certain degree of scalability, manageability, and availability as well as consistent, deterministic performance. The NetApp reference architecture is designed for HA scale out and will provide SLAs with deterministic performance. In short it is an Open enterprise grade scale out solution. 1.1 Benefits Key benefits of the NetApp solutions for Hadoop are: Enterprise-class implementation of Hadoop with lower cluster downtime, higher data availability, and linear scalability Fast time to value with the validated, presized configurations that are ready to deploy, thereby reducing risks that are traditionally associated with Hadoop High storage efficiency and lower operational expenses for Hadoop Open solution built with NetApp and best of breed products from partners, providing choice without vendor lock-in or performance compromise NetApp has partnered with major Hadoop distribution providers and other analytic platform vendors to deliver reliable and efficient storage for Hadoop solutions. These validated solutions when implemented in a Hadoop cluster provide the capabilities for ingesting, storing, and managing large datasets with high efficiency, reliability and performance. 1.2 Brief Hadoop Overview Readers familiar with Hadoop can skip this section and can either scroll to section 1.3, Basic Hadoop Architecture, or section 2, Hadoop Challenges. Hadoop is an open source analytical framework and an ecosystem of products and technologies that contain, among other things, databases, scripting languages, and management tools. Hadoop is designed to run on a large number of parallel machines. When you want to load all of your organization s data into Hadoop, the software makes three copies of the data, then breaks that data into pieces and spreads it across different servers that are set up to run in parallel. There is no one place where you can go to manage all of your data; Hadoop NameNode keeps track of where the data resides. And because there are multiple copy stores, data stored on a server that goes offline or dies can be automatically replicated from a known good copy. The two main components of Hadoop are MapReduce, which is the framework for parallel processing. The other one is Hadoop Distributed File System (HDFS), which is the distributed file system that provides petabyte-size storage and data management. 4 NetApp Solutions for Hadoop Reference Architecture

1.3 Basic Hadoop Architecture The basic Hadoop architecture is built around commodity servers and internal DAS storage and networking. Following are the basic requirements of this architecture. A conventional Hadoop cluster consists of a few basic components: NameNode/active NameNode. This manages the HDFS namespace. It runs all of the services needed to manage the HDFS data storage and MapReduce tasks in the cluster, including data and task distribution and tracking. This is sometimes called the NameNode or active NameNode. Checkpoint node. It is a secondary NameNode that manages the on-disk representation of the NameNode metadata. In active/standby HA, this node runs a second NameNode process, usually called the standby NameNode process. In quorum-based HA mode, this node runs the second journal node. JobTracker node. It manages all of the jobs submitted to the Hadoop cluster and facilitates job and task scheduling. DataNodes. These are slave nodes that provide both TaskTracker functionality and DataNode functionality. These nodes perform all of the real processing work done by the cluster. As DataNodes, they store all of the data blocks that make up the file system, and they serve I/O requests. As TaskTrackers, they perform the job tasks assigned to them by the JobTracker. Figure 1) Basic Hadoop architecture. 2 Hadoop Challenges Hadoop was incubated and developed at Internet companies including Google and Yahoo. Their goal was to practically analyze billions of files and petabytes of data using thousands of low-cost servers. They achieved this in an environment where development and operations coexist. As Hadoop gains traction in enterprise data centers, operational requirements change. A cluster with a few hundred nodes suffices for most businesses; enterprise customers do not have active developers working on Hadoop. Enterprises are driven by very strict SLAs. Any solution put into such environments must meet enterprise operational requirements and data center standards and SLAs. Following are some common challenges that are faced while implementing Hadoop in an enterprise: Operational 5 NetApp Solutions for Hadoop Reference Architecture

Limited flexibility when scaling, cannot independently add servers or storage (for example: when adding servers, storage has to be added as well). Not highly available. Jobs get interrupted during failures, and recovering these jobs from drive failures can be time consuming. Might not meet enterprise-level SLAs. Efficiency Three copies of data (also called triple mirroring). Network congestion and overdrain of system resources. Poor job efficiency due to storage fragmentation over a period of time. Management Requires skilled resources. Flexibility of changing architectures. Data center footprint. 2.1 HDFS Replication Trade-Offs Although HDFS is distributed, feature rich, and flexible, there are both throughput and operational costs associated with server-based replication for ingest, redistribution of data following recovery of a failed DataNode, and capacity utilization. A standard Hadoop install implements at least three copies of every piece of data stored. Costs of triple replication include: For each usable terabyte (TB) of data, 3TB of raw capacity are required. Server-based triple replication creates a significant load on the servers themselves. At the beginning of an ingest task, before any data blocks are actually written to a DataNode, the NameNode creates a list of DataNodes where all three of the first block replicas are written, forming a pipeline and allocating blocks to those nodes. That DataNode forwards the same block to the next node in the pipeline and the preceding procedure is repeated. Finally, this second DataNode sends the block to the third DataNode, and the write to storage follows the same I/O path as the former writes. This results in a significant load on server resources. Server-based replication also creates a significant load on the network. For ingest, three network trips are required to accomplish a single block write. All writes compete for and consume the same network, CPU, memory, and storage resources allocated to process the Hadoop analysis tasks. The NetApp solution can provide better reliability with 1/3 reduction in replication that is the NetApp architecture needs only two copies of data with our strictly shared nothing architecture. Anything done to reduce the consumption of server and network resources during the loading of data increases cluster efficiency. 2.2 When Drives Fail The most fault prone component in the Hadoop architecture are the disk drives, it is simply a matter of time when a drive will fail, and the larger the cluster, the higher the possibility that a drive failure will occur. If one or more drives fail, all tasks accessing data on that drive will fail and be reassigned to other DataNodes in the cluster. This results in a severe degradation of completion time for running jobs. In addition, the drive must be physically replaced, mounted, partitioned, formatted, and reloaded with data from the nearby copies before HDFS can make full use of that drive. Repopulating the new disk with data can be accomplished by using the Hadoop balancer, which redistributes HDFS data across all DataNodes by moving blocks from overutilized to underutilized DataNodes. For every full 3TB SATA disk that fails, several hours are needed for rebalancing the distribution of blocks following replacement of that disk, depending on the tolerance for network resource consumption in the 6 NetApp Solutions for Hadoop Reference Architecture

cluster. The trade-off is between rebalancing time, network utilization, and disk I/O bandwidth utilization. Remember that higher resource utilization affects MapReduce job performance, especially during data ingest operations. In some cases, rebalancing might also require some manual movement of blocks between disk drives. 3 NetApp Solutions for Hadoop Architecture 3.1 High-Level Architecture Figure 2 is the general architecture diagram for NetApp solutions for Hadoop reference designs. Figure 2) Architecture diagram for NetApp solutions for Hadoop. 3.2 Technical Advantages of the NetApp Solutions for Hadoop Lower cluster downtime with transparent RAID operations and faster rebuild of failed media Less storage needed with Hadoop replication count of two Enhanced robustness of HDFS through reliable NFS server-based metadata protection Disk drives can be replaced while the DataNode is running; this operation cannot be done with the internal DAS-based DataNode configurations Performance for the next generation of DataNodes (SMP) and networks (10GbE) Highly reliable and manageable Hadoop platform Server, storage, and networking hardware preconfigured and validated with Apache-compatible Hadoop distributions 3.3 Validated Base Configuration The NetApp solution for Hadoop has a validated base reference architecture that constitutes the following hardware, software, and networking. This recommended configuration has been tested, validated, and optimized for running a Hadoop cluster. 7 NetApp Solutions for Hadoop Reference Architecture

Servers Supports all of the Intel Xeon E5-2400 series processors Up to two 2.4GHz (6-core) or 2.3GHz (8-core) processors Memory: 12 DIMMs (up to 192GB), up to 1600Mhz Hard disks: 2 x 3.5" HS HDDs Network: 2x1Gb/sec Ethernet ports, 10GbE network port LSI 9207-8e SAS HBA or compatible ISCSI/FC HBA needed SFF-8088 external mini-sas cable (for maximum signal integrity, cable should not exceed 5m) Storage NetApp E5460 storage array 1 DE6600 4U chassis with rail kit and power cords 2 E-5460 series controllers, each configured with dual SAS ports, and SAS disk drives 60 3TB, 7.2K-RPM, 3.5" near-line NL-SAS disks NetApp Engenio operating system (EOS CFW 10.80 or later release) Base SANtricity Storage Manager Turbo feature enabled Optional: NetApp FAS2220 NFS storage system NetApp FAS2220 is recommended if customers require high-availability (HA) protection of the NameNode or need simplification in managing Hadoop clusters. Network 10GbE nonblocking network switch GbE network switch Software Red Hat Enterprise Linux 6.2 or later One of the following Hadoop distributions: Cloudera Distribution 4.1.3 or later; this includes Cloudera Manager and Cloudera Enterprise Hortonworks Data Platform HDP 1.3 or later NetApp SANtricity 10.84 or later storage management (bundled with NetApp E-Series storage arrays) 3.4 Other Supported Configurations Apart from the validated based configuration, this solution offers flexibility in choosing other models and versions of each component. You might be able to choose a lower or higher version of storage arrays based on your requirements. You might also consider a high-availability pair of network switches. Some users prefer to go after a bonded GbE NIC pair instead of a 10GbE network. Storage to DataNode connectivity also offers multiple options from direct connect SAS, FCP, or ISCSI. Section 6, Other Supported Features, Products, and Protocols, of this report provides a detailed list of such available options. 4 Solution Architecture Details In Figure 2, there are two pieces of NetApp storage in the reference design. The first is the NetApp E- Series storage array, which provides the data storage services with one E-Series dedicated to four 8 NetApp Solutions for Hadoop Reference Architecture

Hadoop DataNodes. The second optional piece is the FAS2220, which provides enterprise-grade protection of the NameNode metadata, if that is needed for the cluster. Each DataNode has its own dedicated, nonshared set of disks. This exactly follows the way in which Hadoop DataNodes configure capacity when these drives are physically colocated in the chassis, as in the internal JBOD configurations. In practical terms, Linux is not able to distinguish between a set of logical unit numbers (LUNs) presented by the LSI 1068 controller chip on the HBA and a set of disks presented by either an embedded SAS or a SATA controller. There is no logical difference. However, there are differences in performance and reliability. This set of disks is configured as two blocks, each using one of two parity-based protection schemes. Eight volume groups are configured in the E-Series, and they are configured so that each DataNode sees only its set of two private blocks. This design improves just a bunch of disks (JBOD) DAS media for four DataNodes by offering better performance, reliability, data availability, uptime, and manageability. Each block is a RAID-protected volume group that contains a single virtual logical volume, which can be created and exported to a given DataNode as a LUN. These LUNs appear as physical disks in Red Hat Enterprise Linux (RHEL). They can be partitioned, and file systems can be created and mounted to store HDFS blocks. (HDFS blocks can be 64MB, 128MB, 256MB, or even 512MB in size.) 4.1 Network Architecture The network architecture of a Hadoop cluster is critical. With a default replication count of three, more than two-thirds of all the I/O traffic must traverse the network during ingest, and perhaps a third of all MapReduce I/O during processing runs could originate from other nodes. The NetApp E-Series storage modules provide a level of performance that is significantly better than that provided by JBOD SATA. The storage array offers a performance platform based on four 6Gb/sec SAS ports, supported by proven caching and best-in-class hardware RAID. The solution uses two network backplanes. The core of this solution is the HDFS network, which is a 10GbE network backplane served by an extremely high-performance network 10/40GbE switch solution for a top-of-rack (ToR) switch. The other network is a GbE network, which caters to the administration network and also connects the NameNode, JobTracker, and NetApp FAS storage systems (if a FAS is used for NameNode HA). On the 10GbE network, all components should be configured for jumbo frames. This requirement also applies to the switch ports to which the components are connected. Any ports on the 10GbE switches that are configured for GbE connections (such as FAS NFS and uplinks) should not have jumbo frames enabled. 9 NetApp Solutions for Hadoop Reference Architecture

Figure 3) Network architecture. Recommended Network Topology Data network (10GbE) Private interconnect between all DataNodes 10GbE nonblocking switch 40GbE uplink between the racks Dedicated nonroutable VLAN Primary purpose: Data ingest and movement within cluster Use dedicated separate VLAN for iscsi network, if opted Management network (GbE) System administration network Publically routable subnet NameNode and JobTracker also use GbE All nodes have a public GbE interface for administration and data transfer purpose E-Series storage systems also need two GbE ports for management purpose only 4.2 Storage Architecture Each NetApp E-Series storage array is configured as eight independent RAID groups of seven disks that can be set up in a RAID 5 (2 x 6+1) configuration. This consumes 56 (8 x 7) disks. The remaining four disks are global hot spares. If customers deploy all of their files with a replication count of two, then using a single-parity drive over six spindles provides a good balance between storage space utilization and 10 NetApp Solutions for Hadoop Reference Architecture

RAID protection. As a result, there are constantly two forms of protection when the E-Series storage arrays are running with files set to a replication count of two. Users have an option to select two available fans in ratios based on their requirements. The more common one is having one E-Series storage server with 4 DataNodes, as shown in Figure 3. For workloads where lower storage density is enough, a lighter configuration of single E-Series storage serving 8 DataNodes may be considered. This is depicted in Figure 5. Figure 4) E-Series configured with four hosts per array. Figure 5) E-Series configured with eight hosts per array. 4.3 NameNode Metadata Protection Prior to Hadoop 2.0, the NameNode was a single point of failure (SPOF) in an HDFS cluster. Each cluster had a single NameNode, and if that machine or process became unavailable, the cluster as a whole would be unavailable until the NameNode was either restarted or brought up on a separate machine. There are two ways to mitigate such a risk of NameNode failure and the data loss: backup/restore if using Hadoop 1.x or using the HDFS high-availability feature in Hadoop 2.0 or later versions. Option 1: Backup/Restore Method If you are planning to go with the traditional Hadoop method, then the NetApp solution offers a unique feature of NameNode metadata protection. This protection is offered by introducing a NetApp FAS NFS storage system. The NameNode is configured to keep a copy of its HDFS metadata (FSImage) in the NFS mounted storage. The same NFS mounted storage is also made available to the secondary NameNode. In the case of a NameNode failure, the critical metadata is still safe on the NetApp FAS storage system. At this point, we can recover the metadata and restart services on the secondary NameNode to get the HDFS back and running promptly. 11 NetApp Solutions for Hadoop Reference Architecture

Figure 6) NameNode backup configuration using NFS. Option 2: HDFS High Availability Versions 2.0 and later of HDFS offer a high-availability NameNode feature to mitigate NameNode failures. The HDFS high-availability feature addresses the preceding problems by providing the option of running two redundant NameNodes in the same cluster in an active-passive configuration with a hot standby. This allows a fast failover to a new NameNode in the case that a machine crashes or a graceful administrator-initiated failover for the purpose of planned maintenance. In order for the standby node to keep its state synchronized with the active node, the current implementation requires that the two nodes both have access to a directory on a shared storage device. This is facilitated by a NetApp FAS NFS storage device in the solution, as stated in the following section. Figure 7) NameNode HA configuration using NFS storage. For more details about this configuration, visit http://hadoop.apache.org/docs/current2/hadoopyarn/hadoop-yarn-site/hdfshighavailabilitywithnfs.html. For either option, FAS is recommended because it helps manage the Hadoop cluster by storing boot images and binaries simplifying administration tasks. 12 NetApp Solutions for Hadoop Reference Architecture

4.4 Rack Awareness Implementation Implementing rack awareness in our Hadoop solution offers important benefits. Using this feature allows us to configure Hadoop and know the topology of the network: Rack awareness enables Hadoop to maximize network utilization by favoring block transfers within a rack over transfers between racks. The JobTracker is able to optimize MapReduce job performance by assigning tasks to nodes that are closer to their data in terms of network topology. By default, during HDFS writes, the NameNode assigns the second and third block replicas to nodes in a different rack from the first replica. This provides data protection even against rack failure; however, this is possible only if Hadoop has been configured with knowledge of its rack configuration. By utilizing the rack awareness feature, we can also make sure that the second copy of each block always gets stored on a different E-Series-based DataNode. This offers tolerance to storage array and DataNode failures. Care must be taken to make sure that for each HDFS block, the corresponding replica is stored on another E-Series controller. In multirack clusters, providing the actual rack location of each DataNode will accomplish this. If all the DataNodes are located in the same rack, rack mapping must be extended to reflect the E-Series enclosure or E-Series controller used for DataNode storage. Basically, use the HDFS rack awareness configuration to prevent data loss in the unlikely event of storage controller failure. 5 Key Components of Validated Configuration Table 1 summarizes the recommended products for the validated reference architecture. Table 1) Recommended products for the validated reference architecture. Component Product or Solution Details Storage NetApp E-Series 5460 storage array FAS2220 (if using for NameNode protection) E5460: 60 3TB drives for 180TB capacity 4 hot spares 8 volume groups; 8 volumes RAID 5 volume groups, 7 disks each Servers Intel Xeon E5-2400 series processors Up to two 2.4GHz (6-core) or 2.3GHz (8- core) processors Memory: 12 DIMMs (up to 192GB), up to 1600Mhz Hard disks: 2 x 3.5" HS HDDs 2 x 1Gb/sec Ethernet ports, 10GbE network port LSI 9207-8e SAS HBA Networking 10GbE nonblocking network switch GbE network switch OS Red Hat Enterprise Linux Server 6.2 (x86_64) or later Hadoop typically requires a Linux distribution Hadoop distribution Cloudera Distribution for Hadoop Cloudera Enterprise Core 4.1.3 Cloudera Manager 4.1.3 13 NetApp Solutions for Hadoop Reference Architecture

Component Product or Solution Details Hortonworks Data Platform 1.3 Management software NetApp SANtricity 10.84 SANtricity software is bundled with E- Series storage arrays 6 Other Supported Features, Products, and Protocols Although NetApp recommends the validated configuration in Table 1, individual customer requirements might call for different components. The components that are supported by NetApp and its partners are listed in Table 2. Table 2) Components supported by NetApp and its partners. Component or Feature Other Supported Options Details Storage arrays E5412 E5424 E27xx E55xx EF5xx Other E5400 storage arrays are also supported, as are the E2700 and E5500 series. Disk drives and type 12, 24, 60 1TB, 2TB 12 and 24 drives are supported as well as 1TB, 2TB, and 4TB drives. Protocols Other Hadoop distributions Fibre Channel ISCSI InfiniBand MapR BigInsights Intel Any Apache-compatible Hadoop distribution is supported. 7 Conclusion With big data growing from its roots in Internet companies to becoming more established in data centers, businesses are turning to Hadoop to help them ingest, store, manage, and analyze all the various data they are collecting. Typically, organizations that are deploying or considering Hadoop use traditional server-based storage in an internal DAS configuration. However, this can come with challenges in data centers because internal DAS might be not as reliable for an organization s SLAs, harder to scale, and not as flexible. The NetApp Solutions for Hadoop reference design employs the external or managed DAS storage, which provides dedicated storage, with higher scalability and reliability. There is also architectural flexibility by decoupling compute nodes from storage to scale one without needing to scale the other, being able to choose any NetApp E-Series and being able to use most servers and networking technologies. There is also freedom to choose The NetApp solutions for Hadoop reference architectures are validated to help customers get started with Hadoop or deploy Hadoop in production so they can begin mining their data for insights. 14 NetApp Solutions for Hadoop Reference Architecture

Additional Information To learn about NetApp Hadoop solutions, visit http://www.netapp.com/hadoop. Additional information about the base reference design is discussed on this website, and it also provides information about FlexPod Select with Hadoop, which is the validated base reference architecture with Cisco servers and networking. There are two Cisco Validated Designs for FlexPod Select with Hadoop: FlexPod Select with Cloudera Enterprise FlexPod Select with Hortonworks Data Platform If you want to contact NetApp sales or a partner to either investigate NetApp solutions further or start a pilot or proof of concept, click here. More information on the Hadoop project, technologies, and ecosystem can be found at www.hadoop.apache.org. You can also contact the authors at iyerv@netapp.com or gustav.horn@netapp.com. Version History Version Date Document Version History Version 1.0 April 2014 Initial release Refer to the Interoperability Matrix Tool (IMT) on the NetApp Support site to validate that the exact product and feature versions described in this document are supported for your specific environment. The NetApp IMT defines the product components and versions that can be used to construct configurations that are supported by NetApp. Specific results depend on each customer's installation in accordance with published specifications. NetApp provides no representations or warranties regarding the accuracy, reliability, or serviceability of any information or recommendations provided in this publication, or with respect to any results that may be obtained by the use of the information or observance of any recommendations provided herein. The information in this document is distributed AS IS, and the use of this information or the implementation of any recommendations or techniques herein is a customer s responsibility and depends on the customer s ability to evaluate and integrate them into the customer s operational environment. This document and the information contained herein may be used solely in connection with the NetApp products discussed in this document. 2014 NetApp, Inc. All rights reserved. No portions of this document may be reproduced without prior written consent of NetApp, Inc. Specifications are subject to change without notice. NetApp, the NetApp logo, Go further, faster, FlexPod, and SANtricity are trademarks or registered trademarks of NetApp, Inc. in the United States and/or other countries. Intel and Xeon are registered 15 NetApp Solutions for trademarks Hadoop of Reference Intel Corporation. Architecture Linux is a registered trademark of Linus Torvalds. Cisco is a registered trademark of Cisco Systems, Inc. All other brands or products are trademarks or registered trademarks of their respective holders and should be treated as such. WP-7196-0414