OSG Hadoop is packaged into rpms for SL4, SL5 by Caltech BeStMan, gridftp backend



Similar documents
Michael Thomas, Dorian Kcira California Institute of Technology. CMS Offline & Computing Week

Investigation of storage options for scientific computing on Grid and Cloud facilities

Hadoop Architecture. Part 1

Investigation of storage options for scientific computing on Grid and Cloud facilities

Next Generation Tier 1 Storage

Roadmap for Applying Hadoop Distributed File System in Scientific Grid Computing

Use of Hadoop File System for Nuclear Physics Analyses in STAR

Optimize the execution of local physics analysis workflows using Hadoop

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute

CERN Cloud Storage Evaluation Geoffray Adde, Dirk Duellmann, Maitane Zotes CERN IT

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007

Preview of a Novel Architecture for Large Scale Storage

HADOOP, a newly emerged Java-based software framework, Hadoop Distributed File System for the Grid

Welcome to the unit of Hadoop Fundamentals on Hadoop architecture. I will begin with a terminology review and then cover the major components

Testing of several distributed file-system (HadoopFS, CEPH and GlusterFS) for supporting the HEP experiments analisys. Giacinto DONVITO INFN-Bari

Hadoop: Embracing future hardware

Hadoop s Entry into the Traditional Analytical DBMS Market. Daniel Abadi Yale University August 3 rd, 2010

Storage strategy and cloud storage evaluations at CERN Dirk Duellmann, CERN IT

GraySort and MinuteSort at Yahoo on Hadoop 0.23

Analisi di un servizio SRM: StoRM

GraySort on Apache Spark by Databricks

Hadoop Distributed File System. T Seminar On Multimedia Eero Kurkela

Benchmarking Hadoop & HBase on Violin

Integration of Virtualized Workernodes in Batch Queueing Systems The ViBatch Concept

CMS Tier-3 cluster at NISER. Dr. Tania Moulik

KIT Site Report. Andreas Petzold. STEINBUCH CENTRE FOR COMPUTING - SCC

Detailed Outline of Hadoop. Brian Bockelman

BookKeeper. Flavio Junqueira Yahoo! Research, Barcelona. Hadoop in China 2011

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.


Performance Analysis of Mixed Distributed Filesystem Workloads

Panasas at the RCF. Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory. Robert Petkus Panasas at the RCF

Enterprise Edition. Hardware Requirements

Case Study : 3 different hadoop cluster deployments

Architecting for the next generation of Big Data Hortonworks HDP 2.0 on Red Hat Enterprise Linux 6 with OpenJDK 7

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. dcache Introduction

Running a typical ROOT HEP analysis on Hadoop/MapReduce. Stefano Alberto Russo Michele Pinamonti Marina Cobal

CSE-E5430 Scalable Cloud Computing Lecture 2

KIT Site Report. Andreas Petzold. STEINBUCH CENTRE FOR COMPUTING - SCC

Large Scale Storage. Orlando Richards, Information Services LCFG Users Day, University of Edinburgh 18 th January 2013

Scientific Storage at FNAL. Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015

NoSQL Performance Test In-Memory Performance Comparison of SequoiaDB, Cassandra, and MongoDB

Distributed File System. MCSN N. Tonellotto Complements of Distributed Enabling Platforms

BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB

Weekly Report. Hadoop Introduction. submitted By Anurag Sharma. Department of Computer Science and Engineering. Indian Institute of Technology Bombay

Design and Evolution of the Apache Hadoop File System(HDFS)

Lessons learned from parallel file system operation

Storage Architectures for Big Data in the Cloud

Hadoop IST 734 SS CHUNG

Chapter 7. Using Hadoop Cluster and MapReduce

Unstructured Data Accelerator (UDA) Author: Motti Beck, Mellanox Technologies Date: March 27, 2012

Maurice Askinazi Ofer Rind Tony Wong. Cornell Nov. 2, 2010 Storage at BNL

Virtuoso and Database Scalability

The Top 10 7 Hadoop Patterns and Anti-patterns. Alex

Take An Internal Look at Hadoop. Hairong Kuang Grid Team, Yahoo! Inc

Enabling High performance Big Data platform with RDMA

Performance and scalability of a large OLTP workload

Introduction to HDFS. Prasanth Kothuri, CERN

XTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines. A.Zydroń 18 April Page 1 of 12

Low-cost

HDFS Users Guide. Table of contents

Cost-Effective Business Intelligence with Red Hat and Open Source

Comparison of the Frontier Distributed Database Caching System with NoSQL Databases

How To Run Apa Hadoop 1.0 On Vsphere Tmt On A Hyperconverged Network On A Virtualized Cluster On A Vspplace Tmter (Vmware) Vspheon Tm (

GeoGrid Project and Experiences with Hadoop

InfoScale Storage & Media Server Workloads

PARALLELS CLOUD STORAGE

Oracle Database Scalability in VMware ESX VMware ESX 3.5

Building & Optimizing Enterprise-class Hadoop with Open Architectures Prem Jain NetApp

Open Cloud System. (Integration of Eucalyptus, Hadoop and AppScale into deployment of University Private Cloud)

Accelerating and Simplifying Apache

Understanding Hadoop Performance on Lustre

THE HADOOP DISTRIBUTED FILE SYSTEM

High Performance Data-Transfers in Grid Environment using GridFTP over InfiniBand

Leveraging BlobSeer to boost up the deployment and execution of Hadoop applications in Nimbus cloud environments on Grid 5000

Building a Private Cloud with Eucalyptus

Apache Hadoop Cluster Configuration Guide

HDFS Space Consolidation

Virtualization Performance on SGI UV 2000 using Red Hat Enterprise Linux 6.3 KVM

Hadoop Distributed File System (HDFS) Overview

Apache Hadoop. Alexandru Costan

VMware Best Practice and Integration Guide

Can High-Performance Interconnects Benefit Memcached and Hadoop?

The Simple High Available Linux File Server. Schlomo Schapiro Principal Consultant Leitung Virtualisierung & Open Source

Transcription:

Hadoop on HEPiX storage test bed at FZK Artem Trunov Karlsruhe Institute of Technology Karlsruhe, Germany KIT The cooperation of Forschungszentrum Karlsruhe GmbH und Universität Karlsruhe (TH) www.kit.edu

Motivation Hadoop is a distributed file system with map/reduce framework, designed to run on commodity cluster and make use of their local hard drives Has potential for high parallel performance, scalable with the number of nodes Extremely fault tolerant against loss of cluster nodes Already in use at a number of OSG site in production Packaged in OSG, supported. Reference installation at FNAL OSG Hadoop is packaged into rpms for SL4, SL5 by Caltech BeStMan, gridftp backend There is a native ROOT I/O plugin for Hadoop Exists as a patch, by Brian Bockelman HEPiX storage WG has a test bed and reference data for performance comparison 2 25.09.2009

2x4G FC Test bed at FZK and previous tests IBM 3550 8 core Xeon DDN 9550 8Tiers ( 8 data disks each) E4 8 core Xeon 16 internal disks with RAID5 or RAID6 10GB Best Results with DDN disks Admin nodes (dcache, AFS) 10 Worker nodes AFS/XFS 73 MB/sec 116 MB/sec 116 MB/sec 113 MB/sec NATIVE 145647 evs 229958 evs 225194 evs 214410 evs GPFS 171 MB/sec 307 MB/sec 398 MB/sec 439 MB/sec 168381 evs 297807 evs 380273 evs 420741 evs AFS/LU 134 MB/sec 251 MB/sec 341 MB/sec 394 MB/sec VIA VICE 161485 evs 297370 evs 401725 evs 445878 evs LUSTRE 146 MB/sec 256 MB/sec 358 MB/sec 399 MB/sec NATIVE 174865 evs 308427 evs 423087 evs 470939 evs Basic setup of tests with DDN (Andrei Maslennikov) Server 8 cores E5335 @ 2.00GHz RAM: 16 GB FC: 2 x Qlogic 2462 dual 4Gb Network: Quad GE card in ALB bonding mode=6. miimon=100 measured 450 MB/sec memory-memory in both directions OS: SuSE SLES10 SP2 64 bit Kernel: 2.6.16.60-0.27_lustre.1.6.6-smp Disk system: DDN 9550 Direct attach with 2 x 4 Gb FC 4 Luns each composed of two tiers 16 data disks per Lun Lun block size - 4MB Cache: disabled Clients 10 x 8 cores E5430 @ 2.66GHz 16 GB RAM OS: RHEL4 64 bit Kernel: 2.6.9-78.0.13.ELsmp non-modified + Lustre modules Test jobs CMSSW 1_6_6 2,4,6,8 jobs per node, 10 nodes 40 minutes run 3 25.09.2009

Testbed for Hadoop IBM 3550 8 core Xeon DDN 9550 8Tiers ( 8 data disks each) E4 8 core Xeon 16 internal disks with RAID5 or RAID6 10GB 10 Worker nodes External servers and storage are not used Make use of worker nodes two internal 250 GB SATA hard drives On each allocate ~200MB partitions, format with ext3 Total of ~4TB of free space Hadoop setup Version 0.19 SL4 x86_64 from Caltech repo 10 datanodes + 1 namenode Test Jobs were run on 9 data nodes and the name node Fuse interface to HDFS, mounted on each node Slight complication: due to high sensitivity of Hadoop to performance of hard drives, had to reject one data node and use one of admin nodes as data nodes This had little impact on the test result. Hadoop settings: block size 64M replication factor 1 java heap size 512MB fuse settings: 'ro,rdbuffer=65536,allow_other network settings (pretty standard): net.core.netdev_max_backlog = 30000 net.core.rmem_max = 16777216 net.core.wmem_max = 16777216 net.ipv4.tcp_rmem = 4096 87380 16777216 net.ipv4.tcp_wmem = 4096 65536 16777216 block device settings: echo 32767 > /sys/block/sd${dr}/queue/max_sectors_kb echo 16384 > /sys/block/sd${dr}/queue/read_ahead_kb echo 32 > /sys/block/sd${dr}/queue/iosched/quantum echo 128 > /sys/block/sd${dr}/queue/nr_requests Vary during the tests block devise settings: read_ahead_kb: from 1 to 32 MB nr_requests from 32 to 512 fuse read ahead buffer: rdbuffer: from 16k to 256k Optimum was found at the following: read_ahead_kb: 16 MB nr_requests: 128 fuse rdbuffer: 128k Measure Total event count for 40 minute test jobs Read rate from disk 4 25.09.2009

Best results 20 threads 40 threads 60 threads 80 threads Hadoop 116 MB/sec 218 MB/sec 270 MB/sec 369 MB/sec 161103 evs 298243 evs 412836 evs 481841 evs LUSTRE 146 MB/sec 256 MB/sec 358 MB/sec 399 MB/sec DDN DISK 174865 evs 308427 evs 423087 evs 470939 evs 600 Events rate 450 Throughput 500 400 350 events, 10^3 400 300 200 Hadoop Lustre DDN disk MB/s 300 250 200 150 100 100 50 0 0 20 40 60 80 jobs 20 40 60 80 jobs 5 25.09.2009

Discussion Hadoop in this test bed is close to Lustre and outperforms it in the maximum load test. 8 jobs on 8 core machine is a standard batch setup Some other considerations are also taken into account when selecting storage Cost of administration, ease of deployment, capacity scaling, support for large name spaces Hard to think it s a main HEP T1 storage solution, or it needs a lot of additional testing and careful deployment. As T2/T3 storage should be very interesting to WLCG sites Cost and maintenance factor is very favorable to small sites 6 25.09.2009

Future plans HEPiX test bed at FZK moved to a dedicated rack, RH5 Hadoop 0.20 or 0.21, all 64bit Newer CMS and Atlas software Check performance with replication factor >1 Check with various chunk sizes Test on a high end storage? 7 25.09.2009

Acknowledgments Andrei Maslennikov test suite setup, valuable guidance and comments Brian Bockelman, Terrence Martin (OSG) Hadoop wiki, tuning tips FZK team Jos van Wezel, Manfred Alef, Bruno Hoeft, Marco Stanossek, Bernhard Verstege 8 25.09.2009