CERN Cloud Storage Evaluation Geoffray Adde, Dirk Duellmann, Maitane Zotes CERN IT



Similar documents
Storage strategy and cloud storage evaluations at CERN Dirk Duellmann, CERN IT

Using S3 cloud storage with ROOT and CernVMFS. Maria Arsuaga-Rios Seppo Heikkila Dirk Duellmann Rene Meusel Jakob Blomer Ben Couturier

DSS. High performance storage pools for LHC. Data & Storage Services. Łukasz Janyst. on behalf of the CERN IT-DSS group

Next Generation Tier 1 Storage

DSS. Diskpool and cloud storage benchmarks used in IT-DSS. Data & Storage Services. Geoffray ADDE

CERNBox + EOS: Cloud Storage for Science

Analisi di un servizio SRM: StoRM

Alternative models to distribute VO specific software to WLCG sites: a prototype set up at PIC

Michael Thomas, Dorian Kcira California Institute of Technology. CMS Offline & Computing Week

Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle

High Performance Computing OpenStack Options. September 22, 2015

How swift is your Swift? Ning Zhang, OpenStack Engineer at Zmanda Chander Kant, CEO at Zmanda

Scientific Storage at FNAL. Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015

Panasas at the RCF. Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory. Robert Petkus Panasas at the RCF

OSG Hadoop is packaged into rpms for SL4, SL5 by Caltech BeStMan, gridftp backend

Development of Monitoring and Analysis Tools for the Huawei Cloud Storage

Maurice Askinazi Ofer Rind Tony Wong. Cornell Nov. 2, 2010 Storage at BNL

POSIX and Object Distributed Storage Systems

Quantum StorNext. Product Brief: Distributed LAN Client

Reference Design: Scalable Object Storage with Seagate Kinetic, Supermicro, and SwiftStack

ENABLING GLOBAL HADOOP WITH EMC ELASTIC CLOUD STORAGE

Scala Storage Scale-Out Clustered Storage White Paper

Big Science and Big Data Dirk Duellmann, CERN Apache Big Data Europe 28 Sep 2015, Budapest, Hungary

Investigation of storage options for scientific computing on Grid and Cloud facilities

BlueArc unified network storage systems 7th TF-Storage Meeting. Scale Bigger, Store Smarter, Accelerate Everything

Optimize the execution of local physics analysis workflows using Hadoop

HDFS Under the Hood. Sanjay Radia. Grid Computing, Hadoop Yahoo Inc.

Lustre * Filesystem for Cloud and Hadoop *

Archival Storage At LANL Past, Present and Future

The Design and Implementation of the Zetta Storage Service. October 27, 2009

Object storage in Cloud Computing and Embedded Processing

IBM Spectrum Protect in the Cloud

Data storage at CERN

Testing of several distributed file-system (HadoopFS, CEPH and GlusterFS) for supporting the HEP experiments analisys. Giacinto DONVITO INFN-Bari

Building Storage as a Service with OpenStack. Greg Elkinbard Senior Technical Director

THE EMC ISILON STORY. Big Data In The Enterprise. Copyright 2012 EMC Corporation. All rights reserved.

Aspera Direct-to-Cloud Storage WHITE PAPER

StorReduce Technical White Paper Cloud-based Data Deduplication

Use of Hadoop File System for Nuclear Physics Analyses in STAR

THE EXPAND PARALLEL FILE SYSTEM A FILE SYSTEM FOR CLUSTER AND GRID COMPUTING. José Daniel García Sánchez ARCOS Group University Carlos III of Madrid

High Availability with Windows Server 2012 Release Candidate

Big data management with IBM General Parallel File System

Data storage services at CC-IN2P3

Storage Architectures for Big Data in the Cloud

AIX NFS Client Performance Improvements for Databases on NAS

Lessons learned from parallel file system operation

Cisco Wide Area Application Services Optimizes Application Delivery from the Cloud

Managed GRID or why NFSv4.1 is not enough. Tigran Mkrtchyan for dcache Team

February, 2015 Bill Loewe

Distributed Data Storage Based on Web Access and IBP Infrastructure. Faculty of Informatics Masaryk University Brno, The Czech Republic

Cray DVS: Data Virtualization Service

Mass Storage System for Disk and Tape resources at the Tier1.

The dcache Storage Element

Red Hat Storage Server

XtreemFS a Distributed File System for Grids and Clouds Mikael Högqvist, Björn Kolbeck Zuse Institute Berlin XtreemFS Mikael Högqvist/Björn Kolbeck 1

IBM ELASTIC STORAGE SEAN LEE

EMC IRODS RESOURCE DRIVERS

Patrick Fuhrmann. The DESY Storage Cloud

Take An Internal Look at Hadoop. Hairong Kuang Grid Team, Yahoo! Inc

POWER ALL GLOBAL FILE SYSTEM (PGFS)

Maginatics Cloud Storage Platform for Elastic NAS Workloads

XtreemStore A SCALABLE STORAGE MANAGEMENT SOFTWARE WITHOUT LIMITS YOUR DATA. YOUR CONTROL

RAID Storage, Network File Systems, and DropBox

Hadoop: Embracing future hardware

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

Best Practices for Deploying Citrix XenDesktop on NexentaStor Open Storage

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing

High Availability Databases based on Oracle 10g RAC on Linux

Getting performance & scalability on standard platforms, the Object vs Block storage debate. Copyright 2013 MPSTOR LTD. All rights reserved.

The Evolution of Cloud Storage - From "Disk Drive in the Sky" to "Storage Array in the Sky" Allen Samuels Co-founder & Chief Architect

New Storage System Solutions

The BIG Data Era has. your storage! Bratislava, Slovakia, 21st March 2013

RED HAT STORAGE SERVER TECHNICAL OVERVIEW

Distributed File System Choices: Red Hat Storage, GFS2 & pnfs

WHITE PAPER. Software Defined Storage Hydrates the Cloud

Introduction to Cloud : Cloud and Cloud Storage. Lecture 2. Dr. Dalit Naor IBM Haifa Research Storage Systems. Dalit Naor, IBM Haifa Research

Technology Insight Series

Designing a Cloud Storage System

Virtual Provisioning. Management. Capacity oversubscription Physical allocation on the fly to logical size. With Thin Provisioning enabled

Disaster Recovery Checklist Disaster Recovery Plan for <System One>

Service Description Cloud Storage Openstack Swift

Investigating Private Cloud Storage Deployment using Cumulus, Walrus, and OpenStack/Swift

TECHNICAL WHITE PAPER: ELASTIC CLOUD STORAGE SOFTWARE ARCHITECTURE

Hadoop Architecture. Part 1

Scalable NAS for Oracle: Gateway to the (NFS) future

SCALABLE FILE SHARING AND DATA MANAGEMENT FOR INTERNET OF THINGS

Transcription:

SS Data & Storage CERN Cloud Storage Evaluation Geoffray Adde, Dirk Duellmann, Maitane Zotes CERN IT HEPiX Fall 2012 Workshop October 15-19, 2012 Institute of High Energy Physics, Beijing, China

SS Outline What are the interests in cloud storage implementations and protocols? How can we insure they can also be realised in the HEP environment? Test plans for two S3 implementations OpenStack/Swift Openlab collaboration with Huawei Results after first testing phase important contributions from Lu Wang (IHEP) Test plans for upcoming period 2

SS What cloud storage? Storage used by jobs running in a computational cloud network latency impacts what is usable depending on type of applications Storage build from (computational) cloud node resources storage life time = life time of node Storage service by commercial (or private) providers exploiting similar scaling concepts as computational clouds clustered storage with remote access with cloud protocol and modified storage semantic Term may be valid for all of the above but here I will refer to the last group of storage solutions 3

SS Why Cloud Storage & S3 Protocol? Cloud computing and storage gain rapidly in popularity Both as private infrastructure and as commercial service Several investigations are taking place also in the HEP and broader science community Price point of commercial offerings may not (yet?) be comparable with services we run at CERN or WLCG sites, but Changes in semantics, protocols, deployment model promise increased scalability with reduced deployment complexity (TCO) Market is growing rapidly and we need to understand if promises can be confirmed with HEP work loads Need to understand how cloud storage will integrate with (or change) current HEP computing models 4

SS S3 Semantics and Protocol Simple Storage Service (Amazon S3) just a storage service in contrast to eg Hadoop, which comes with a distributed computation model exploiting data locality (Hadoop also being evaluated in IT-DSS - but not reported in here) uses a language independent REST API http(s) for transport Provide additional scalability by focussing on a defined subset of posix functionality partitioning of namespace into independent buckets S3 protocol alone does not provide scalability eg if added naively on top of a traditional storage system scalability gains to be proven for each S3 implementation 5

SS CERN Interest in Cloud Storage Main Interest: Scalability and TCO can we run cloud storage systems to complement or consolidate existing storage services? Focus: storage for physics and infrastructure near-line archive pools, analysis disk pools, home directories, virtual machine image storage Which areas of the storage phase space can be covered well? First steps: 6 setup and run a cloud storage service of PB scale confirm scalability and/or deployment gains

SS Potential Interest for WLCG S3 Protocol could be a standard interface for access, placement or federation of physics data Allowing to provide (or buy) storage services without change to user application large sites may provide private clouds storage on acquired hardware smaller sites may buy S3 or rent capacity on demand First Steps 7 successful deployment at one site (eg CERN) demonstrate data distribution across sites (S3 implementations) according to experiment computing models

Component Layering in current SEs User Protocol Layer local & WAN efficiency, federation support, identity & role mapping random client I/O sequential p-2-p put / get Grid to Local User & Role mapping DM Admin Access rfio/xroot gridftp SRM xroot gridftp Bestman Cluster Layer scaling for large numbers of concurrent clients Clustering & Scheduling {replicated} File Cluster Catalogue & Meta Data DB CASTOR / DPM EOS {reliable} Media Raw Media {reliable} Media Raw Media {reliable} Media Raw Media (RAIDed) disk servers JBOD Media Layer Stable manageable storage, scaling in volume per $ (including ops effort) 8

Poten5al Future Scenarios integrate cloud market buy or build and integrate via standard Protocol Layer xroot http S3 Cluster Layer EOS Cloud Storage Media Layer Cloud Storage need to prove S3 TCO gains S3 alone functionally sufficient? 9

SS Common Work Items Common items for OpenStack/Swift and Huawei systems Define the main I/O pattern of interest based on measured I/O patterns in current storage services (eg archive, analysis pool, home dir, grid home?) Define implement and test a S3 functionality test define S3 API areas of main importance develop a S3 stress / scalability test scale up to several hundred concurrent clients (planned for August) copy-local and remote access scenarios Define key operational use cases and classify human effort and resulting service unavailability add remove disk servers (incl. draining) h/w intervention by vendor (does not apply to Huawei) s/w upgrade power outage Compare performance and TCO metrics to existing 10services at CERN

SS Work Items - Huawei Appliance Support commissioning of Huawei storage system 0.8 PB system in place at CERN Share CERN test suite and results with Huawei Tests (including ROOT based ones) are regular being run by Huawei development team Perform continuous functional and performance tests to validate new Huawei s/w releases maintain a list of unresolved operational or performance issues Schedule and execute large scale stress test 12 S3 test suite Hammercloud with experiment applications

SS OpenStack/SWIFT Actively participate in fixing of CERN issues with released software build up internal knowledge about SWIFT implementation and defects probe our ability to contribute and influence the release content from the OpenStack foundation Run the same functionality and stress tests as for the Huawei system Visit larger scale SWIFT deployments to get in-depth knowledge about level of overlap between open software and in-house deployed additions / improvements compare I/O pattern in typical deployments with CERN services 13

SS S3 plug-in for ROOT Based on the h;p plug- in Adapts to S3 protocol Supports vector read requests issued by TTree cache Huawei added multi-range read to S3 implementation Integrated with the distributed I/O test framework $ Na=ve$ File$I/O$ $ Lustre/ NFS/ GPFS $ ROOT$$ TTreeCache$ Virtual$File$Layer:$ Open/ReadBuffer/ReadBuffers/close$$ RFIO$ plugfin$ CASTOR$ Xrootd$$ plugfin$ EOS$ hgp$ plugfin$ Web$$ Server$ S3$ Plugin$ Cloud$ Storage$ 14

SS Fuse-based file system Open source prototype S3FS h;p://code.google.com/p/s3fs/ Current limita5ons: Can only mount one single bucket instead of the whole system Instead of remote I/O, file is downloaded to local cache during open df returns not relevant informa5on 15

SS Test Job A real ATLAS ROOT file 793MB on disk, 2.11 GB aver decompression 11918 entries, 5860 branches,cache size=30mb Reads in entries sequen5ally in physics tree 16 Reading Offset vs Entry Distribution of read Size

SS Test Result 120" Wall(Time( 100" Seconds( 80" 60" 40" 100%"content" 10%"content" 20" 0" EOS" OpenStack" Huawei"UDS" Huawei"UDS(mount" through"s3fs)" local"ext3" 17 Single client performance reach similar performance than local fs access already with previous Huawei release

SS Preliminary Results OpenStack/Swift and Huawei reach similar performance as EOS or local filesystem for full file access or 10% fraction of file Analysis type access using the ROOT S3 plugin naive use (no TTreeCache) of both S3 implementations shows significant overhead with enabled ROOT cache and vector read this overhead is removed S3 filesystem reaches performance of S3 plugin based access assuming that local cache space (/tmp) is available No authentication and authorisation yet for S3 storage 18 not yet mapped from certificates used in WLCG

SS Morning test - Long term stability files/sec 8000 7300 6000 download issue solved 4000 2000 2500 19 0 24/04/2012 20/09/2012 date

SS 100MB Upload - Huawei MB/sec 600 588,39 450 300 150 122 0 0 25 50 75 100 nº processes 20 No problem to reach bandwidth limit of 5Gb (from 5 client boxes - up to 100 processes total)

SS 4KB Download - Huawei files/sec 4000 3751 3000 2000 1000 682 21 0 0 25 50 75 100 Performance after upgrade: five times better nº processes

SS 4KB Download - Huawei files/sec 20000 18000 15000 Hot cache on backend nodes! 10000 7250 5000 3751 22 0 0 100 200 300 400 Close to linear metadata scaling (up to 300 procs) 682 nº processes

SS 100MB Download - Huawei MB/sec 3000 2250 2200 1500 960 750 560 0 0 100 200 300 400 nº processes 23 20 clients reach bandwidth limit of 18Gb

SS 100MB Download - OpenStack MB/sec Huawei OpenStack 3,000 2,250 1,500 OpenStack : 1 x 10Gb uplink 750 0 nº processes 0 75 150 225 300 24

SS Summary Cloud storage evaluation with two S3 implementations has started this year Client performance of local S3 based storage looks comparable to current production systems Achieved expected stability and aggregate performance (Huawei) Metadata performance up to 8k files/s proved expected scalability of the system Total throughput up to 18 Gb/s fully maxed out the 2 fibres available balanced system with 350MB/s per OSC Minor technical problems found and resolved rapidly productive collaboration with Huawei in context of CERN openlab OpenStack/Swift looks promising, but need to complete test suite Realistic TCO estimation can not yet be done in a small (1PB) test system w/o real users access 25

SS Future plans 2012 - short term Further increase scalability range with additional network and client resources Analyse performance impact of cache(s) and journal Collect feedback ATLAS workload management system Exercise transparent upgrade procedure (Huawei) Complete/compare OpenStack/Swift measurements 2013 - next year Multiple datacenter tests (eg in collaboration with IHEP) Erasure code impact on performance and space overhead Prove transparent failure recovery with consumer disks Goal for 2013 Evaluate TCO gains of S3 storage as part of a production service Candidate services being evaluated 26 CVMfs on S3 storage (interest from CVMfs team) AFS on S3 storage (prototype developed by Rainer Többicke)

SS Thank you! 27