CERNBox + EOS: Cloud Storage for Science

Similar documents
Prototyping a file sharing and synchronisation platform with owncloud

DSS. High performance storage pools for LHC. Data & Storage Services. Łukasz Janyst. on behalf of the CERN IT-DSS group

CERN Cloud Storage Evaluation Geoffray Adde, Dirk Duellmann, Maitane Zotes CERN IT

DSS. The Data Storage Services (DSS) Strategy at CERN. Jakub T. Moscicki. (Input from J. Iven, M. Lamanna A. Pace, A. Peters and A.

Book of Abstracts CS3 Workshop

owncloud Enterprise Edition on IBM Infrastructure

Storage strategy and cloud storage evaluations at CERN Dirk Duellmann, CERN IT

Scala Storage Scale-Out Clustered Storage White Paper

Data and Storage Services

DSS. Diskpool and cloud storage benchmarks used in IT-DSS. Data & Storage Services. Geoffray ADDE

Implementing Internet Storage Service Using OpenAFS. Sungjin Dongguen Arum

Storage Virtualization. Andreas Joachim Peters CERN IT-DSS

owncloud Architecture Overview

Panasas at the RCF. Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory. Robert Petkus Panasas at the RCF

owncloud Architecture Overview

Patrick Fuhrmann. The DESY Storage Cloud

Analisi di un servizio SRM: StoRM

GeoCloud Project Report USGS/EROS Spatial Data Warehouse Project

EMC SYNCPLICITY FILE SYNC AND SHARE SOLUTION

Introduction to Gluster. Versions 3.0.x

The cloud storage service bwsync&share at KIT

Scientific Storage at FNAL. Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015

WOS Cloud. ddn.com. Personal Storage for the Enterprise. DDN Solution Brief

Business and enterprise cloud sync, backup and sharing solutions

Investigation of storage options for scientific computing on Grid and Cloud facilities

HSS: A simple file storage system for web applications

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

Reference Design: Scalable Object Storage with Seagate Kinetic, Supermicro, and SwiftStack

IBM ELASTIC STORAGE SEAN LEE

Parallels Plesk Automation

Google File System. Web and scalability

Big Science and Big Data Dirk Duellmann, CERN Apache Big Data Europe 28 Sep 2015, Budapest, Hungary

Introduction to the EIS Guide

Storage solutions for a. infrastructure. Giacinto DONVITO INFN-Bari. Workshop on Cloud Services for File Synchronisation and Sharing

HDFS Users Guide. Table of contents


The Ultimate Business & Enterprise Hosting Solutions.


Self service for software development tools

Datacenter Operating Systems


Hadoop & its Usage at Facebook

Object storage in Cloud Computing and Embedded Processing

AIX NFS Client Performance Improvements for Databases on NAS

White Paper. Anywhere, Any Device File Access with IT in Control. Enterprise File Serving 2.0

Ignify ecommerce. Item Requirements Notes

File Sharing and Network Marketing

Distributed Data Storage Based on Web Access and IBP Infrastructure. Faculty of Informatics Masaryk University Brno, The Czech Republic

Hadoop Distributed File System. T Seminar On Multimedia Eero Kurkela

Enterprise Private Cloud Storage

files without borders

Technical Overview Simple, Scalable, Object Storage Software

Michael Thomas, Dorian Kcira California Institute of Technology. CMS Offline & Computing Week

HP reference configuration for entry-level SAS Grid Manager solutions

Oracle TimesTen In-Memory Database on Oracle Exalogic Elastic Cloud

Migration and Building of Data Centers in IBM SoftLayer with the RackWare Management Module

Performance Analysis and Capacity Planning Whitepaper

SYSTEM SETUP FOR SPE PLATFORMS

Apache Hadoop FileSystem and its Usage in Facebook

Migration and Building of Data Centers in IBM SoftLayer with the RackWare Management Module

The OpenStack TM Object Storage system

Server Installation Manual 4.4.1

Large Scale file storage with MogileFS. Stuart Teasdale Lead System Administrator we7 Ltd

Deploying Business Virtual Appliances on Open Source Cloud Computing

WHITE PAPER: BEST PRACTICES. Sizing and Scalability Recommendations for Symantec Endpoint Protection. Symantec Enterprise Security Solutions Group

How swift is your Swift? Ning Zhang, OpenStack Engineer at Zmanda Chander Kant, CEO at Zmanda

Data storage services at CC-IN2P3

DESYcloud: an owncloud & dcache update

Floating on a Hybrid Cloud: SQL Server 2012 & Microsoft Azure. Timothy P. McAliley Microsoft Premier Field Engineer SQL Server

Cloud Optimize Your IT

SQL Server Hybrid Features. Luis Vargas, Senior Program Manager Lead Xin Jin, Program Microsoft

Server Scalability and High Availability

Alfresco Enterprise on AWS: Reference Architecture

MDMCenter Hardware Specification

SHAREPOINT 2010 REMOTE BLOB STORES WITH EMC ISILON NAS AND METALOGIX STORAGEPOINT

Four Reasons To Start Working With NFSv4.1 Now

NoSQL Performance Test In-Memory Performance Comparison of SequoiaDB, Cassandra, and MongoDB

THE ATLAS DISTRIBUTED DATA MANAGEMENT SYSTEM & DATABASES

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database

ViewBox: Integrating Local File System with Cloud Storage Service

Scientific Computing Data Management Visions

OIS. Update on Windows 7 at CERN & Remote Desktop Gateway. Operating Systems & Information Services CERN IT-OIS

BlackBerry Enterprise Server for Microsoft Exchange. Version: 5.0 Service Pack: 4. Upgrade Guide

Hadoop IST 734 SS CHUNG

Best Practices Guide. McAfee epolicy Orchestrator Software


Windows Server 2012 授 權 說 明

Experience with Server Self Service Center (S3C)

Software Environment. Options. Service guarantee:. 24/7 Hardware Support. 99% uptime

SAP HANA Operation Expert Summit BUILD - High Availability & Disaster Recovery

Experience in integrating enduser cloud storage for CMS Analysis

msuite5 & mdesign Installation Prerequisites

IRODS use case : Ciment, the Univ. Grenoble-Alpes HPC center. B.Bzeznik / X.Briand Irods users group meeting 11/06/2015

Fileweave. Large File Transfer. Seamless Microsoft Outlook add-in. Simple drag and drop functionality

Storage Made Easy Enterprise File Share and Sync (EFSS) Cloud Control Gateway Architecture

Integrated Challan cum Return (ICR) - Scope of Work for ICR Platform

Pre-Migration Assessment Report SAMPLE REPORT. Number and type of source servers: 6 Windows Number of type of destination servers: 3 windows

CLOUD BASED SERVICE (CBS STORAGE)

Building low cost disk storage with Ceph and OpenStack Swift

Microsoft SharePoint Server 2010

Transcription:

Data & Storage Services CERNBox + EOS: Cloud Storage for Science CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/it Presenter: Luca Masce. Thanks to: Jakub T. Mościcki, Andreas J. Peters, Hugo G. Labrador, Massimo Lamanna CERN/IT- DSS

Content What we have done What we do What we will do CERNBox 2

The origins of the CERNBox project Missing link? CERNBox 4500 dissnct IPs in DNS from cern.ch to *.dropbox.com (daily...) What we are missing easy access cloud storage for end users files go automascally to the cloud and are available always everywhere broken laptop data lost offline access to data work on the plane and rsync when back online keep files in sync across devices access on mobile clients (easy) sharing of files with colleagues ssll surprisingly difficult Can we have this? for documents (small files, oben ppts, text, ) for science data (integrated into data processing workflows and exissng infrastructure) 3

Original architecture (CERNBox beta service) USER Sync client (webdav) Web access (hgps) HTTPS LB Data flow Metadata flow Apache, PHP 5.4 (SCL1.0) mod_proxy_balancer 64 core, 64GB RAM AS OC AS OC AS OC Setup 100% RH6 on standard hardware Based on owncloud Guaranteed failover (redundant nodes) Image courtesy of www.phdcomics.com Keeps track of sync state for every file in the system SQL overheads (Hz metadata ops) DB MySQL server 48GB RAM filesystem (POSIX) Files not exposed directly to the user STORAGE NFS servers, async, SW RAID 1 IniSal space: 20 TB

Usage of the beta service CERNBox Beta 2014 March April May June October users 190 (*) 285 361 429 720 files 191K 907K 1.6M 2.7M 6.4M size 480GB 1TB 1.5TB 1.9TB 3.4TB (*) users inherited from the inisal prototype deployment 15% 1% Size per user 84% Avg ~5GB <10GB >10GB up to 100GB Files per user 1% 5% 94% Avg ~10K files < 5K 5K- 20K up to 100K 5

File access patterns GET/PUT raso: 2/1 File type distribuson: 1200 different file extensions! 30%.c.h.C 30%.jpg.png 15% no extension (UNIX world!) 25% other:.pdf,.txt,.ppt,.docx,.root,.py,.eps,.tex ~100 URL shares, ~40 synced shares UNICODE filenames: greek, russian, thai(?) 6

Pilot limitations Move On the origin client move is propagated to the server On the other clients it is propagated as COPY/DELETE (subopsmal) Symlinks are not supported Ignored files:, :? * < > We currently recommend one sync folder setup: ~/cernbox High per- file overhead Expect 2-5Hz PUT Expect ~10Hz GET Transfer rates Expect 10-30MB download Expect 5-10MB upload Larger files: 400MB file on standard desktop hgps/upload: ~25MB/s, hgps/download: ~60MB/s For wireless devices, laptops, phones do we care about transfer rates? 7

Towards large-scale data sync and share Currently deployed CERNBox beta works OK so far for the classical Dropbox use- case low- frequency document sync and share But can we bring this system to the next level? Our core- business and large- scale workloads expose PBs of exis%ng data from day 1 integrason into physics data processing eco- system central services: batch, interacsve data analysis applicasons sync higher data volumes at higher rates Can we ssll keep the simplicity of cloud storage access? 8

Massive scaling at reduced cost? No need to keep track of all files and directories in the database avoids explosive growth of your DB infrastructure Our file number essmate? With 10K users we have 2.5 billion files in AFS already! What is your number for 100K users? Before we start throwing hardware at the problem consider the cost of running the service Fixed: hardware purchase, service deployment, infrastructure Scaling: hardware incidents, user support; backup; integrity checks; upgrades Infrastructure: space, electricity and cooling in the data center For massive scaling we need to keep TCO under control profit from exissng large- scale operasons and support of our storage services exploit economies of scale 9

Integration Started in May 2014 FuncSonality Enable sync and share for exissng data in EOS Without exporsng data to another storage Direct access to data with efficient sync behind OperaSons NFS/async backend server is a temporary soluson EOS offer virtually unlimited cloud storage for end- users Fold- in the operason cost into EOS But: Integrate as transparently as possible most users don t care about storage backend Fully working soluson compasble with owncloud clients we don t want to end up with half- working CERN- specific soluson 10

EOS Integration Details Understanding sync protocol and underlying semanscs. Add a few consistency features to EOS (e.g. atomic upload) Adding few new features to EOS or libing restricsons (e.g. UTF8 support) Beef- up the webdav endpoint to allow owncloud clients to talk directly to it Integrate web- access and sharing funcsonality Web fronted: develop new plugins Nice integrason of trashbin, versions and sharing: Fusion between owncloud model and EOS model (Hugo G. Labrador) Making more robust less stressed parts in EOS (hgp/webdav) Lots, lots of tessng. 11

CERNBox 2.0 Architecture Sync client (webdav) Web access (hgps) HTTPS HTTPS LB HTTPS LB HTTPS LB LB Data flow Metadata flow Data directly accessible by the user USER hgp (public data) hgps (private data) hgp (internal) KHz metadata ops OC fuse All sync state as metadata in the storage STORAGE (EOS) Files wriyen with USER credenzals disk servers (1000s) IO redirect namespace

Prototype deployment on EOSPPS /eos/user/<u>/<username> this is the default sync and web- enabled folder as an advanced user you may add arbitrary folder from EOS very easy to implement a folder shared by an e- group We can also allow transparent access to different instances 13

First performance numbers User- perceived performance (client) Metadata operason (pycurl with SSL sessions) PROPFIND with 1 entry: 90 Hz PROPFIND with 1K entries: 8.5 KHz PROPFIND with 10K entries: 10KHz nice speed e.g. kernel src tree upload (50K files, 500MB) ~ 1h from laptop/wifi at home, download ~20 min ops/s 200 200 Small files (10KB) pycurl seq 150 100 50 0 125 70 60 50 65 57 30 30 20 11 Download Upload Delete pycurl P=10 pycurl P=50 owncloud sync client 14

Summary Working and usable beta service Useful for ge.ng experience, user feedback and understanding what we want / don t want in the final producson system based on EOS CERNBox Advanced integrason into EOS will open up new possibilises but there is no free lunch: we will have to adapt to evolving owncloud clients, etc. heading towards large sync and share layer for science research all our data exposed from day 1 massive scalability, high performance integrated into exissng workflows - new capabilises! small overhead on top of our exissng operasons and development TCO control and ssll as easy to use as Dropbox.com 15

Integrated storage ecosystem for scientific research sync / share / offline access 2.0 USER webdav & hgps:// online file- system access fuse CERNBox Analysis cluster Central Services high- performance applicason access xrootd:// batch access xrdcopy LARGE- SCALE STORAGE

agenda full ~35 parscipants Tracks Keynote B.Pierce Technology Users Site reports Vendor talks IBM Powerfolder SeaFile PyDio Owncloud 17

CERNBox 2.0 some numbers Advanced prototype stage Adapted exissng webdav interface in EOS to be compasble with owncloud sync clients Test environment (EOSPPS) standard hardware namespace node with Xeon 2.2GHz, 16 cores, 24GB RAM 50 disk servers: cheap JBODs (1000 disks), total 800TB usable space Storage layout: 2 replicas in RAIN mode à every file PUT = 2 copies of the file on two independent storage nodes (with adler32 checksums of content) Event- based hgp(s) load- balancer (nginx) Underlying storage scalability (EOS Prod) Max observed IO: ~40GB/s on a single instance (eosatlas => ) Max observed file stats: 10s KHz Thousands of connected clients Server should never be a bogleneck for CERNBox 18