Data storage services at CC-IN2P3



Similar documents
irods at CC-IN2P3: managing petabytes of data

KIT Site Report. Andreas Petzold. STEINBUCH CENTRE FOR COMPUTING - SCC

Maurice Askinazi Ofer Rind Tony Wong. Cornell Nov. 2, 2010 Storage at BNL

Report from SARA/NIKHEF T1 and associated T2s

Scientific Storage at FNAL. Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. dcache Introduction

Managed GRID or why NFSv4.1 is not enough. Tigran Mkrtchyan for dcache Team

Evolution of the Italian Tier1 (INFN-T1) Umea, May 2009

Mass Storage System for Disk and Tape resources at the Tier1.

The dcache Storage Element

Preview of a Novel Architecture for Large Scale Storage

Low-cost

KIT Site Report. Andreas Petzold. STEINBUCH CENTRE FOR COMPUTING - SCC

Solution for private cloud computing

Archival Storage At LANL Past, Present and Future

Wrangler: A New Generation of Data-intensive Supercomputing. Christopher Jordan, Siva Kulasekaran, Niall Gaffney

DSS. High performance storage pools for LHC. Data & Storage Services. Łukasz Janyst. on behalf of the CERN IT-DSS group

Storage Virtualization. Andreas Joachim Peters CERN IT-DSS

Tier0 plans and security and backup policy proposals

CERN Cloud Storage Evaluation Geoffray Adde, Dirk Duellmann, Maitane Zotes CERN IT

Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers

Seriously: Tape Only Backup Systems are Dead, Dead, Dead!

SCALABLE FILE SHARING AND DATA MANAGEMENT FOR INTERNET OF THINGS

SLIDE 1 Previous Next Exit

NT1: An example for future EISCAT_3D data centre and archiving?

Big Science and Big Data Dirk Duellmann, CERN Apache Big Data Europe 28 Sep 2015, Budapest, Hungary

A Model for Forecasting Data Centre Infrastructure Costs

Solid State Storage in the Evolution of the Data Center

How To Build A Cloud Storage System

HAMBURG ZEUTHEN. DESY Tier 2 and NAF. Peter Wegner, Birgit Lewendel for DESY-IT/DV. Tier 2: Status and News NAF: Status, Plans and Questions

POSIX and Object Distributed Storage Systems

Integrating a heterogeneous and shared Linux cluster into grids

Panasas at the RCF. Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory. Robert Petkus Panasas at the RCF

Object storage in Cloud Computing and Embedded Processing

Hadoop: Embracing future hardware

Large File System Backup NERSC Global File System Experience

Status of Grid Activities in Pakistan. FAWAD SAEED National Centre For Physics, Pakistan

Tape Cloud Scalable and Cost Efficient Big Data Infrastructure for Cloud Computing. Varun S. Prakash with Yuanfeng Wen Weidong Shi

AFS Usage and Backups using TiBS at Fermilab. Presented by Kevin Hill

Globus and the Centralized Research Data Infrastructure at CU Boulder

Mass Storage at GridKa

NERSC File Systems and How to Use Them

SURFsara Data Services

Data Center Storage Solutions

Solution Brief: Creating Avid Project Archives

Running the scientific data archive

The Rise of Industrial Big Data. Brian Courtney General Manager Industrial Data Intelligence

Software-defined Storage Architecture for Analytics Computing

Protecting Information in a Smarter Data Center with the Performance of Flash

Michał Jankowski Maciej Brzeźniak PSNC

2. COMPUTER SYSTEM. 2.1 Introduction

(Possible) HEP Use Case for NDN. Phil DeMar; Wenji Wu NDNComm (UCLA) Sept. 28, 2015

Migrating NASA Archives to Disk: Challenges and Opportunities. NASA Langley Research Center Chris Harris June 2, 2015

Global Grid User Support - GGUS - in the LCG & EGEE environment

10th TF-Storage Meeting

XenData Product Brief: SX-250 Archive Server for LTO

Analisi di un servizio SRM: StoRM

Quick Introduction to HPSS at NERSC

Hadoop on the Gordon Data Intensive Cluster

Distributed Computing for CEPC. YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep.

Big Fast Data Hadoop acceleration with Flash. June 2013

LDM-129: Data Management Infrastructure Design

Best in Class Storage Solutions for BS2000/OSD

BlueArc unified network storage systems 7th TF-Storage Meeting. Scale Bigger, Store Smarter, Accelerate Everything

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

Using S3 cloud storage with ROOT and CernVMFS. Maria Arsuaga-Rios Seppo Heikkila Dirk Duellmann Rene Meusel Jakob Blomer Ben Couturier

Amazon Cloud Storage Options

Hadoop Distributed File System. T Seminar On Multimedia Eero Kurkela

Utilizing the SDSC Cloud Storage Service

Improvement Options for LHC Mass Storage and Data Management

Performance Analysis of Flash Storage Devices and their Application in High Performance Computing

Application-Focused Flash Acceleration

Moving Virtual Storage to the Cloud. Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage

EMC BACKUP MEETS BIG DATA

How To Backup And Restore A Database With A Powervault Backup And Powervaults Backup Software On A Poweredge Powervalt Backup On A Netvault 2.5 (Powervault) Powervast Backup On An Uniden Power

XenData Product Brief: SX-250 Archive Server for LTO

Computing at the HL-LHC

EMC IRODS RESOURCE DRIVERS

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA

Hitachi NAS Platform and Hitachi Content Platform with ESRI Image

Michael Thomas, Dorian Kcira California Institute of Technology. CMS Offline & Computing Week

Boas Betzler. Planet. Globally Distributed IaaS Platform Examples AWS and SoftLayer. November 9, IBM Corporation

Object Storage, Cloud Storage, and High Capacity File Systems

Big Data: A Storage Systems Perspective Muthukumar Murugan Ph.D. HP Storage Division

DSS. The Data Storage Services (DSS) Strategy at CERN. Jakub T. Moscicki. (Input from J. Iven, M. Lamanna A. Pace, A. Peters and A.

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute

Main Memory Data Warehouses

Storage solutions for a. infrastructure. Giacinto DONVITO INFN-Bari. Workshop on Cloud Services for File Synchronisation and Sharing

OSG Hadoop is packaged into rpms for SL4, SL5 by Caltech BeStMan, gridftp backend

Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle

Extreme Scalability and Flexibility for Content Management Throughout Its Lifecycle

Deploying Flash- Accelerated Hadoop with InfiniFlash from SanDisk

Next Generation Tier 1 Storage

Moving Virtual Storage to the Cloud

2009 Oracle Corporation 1

The DESY Big-Data Cloud System

Software, Computing and Analysis Models at CDF and D0

EMC DATA PROTECTION. Backup ed Archivio su cui fare affidamento

Welcome to the unit of Hadoop Fundamentals on Hadoop architecture. I will begin with a terminology review and then cover the major components

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Transcription:

Centre de Calcul de l Institut National de Physique Nucléaire et de Physique des Particules Data storage services at CC-IN2P3 Jean-Yves Nief

Agenda Hardware: Storage on disk. Storage on tape. Software: Storage services. Pitfalls and lessons learned. Storage needs in the near future. Prospects. Visit of NCSA 27-05-2015 2

Storage at CC-IN2P3: disk Hardware Software Direct Attached Storage servers (DAS): Dell servers (R720xd + MD1200) ~ 240 servers Capacity: 15 PBs Parallel File System: GPFS (1.9 PBs) File servers: xrootd, dcache (14 PBs) Used for High Energy Physics (LHC etc ) Disk attached via SAS: Dell servers ( R620 + MD3260) Capacity: 1.9 PBs Storage Area Network disk arrays (SAN): IBM V7000 and DCS3700, Hitachi HUS 130. Capacity: 240 TBs Mass Storage System: HPSS (600 TBs) Used as a disk cache in front of the tapes. Middlewares: SRM, irods (840 TBs) Databases: mysql, PostGres, Oracle (57 TBs) Visit of NCSA 27-05-2015 3

Storage at CC-IN2P3: tapes Hardware 4 Oracle/STK SL8500 librairies: 40,000 slots (T10K and LTO4) Max capacity: 320 PBs (with T10KD tapes) 111 tape drives 1 IBM TS3500 library: 3500 slots (LTO6) Software Mass Storage System: HPSS 25 PBs Max traffic (from HPSS): 100 TBs / day Interfaced with our disk services Backup service: TSM (1 PB) Visit of NCSA 27-05-2015 4

Storage hardware as of today 4 Oracle/STK SL8500 libraries (4 x 10000 slots) for HPSS and TSM. 1 IBM TS3500 library (3500 slots) for TSM. Tape media: Media Capacity Max Number of drives Bandwidth Service T10K-B 1TB 120 MB/s 59 HPSS T10K-C 5 TB 240 MB/s 21 HPSS T10K-D 8.5 TB 252 MB/s 31 HPSS LTO4 800 GB 120 MB/s 20 TSM LTO6 2.4 TB 160 MB/s 6 TSM Visit of NCSA 27-05-2015 5

How experiments are dealing with the storage? Client apps (local or distant apps) File servers / filesystem (local access) Middlewares (wide area access) xrootd dcache GPFS SRM irods Mass Storage System HPSS TSM Backup Visit of NCSA 27-05-2015 6

Local access: distributed filesystem Purpose: For heavy I/O operations by many batch jobs in //. Not a permanent repository. 60 servers, 1.9 PBs of disk space. 1000 GPFS clients installed on our Used by: up to 25% of the batch farm. 84 groups. Max activity observed: 500 TBs read in 5 days (Planck) To do: Migration of metadata storage to flash technology (NetApp EF560). Reduction of the number of space (35 right now) down to 4. Visit of NCSA 27-05-2015 7

Local access: dcache-xrootd Purpose: Data access by batch jobs (local + grid). Pattern: remote access (random I/O, sequential). Interface with HPSS: Using Treqs (Tape request scheduler). dcache service part of the Atlas and CMS federations: xrootd proxy servers (gateway to our local dcache service). dcache head nodes db: use of SSD. To do: xrootd: lots of tape staging (eg: 20k requests per day) è «automatic» prestaging for some large «productions» needed. service # servers capacity experiments # clients in // access protocol dcache 165 8.5 PiB LCG + «EGEE» VOs (17) 15000 dcap, xrootd, WebDAV, gsiftp xrootd 44 2.5 PiB Alice + HEP + astroparticle (13) 8000 xrootd, http Visit of NCSA 27-05-2015 8

Local access (xrootd/dcache) (4) + (5): dynamic staging (TreqS) HPSS (4) (5) T1.root Data server: Xrootd Data server: Xrootd Data servers (6): random access Client (3) (1) (2) /hpss/in2p3.fr/t1.root? (6) Redirector server: Xrootd (1) + (2): load balancing (etc ) redirectors Visit of NCSA 27-05-2015 9

Mass Storage service: HPSS Purpose: Interface to our back end storage: tapes (main storage). 29 servers (disk + tape movers). Massive activity: 27 PiB on tapes (rank #12 in HPSS world), 597 TiB of disk space (Dell R510, R720 servers). Up to 100 TiB in 22h between HPSS and other storage services. New drive technology every 30 months. Life cycle of a tape generation 8 years: Ramping up in 4 years. 4 more years in production. T10KB decommissionning: 20000 tapes to be repacked until end 2017 ó 5 PiB / year. Represents: 1000 T10K-T2 è T10K-C Or 750 T10K-T2 è T10K-D Tape lifetime: avg 7 years (max 10 years). Redefine the class of services. Refactoring of Treqs (Tape request scheduler) by our dev team. Visit of NCSA 27-05-2015 10

Local access: pitfalls and lessons learned Dynamic staging may result in: Massive staging requests. Long delays to access the files. Tapes capacity is getting bigger! è Increased job inefficiency. To leverage these drawbacks: Big files on tapes (several GBs). Reordering of file requests per tape id (reduce # of mounts/ dismounts): up to 10000 per day. Prestaging. Avoid file scattering on a large amount of tapes (tape families?). Refactoring of Treqs (Tape request scheduler) by our dev team. Visit of NCSA 27-05-2015 11

Middleware: SRM, irods Purpose: Data management (storage virtualization, policies). Data sharing, access, archival etc in a distributed and heterogeneous environment. SRM: Used by Atlas, CMS, LHCb + EGEE/EGI virtual organizations. irods: Managing > 9 PBs of data (HEP, astro, biology, Arts & Humanities): includes data stored on tapes. Web interfaces, HTTP/REST, webdav. Visit of NCSA 27-05-2015 12

Focus: LSST storage usage GPFS: 130 TB (50% being used at the moment) HPSS: 78 TB irods: 90 TB Visit of NCSA 27-05-2015 13

Future storage needs (Hard Sci Fi) Cumulated Disk (TB) Cumulated Tape (TB) Hypothesis: LCG: +10% / year HEP like experiment: ratio disk/ tape = 30% CTA not included Credit: Rachid Lemrani Visit of NCSA 27-05-2015 14

Qserv Used to debug Qserv. 1 interactive node: 1 virtual machine ccqservbuild.in2p3.fr. Mainly used for: the compilation of the Qserv software. the deployement of Qserv on the cluster. Accessible from SLAC. 50 Qserv nodes: Dell / CC-IN2P3 partnership. ccqserv{100..149}.in2p3.fr. Dell servers R620 and R630 Intel Xeon, 16GB RAM 7 TB of usable storage run the Qserv software (1 master, 49 slaves). private subnetwork (nodes accessible from ccqservbuild only). (Credit: Yvan Calas) Visit of NCSA 27-05-2015 15

Storage prospects Tape will still have a major role: Balance between nearline and pure archive storage? Hard drives still preferred to SSD: prices will converge in the next couple of years. SSD usage limited at the moment for key apps (GPFS metadata, HPSS and dcache databases etc ). Heavy I/O still and even heavier: Lots of «random» I/O: still strong needs for seek, read, write. Simple put/get interaction not sufficient. Big metadata challenge (storage system metadata). Data life cycle improvements: > 100 groups, 5000 users (~ 2000 active users). Data Management Plan. archival service beyond simple bit preservation? OAIS, dublin core metadata. Provide interfaces for Open data? Visit of NCSA 27-05-2015 16

Storage prospects Assessment of object storage: Swift: testing it at the moment. Swift extra value wrt what we already have: to be proven. Ceph (but it can be more than an object storage): evaluation starting in the next couple of months. Hadoop, MapReduce type of systems: no plans yet as not much needs expressed by our user community. Metadata: will increase massively with increased storage. Limit the amount of files => increased file size. Visit of NCSA 27-05-2015 17