Mass Storage at GridKa



Similar documents
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. dcache Introduction

The dcache Storage Element

dcache - Managed Storage - LCG Storage Element - HSM optimizer Patrick Fuhrmann, DESY for the dcache Team

dcache, a managed storage in grid

Managed GRID or why NFSv4.1 is not enough. Tigran Mkrtchyan for dcache Team

Patrick Fuhrmann. The DESY Storage Cloud

Scientific Storage at FNAL. Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015

dcache, Software for Big Data

Integrating a heterogeneous and shared Linux cluster into grids

dcache, list of topics

Fermilab s Multi-Petabyte Scalable Mass Storage System

Maurice Askinazi Ofer Rind Tony Wong. Cornell Nov. 2, 2010 Storage at BNL

Data storage services at CC-IN2P3

Report from SARA/NIKHEF T1 and associated T2s

GridKa: Roles and Status

Mass Storage System for Disk and Tape resources at the Tier1.

IPv6 Traffic Analysis and Storage

Flexible Scalable Hardware independent. Solutions for Long Term Archiving

Running the scientific data archive

Data Storage and Data Transfer. Dan Hitchcock Acting Associate Director, Advanced Scientific Computing Research

A Tutorial on Configuring and Deploying GridFTP for Managing Data Movement in Grid/HPC Environments

dcache Firewall Issue

Analisi di un servizio SRM: StoRM

GridKa site report. Manfred Alef, Andreas Heiss, Jos van Wezel. Steinbuch Centre for Computing

The GridKa Installation for HEP Computing

Grid Data Management. Raj Kettimuthu

2011 FileTek, Inc. All rights reserved. 1 QUESTION

The DESY Big-Data Cloud System

XenData Archive Series Software Technical Overview

irods at CC-IN2P3: managing petabytes of data

OSG Hadoop is packaged into rpms for SL4, SL5 by Caltech BeStMan, gridftp backend

Preview of a Novel Architecture for Large Scale Storage

LBNC and IBM Corporation Document: LBNC-Install.doc Date: Path: D:\Doc\EPFL\LNBC\LBNC-Install.doc Version: V1.0

Configuring Celerra for Security Information Management with Network Intelligence s envision

File transfer in UNICORE State of the art

BNL FAX Setup. Hironori Ito Brookhaven National Laboratory

Plateforme de Calcul pour les Sciences du Vivant. SRB & glite. V. Breton.


The EU DataGrid Data Management

Metalogix SharePoint Backup. Advanced Installation Guide. Publication Date: August 24, 2015

Tuning WebSphere Application Server ND 7.0. Royal Cyber Inc.

Data Management in an International Data Grid Project. Timur Chabuk 04/09/2007

LHCb activities at PIC

CMS Tier-3 cluster at NISER. Dr. Tania Moulik

Recommendations for Static Firewall Configuration in D-Grid

Implementing Network Attached Storage. Ken Fallon Bill Bullers Impactdata

Effective Planning and Use of TSM V6 Deduplication

Michał Jankowski Maciej Brzeźniak PSNC

Cornell University Center for Advanced Computing

CERN Cloud Storage Evaluation Geoffray Adde, Dirk Duellmann, Maitane Zotes CERN IT

The glite File Transfer Service

Intershop 7 System Requirements Sheet

CTERA Cloud Onramp for IBM Tivoli Storage Manager

Introduction to NetApp Infinite Volume

Using Globus Toolkit

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago

DSS. The Data Storage Services (DSS) Strategy at CERN. Jakub T. Moscicki. (Input from J. Iven, M. Lamanna A. Pace, A. Peters and A.

Evolution of the Italian Tier1 (INFN-T1) Umea, May 2009

XenData Product Brief: SX-250 Archive Server for LTO

Michael Thomas, Dorian Kcira California Institute of Technology. CMS Offline & Computing Week

HOW TO DEPLOY AN EJB APLICATION IN WEBLOGIC SERVER 11GR1

A Web Services Data Analysis Grid *

Software Infrastructure Supplement for the OSG Proposal

Advanced School in High Performance and GRID Computing November 2008

How To Backup At Qmul

XtreemStore A SCALABLE STORAGE MANAGEMENT SOFTWARE WITHOUT LIMITS YOUR DATA. YOUR CONTROL

glibrary: Digital Asset Management System for the Grid

How To Virtualize A Storage Area Network (San) With Virtualization

Data Management. Network transfers

XenData Product Brief: SX-250 Archive Server for LTO

SDFS Overview. By Sam Silverberg

Storage Resource Managers: Recent International Experience on Requirements and Multiple Co-Operating Implementations

MIGRATING DESKTOP AND ROAMING ACCESS. Migrating Desktop and Roaming Access Whitepaper

KIT Site Report. Andreas Petzold. STEINBUCH CENTRE FOR COMPUTING - SCC

EREBOS: CosmoSim Database. CLUES Research Environment. Harry Enke (Kristin Riebe, Jochen Klar, Adrian Partl) CLUES Meeting 2015, Copenhagen

HPSS Best Practices. Erich Thanhardt Bill Anderson Marc Genty B

Scala Storage Scale-Out Clustered Storage White Paper

LHC schedule: what does it imply for SRM deployment? CERN, July 2007

EMC Disk Library with EMC Data Domain Deployment Scenario

High Availability Databases based on Oracle 10g RAC on Linux

Solution Brief: Creating Avid Project Archives

Research Data Storage Infrastructure (RDSI) Project. DaSh Straw-Man

Configuring Apache Derby for Performance and Durability Olav Sandstå

Big Science and Big Data Dirk Duellmann, CERN Apache Big Data Europe 28 Sep 2015, Budapest, Hungary

Backup and Recovery 1

Solution Brief: Using a XenData Digital Video Archive with Grass Valley STRATUS

Multi-Terabyte Archives for Medical Imaging Applications

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Available Performance Testing Tools

Disk Arrays Fiber Channel connectivity NAS Data Movers Backup/Recover media and software

NAS or iscsi? White Paper Selecting a storage system. Copyright 2007 Fusionstor. No.1

SURFsara Data Services

Development of Monitoring and Analysis Tools for the Huawei Cloud Storage

TSM Studio Server User Guide

BaBar and ROOT data storage. Peter Elmer BaBar Princeton University ROOT Oct. 2002

Implementing an Automated Digital Video Archive Based on the Video Edition of XenData Software

Objectivity Data Migration

Virtual Server and Storage Provisioning Service. Service Description

Distributed File System. MCSN N. Tonellotto Complements of Distributed Enabling Platforms

Performance Management Platform

Transcription:

Mass Storage at GridKa Forschungszentrum Karlsruhe GmbH Institute for Scientific Computing P.O. Box 3640 D-76021 Karlsruhe, Germany Dr. Doris Ressmann http://www.gridka.de 1

Overview What is dcache? Pool Selection mechanism dcache properties LCG connection Forschungszentrum Karlsruhe Introduction Access to dcache connection to CERN Tape Management Conclusion 2

Service Challenge disks 10Gbit gridftp SAN 3

Mass Storage Environment disks 10Gbit gridftp SAN tape library NFS xrootd dcache 4

What is dcache? Developed at DESY and FNAL Disk pool management with or without tape backend Data may be distributed among a huge amount of disk servers. Automatic load balancing by cost metric and inter pool transfers. Data removed only if space is needed Fine grained configuration of pool attraction scheme 5

Pool Selection Mechanism Pool Selection required for: Client dcache Tape dcache dcache dcache dcache Client Pool selection is done in 2 steps Query configuration database : which pools are allowed for requested operation (intern/extern) Query 'allowed pool' for their vital functions : find pool with lowest cost for requested operation 6

LCG Storage Element DESY dcap lib incorporates with CERN GFAL library SRM version ~ 1.1 supported gsiftp supported 7

Multiple access of one file Pool 1 Pool 2 Pool 3 File 1 8

Multiple access of one file Pool 1 Pool 2 Pool 3 File 1 File 1 File 1 9

Intern Mountpoint ls mv rm dcap dccp <source> <destination> dc_open(...) dc_read(...) Preload library Access to dcache Extern Gridftp Problematic when file needs to be staged first SRMCP 10

dcache environment Internal nodes file transfer head node pools tape library file transfer 11

dcache environment Internal nodes file transfer head node srm gsiftp pools tape library file transfer 12

dcache environment Internal nodes file transfer head node srm gsiftp file transfer pools tape library file transfer 13

dcache environment Internal nodes file transfer head node srm srmcp pools file transfer tape library file transfer 14

PNFS Perfectly Normal File System gdbm databases Experiment specific databases Independent access Content of metadata: User file name File name within dcache pool and tape real data 0000000000000000000014F0 000000000000000000001510 0000000000000000000015A0 0000000000000000000017E8 000000000000000000001858 Information about the tape location (storage class ) Pool name where the file is located pnfs database for filenames metadata 15

gsiftp Only registered dcache user!!! grid-proxy-init globus-url-copy dbg \ file:///tmp/file1 \ gsiftp://srm1.fzk.de/grid/fzk.de/mounts/pnfs/cms/file1 dcache gridftp client and server in Java copy direct into available pool node pool: data is precious (can't be deleted) flush into tape data is cached (can be deleted from pool) 16

srmcp Only registered dcache user!!! grid-proxy-init srmcp debug=true \ srm://srm.web.cern.ch:80//castor/cern.ch/grid/dteam/castorfile \ srm://srm1.fzk.de:8443//pnfs/gridka.de/data/ressmann/file2 srmcp debug=true \ srm://srm1.fzk.de:8443//pnfs/gridka.de/data/ressmann/file2 file:////tmp/file2 17

Firewall issues Connection to headnode: Ports 8443 and 2811 Port Range to pool nodes: 20.000 to 50.000 18

SRM Disk Version FNAL is currently developing a standalone SRM Disk version. The client uses a java version of gridftp The server uses a standard globus gridftp. It is far from production ready and needs: SQL Database jdbc driver http://www-isd.fnal.gov/srm/unix-fs-srm/ 19

20

Tape Management Tivoli Storage Manager (TSM) library management TSM is not developed for archive Interruption of TSM archive No control what has been archived 21

Tape Management Tivoli Storage Manager (TSM) library management TSM is not developed for archive Interruption of TSM archive No control what has been archived 22

Tape Management Tivoli Storage Manager (TSM) library management TSM is not developed for archive Interruption of TSM archive No control what has been archived 23

dcache tape access Convenient HSM connectivity (done for Enstore, OSM, TSM, bad for HPSS) Creates a separate session for every file Transparent access Allows transparent maintenance at HSM 24

20 GB Forschungszentrum Karlsruhe dcache pool node 800 GB 1h 25

dcache tape management Precious data is separately collected per 'storage class Each 'storage class queue ' has individual parameters, steering the tape flush operation. Maximum time, a file is allowed to be 'precious' per 'storage class'. Maximum number of precious bytes per 'storage class Maximum number of precious files per 'storage class Maximum number of simultaneous tape flush' operations can be configured 26

Conclusion and Future Work Low cost read pools Reliable write pools Write once never change a dcache file Single point of failure Working SRM connection between CERN and FZK Connection to openlab at CERN Adding 15 Pool nodes for the 10 Gbit test from SRM to SRM Adding tape drives to increase throughput More at www.dcache.org 27

28