Mass Storage System for Disk and Tape resources at the Tier1.



Similar documents
Evolution of the Italian Tier1 (INFN-T1) Umea, May 2009

Maurice Askinazi Ofer Rind Tony Wong. Cornell Nov. 2, 2010 Storage at BNL

High Availability Databases based on Oracle 10g RAC on Linux

Tier0 plans and security and backup policy proposals

CERN Cloud Storage Evaluation Geoffray Adde, Dirk Duellmann, Maitane Zotes CERN IT

Implementing a Digital Video Archive Based on XenData Software

Improvement Options for LHC Mass Storage and Data Management

Implementing a Digital Video Archive Using XenData Software and a Spectra Logic Archive

DSS. High performance storage pools for LHC. Data & Storage Services. Łukasz Janyst. on behalf of the CERN IT-DSS group

Data storage services at CC-IN2P3

Preview of a Novel Architecture for Large Scale Storage

Implementing an Automated Digital Video Archive Based on the Video Edition of XenData Software

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute

Analisi di un servizio SRM: StoRM

Scientific Storage at FNAL. Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015

SAN TECHNICAL - DETAILS/ SPECIFICATIONS

Implementing Offline Digital Video Storage using XenData Software

Extreme Scalability and Flexibility for Content Management Throughout Its Lifecycle

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

Reference Architecture. EMC Global Solutions. 42 South Street Hopkinton MA

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007

(Scale Out NAS System)

Oracle Database Deployments with EMC CLARiiON AX4 Storage Systems

Large Scale Storage. Orlando Richards, Information Services LCFG Users Day, University of Edinburgh 18 th January 2013

UN 4013 V - Virtual Tape Libraries solutions update...

Microsoft SQL Server 2005 on Windows Server 2003

Data Information and Management System (DIMS) The New DIMS Hardware

Oracle Database Scalability in VMware ESX VMware ESX 3.5

<Insert Picture Here> Refreshing Your Data Protection Environment with Next-Generation Architectures

Optimizing Large Arrays with StoneFly Storage Concentrators

IT-INFN-CNAF Status Update

The safer, easier way to help you pass any IT exams. Exam : Storage Sales V2. Title : Version : Demo 1 / 5

Dennis Waldron, CERN IT/DM/DA CASTOR Face-to-Face Meeting, Feb 19 th CERN IT Department CH-1211 Genève 23 Switzerland

Scalable NAS for Oracle: Gateway to the (NFS) future

WHITE PAPER BRENT WELCH NOVEMBER

XenData Video Edition. Product Brief:

Understanding Disk Storage in Tivoli Storage Manager

Managing managed storage

IBM Global Technology Services November Successfully implementing a private storage cloud to help reduce total cost of ownership

CMS Tier-3 cluster at NISER. Dr. Tania Moulik

SMB Direct for SQL Server and Private Cloud

Brocade and EMC Solution for Microsoft Hyper-V and SharePoint Clusters

EMC Disk Library with EMC Data Domain Deployment Scenario

Storage Virtualization. Andreas Joachim Peters CERN IT-DSS

STORAGE CENTER. The Industry s Only SAN with Automated Tiered Storage STORAGE CENTER

Using VMware VMotion with Oracle Database and EMC CLARiiON Storage Systems

Beyond High Performance Computing: What Matters to CERN

An Oracle White Paper September Oracle Exadata Database Machine - Backup & Recovery Sizing: Tape Backups

Solution Brief: Creating Avid Project Archives

VTrak SATA RAID Storage System

Staying afloat in today s Storage Pools. Bob Trotter IGeLU 2009 Conference September 7-9, 2009 Helsinki, Finland

Quantum StorNext. Product Brief: Distributed LAN Client

Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle

Future technologies for storage networks. Shravan Pargal Director, Compellent Consulting

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

10th TF-Storage Meeting

LAN/Server Free* Backup Experiences

OSG Hadoop is packaged into rpms for SL4, SL5 by Caltech BeStMan, gridftp backend

SAN Conceptual and Design Basics

The IntelliMagic White Paper: Storage Performance Analysis for an IBM Storwize V7000

ClearPath Storage Update Data Domain on ClearPath MCP

IOmark- VDI. Nimbus Data Gemini Test Report: VDI a Test Report Date: 6, September

WHITEPAPER: Understanding Pillar Axiom Data Protection Options

How To Back Up A Computer To A Backup On A Hard Drive On A Microsoft Macbook (Or Ipad) With A Backup From A Flash Drive To A Flash Memory (Or A Flash) On A Flash (Or Macbook) On

Archival Storage At LANL Past, Present and Future

CONFIGURATION GUIDELINES: EMC STORAGE FOR PHYSICAL SECURITY

EMC Unified Storage for Microsoft SQL Server 2008

Choosing the best architecture for data protection in your Storage Area Network

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. dcache Introduction

EMC DATA DOMAIN OVERVIEW. Copyright 2011 EMC Corporation. All rights reserved.

Ultimate Guide to Oracle Storage

List of Figures and Tables

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.

EMC MIGRATION OF AN ORACLE DATA WAREHOUSE

IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE

EMC DATA PROTECTION. Backup ed Archivio su cui fare affidamento

Database Services for CERN

CERN local High Availability solutions and experiences. Thorsten Kleinwort CERN IT/FIO WLCG Tier 2 workshop CERN

Best in Class Storage Solutions for BS2000/OSD

HP reference configuration for entry-level SAS Grid Manager solutions

XenData Product Brief: SX-550 Series Servers for LTO Archives

Storage strategy and cloud storage evaluations at CERN Dirk Duellmann, CERN IT

Step by Step Guide To vstorage Backup Server (Proxy) Sizing

August Transforming your Information Infrastructure with IBM s Storage Cloud Solution

June Blade.org 2009 ALL RIGHTS RESERVED

The Modern Virtualized Data Center

Moving Virtual Storage to the Cloud

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

Scala Storage Scale-Out Clustered Storage White Paper

Experience and Lessons learnt from running High Availability Databases on Network Attached Storage

Implementing Enterprise Disk Arrays Using Open Source Software. Marc Smith Mott Community College - Flint, MI Merit Member Conference 2012

New Storage System Solutions

The GridKa Installation for HEP Computing

Overview of I/O Performance and RAID in an RDBMS Environment. By: Edward Whalen Performance Tuning Corporation

Vicom Storage Virtualization Engine. Simple, scalable, cost-effective storage virtualization for the enterprise

PolyServe Matrix Server for Linux

Data storage considerations for HTS platforms. George Magklaras -- node manager

Open-E Data Storage Software and Intel Modular Server a certified virtualization solution

EMC Virtual Infrastructure for Microsoft Applications Data Center Solution

Virtualization, Business Continuation Plan & Disaster Recovery for EMS -By Ramanj Pamidi San Diego Gas & Electric

Transcription:

Mass Storage System for Disk and Tape resources at the Tier1. Ricci Pier Paolo et al., on behalf of INFN TIER1 Storage pierpaolo.ricci@cnaf.infn.it ACAT 2008 November 3-7, 2008 Erice

Summary Tier1 Disk and Tape resources Castor status Disk SAN and GPFS TSM (tape backend for GPFS) GPFS and TSM first results 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 2

Tier1 Disk and Tape resources Here is what we have in production: Disk (SAN): ~1250 TB RAW (ATA RAID-5) 9 Infortrends A16F-R1211-M2 50TB 2 SUN STK Bladestore 80TB 4 IBM FastT900 (DS 4500) 160TB 5 SUN STK FLX600 290TB 3 DELL EMC CX-380 670TB Installation of additional 8 DELL EMC 1600TB in progress NEXT MONTH => 2.5 PBYTE 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 3

Tier1 Disk and Tape resources Tape: 2 tape robot libraries in production 1 SUN STK L5500 partitioned in 2000 slots LTO-2 (200GB) and 3500 slots 9940B (200GB) 6 LTO-2 Drives (20-30 MB/s each) 10 9940B Drives (25-30 MB/s each) 1 Pbyte Capacity 1 SUN SL8500 with 7000 slots T1000 slot (4000 tapes) 8 T1000A Drives (500GB/tape capacity and 110 MB/s bandwidth) in production 2 Pbyte Actual Capacity UPGRADE to 10000 slots and 20 T1000B Drives (1TB/tape capacity) at end 2008 => 10 Pbyte capacity 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 4

HSM (3PB) CASTOR-2 HSM Castor services servers and tapeservers STK L5500 robot (5500 TSM HSM services slots) 6 IBM LTO-2, 10 STK 9940B drives TIER1 INFN CNAF Storage STK SL8500 robot (7000 slots) 8 SUN T1000A drives Worker Nodes (LSF Batch System) Farm nodes for 9000KSPI2k RFIO RFIO,GPFS, Xroot WAN or TIER1 LAN Fibre Channel (TSM drives) SAN (~ 1250TB RAW -15/25% for NET SPACE => 1000TB) ~90 Diskservers with Qlogic FC (HBA 2340 and 2462) Fibre Channel SAN Fibre Channel 56TB RAW 32TB RAW 200TB RAW 290TB RAW 670TB RAW 4 Infortrend 5 Infortrend A16F-R1A2-M1 A16F-R1211-M2 + 4 x 3200 GByte SATA JBOD 2 x 2Gb FC interfaces 5 x 6400 GByte SATA each 2 x 2Gb FC interfaces each 2 SUN STK BladeStore 1x 24000 GByte 1250 SATA Blades 4 x 2Gb FC interfaces 4 IBM FastT900 (DS 4500) 4x43000Gbyte SATA 4 x 2Gb FC interfaces each 5 SUN STK FLX680 3 EMC CX380 5 x 46000 Gbyte 500GB FATA disks 500GB SATA Blades with 750GB FATA 4 x 2Gb FC interfaces disks 1TB SATA upgrade each 8 x 4Gb FC intefaces each 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 5

Oracle Database Service Main goals: high availability, scalability, reliability Achieved through a modular architecture based on the following building blocks: Oracle ASM volume manager for storage management implementation of redundancy and striping in an Oracle oriented way Oracle Real Application Cluster (RAC) the database is shared across several nodes with failover and load balancing capabilities (Castor with 5 instances, LCG File Catalog Atlas LHCB, Lemon, SRM) Oracle Streams geographical data redundancy for LHCB conditions database 32 server, 24 of them configured in 12 cluster 30 database instances Storage: 5TB FC Array dedicated (20TB raw) UPGRADE TO 40TB raw (installing now...) Availability rate: 98,7% in 2007 Availability (%) = Uptime/(Uptime + Target Downtime + Agent Downtime) RAC ASM 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 6

CASTOR CASTOR 2.1.7-17 deployment deployment Core services are on machines with SCSI disks, hardware RAID1, redundant power supplies SLC4 32 bit tape servers and disk servers have lower level hardware, like WNs ~ 40 CASTOR disk servers attached to a SAN full redundancy FC 2Gb/s or 4Gb/s connections (dual controller HW and Qlogic SANsurfer Path Failover SW or Vendor Specific Software) STK L5500 silos (5500 slots, 200GB cartridges, capacity ~1.1 PB ) + SL8000 silos (7000 slots, 500GB/1TB cartridges, actual capacity ~2 PB ) 24 tape drives, 3 Oracle databases (DLF, Stager, Nameserver) on ORACLE Real Application Cluster LSF plug-in for scheduling SRM v2 (2 front-ends), SRM v1 (phasing out) 15 tape servers SAN STK FlexLine 600... 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 7

CASTOR setup - SUPPORTED VO TAPE CAPACITY alice-lcg CAPACITY 20.70TB FREE 14.63TB ( 70.7%) ams CAPACITY 21.29TB FREE 200.00GB ( 0.9%) argo CAPACITY 151.46TB FREE 900.00GB ( 0.6%) argo-raw CAPACITY 49.41TB FREE 18.81TB ( 38.1%) argo-reco CAPACITY 15.62TB FREE 2.87TB ( 18.4%) atlas-lcg CAPACITY 193.36TB FREE 450.44GB ( 0.2%) cdf CAPACITY 14.84TB FREE 6.42TB ( 43.3%) cms-t1-csa07 CAPACITY 28.81TB FREE 0B ( 0.0%) cms-lcg CAPACITY 83.01TB FREE 0B ( 0.0%) cms-lcg-raw CAPACITY 58.79TB FREE 58.10GB ( 0.1%) cms-lcg-reco CAPACITY 98.05TB FREE 2.14GB ( 0.0%) lhcb-lcg CAPACITY 105.27TB FREE 443.23GB ( 0.4%) magic CAPACITY 16.60TB FREE 9.77TB ( 58.8%) pamela CAPACITY 23.44TB FREE 613.15GB ( 2.6%) POOL pamela1 CAPACITY 3.59T FREE 397.87G(10%) 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 8 virgo CAPACITY 50.98TB FREE 2.91TB ( 5.7%) - ~40 disk servers 350 TB net disk space staging area - about 5-6 fs per node, both XFS and EXT3 used, typical size 1.5-2 TB - LSF software distributed via NFS (exported by the LSF Master node) - # LSF slots: from 30 to 450, modified many times.(lower or highter values only for test ) - Many servers are used both for file transfers and for job reco/analysis => max slots limitation not very useful in such a case -SUPPORTED VO DISK STAGING CAPACITY POOL alice1 CAPACITY 25.26T FREE 21.78T(86%) POOL ams1 CAPACITY 3.53T FREE 797.32G(22%) POOL archive1 CAPACITY 94.20T FREE 69.35T(73%) POOL argo1 CAPACITY 35.02T FREE 13.18T(37%) POOL atlas1 CAPACITY 25.42T FREE 2.54T(10%) POOL atlas2 CAPACITY 14.61T FREE 4.40T(30%) POOL cms1 CAPACITY 135.74T FREE 29.60T(21%) POOL lhcb1 CAPACITY 2.69T FREE 2.55T(94%) POOL lhcb_raw1 CAPACITY 8.84T FREE 7.98T(90%)

Castor Monitoring (Lemon) Lemon is in production as a Monitoring Tool Lemon is the CERN suggested monitoring tool, strong integration with Castor v.2 Oracle10 on Real Application Cluster as database backend 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 9

STORAGE AREA NETWORK All Disk Hardware at our Tier1 is on Storage Area Network. SAN give some good advantages: diskservers could implement a No Single Point of Failure system where every component of the storage system is rendundant (storage array controllers, SAN switches, and server HBA). If software supports it, a cluster approach is possible The SAN give the best flexibility, we can dinamically assign new volumes or disk storage arrays to diskservers Monitoring tool on SAN could help to monitor i/o bandwidth on devices LAN free systems for archiving and backup purpose to the tape facilities is possible 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 10

DISK access typical case (NSPF) 12 Diskserver Dell 1950 Dual Core Biprocessors 2 x 1.6Ghz 4MB L2 Cache, 4 GByte RAM, 1066 MHz FSB SL 3.0 or 4.0 OS, Hardware Raid1 on system disks and redundant power supply LAN Gb Ethernet Connections LUN0 LUN1... 2 x 4Gb Qlogic 2460 FC redundand connections every Diskserver SAN ZONING: Each diskserver => 4 paths to the storage EMC PowerPath for Load- Balancing and Failover on the 4 paths LUN0 => /dev/sda LUN1 => /dev/sdb... Example of Application High Avaliability: GPFS with configuration Network Shared Disk 4Gb FC connections 2 Storage Processor (A e B) A1 A2 A3 A4 B1 B2B3 B4 220TB EMC CX3-80 Dual redundant Controllers (Storage Processors A,B) 4 Ouput for each SP (1,2,3,4) SUSTAINED PERFORMANCE R/W 800MByte/s (each EMC CX3-80 sys.) 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 11

GPFS implementation The idea of GPFS is to provide a fast and reliable (NSPF) diskpool storage with direct access (posix file protocol) from the Worker Nodes Farm using Block level I/O interface over network GPFS Network Shared Disk (NSD) and parallel access GPFS is a cluster, with a SAN hardware a true full NSPF is possible (diskservers failures just decrease the theorical bandwidth but the filesystem is still avaliable) One single "big filesystem" for each VO could be possible (strongly preferred by users) GPFS is widely used at our TIER1, GPFS filesystems are directly accessible from ALL the worker node in the TIER1 FARM GPFS filesystem uses parallel i/o, drastically increase end optimize the disk performances compared to single filesystem (like Castor diskpool) In GPFS v.3.2 concept of external storage pool extends use of policy driven migration/recall system to/from tape storage. GPFS is SRM v.2 compliant using INFN STORM (Storm http://storm.forge.cnaf.infn.it/) SRM interface for parallel file systems 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 12

Physical layout GPFS on SAN Farm nodes Logical layout Diskservers All diskservers accessing all disks All farm nodes accessing using LAN and NSD gpfs configuration Additional servers (i.e. front-end like gridftp) can easily be added Failure of a single server will only reduce available bandwidth to storage by factor N-1/N (N number of diskservers) Up to 8 diskserver could be assigned to a single device i.e. the filesystem will be online as long as at 1 out of 8 servers is up Bandwidth to disks could be optimized using filesystem striped over different piece of hardware Long experience at CNAF (> 3 years), ~ 27 GPFS file systems in production at CNAF (~ 720 net TB) mounted on all farm WNs 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 13

GPFS Tape Extension In GPFS v.3.2 concept of external storage pool extends use of policy driven migration/recall system to/from tape storage. The "natural" choice for managing tape storage extension for GPFS is Tivoli Storage Manager (TSM also from IBM). External pool rule defines script to call to migrate/recall/etc. files to/from the external storage manager (TSM in our case). GPFS policy engine automatically builds candidate lists and passes them to external pool scripts. External storage manager (TSM) actually moves the data. TSM installation has been under test for more than one year at CNAF TIER1 and a LHCb production testbed is in use from Spring 2008. This "Long pre-production" is due some features lacks in recall and migration policies, which is under development right now GPFS with an efficient TSM tape extension could be seen as a true Hierarchical Storage Manager facility. 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 14

TSM Agreemen with IBM to use the software until ready for full production, strong collaboration with the development team for the migration/recall optimization features Running Server Version 5.5, also beta version 6.1.0 client is installed for test purpose (better recall policies with "intelligent" queue and sorting optimization) LAN-free migration/recall to/from tape is possible. Drive should be connected to a dedicated SAN portion (Tape Area Network or TAN) TSM could also be easily used as a standard backup system for replacing our Legato Networker system TSM uses an internal database for storing filesytem metadata that could be easily duplicated. So TSM central services could be made rendundant 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 15

LHC Storage Classes at CNAF Implementation of 3 Storage Classes needed for LHC Disk0Tape1 (D0T1) CASTOR Space managed by system Data migrated to tapes and deleted from when staging area is full Disk1tape0 (D1T0) GPFS/StoRM (in production) Space managed by VO Disk1tape1 (D1T1) CASTOR (production), GPFS/StoRM (production prototype for LCHb only) Space managed by VO (i.e. if disk is full, copy fails) Large permanent buffer of disk with tape back-end and no gc 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 16

GPFS/TSM Prototype 40TB GPFS File system (v.3.2.0-3) served by 4 I/O NSD servers (SAN devices are EMC CX3-80) FC (4Gbit/s) interconnection between servers and disks array TSM v.5.5 2 servers (1Gb Ethernet) TSM front-ends each one acting as: GPFS client (reads and writes on the file-system via LAN) TSM client (reads and writes from/to tapes via FC) 3 LTO-2 tape drives Sharing of the tape library (STK L5500) between Castor e TSM i.e. working together with the same tape library direct access using TAN (tape area network) for LAN free migration/recall (using TSM storage agent) will be possible (not tested yet...) In the next slides we'll see the prototype test and the following production results 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 17

LHCb GPFS/TSM prototype and production layout GPFS FC FC SAN SAN Direct Access for TSM LAN- Free client (future test) FC FC TAN TAN GPFS Server GPFS/TSM client gridftp Server GPFS Server 1/10 Gbps Ethernet 2/4 Gbps FC 2/4 Gbps FC (Future testing) LAN TSM Server DB 2 EMC CX3-80 4 GPFS server 2 StoRM servers 2 Gridftp Servers 2 GPFS/TSM frontend nodes 3 Tape Drive LTO-2 1 TSM server 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 18

GPFS/TSM Prototype LHCb Test File size distribution Data transfer of LHCb files from CERN Castor-disk to CNAF StoRM/GPFS using the File Transfer Service (FTS) Automatic migration of the data files from GPFS to TSM while the data was being transferred by FTS This is a realistic scenario! Most of the files are of 4 and 2 GB size, with a bit of other sizes in addition data files are LHCb stripped DST 2477 files 8 TB in total 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 19

GPFS/TSM Prototype LHCb Test Red curve: net data throughput from GPFS to TSM Just two LTO-2 drives A third LTO-2 drive was added A drive was removed Black curve: net data throughput from CERN to CNAF vs. time Zero tape migration failures Zero retrials 8 TB in total were transferred to tape in 150k seconds (almost 2 days) from CERN FTS transfers were temporarily interrupted About 50 MB/s to tape with two LTO-2 drives and 65 MB/s with three LTO-2 drives 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 20

GPFS/TSM Prototype LHCb Test Retention time on disk (time since file is written until it is migrated to tape) Most of the files were migrated within less than 3 hours with a tail up to 8 hours The tail comes from the fact that at some point the CERN-to-CNAF throughput raised to 80 MiB/s, overcoming max performance of tape migration at that time. So, GPFS/TSM accumulated a queue of files with respect to the FTS transfers 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 21

GPFS/TSM LHCb Production After the good results from the test phase described in the previous slides, we decide to run the prototype in production. 40 Tbyte of D1T1 LHCb production data successfully stored About 70 MByte/s sustained No tape migration failures detected A test of complete deletion of portion of the Disk Filesystem and successive full recovery from TSM tape has been made (using the TSM metadata db) A very promising starting! 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 22

Conclusion and "What's next"? This presentation contains a site report from the INFN CNAF Tier1 Storage Group activities focusing on Database, Castor, SAN and GPFS usage at our site. In addition the presentation briefly summarizes the promising implementation of the new GPFS/TSM prototype. The GPFS/TSM prototype with the SRM StoRM interface proves itself as a good and realiable D1T1 system, LHCb is still using this system in production. Next Steps will be: A D0T1 storage class implementation of the system in collaboration with the IBM development team. Since operation of recalls becomes crucial in D0T1 systems, optimization in accessing data stored on tapes becomes of primary importance Also LAN-Free migration/recall to/from the tape facilities should be carefully tested. Using the SAN/TAN for migrating and read the data between the GPFS and TSM layers could seriously improve the performance and decrease the LAN data troughput request Thank you for the attention! 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 23

Abstract Title: Mass Storage System for Disk and Tape resources at the Tier1. Abstract: The activities in the last 5 years for the storage access at the INFN CNAF Tier1 can be enlisted under two different solutions efficiently used in production: the CASTOR software, developed by CERN, for Hierarchical Storage Manager (HSM), and the General Parallel File System (GPFS), by IBM, for the disk resource management. In addition, since last year, a promising alternative solution for the HSM, using Tivoli Storage Manager (TSM) and GPFS, has been under intensive test. This paper reports the description of the current hardware and software installation with an outlook on the last GPFS and TSM tests results. 3-Nov-08 ACAT 2008 ERICE pierpaolo.ricci@cnaf.infn.it 24