NERSC Archival Storage: Best Practices



Similar documents
Quick Introduction to HPSS at NERSC

Research Technologies Data Storage for HPC

HPSS Best Practices. Erich Thanhardt Bill Anderson Marc Genty B

Introduction to Archival Storage at NERSC

Archival Storage At LANL Past, Present and Future

NERSC File Systems and How to Use Them

Storage Systems: 2014 and beyond. Jason Hick! Storage Systems Group!! NERSC User Group Meeting! February 6, 2014

Data Deduplication in Tivoli Storage Manager. Andrzej Bugowski Spała

Data Management in the Cloud: Limitations and Opportunities. Annies Ductan

Understanding Disk Storage in Tivoli Storage Manager

Storage Capacity Expansion Plan (initial)

GPN - What is theGPFS HSI HTAR ISH?

High Performance Storage System. Overview

Welcome Introduc.ons Poll members: who has licensed vsphere, who has dedicated storage array. Who has HyperV

Using RDBMS, NoSQL or Hadoop?

Reinvent your storage infrastructure for e-business

CHAPTER 17: File Management

Offensive & Defensive & Forensic Techniques for Determining Web User Iden<ty

Large File System Backup NERSC Global File System Experience

AFS Usage and Backups using TiBS at Fermilab. Presented by Kevin Hill

Data Warehousing. Yeow Wei Choong Anne Laurent

Amazon Cloud Storage Options

XenData Archive Series Software Technical Overview

OPTIMIZING EXCHANGE SERVER IN A TIERED STORAGE ENVIRONMENT WHITE PAPER NOVEMBER 2006

The Archival Upheaval Petabyte Pandemonium Developing Your Game Plan Fred Moore President

XenData Product Brief: SX-550 Series Servers for LTO Archives

Map- reduce, Hadoop and The communica3on bo5leneck. Yoav Freund UCSD / Computer Science and Engineering

YOUR PROCESS MANAGEMENT AND CONTROLLING SUITE FOR MULTI-CHANNEL ONLINE MARKETING.!

Introduction Disks RAID Tertiary storage. Mass Storage. CMSC 412, University of Maryland. Guest lecturer: David Hovemeyer.

Kaseya Fundamentals Workshop DAY THREE. Developed by Kaseya University. Powered by IT Scholars

Intro to AWS: Storage Services

Everything you need to know about flash storage performance

Data Preservation Storage Systems HAMBURG ZEUTHEN

Neil Stobart Cloudian Inc. CLOUDIAN HYPERSTORE Smart Data Storage

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. dcache Introduction

Data Management using Hierarchical Storage Management (HSM) with 3-Tier Storage Architecture

Media Cloud Building Practicalities

Data Center Evolu.on and the Cloud. Paul A. Strassmann George Mason University November 5, 2008, 7:20 to 10:00 PM

Data Management. Network transfers

Tape or Disk Backup Which is Best?

Disk Library for mainframe - DLm6000 Product Overview

Michał Jankowski Maciej Brzeźniak PSNC

Update on the Cloud Demonstration Project

HSI Best Practices for NERSC Users

Storing Data: Disks and Files

Long term retention and archiving the challenges and the solution

Sibling Rivalries: Integrating Data Management Technologies

U"lizing the SDSC Cloud Storage Service

Data storage services at CC-IN2P3

XenData Product Brief: SX-250 Archive Server for LTO

State of Tennessee. Enterprise Storage, Backup / Recovery, and Tape Virtualization. Category: Business Continuity and Disaster Recovery

AXF Archive exchange Format: Interchange & Interoperability for Operational Storage and Long-Term Preservation

OLCF Best Practices. Bill Renaud OLCF User Assistance Group

Creating a Catalog for ILM Services. Bob Mister Rogers, Application Matrix Paul Field, Independent Consultant Terry Yoshii, Intel

UNINETT Sigma2 AS: architecture and functionality of the future national data infrastructure

AppLogic and the Mainframe: The Ul7mate Private Cloud

EMC DISK LIBRARY FOR MAINFRAME

Faster Cloud Backup: How It's Done

Top Ten Questions. to Ask Your Primary Storage Provider About Their Data Efficiency. May Copyright 2014 Permabit Technology Corporation

IBM TotalStorage IBM TotalStorage Virtual Tape Server

IBM Tivoli Storage Manager Version Introduction to Data Protection Solutions IBM

Best Prac*ces for Deploying Oracle So6ware on Virtual Compute Appliance

Cloud Bursting. HPC for Wall Street Conference April 4, 2016 Scott Jeschonek, Director of Cloud Products at Avere Systems.

In-memory Tables Technology overview and solutions

Virtual Tape Systems for IBM Mainframes A comparative analysis

PetaLibrary Storage Service MOU

Clusters in the Cloud

Data Management Best Practices

The DATA Difference Targe.ng for Stronger ROI!

High Performance Storage System Interfaces. Harry Hulen

FTC Data Security Standard

IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE

Introduc)on of Pla/orm ISF. Weina Ma

Mass Storage Structure

CS 91: Cloud Systems & Datacenter Networks Failures & Replica=on

Architec;ng Splunk for High Availability and Disaster Recovery

Engagement Strategies for Emerging Big Data Collaborations

ILM: Tiered Services & The Need For Classification

PARALLELS CLOUD STORAGE

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

Migrating to Hosted Telephony. Your ultimate guide to migrating from on premise to hosted telephony.

<Insert Picture Here> RMAN Configuration and Performance Tuning Best Practices

Flash 101. Violin Memory Switzerland. Violin Memory Inc. Proprietary 1

Outline. File Management Tanenbaum, Chapter 4. Files. File Management. Objectives for a File Management System

Hadoop Distributed File System. Dhruba Borthakur Apache Hadoop Project Management Committee

Introduction to Optical Archiving Library Solution for Long-term Data Retention

Ubuntu, FEAP, and Virtualiza3on. Jonathan Wong Lab Mee3ng 11/08/10

AFS Usage and Backups using TiBS at Fermilab. Presented by Kevin Hill

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007

Performance Management in Big Data Applica6ons. Michael Kopp, Technology

Practical Solutions for Big Data Analytics

Effective Planning and Use of TSM V6 Deduplication

Lustre* is designed to achieve the maximum performance and scalability for POSIX applications that need outstanding streamed I/O.

DNS must be up and running. Both the Collax server and the clients to be backed up must be able to resolve the FQDN of the Collax server correctly.

Scientific Storage at FNAL. Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015

QStar White Paper. Tiered Storage

ATA DRIVEN GLOBAL VISION CLOUD PLATFORM STRATEG N POWERFUL RELEVANT PERFORMANCE SOLUTION CLO IRTUAL BIG DATA SOLUTION ROI FLEXIBLE DATA DRIVEN V

Texas Digital Government Summit. Data Analysis Structured vs. Unstructured Data. Presented By: Dave Larson

File Management. Chapter 12

Physical Data Organization

Transcription:

NERSC Archival Storage: Best Practices Lisa Gerhardt! NERSC User Services! Nick Balthaser! NERSC Storage Systems! Joint Facilities User Forum on Data Intensive Computing! June 18, 2014

Agenda Introduc)on to Archival Storage Archive vs. backup Data lifecycle management and 9ered storage Features of the NERSC archive NERSC user case study Access methods (clients) Op)mizing Archival Storage Tape IO characteris9cs Storage and retrieval strategies Data Redundancy and Integrity 3-2- 1 Rule Checksums - 2 -

What is an archive? Long- term data storage Data that is no longer modified or regularly accessed OKen the only copy of the data Valuable data to be kept for the long- term An archive is not a backup A backup is a copy of produc9on data If a backup is lost, produc9on data is s9ll online Value and reten9on of backup data is short- term Main purpose of a backup is fast recovery NERSC archive has files da)ng back to the 1970s NERSC began using HPSS storage system sokware in 1998 Data migrated from previous archive systems including CFS and Unitree - 3 -

Why should I use an archive? Data growth is exponen)al File system space is finite 80% of stored data is never accessed aker 90 days Cost of storing infrequently accessed data on flash or spinning disk is prohibi9ve Important, but less frequently accessed data should be stored in an archive to free higher performing resources for processing workload - 4 -

Data Lifecycle Management Manage data according to access palerns and media cost At NERSC Scratch File System High capacity, fast access, high IO throughput Temporary storage of large data files and compute output Regular purges, not backed up Project File System Medium capacity, medium- term storage of shared data Designed for file sharing - 5 - HPSS Tape- backed storage system NERSC users are responsible for their own data management

Features of the NERSC archive NERSC implements an online or ac)ve archive Parallel high- speed transfer and fast data access Data is transferred over parallel connec9ons on the NERSC internal 10Gb network Access to first byte in seconds or minutes as opposed to hours or days Tiered internal storage facilitates high speed data access: Ini9al data ingest to high- performance disk cache Data migrated to automated enterprise tape system and managed by HSM sokware (HPSS) based on file age and usage Indefinite data reten9on policy The archive is a shared mul)- user system No batch system. Inefficient use affects others. Session limits enforced - 6 -

The Archive is an HSM The NERSC archive is a Hierarchical Storage Management system (HSM) Highest performance requirements and access characteris)cs at top level Lowest cost, greatest capacity at lower levels Migra)on between levels is automa)c based on access palerns Local Disk or Tape Fast Disk High Capacity Disk Remote Disk or Tape Capacity HPSS performs differently than a file system Metadata in transac9onal DB (IBM DB2) for integrity Time to move data between hierarchies - 7 -

NERSC Archive: HPSS Currently holds about 60 PB. Total capacity is 240 PB. - 8 -

HPSS is Heavily Used - 9 -

NERSC User Case Study: Daya Bay Neutrino oscilla)on experiment in China High precision measurement of neutrino oscilla)on parameter Science Magazine s top ten breakthroughs of 2012 Data analysis and simula)on done primarily on PDSF (HEP / NP cluster) at NERSC - 10 -

NERSC User Case Study: Daya Bay NERSC is their US Tier 1 facility Archive of raw data on HPSS Data copied to HPSS within 15 20 min. of a run finishing Yearly data analysis Constant upload rate of about 220 GB / day - 11 -

Accessing NERSC HPSS HSI Fast, parallel transfers, unix shell- like interface HTAR Parallel, put tar file directly into HPSS, excellent for groups of small files FTP/PFTP Standard and high- performance parallel FTP (NERSC plahorms) gridftp Grid (GSI) authen9ca9on Enables 3 rd - party transfer Globus Web- enabled reliable transfer - 12 -

Optimizing Archival Storage - 13 -

Tape IO Characteristics Ul)mately all data in the archive is wrilen to tape Tape is sequen)al (linear) media Behaves differently than disk media: Data cannot be re- wrijen in place, it is appended aker the end of exis9ng data Reading and wri9ng are sequen9al opera9ons no random access Tape drives behave differently than disk drives Take 9me to seek to file loca9ons on tape Take 9me to ramp up to full speed Tape drives pause aker reading or wri9ng each file (file sync) HPSS does not respond like a normal file system Presents itself as one, but some opera9ons can have unexpected results - 14 -

Reading from Tape Loading a tape into a drive and posi)oning to the beginning of data are the slowest system ac)vi)es Average 9me to data on current drive technology is 45s, +15 30 sec cartridge load Reading a few large files from tape is rela9vely quick up to 400MB/sec Reading many small files stored on mul9ple tapes is slow Minimize tape mounts and posi)oning ac)vity for best read performance - 15 -

Size Matters Sweet Spot Tape drives perform best when opera9ng at full rate ( streaming ) for long dura9ons Large file IO is best for tape drive performance Small file IO causes frequent pauses and low- speed opera9ons File bundles in the 100s of GB currently provide best performance Aggregate (bundle) small files for op)mal storage Use HTAR, GNU tar, and/or zip to bundle groups of small files together to op9mize tape and network performance There is such a thing as too big Long- running transfers can be failure- prone Files spanning mul9ple tapes may incur tape mount delays - 16 -

Tape Ordering for Non-optimal Retrieval Workloads When retrieving data from the archive: Minimize tape mounts Retrieve all files on a par9cular tape while it is mounted, before retrieving files on subsequent tapes Minimize tape posi)oning Retrieve files resident on a par9cular tape in order of tape posi9on Contact NERSC Consul)ng for details of procedure - 17 -

Using a Shared Storage Resource No exclusive access to the archive No batch system Inefficient use affects performance for everyone The archive relies on mechanical devices Robots, tape drives, tape cartridges Limited number of drives and robots to serve requests Avoid administra)ve ac)on Bundle small files together/avoid excessive writes Order your retrievals Avoid excessive transfer failures - 18 -

Data Redundancy Use the 3-2- 1 Rule for cri)cal data: 3 copies of the data (in different places) 2 different formats 1 copy off- site Data loss in the NERSC archive is rare but it can happen: No offsite backup for NERSC archive data Archive data is single- copy by default No Trash Can deleted files are gone! Make mul)ple copies of data you care about If off- site copies are not possible at least store mul9ple archive copies several days apart (different tapes) Contact NERSC Consul9ng if dual- copy is a persistent requirement (data charges apply) We take great care to preserve archival data But data loss can s9ll happen! - 19 -

Data Integrity Corrup)on Any unintended change to data Rare but it can happen any point during reading, wri9ng, transfer, or processing is subject to unintended change Bad network interface, disk controller, computer memory, etc. at any point in data life9me can introduce an error Corrup9on is data loss make mul9ple copies Use checksums for cri)cal data Record checksums before storing cri9cal data in the archive, check aker retrieving Many checksum methods including HSI op9ons Checksums incur performance penalty due to CPU load check with system administrators before running checksums Interrupted/failed transfers are not data corrup)on Check transfer return codes! - 20 -

Asking Questions, Reporting Problems Contact NERSC Consul)ng Toll- free 800-666- 3772 510-486- 8611, #3 Email consult@nersc.gov. - 21 -

Further Reading NERSC Website Archive documenta9on: h/p://www.nersc.gov/users/data- and- file- systems/hpss/ge:ng- started/ Data management strategy and policies: h/p://www.nersc.gov/users/data- and- file- systems/policies/ Accessing HPSS h/p://www.nersc.gov/users/data- and- file- systems/hpss/ge:ng- started/accessing- hpss/ HSI and HTAR man pages are installed on NERSC compute pladorms Gleicher Enterprises Online Documenta)on (HSI, HTAR) h/p://www.mgleicher.us/index.html/hsi/ h/p://www.mgleicher.us/index.html/htar/ HSI Best Prac-ces for NERSC Users, LBNL Report #LBNL- 4745E h/p://www.nersc.gov/assets/pubs_presos/hsibestpraceces- Balthaser- Hazen- 2011-06- 09.pdf - 22 -

Thank you. - 23 -

Tape Ordering for Non-optimal Workloads Minimize tape mounts Retrieve all files on a par9cular tape while it is mounted, before retrieving files on subsequent tapes Minimize tape posi)oning Retrieve files resident on a par9cular tape in order of tape posi9on Find volume name and tape posi)on for every file: HSI ls - X or ls - P arguments: A:/home/n/nickb- > ls - P z.tar z.tar.idx FILE /home/n/nickb/z.tar 464566784 464566784 3095+0 EP251400 5 0 1 07/03/2013 15:00:57 07/03/2013 15:01:01 FILE /home/n/nickb/z.tar.idx 90912 90912 608+0 EF202900 4 0 1 07/03/2013 15:01:01 07/03/2013 15:01:02 Posi)on Volume - 24 -

Retrieve Files in Tape and Position Order Generate per- volume file lists in tape posi)on order Using HSI, generate lists of files per tape and sort in ascending posi9on Put file path names in HSI command file using HSI get syntax: bash$ cat./ef2092.cmd get z.tar.idx : /home/n/nickb/z.tar.idx get x.tar : /home/n/nickb/x.tar get x.tar.idx : /home/n/nickb/x.tar.idx quit Retrieve files in per- volume lists using HSI command file Use HSI in cmd_file syntax: hsi q "in./ef2092.cmd Username: nickb UID: 33065 Acct: 33065(33065) Copies: 1 Firewall: on [hsi.4.0.1.2 Thu Oct 25 16:31:52 PDT 2012][V4.0.1.2_2012_10_22.02] get z.tar.idx : /home/n/nickb/z.tar.idx get 'z.tar.idx' : '/home/n/nickb/z.tar.idx' (2013/07/03 15:01:02 90912 bytes, 1523.1 KBS ) get x.tar : /home/n/nickb/x.tar get 'x.tar' : '/home/n/nickb/x.tar' (2013/07/05 13:57:06 63488 bytes, 1563.4 KBS ) Contact NERSC Consul)ng for full procedure - 25 -