NERSC File Systems and How to Use Them



Similar documents
Data Analytics at NERSC. Joaquin Correa NERSC Data and Analytics Services

Large File System Backup NERSC Global File System Experience

New Storage System Solutions

Research Technologies Data Storage for HPC

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

Current Status of FEFS for the K computer

Lessons learned from parallel file system operation

Lustre SMB Gateway. Integrating Lustre with Windows

Hadoop Distributed File System. T Seminar On Multimedia Eero Kurkela


NERSC Archival Storage: Best Practices

Quick Introduction to HPSS at NERSC

Cluster Implementation and Management; Scheduling

Data storage services at CC-IN2P3

Distributed Data Storage Based on Web Access and IBP Infrastructure. Faculty of Informatics Masaryk University Brno, The Czech Republic

Globus and the Centralized Research Data Infrastructure at CU Boulder

Kriterien für ein PetaFlop System

Cray DVS: Data Virtualization Service

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute

LLNL Production Plans and Best Practices

Mendel at NERSC: Multiple Workloads on a Single Linux Cluster

Welcome to the unit of Hadoop Fundamentals on Hadoop architecture. I will begin with a terminology review and then cover the major components

Using the Yale HPC Clusters

Data Movement and Storage. Drew Dolgert and previous contributors

JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert

NetApp High-Performance Computing Solution for Lustre: Solution Guide

Monitoring Tools for Large Scale Systems

Maurice Askinazi Ofer Rind Tony Wong. Cornell Nov. 2, 2010 Storage at BNL

Hadoop Architecture. Part 1

File Transfer Best Practices

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007

Hadoop: Embracing future hardware

Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar

KIT Site Report. Andreas Petzold. STEINBUCH CENTRE FOR COMPUTING - SCC

Running on Blue Gene/Q at Argonne Leadership Computing Facility (ALCF)

High Performance Computing OpenStack Options. September 22, 2015

Parallel Processing using the LOTUS cluster

Architecting a High Performance Storage System

Archival Storage At LANL Past, Present and Future

Installing MooseFS Step by Step Tutorial

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

Introduction to Archival Storage at NERSC

Cray Lustre File System Monitoring

Analisi di un servizio SRM: StoRM

New and Improved Lustre Performance Monitoring Tool. Torben Kling Petersen, PhD Principal Engineer. Chris Bloxham Principal Architect

Science Gateway Services for NERSC Users

SMB Direct for SQL Server and Private Cloud

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing

Michael Thomas, Dorian Kcira California Institute of Technology. CMS Offline & Computing Week

Getting Started with HPC

HPCHadoop: MapReduce on Cray X-series

Quantum StorNext. Product Brief: Distributed LAN Client

A Survey of Shared File Systems

Vital-IT Storage Guidelines

Sawmill Log Analyzer Best Practices!! Page 1 of 6. Sawmill Log Analyzer Best Practices

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

NERSC Data Efforts Update Prabhat Data and Analytics Group Lead February 23, 2015

OLCF Best Practices. Bill Renaud OLCF User Assistance Group

Panasas at the RCF. Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory. Robert Petkus Panasas at the RCF

Partek Flow Installation Guide

Biowulf2 Training Session

Hadoop on the Gordon Data Intensive Cluster

Scalable Cloud Computing Solutions for Next Generation Sequencing Data

10th TF-Storage Meeting

Hadoop Distributed File System. Dhruba Borthakur June, 2007

CRIBI. Calcolo Scientifico e Bioinformatica oggi Università di Padova 13 gennaio 2012

HPC Advisory Council

HPCHadoop: A framework to run Hadoop on Cray X-series supercomputers

Integrated Grid Solutions. and Greenplum

An Introduction to High Performance Computing in the Department

Lustre tools for ldiskfs investigation and lightweight I/O statistics

Inside Lustre HSM. An introduction to the newly HSM-enabled Lustre 2.5.x parallel file system. Torben Kling Petersen, PhD.

Lustre* is designed to achieve the maximum performance and scalability for POSIX applications that need outstanding streamed I/O.

Introduction to Supercomputing with Janus

TELE 301 Lecture 7: Linux/Unix file

The CNMS Computer Cluster

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

Using the Windows Cluster

Reference Design: Scalable Object Storage with Seagate Kinetic, Supermicro, and SwiftStack

A parallel file system made in Germany Tiered Storage and HSM

LUSTRE FILE SYSTEM High-Performance Storage Architecture and Scalable Cluster File System White Paper December Abstract

Hadoop Hands-On Exercises

Linux Powered Storage:

Deploying and managing a Visualization Onera

University of Birmingham & CLIMB GPFS User Experience. Simon Thompson Research Support, IT Services University of Birmingham

The safer, easier way to help you pass any IT exams. Exam : E Backup Recovery - Avamar Expert Exam for Implementation Engineers.

Storage Challenges for Petascale Systems

A Survey of Shared File Systems

Transcription:

NERSC File Systems and How to Use Them David Turner! NERSC User Services Group! Joint Facilities User Forum on Data- Intensive Computing! June 18, 2014

The compute and storage systems 2014 Hopper: 1.3PF, 212 TB RAM 2.2 PB Local Scratch 70 GB/s Cray XE6, 150K Cores Edison: 2.5PF, 357 TB RAM 80 GB/s 16 x QDR IB 6.4 PB Local Scratch 140 GB/s 16 x FDR IB /global/ scratch Vis & Analy'cs, Data Transfer Nodes, Adv. Arch., Science Gateways Ethernet & IB Fabric /project 5 GB/s /home 250 TB NetApp 5460 12 GB/s HPSS 65 PB stored, 240 PB capacity, 40 years of community data Science Friendly Security ProducKon Monitoring Power Efficiency WAN - 2-5 PB DDN9900 & NexSAN 50 GB/s Cray XC30, 130K Cores Produc'on Clusters Carver, PDSF, JGI,KBASE,HEP 14x QDR 3.6 PB 5 x SFA12KE 2 x 10 Gb 1 x 100 Gb Science Data Network

Overview Focus on user- writable file systems Global file systems Local file systems Policies Performance PlaAorm summary Edison, Hopper, Carver - 3 -

Protect Your Data! Some file systems are backed up Some file systems are not backed up RestoraFon of individual files/directories may not be possible Hardware failures and human errors will happen BACK UP YOUR FILES TO HPSS! - 4 -

Global File Systems NERSC Global Filesystem (NGF) Based on IBM s General Parallel File System (GPFS) Architected and managed by NERSC s Storage Systems Group Provides directories for home, global scratch, and project Also provides /usr/common NERSC- supported sorware - 5 -

Global Homes File System Overview Provided by two ~100 TB file systems /global/u1 /global/u2 5 GB/s aggregate bandwidth Low- level name /global/u1/d/dpturner /global/u2/d/dpturner -> /global/u1/d/dpturner Be^er name /global/homes/d/dpturner Best name $HOME - 6 -

Global Homes Use Shared across all plaaorms $HOME/edison, $HOME/hopper, etc. dot files (.bashrc,.cshrc.ext, etc.) might contain plazorm- specific clauses if ($NERSC_HOST == edison ) then... endif Tuned for small file access Compiling/linking Job submission Do not run batch jobs in $HOME! - 7 -

Global Homes Policies Quotas enforced 40 GB 1,000,000 inodes Quota increases rarely (i.e., never) granted Monitor with myquota command Permanent storage No purging Backed up Hardware failures and human errors will happen BACK UP YOUR FILES TO HPSS! - 8 -

Global Scratch File System Overview Provides 3.6 PB high- performance disk 80 GB/s aggregate bandwidth Primary scratch file system for Carver Also mounted on Edison, Hopper, Datatran, etc. Low- level name /global/scratch2/sd/dpturner Be^er name $GSCRATCH AKA $SCRATCH on Carver and Datatran - 9 -

Global Scratch Use Shared across many plaaorms $GSCRATCH/carver, $GSCRATCH/edison, etc. Tuned for large streaming file access Running I/O intensive batch jobs Data analysis/visualiza'on - 10 -

Global Scratch Policies Quotas enforced 20 TB 4,000,000 inodes Quota increases may be requested Monitor with myquota command Temporary storage Bi- weekly purges of all files that have not been accessed in over 12 weeks List of purged files in $GSCRATCH/purged.<'mestamp> Hardware failures and human errors will happen BACK UP YOUR FILES TO HPSS! - 11 -

Project File System Overview Provides 5.1 PB high- performance disk 50 GB/s aggregate bandwidth Widely available Intended for sharing data between plaaorms, between users, or with the outside world Prior to AY14 Must be requested /project/projectdirs/bigsci Beginning AY14 Every MPP repo gets project directory /project/projectdirs/m9999-12 -

Project Use Tuned for large streaming file access Running I/O intensive batch jobs Data analysis/visualiza'on Access controlled by Unix file groups Group name usually same as directory Requires administrator (usually the PI or PI Proxy) Can also use access control list (ACL) - 13 -

Project Policies Quotas enforced 1 TB 1,000,000 inodes Quota increases may be requested Monitor with prjquota command % prjquota bigsci Permanent storage No purging Backed up if quota <= 5 TB Hardware failures and human errors will happen BACK UP YOUR FILES TO HPSS! - 14 -

Science Gateways on Project Make data available to outside world mkdir /project/projectdirs/bigsci/www chmod o+x /project/projectdirs/bigsci chmod o+rx /project/projectdirs/bigsci/www Access with web browser http://portal.nersc.gov/project/bigsci - 15 -

Local File Systems on Edison Edison scratch file systems /scratch1 /scratch2 Each has 2.1 PB Each has 48 GB/s aggregate bandwidth /scratch3 3.2 PB 72 GB/s aggregate bandwidth Provided by Cray, based on Lustre - 16 -

Edison Scratch Use Each user gets a scratch directory in /scratch1 or /scratch2 /scratch2/scratchdirs/dpturner $SCRATCH Access to /scratch3 must be requested Large datasets High bandwidth Tuned for large streaming file access Running I/O intensive batch jobs Data analysis/visualiza'on - 17 -

Edison Scratch Policies Quotas enforced in $SCRATCH by submit filter 10 TB 10,000,000 inodes Quota increases may be requested Monitor with myquota command No quota enforcement in /scratch3 Temporary storage Daily purges of all files that have not been accessed in over 12 weeks List of purged files in $SCRATCH/purged.<'mestamp> Hardware failures and human errors will happen BACK UP YOUR FILES TO HPSS! - 18 -

Local File Systems on Hopper Hopper scratch file systems /scratch /scratch2 Each has 1.0 PB Each has 35 GB/s aggregate bandwidth Provided by Cray, based on Lustre - 19 -

Hopper Scratch Use Each user gets a scratch directory in /scratch1 and /scratch2 /scratch/scratchdirs/dpturner $SCRATCH /scratch2/scratchdirs/dpturner $SCRATCH2 Tuned for large streaming file access Running I/O intensive batch jobs Data analysis/visualiza'on - 20 -

Hopper Scratch Policies Quotas enforced by submit filter Combined (scratch/scratch2) quotas 5 TB 5,000,000 inodes Quota increases may be requested Monitor with myquota command Temporary storage Daily purges of all files that have not been accessed in over 12 weeks List of purged files in $SCRATCH/purged.<'mestamp> Hardware failures and human errors will happen BACK UP YOUR FILES TO HPSS! - 21 -

Long-Term File Systems Global home directories Source/object/executable files, batch scripts, input files, configura'on files, batch job summaries (not for running jobs) Backed up 40 GB permanent quota $HOME Global project directories Sharing data between people and/or systems All MPP repos have one Backed up if quota less than or equal to 5 TB 1 TB default quota - 22 -

Short-Term File Systems Local scratch directories Cray (Edison, Hopper) only Large, high- performance parallel Lustre file system Not backed up; files purged arer 12 weeks Hopper: 5 TB default quota; Edison: 10 TB default quota $SCRATCH, $SCRATCH2 Global scratch directories All systems Large, high- performance parallel GPFS file system Not backed up; files purged arer 12 weeks 20 TB default quota $GSCRATCH - 23 -

File System Suggestions DO NOT RUN BATCH JOBS IN $HOME Use $SCRATCH for running Edison/Hopper batch Use $GSCRATCH for running Carver batch Performance can be limited by metadata Do not store 1000s of files in single directory Use tar to conserve inodes Use HPSS to archive important data Protec'on against hardware failure Quota management DO NOT USE /tmp! - 24 -

File Systems Summary File System Path Type Default Quota Backups Purge Policy Global Homes $HOME GPFS 40 GB / 1M inodes Yes Not purged Global Scratch $GSCRATCH GPFS 20 TB / 4M inodes No 12 weeks from last access Global Project /project/ projectdirs/ projectname Hopper Scratch $SCRATCH and $SCRATCH2 GPFS 1 TB / 1M inodes Yes, if quota less than or equal to 5TB Lustre 5 TB / 5M inodes (combined) No Not purged 12 weeks from last access Edison Scratch $SCRATCH Lustre 10 TB / 5M inodes (none in /scratch3) No 12 weeks from last access - 25 -

Resources h^p://www.nersc.gov/users/data- and- file- systems/ h^p://www.nersc.gov/users/data- and- file- systems/ file- systems/ h^p://www.nersc.gov/users/computafonal- systems/ edison/file- storage- and- i- o/ h^p://www.nersc.gov/users/computafonal- systems/ hopper/file- storage- and- i- o/ - 26 -

Thank you. - 27 -