NERSC File Systems and How to Use Them
|
|
|
- Ira Owens
- 10 years ago
- Views:
Transcription
1 NERSC File Systems and How to Use Them David Turner! NERSC User Services Group! Joint Facilities User Forum on Data- Intensive Computing! June 18, 2014
2 The compute and storage systems 2014 Hopper: 1.3PF, 212 TB RAM 2.2 PB Local Scratch 70 GB/s Cray XE6, 150K Cores Edison: 2.5PF, 357 TB RAM 80 GB/s 16 x QDR IB 6.4 PB Local Scratch 140 GB/s 16 x FDR IB /global/ scratch Vis & Analy'cs, Data Transfer Nodes, Adv. Arch., Science Gateways Ethernet & IB Fabric /project 5 GB/s /home 250 TB NetApp GB/s HPSS 65 PB stored, 240 PB capacity, 40 years of community data Science Friendly Security ProducKon Monitoring Power Efficiency WAN PB DDN9900 & NexSAN 50 GB/s Cray XC30, 130K Cores Produc'on Clusters Carver, PDSF, JGI,KBASE,HEP 14x QDR 3.6 PB 5 x SFA12KE 2 x 10 Gb 1 x 100 Gb Science Data Network
3 Overview Focus on user- writable file systems Global file systems Local file systems Policies Performance PlaAorm summary Edison, Hopper, Carver - 3 -
4 Protect Your Data! Some file systems are backed up Some file systems are not backed up RestoraFon of individual files/directories may not be possible Hardware failures and human errors will happen BACK UP YOUR FILES TO HPSS! - 4 -
5 Global File Systems NERSC Global Filesystem (NGF) Based on IBM s General Parallel File System (GPFS) Architected and managed by NERSC s Storage Systems Group Provides directories for home, global scratch, and project Also provides /usr/common NERSC- supported sorware - 5 -
6 Global Homes File System Overview Provided by two ~100 TB file systems /global/u1 /global/u2 5 GB/s aggregate bandwidth Low- level name /global/u1/d/dpturner /global/u2/d/dpturner -> /global/u1/d/dpturner Be^er name /global/homes/d/dpturner Best name $HOME - 6 -
7 Global Homes Use Shared across all plaaorms $HOME/edison, $HOME/hopper, etc. dot files (.bashrc,.cshrc.ext, etc.) might contain plazorm- specific clauses if ($NERSC_HOST == edison ) then... endif Tuned for small file access Compiling/linking Job submission Do not run batch jobs in $HOME! - 7 -
8 Global Homes Policies Quotas enforced 40 GB 1,000,000 inodes Quota increases rarely (i.e., never) granted Monitor with myquota command Permanent storage No purging Backed up Hardware failures and human errors will happen BACK UP YOUR FILES TO HPSS! - 8 -
9 Global Scratch File System Overview Provides 3.6 PB high- performance disk 80 GB/s aggregate bandwidth Primary scratch file system for Carver Also mounted on Edison, Hopper, Datatran, etc. Low- level name /global/scratch2/sd/dpturner Be^er name $GSCRATCH AKA $SCRATCH on Carver and Datatran - 9 -
10 Global Scratch Use Shared across many plaaorms $GSCRATCH/carver, $GSCRATCH/edison, etc. Tuned for large streaming file access Running I/O intensive batch jobs Data analysis/visualiza'on
11 Global Scratch Policies Quotas enforced 20 TB 4,000,000 inodes Quota increases may be requested Monitor with myquota command Temporary storage Bi- weekly purges of all files that have not been accessed in over 12 weeks List of purged files in $GSCRATCH/purged.<'mestamp> Hardware failures and human errors will happen BACK UP YOUR FILES TO HPSS!
12 Project File System Overview Provides 5.1 PB high- performance disk 50 GB/s aggregate bandwidth Widely available Intended for sharing data between plaaorms, between users, or with the outside world Prior to AY14 Must be requested /project/projectdirs/bigsci Beginning AY14 Every MPP repo gets project directory /project/projectdirs/m
13 Project Use Tuned for large streaming file access Running I/O intensive batch jobs Data analysis/visualiza'on Access controlled by Unix file groups Group name usually same as directory Requires administrator (usually the PI or PI Proxy) Can also use access control list (ACL)
14 Project Policies Quotas enforced 1 TB 1,000,000 inodes Quota increases may be requested Monitor with prjquota command % prjquota bigsci Permanent storage No purging Backed up if quota <= 5 TB Hardware failures and human errors will happen BACK UP YOUR FILES TO HPSS!
15 Science Gateways on Project Make data available to outside world mkdir /project/projectdirs/bigsci/www chmod o+x /project/projectdirs/bigsci chmod o+rx /project/projectdirs/bigsci/www Access with web browser
16 Local File Systems on Edison Edison scratch file systems /scratch1 /scratch2 Each has 2.1 PB Each has 48 GB/s aggregate bandwidth /scratch3 3.2 PB 72 GB/s aggregate bandwidth Provided by Cray, based on Lustre
17 Edison Scratch Use Each user gets a scratch directory in /scratch1 or /scratch2 /scratch2/scratchdirs/dpturner $SCRATCH Access to /scratch3 must be requested Large datasets High bandwidth Tuned for large streaming file access Running I/O intensive batch jobs Data analysis/visualiza'on
18 Edison Scratch Policies Quotas enforced in $SCRATCH by submit filter 10 TB 10,000,000 inodes Quota increases may be requested Monitor with myquota command No quota enforcement in /scratch3 Temporary storage Daily purges of all files that have not been accessed in over 12 weeks List of purged files in $SCRATCH/purged.<'mestamp> Hardware failures and human errors will happen BACK UP YOUR FILES TO HPSS!
19 Local File Systems on Hopper Hopper scratch file systems /scratch /scratch2 Each has 1.0 PB Each has 35 GB/s aggregate bandwidth Provided by Cray, based on Lustre
20 Hopper Scratch Use Each user gets a scratch directory in /scratch1 and /scratch2 /scratch/scratchdirs/dpturner $SCRATCH /scratch2/scratchdirs/dpturner $SCRATCH2 Tuned for large streaming file access Running I/O intensive batch jobs Data analysis/visualiza'on
21 Hopper Scratch Policies Quotas enforced by submit filter Combined (scratch/scratch2) quotas 5 TB 5,000,000 inodes Quota increases may be requested Monitor with myquota command Temporary storage Daily purges of all files that have not been accessed in over 12 weeks List of purged files in $SCRATCH/purged.<'mestamp> Hardware failures and human errors will happen BACK UP YOUR FILES TO HPSS!
22 Long-Term File Systems Global home directories Source/object/executable files, batch scripts, input files, configura'on files, batch job summaries (not for running jobs) Backed up 40 GB permanent quota $HOME Global project directories Sharing data between people and/or systems All MPP repos have one Backed up if quota less than or equal to 5 TB 1 TB default quota
23 Short-Term File Systems Local scratch directories Cray (Edison, Hopper) only Large, high- performance parallel Lustre file system Not backed up; files purged arer 12 weeks Hopper: 5 TB default quota; Edison: 10 TB default quota $SCRATCH, $SCRATCH2 Global scratch directories All systems Large, high- performance parallel GPFS file system Not backed up; files purged arer 12 weeks 20 TB default quota $GSCRATCH
24 File System Suggestions DO NOT RUN BATCH JOBS IN $HOME Use $SCRATCH for running Edison/Hopper batch Use $GSCRATCH for running Carver batch Performance can be limited by metadata Do not store 1000s of files in single directory Use tar to conserve inodes Use HPSS to archive important data Protec'on against hardware failure Quota management DO NOT USE /tmp!
25 File Systems Summary File System Path Type Default Quota Backups Purge Policy Global Homes $HOME GPFS 40 GB / 1M inodes Yes Not purged Global Scratch $GSCRATCH GPFS 20 TB / 4M inodes No 12 weeks from last access Global Project /project/ projectdirs/ projectname Hopper Scratch $SCRATCH and $SCRATCH2 GPFS 1 TB / 1M inodes Yes, if quota less than or equal to 5TB Lustre 5 TB / 5M inodes (combined) No Not purged 12 weeks from last access Edison Scratch $SCRATCH Lustre 10 TB / 5M inodes (none in /scratch3) No 12 weeks from last access
26 Resources h^p:// and- file- systems/ h^p:// and- file- systems/ file- systems/ h^p:// systems/ edison/file- storage- and- i- o/ h^p:// systems/ hopper/file- storage- and- i- o/
27 Thank you
Data Analytics at NERSC. Joaquin Correa [email protected] NERSC Data and Analytics Services
Data Analytics at NERSC Joaquin Correa [email protected] NERSC Data and Analytics Services NERSC User Meeting August, 2015 Data analytics at NERSC Science Applications Climate, Cosmology, Kbase, Materials,
Large File System Backup NERSC Global File System Experience
Large File System Backup NERSC Global File System Experience M. Andrews, J. Hick, W. Kramer, A. Mokhtarani National Energy Research Scientific Computing Center at Lawrence Berkeley National Laboratory
New Storage System Solutions
New Storage System Solutions Craig Prescott Research Computing May 2, 2013 Outline } Existing storage systems } Requirements and Solutions } Lustre } /scratch/lfs } Questions? Existing Storage Systems
Research Technologies Data Storage for HPC
Research Technologies Data Storage for HPC Supercomputing for Everyone February 17-18, 2014 Research Technologies High Performance File Systems [email protected] Indiana University Intro to HPC on Big
How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx
Current Status of FEFS for the K computer
Current Status of FEFS for the K computer Shinji Sumimoto Fujitsu Limited Apr.24 2012 LUG2012@Austin Outline RIKEN and Fujitsu are jointly developing the K computer * Development continues with system
Lessons learned from parallel file system operation
Lessons learned from parallel file system operation Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association
Lustre SMB Gateway. Integrating Lustre with Windows
Lustre SMB Gateway Integrating Lustre with Windows Hardware: Old vs New Compute 60 x Dell PowerEdge 1950-8 x 2.6Ghz cores, 16GB, 500GB Sata, 1GBe - Win7 x64 Storage 1 x Dell R510-12 x 2TB Sata, RAID5,
Hadoop Distributed File System. T-111.5550 Seminar On Multimedia 2009-11-11 Eero Kurkela
Hadoop Distributed File System T-111.5550 Seminar On Multimedia 2009-11-11 Eero Kurkela Agenda Introduction Flesh and bones of HDFS Architecture Accessing data Data replication strategy Fault tolerance
www.thinkparq.com www.beegfs.com
www.thinkparq.com www.beegfs.com KEY ASPECTS Maximum Flexibility Maximum Scalability BeeGFS supports a wide range of Linux distributions such as RHEL/Fedora, SLES/OpenSuse or Debian/Ubuntu as well as a
NERSC Archival Storage: Best Practices
NERSC Archival Storage: Best Practices Lisa Gerhardt! NERSC User Services! Nick Balthaser! NERSC Storage Systems! Joint Facilities User Forum on Data Intensive Computing! June 18, 2014 Agenda Introduc)on
Quick Introduction to HPSS at NERSC
Quick Introduction to HPSS at NERSC Nick Balthaser NERSC Storage Systems Group [email protected] Joint Genome Institute, Walnut Creek, CA Feb 10, 2011 Agenda NERSC Archive Technologies Overview Use Cases
Cluster Implementation and Management; Scheduling
Cluster Implementation and Management; Scheduling CPS343 Parallel and High Performance Computing Spring 2013 CPS343 (Parallel and HPC) Cluster Implementation and Management; Scheduling Spring 2013 1 /
Data storage services at CC-IN2P3
Centre de Calcul de l Institut National de Physique Nucléaire et de Physique des Particules Data storage services at CC-IN2P3 Jean-Yves Nief Agenda Hardware: Storage on disk. Storage on tape. Software:
Distributed Data Storage Based on Web Access and IBP Infrastructure. Faculty of Informatics Masaryk University Brno, The Czech Republic
Distributed Data Storage Based on Web Access and IBP Infrastructure Lukáš Hejtmánek Faculty of Informatics Masaryk University Brno, The Czech Republic Summary New web based distributed data storage infrastructure
Globus and the Centralized Research Data Infrastructure at CU Boulder
Globus and the Centralized Research Data Infrastructure at CU Boulder Daniel Milroy, [email protected] Conan Moore, [email protected] Thomas Hauser, [email protected] Peter Ruprecht,
Kriterien für ein PetaFlop System
Kriterien für ein PetaFlop System Rainer Keller, HLRS :: :: :: Context: Organizational HLRS is one of the three national supercomputing centers in Germany. The national supercomputing centers are working
Cray DVS: Data Virtualization Service
Cray : Data Virtualization Service Stephen Sugiyama and David Wallace, Cray Inc. ABSTRACT: Cray, the Cray Data Virtualization Service, is a new capability being added to the XT software environment with
PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute
PADS GPFS Filesystem: Crash Root Cause Analysis Computation Institute Argonne National Laboratory Table of Contents Purpose 1 Terminology 2 Infrastructure 4 Timeline of Events 5 Background 5 Corruption
LLNL Production Plans and Best Practices
LLNL Production Plans and Best Practices Lustre User Group, 2014. Miami, FL April 8, 2014 LLNL-PRES-652453 This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore
Mendel at NERSC: Multiple Workloads on a Single Linux Cluster
Mendel at NERSC: Multiple Workloads on a Single Linux Cluster Larry Pezzaglia NERSC Computational Systems Group [email protected] CUG 2013 (May 9, 2013) Snapshot of NERSC Located at LBNL, NERSC is the
Welcome to the unit of Hadoop Fundamentals on Hadoop architecture. I will begin with a terminology review and then cover the major components
Welcome to the unit of Hadoop Fundamentals on Hadoop architecture. I will begin with a terminology review and then cover the major components of Hadoop. We will see what types of nodes can exist in a Hadoop
Using the Yale HPC Clusters
Using the Yale HPC Clusters Stephen Weston Robert Bjornson Yale Center for Research Computing Yale University Oct 2015 To get help Send an email to: [email protected] Read documentation at: http://research.computing.yale.edu/hpc-support
Data Movement and Storage. Drew Dolgert and previous contributors
Data Movement and Storage Drew Dolgert and previous contributors Data Intensive Computing Location Viewing Manipulation Storage Movement Sharing Interpretation $HOME $WORK $SCRATCH 72 is a Lot, Right?
JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert
Mitglied der Helmholtz-Gemeinschaft JUROPA Linux Cluster An Overview 19 May 2014 Ulrich Detert JuRoPA JuRoPA Jülich Research on Petaflop Architectures Bull, Sun, ParTec, Intel, Mellanox, Novell, FZJ JUROPA
NetApp High-Performance Computing Solution for Lustre: Solution Guide
Technical Report NetApp High-Performance Computing Solution for Lustre: Solution Guide Robert Lai, NetApp August 2012 TR-3997 TABLE OF CONTENTS 1 Introduction... 5 1.1 NetApp HPC Solution for Lustre Introduction...5
Monitoring Tools for Large Scale Systems
Monitoring Tools for Large Scale Systems Ross Miller, Jason Hill, David A. Dillow, Raghul Gunasekaran, Galen Shipman, Don Maxwell Oak Ridge Leadership Computing Facility, Oak Ridge National Laboratory
Maurice Askinazi Ofer Rind Tony Wong. HEPIX @ Cornell Nov. 2, 2010 Storage at BNL
Maurice Askinazi Ofer Rind Tony Wong HEPIX @ Cornell Nov. 2, 2010 Storage at BNL Traditional Storage Dedicated compute nodes and NFS SAN storage Simple and effective, but SAN storage became very expensive
Hadoop Architecture. Part 1
Hadoop Architecture Part 1 Node, Rack and Cluster: A node is simply a computer, typically non-enterprise, commodity hardware for nodes that contain data. Consider we have Node 1.Then we can add more nodes,
File Transfer Best Practices
File Transfer Best Practices David Turner User Services Group NERSC User Group Meeting October 2, 2008 Overview Available tools ftp, scp, bbcp, GridFTP, hsi/htar Examples and Performance LAN WAN Reliability
Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007
Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms Cray User Group Meeting June 2007 Cray s Storage Strategy Background Broad range of HPC requirements
Hadoop: Embracing future hardware
Hadoop: Embracing future hardware Suresh Srinivas @suresh_m_s Page 1 About Me Architect & Founder at Hortonworks Long time Apache Hadoop committer and PMC member Designed and developed many key Hadoop
Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar
Computational infrastructure for NGS data analysis José Carbonell Caballero Pablo Escobar Computational infrastructure for NGS Cluster definition: A computer cluster is a group of linked computers, working
KIT Site Report. Andreas Petzold. www.kit.edu STEINBUCH CENTRE FOR COMPUTING - SCC
KIT Site Report Andreas Petzold STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association www.kit.edu GridKa Tier 1 - Batch
Running on Blue Gene/Q at Argonne Leadership Computing Facility (ALCF)
Running on Blue Gene/Q at Argonne Leadership Computing Facility (ALCF) ALCF Resources: Machines & Storage Mira (Production) IBM Blue Gene/Q 49,152 nodes / 786,432 cores 768 TB of memory Peak flop rate:
High Performance Computing OpenStack Options. September 22, 2015
High Performance Computing OpenStack PRESENTATION TITLE GOES HERE Options September 22, 2015 Today s Presenters Glyn Bowden, SNIA Cloud Storage Initiative Board HP Helion Professional Services Alex McDonald,
Parallel Processing using the LOTUS cluster
Parallel Processing using the LOTUS cluster Alison Pamment / Cristina del Cano Novales JASMIN/CEMS Workshop February 2015 Overview Parallelising data analysis LOTUS HPC Cluster Job submission on LOTUS
Architecting a High Performance Storage System
WHITE PAPER Intel Enterprise Edition for Lustre* Software High Performance Data Division Architecting a High Performance Storage System January 2014 Contents Introduction... 1 A Systematic Approach to
Archival Storage At LANL Past, Present and Future
Archival Storage At LANL Past, Present and Future Danny Cook Los Alamos National Laboratory [email protected] Salishan Conference on High Performance Computing April 24-27 2006 LA-UR-06-0977 Main points of
Installing MooseFS Step by Step Tutorial
Installing MooseFS Step by Step Tutorial Michał Borychowski MooseFS Support Manager [email protected] march 2010 Gemius SA Overview... 3 MooseFS install process on dedicated machines... 3 Master server
GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"
GPFS Storage Server Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " Agenda" GPFS Overview" Classical versus GSS I/O Solution" GPFS Storage Server (GSS)" GPFS Native RAID
Introduction to Archival Storage at NERSC
Introduction to Archival Storage at NERSC Nick Balthaser NERSC Storage Systems Group [email protected] NERSC User Training March 8, 2011 Agenda NERSC Archive Technologies Overview Use Cases for the Archive
Cray Lustre File System Monitoring
Cray Lustre File System Monitoring esfsmon Jeff Keopp OSIO/ES Systems Cray Inc. St. Paul, MN USA [email protected] Harold Longley OSIO Cray Inc. St. Paul, MN USA [email protected] Abstract The Cray Data Management
Analisi di un servizio SRM: StoRM
27 November 2007 General Parallel File System (GPFS) The StoRM service Deployment configuration Authorization and ACLs Conclusions. Definition of terms Definition of terms 1/2 Distributed File System The
New and Improved Lustre Performance Monitoring Tool. Torben Kling Petersen, PhD Principal Engineer. Chris Bloxham Principal Architect
New and Improved Lustre Performance Monitoring Tool Torben Kling Petersen, PhD Principal Engineer Chris Bloxham Principal Architect Lustre monitoring Performance Granular Aggregated Components Subsystem
Science Gateway Services for NERSC Users
Science Gateway Services for NERSC Users Shreyas Cholia NERSC User Group Meeting October 7, 2009 Science Gateways at NERSC Web access methods to NERSC resources Much is possible beyond yesterday s ssh+pbs
SMB Direct for SQL Server and Private Cloud
SMB Direct for SQL Server and Private Cloud Increased Performance, Higher Scalability and Extreme Resiliency June, 2014 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server
IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez
IT of SPIM Data Storage and Compression EMBO Course - August 27th Jeff Oegema, Peter Steinbach, Oscar Gonzalez 1 Talk Outline Introduction and the IT Team SPIM Data Flow Capture, Compression, and the Data
An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing
An Alternative Storage Solution for MapReduce Eric Lomascolo Director, Solutions Marketing MapReduce Breaks the Problem Down Data Analysis Distributes processing work (Map) across compute nodes and accumulates
Michael Thomas, Dorian Kcira California Institute of Technology. CMS Offline & Computing Week
Michael Thomas, Dorian Kcira California Institute of Technology CMS Offline & Computing Week San Diego, April 20-24 th 2009 Map-Reduce plus the HDFS filesystem implemented in java Map-Reduce is a highly
Getting Started with HPC
Getting Started with HPC An Introduction to the Minerva High Performance Computing Resource 17 Sep 2013 Outline of Topics Introduction HPC Accounts Logging onto the HPC Clusters Common Linux Commands Storage
HPCHadoop: MapReduce on Cray X-series
HPCHadoop: MapReduce on Cray X-series Scott Michael Research Analytics Indiana University Cray User Group Meeting May 7, 2014 1 Outline Motivation & Design of HPCHadoop HPCHadoop demo Benchmarking Methodology
Quantum StorNext. Product Brief: Distributed LAN Client
Quantum StorNext Product Brief: Distributed LAN Client NOTICE This product brief may contain proprietary information protected by copyright. Information in this product brief is subject to change without
A Survey of Shared File Systems
Technical Paper A Survey of Shared File Systems Determining the Best Choice for your Distributed Applications A Survey of Shared File Systems A Survey of Shared File Systems Table of Contents Introduction...
Vital-IT Storage Guidelines
Introduction This document describes the current storage organization of Vital-IT and defines some rules about its usage. We need to re-specify the usage of the different parts of the infrastructure as
Sawmill Log Analyzer Best Practices!! Page 1 of 6. Sawmill Log Analyzer Best Practices
Sawmill Log Analyzer Best Practices!! Page 1 of 6 Sawmill Log Analyzer Best Practices! Sawmill Log Analyzer Best Practices!! Page 2 of 6 This document describes best practices for the Sawmill universal
Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC
HPC Architecture End to End Alexandre Chauvin Agenda HPC Software Stack Visualization National Scientific Center 2 Agenda HPC Software Stack Alexandre Chauvin Typical HPC Software Stack Externes LAN Typical
NERSC Data Efforts Update Prabhat Data and Analytics Group Lead February 23, 2015
NERSC Data Efforts Update Prabhat Data and Analytics Group Lead February 23, 2015-1 - A little bit about myself Computer Scien.st Brown, IIT Delhi Real- 3me Graphics, Virtual Reality, HCI Computa3onal
OLCF Best Practices. Bill Renaud OLCF User Assistance Group
OLCF Best Practices Bill Renaud OLCF User Assistance Group Overview This presentation covers some helpful information for users of OLCF Staying informed Some aspects of system usage that may differ from
Panasas at the RCF. Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory. Robert Petkus Panasas at the RCF
Panasas at the RCF HEPiX at SLAC Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory Centralized File Service Single, facility-wide namespace for files. Uniform, facility-wide
Partek Flow Installation Guide
Partek Flow Installation Guide Partek Flow is a web based application for genomic data analysis and visualization, which can be installed on a desktop computer, compute cluster or cloud. Users can access
Biowulf2 Training Session
Biowulf2 Training Session 9 July 2015 Slides at: h,p://hpc.nih.gov/docs/b2training.pdf HPC@NIH website: h,p://hpc.nih.gov System hardware overview What s new/different The batch system & subminng jobs
Hadoop on the Gordon Data Intensive Cluster
Hadoop on the Gordon Data Intensive Cluster Amit Majumdar, Scientific Computing Applications Mahidhar Tatineni, HPC User Services San Diego Supercomputer Center University of California San Diego Dec 18,
Scalable Cloud Computing Solutions for Next Generation Sequencing Data
Scalable Cloud Computing Solutions for Next Generation Sequencing Data Matti Niemenmaa 1, Aleksi Kallio 2, André Schumacher 1, Petri Klemelä 2, Eija Korpelainen 2, and Keijo Heljanko 1 1 Department of
Netapp @ 10th TF-Storage Meeting
Netapp @ 10th TF-Storage Meeting Wojciech Janusz, Netapp Poland Bogusz Błaszkiewicz, Netapp Poland Ljubljana, 2012.02.20 Agenda Data Ontap Cluster-Mode pnfs E-Series NetApp Confidential - Internal Use
Hadoop Distributed File System. Dhruba Borthakur June, 2007
Hadoop Distributed File System Dhruba Borthakur June, 2007 Goals of HDFS Very Large Distributed File System 10K nodes, 100 million files, 10 PB Assumes Commodity Hardware Files are replicated to handle
HPC @ CRIBI. Calcolo Scientifico e Bioinformatica oggi Università di Padova 13 gennaio 2012
HPC @ CRIBI Calcolo Scientifico e Bioinformatica oggi Università di Padova 13 gennaio 2012 what is exact? experience on advanced computational technologies a company lead by IT experts with a strong background
HPC Advisory Council
HPC Advisory Council September 2012, Malaga CHRIS WEEDEN SYSTEMS ENGINEER WHO IS PANASAS? Panasas is a high performance storage vendor founded by Dr Garth Gibson Panasas delivers a fully supported, turnkey,
HPCHadoop: A framework to run Hadoop on Cray X-series supercomputers
HPCHadoop: A framework to run Hadoop on Cray X-series supercomputers Scott Michael, Abhinav Thota, and Robert Henschel Pervasive Technology Institute Indiana University Bloomington, IN, USA Email: [email protected]
Integrated Grid Solutions. and Greenplum
EMC Perspective Integrated Grid Solutions from SAS, EMC Isilon and Greenplum Introduction Intensifying competitive pressure and vast growth in the capabilities of analytic computing platforms are driving
An Introduction to High Performance Computing in the Department
An Introduction to High Performance Computing in the Department Ashley Ford & Chris Jewell Department of Statistics University of Warwick October 30, 2012 1 Some Background 2 How is Buster used? 3 Software
Lustre tools for ldiskfs investigation and lightweight I/O statistics
Lustre tools for ldiskfs investigation and lightweight I/O statistics Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State Roland of Baden-Württemberg Laifer Lustre and tools
Inside Lustre HSM. An introduction to the newly HSM-enabled Lustre 2.5.x parallel file system. Torben Kling Petersen, PhD.
Inside Lustre HSM Technology Paper An introduction to the newly HSM-enabled Lustre 2.5.x parallel file system Torben Kling Petersen, PhD Introduction Hierarchical Storage Management (HSM) has been the
Lustre* is designed to achieve the maximum performance and scalability for POSIX applications that need outstanding streamed I/O.
Reference Architecture Designing High-Performance Storage Tiers Designing High-Performance Storage Tiers Intel Enterprise Edition for Lustre* software and Intel Non-Volatile Memory Express (NVMe) Storage
Introduction to Supercomputing with Janus
Introduction to Supercomputing with Janus Shelley Knuth [email protected] Peter Ruprecht [email protected] www.rc.colorado.edu Outline Who is CU Research Computing? What is a supercomputer?
TELE 301 Lecture 7: Linux/Unix file
Overview Last Lecture Scripting This Lecture Linux/Unix file system Next Lecture System installation Sources Installation and Getting Started Guide Linux System Administrators Guide Chapter 6 in Principles
The CNMS Computer Cluster
The CNMS Computer Cluster This page describes the CNMS Computational Cluster, how to access it, and how to use it. Introduction (2014) The latest block of the CNMS Cluster (2010) Previous blocks of the
Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014
Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet Anand Rangaswamy September 2014 Storage Developer Conference Mellanox Overview Ticker: MLNX Leading provider of high-throughput,
Using the Windows Cluster
Using the Windows Cluster Christian Terboven [email protected] aachen.de Center for Computing and Communication RWTH Aachen University Windows HPC 2008 (II) September 17, RWTH Aachen Agenda o Windows Cluster
Reference Design: Scalable Object Storage with Seagate Kinetic, Supermicro, and SwiftStack
Reference Design: Scalable Object Storage with Seagate Kinetic, Supermicro, and SwiftStack May 2015 Copyright 2015 SwiftStack, Inc. swiftstack.com Page 1 of 19 Table of Contents INTRODUCTION... 3 OpenStack
A parallel file system made in Germany Tiered Storage and HSM
A parallel file system made in Germany Tiered Storage and HSM Mai 7th 2012 HLRS Stuttgart Franz-Josef Pfreundt Competence Center for HPC Sven Breuner Fraunhofer Institut for Industrial Mathematics Mathematical
LUSTRE FILE SYSTEM High-Performance Storage Architecture and Scalable Cluster File System White Paper December 2007. Abstract
LUSTRE FILE SYSTEM High-Performance Storage Architecture and Scalable Cluster File System White Paper December 2007 Abstract This paper provides basic information about the Lustre file system. Chapter
Hadoop Hands-On Exercises
Hadoop Hands-On Exercises Lawrence Berkeley National Lab July 2011 We will Training accounts/user Agreement forms Test access to carver HDFS commands Monitoring Run the word count example Simple streaming
Linux Powered Storage:
Linux Powered Storage: Building a Storage Server with Linux Architect & Senior Manager [email protected] June 6, 2012 1 Linux Based Systems are Everywhere Used as the base for commercial appliances Enterprise
Deploying and managing a Visualization Farm @ Onera
Deploying and managing a Visualization Farm @ Onera Onera Scientific Day - October, 3 2012 Network and computing department (DRI), Onera P.F. Berte [email protected] Plan Onera global HPC
University of Birmingham & CLIMB GPFS User Experience. Simon Thompson Research Support, IT Services University of Birmingham
University of Birmingham & CLIMB GPFS User Experience Simon Thompson Research Support, IT Services University of Birmingham University of Birmingham Research intensive University ~19000 Undergraduate Students
The safer, easier way to help you pass any IT exams. Exam : E20-895. Backup Recovery - Avamar Expert Exam for Implementation Engineers.
http://www.51- pass.com Exam : E20-895 Title : Backup Recovery - Avamar Expert Exam for Implementation Engineers Version : Demo 1 / 7 1.An EMC Avamar customer is currently using a 2 TB Avamar Virtual Edition
Storage Challenges for Petascale Systems
Storage Challenges for Petascale Systems Dilip D. Kandlur Director, Storage Systems Research IBM Research Division 2004 IBM Corporation Outline Storage Technology Trends Implications for high performance
A Survey of Shared File Systems
Technical Paper A Survey of Shared File Systems Determining the Best Choice for your Distributed Applications Updated 22 Oct 2014 A Survey of Shared File Systems A Survey of Shared File Systems Table of
