Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007



Similar documents
Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms

BlueArc unified network storage systems 7th TF-Storage Meeting. Scale Bigger, Store Smarter, Accelerate Everything

June Blade.org 2009 ALL RIGHTS RESERVED

Cray DVS: Data Virtualization Service

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

SGI Solutions for RDSI/CAUDIT 2013 SGI

RFP-MM Enterprise Storage Addendum 1

New Storage System Solutions

Lessons learned from parallel file system operation

EMC Backup and Recovery for Microsoft Exchange 2007 SP2

SoftLayer Fundamentals. Storage and Backup. August, 2014

Cluster Implementation and Management; Scheduling

Storage Systems Performance Testing

PARALLELS CLOUD STORAGE

Large Scale Storage. Orlando Richards, Information Services LCFG Users Day, University of Edinburgh 18 th January 2013

Addendum No. 1 to Packet No Enterprise Data Storage Solution and Strategy for the Ingham County MIS Department

Scalable NAS for Oracle: Gateway to the (NFS) future

How To Back Up A Computer To A Backup On A Hard Drive On A Microsoft Macbook (Or Ipad) With A Backup From A Flash Drive To A Flash Memory (Or A Flash) On A Flash (Or Macbook) On

The Panasas Parallel Storage Cluster. Acknowledgement: Some of the material presented is under copyright by Panasas Inc.

Product Overview. UNIFIED COMPUTING Managed Hosting - Storage Data Sheet

Application Brief: Using Titan for MS SQL

Building Storage Service in a Private Cloud

Maurice Askinazi Ofer Rind Tony Wong. Cornell Nov. 2, 2010 Storage at BNL

(Scale Out NAS System)

Archival Storage At LANL Past, Present and Future

The Modern Virtualized Data Center

Virtualization, Business Continuation Plan & Disaster Recovery for EMS -By Ramanj Pamidi San Diego Gas & Electric

Optimizing Large Arrays with StoneFly Storage Concentrators

Server and Storage Virtualization with IP Storage. David Dale, NetApp

High Performance Computing Specialists. ZFS Storage as a Solution for Big Data and Flexibility

Scala Storage Scale-Out Clustered Storage White Paper

HP StorageWorks Enterprise File Services Clustered Gateway. Technical Product Review. Sprawl Makes Inexpensive Servers Expensive

Distributed File System Choices: Red Hat Storage, GFS2 & pnfs

10th TF-Storage Meeting

Moving Virtual Storage to the Cloud. Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage

Panasas at the RCF. Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory. Robert Petkus Panasas at the RCF

Ultimate Guide to Oracle Storage

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

Four Reasons To Start Working With NFSv4.1 Now

Sun Storage Perspective & Lustre Architecture. Dr. Peter Braam VP Sun Microsystems

High Performance Server SAN using Micron M500DC SSDs and Sanbolic Software

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.

Enhancements of ETERNUS DX / SF

List of Figures and Tables

Software to Simplify and Share SAN Storage Sanbolic s SAN Storage Enhancing Software Portfolio

<Insert Picture Here> Refreshing Your Data Protection Environment with Next-Generation Architectures

THE EMC ISILON STORY. Big Data In The Enterprise. Copyright 2012 EMC Corporation. All rights reserved.

Moving Virtual Storage to the Cloud

EMC Unified Storage for Microsoft SQL Server 2008

Evaluation of Enterprise Data Protection using SEP Software

Sun in HPC. Update for IDC HPC User Forum Tucson, AZ, Sept 2008

How to Choose your Red Hat Enterprise Linux Filesystem

Storage and File Services Overview

Sanbolic s SAN Storage Enhancing Software Portfolio

SAN Conceptual and Design Basics

Enterprise Class Storage System, Unified NAS. Enterprise Tape Library & Virtual Tape Library. SAN Fabric Management along with SAN Switches

Pricing - overview of available configurations

IBM System x GPFS Storage Server

SAM-FS - Advanced Storage Management Solutions for High Performance Computing Environments

Lab Validation Report

Best Practices for Data Sharing in a Grid Distributed SAS Environment. Updated July 2010

Getting performance & scalability on standard platforms, the Object vs Block storage debate. Copyright 2013 MPSTOR LTD. All rights reserved.

Scientific Storage at FNAL. Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015

Hitachi NAS Platform and Hitachi Content Platform with ESRI Image

The safer, easier way to help you pass any IT exams. Exam : Storage Sales V2. Title : Version : Demo 1 / 5

Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle

DAS, NAS or SAN: Choosing the Right Storage Technology for Your Organization

THE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC

How To Design A Data Centre

Key Messages of Enterprise Cluster NAS Huawei OceanStor N8500

Data management challenges in todays Healthcare and Life Sciences ecosystems

Data storage considerations for HTS platforms. George Magklaras -- node manager

OPTIMIZING PRIMARY STORAGE WHITE PAPER FILE ARCHIVING SOLUTIONS FROM QSTAR AND CLOUDIAN

The IntelliMagic White Paper: Storage Performance Analysis for an IBM Storwize V7000

EMC Backup and Recovery for Microsoft SQL Server

FlexArray Virtualization

VMware vsphere 5.1 Advanced Administration

Taking Linux File and Storage Systems into the Future. Ric Wheeler Director Kernel File and Storage Team Red Hat, Incorporated

Overview of I/O Performance and RAID in an RDBMS Environment. By: Edward Whalen Performance Tuning Corporation

Enterprise Storage Solution for Hyper-V Private Cloud and VDI Deployments using Sanbolic s Melio Cloud Software Suite April 2011

EMC Business Continuity for Microsoft SQL Server 2008

VMware vsphere 5.0 Boot Camp

Milestone Solution Partner IT Infrastructure Components Certification Summary

The BIG Data Era has. your storage! Bratislava, Slovakia, 21st March 2013

The Methodology Behind the Dell SQL Server Advisor Tool

Scalable filesystems boosting Linux storage solutions

HITACHI VIRTUAL STORAGE PLATFORM FAMILY MATRIX

Measuring Interface Latencies for SAS, Fibre Channel and iscsi

Future technologies for storage networks. Shravan Pargal Director, Compellent Consulting

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Get Success in Passing Your Certification Exam at first attempt!

Transcription:

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms Cray User Group Meeting June 2007

Cray s Storage Strategy Background Broad range of HPC requirements big file I/O, small file I/O, scalability across multiple dimensions, data management, heterogeneous access Rate of improvement in I/O performance lags significantly behind Moore s Law Direction Move away from one solution fits all approach Use cluster file system for supercomputer scratch space and focus on high performance Use scalable NAS, combined with data management tools and new hardware technologies for shared and managed storage May 07 Cray Technical Review 1

Cray Advanced Storage Architecture (CASA) Servers, Clusters, and Workstations High Performance Scratch Files: Shared file system RAID disk Fast NAS Fast NAS FC fabric Fast NAS Fast NAS 10G/1G Ethernet Network HSM/ Backup Server HSM/ Backup Server Virtual Tape Archive HSM/Backup Tape Robot/ Archive May 07 Cray Technical Review 2

CASA Partners Servers, Clusters, and Workstations High Performance Scratch Files: Shared file system RAID disk CFS Lustre DDN, Engenio RAID disk Fast NAS Fast NAS FC fabric Fast NAS Fast NAS 10G/1G Ethernet Network HSM/ Backup Server HSM/ Backup Server ADIC HSM Virtual Tape Archive TBD backup COPAN VTL (MAID) BlueArc scalable NAS HSM/Backup Tape Robot/ Archive ADIC tapes May 07 Cray Technical Review 3

CASA Lab Chippewa Falls Opteron Cluster Cisco 6509 switch Engenio Storage COPAN MAID May 07 Cray Technical Review 4

CASA Lab CASA Lab Opteron Cluster Related Storage May 07 Cray Technical Review 5

CASA Lab Blue Arc Titan Servers Engenio Storage May 07 Cray Technical Review 6

Blue Arc Titan-2 Dual Heads May 07 Cray Technical Review 7

CASA Lab COPAN Revolution System May 07 Cray Technical Review 8

How will this help? Use a cluster file system for big file (bandwidth) I/O for scalable systems Focus on performance for applications Use commercial NAS products to provide solid storage for home directories and shared files Vendors looking at NFS performance, scalability Use new technologies nearline disk, virtual tape in addition to or instead of physical tape for backup and data migration Higher reliability and performance May 07 Cray Technical Review 9

Major HPC Storage Issue Too many HPC RFPs (esp for supercomputers) treat storage as secondary consideration Storage requirements are incomplete or ill-defined Only performance requirement and/or benchmark is maximum aggregate bandwidth No small files, no IOPS, metadata ops Requires HSM or backup with insufficient details No real reliability requirements Selection criteria don t give credit for a better storage solution Vendor judged on whether storage requirements are met or not Result: vendor proposes the minimum cost solution that meets the storage requirements Gets rewarded for putting the rest of the budget towards TFLOPS May 07 Cray Technical Review 10

Why NFS? NFS is the basis of the NAS storage market (but CIFS important as well) Highly successful, adopted by all storage vendors Full ecosystem of data management and administration tools proven in commercial markets Value propositions ease of install and use, interoperability NAS vendors are now focusing on scaling NAS Various technical approaches for increasing client and storage scalability Major weakness performance Some NAS vendors have been focusing on this We see opportunities for improving this May 07 Cray Technical Review 11

CASA Lab Benchmarking CASA Lab in Chippewa Falls provides testbed to benchmark, configure and test CASA components Opteron cluster (30 nodes) running Suse Linux Cisco 6509 switch BlueArc Titan dual-heads, 6x1 Gigabit Ethernet on each head Dual-fabric Brocade SAN with 4 FC controllers and 1 SATA controller Small Cray XT3 May 07 Cray Technical Review 12

Test and Benchmarking Methodology Used Bringsel tool (J. Kaitschuck see CUG paper) Measure reliability, uniformity, scalability and performance Creates large, symmetric directory trees, varying file sizes, access patterns, block sizes Allows testing of the operational behavior of a storage system: behavior under load, reliability, uniformity of performance Executed nearly 30 separate tests Increasing complexity of access patterns and file distributions Goal was to observe system performance across varying workloads May 07 Cray Technical Review 13

Quick Summary of Benchmarking Results A total of over 400 TB of data has been written without data corruption or access failures There have been no major hardware failures since testing began in August 2006 predictable and relatively uniform. with some exceptions, the BlueArc aggregate performance generally scales with the number of clients Recovery from injected faults was fast and relatively transparent to clients 32 test cases have been prepared, about 28 of varying length have been run, all file checksums to date have been valid Early SLES9 NFS client problems under load, detected and corrected via kernel patch; this led to the use of this patch at Cray s AWE customer site, who experienced the same problem May 07 Cray Technical Review 14

Sequential Write Performance: Varying Block Size May 07 Cray Technical Review 15

Large File Writes: 8K Blocks May 07 Cray Technical Review 16

Large File, Random Writes, Variable Sized Blocks: Performance Approaches 500 MB/second for Single Head May 07 Cray Technical Review 17

Titan Performance at SC06 May 07 Cray Technical Review 18

Summary of Results Performance generally uniform for given load Very small block size combined with random access performed poorly with SLES9 client Much improved performance with SLES10 client Like cluster file systems, NFS performance sensitive to client behavior SLES9 Linux NFS client failed under Bringsel load Tests completed with SLES10 client Cisco link aggregation reduces performance by 30% at low node counts Static assignment of nodes to Ethernet links increases performance This effect goes away for 100s of NFS clients May 07 Cray Technical Review 19

Summary of Results BlueArc SAN backend provides performance baseline The Titan NAS heads cannot deliver more performance than these storage arrays make available Need sufficient storage (spindles, array controllers) to meet IOPS and bandwidth goals Stripe storage for each Titan head across multiple controllers to achieve best performance Test your NFS client with your anticipated workload against your NFS server infrastructure to set baseline performance May 07 Cray Technical Review 20

Summary BlueArc NAS storage meets Cray goals for CASA Performance tuning is a continual effort Next big push: efficient protocols and transfers between NFS server tier and Cray platforms iscsi deployments for providing network disks for login, network, and SIO nodes Export SAN infrastructure from BlueArc to rest of data center Storage Tiers: fast FC block storage, BlueArc FC and SATA, MAID May 07 Cray Technical Review 21

Phase 0: Cluster File System Only Large Cray Supercomputer Lustre or Other Cluster File System All data lands and stays In the cluster file system Backup, HSM, other data management tasks all handled here Data sharing via file transfers May 07 Cray Technical Review 22

Phase 1: Cluster File System and Shared NAS MAID Disk Archive Large Cray Supercomputer Lustre or Other Cluster File System Fast NAS 10G/1G Ethernet Network Fast NAS SGI Sun IBM X Fast Nas Fast NAS Add NAS storage for data sharing between Cray and other machines NAS backup and archive support Long-term, managed data MAID for backup Separate storage networks for NAS and CFS stores GridFTP, other software, for sharing and data migration May 07 Cray Technical Review 23

Phase 2: Integrate NAS with Cray Platform Large Cray Supercomputer Lustre or Other Cluster File System 10G/1G Ethernet Network Integrated Fast NAS X MAID Disk Archive SGI Sun IBM Fast Nas Fast NAS Integrate fast NAS with Cray network: reduced NFS overhead, compute node access to shared NFS store Single file system name space: all NFS blades share same name space internal and external MAID for backup and storage tier underneath NAS: FC versus ATA Separate storage networks for NAS and CFS May 07 Cray Technical Review 24

Phase 3: Integrated SANs Large Cray Supercomputer Lustre or Other Cluster File System Storage Partition X 10G/1G Ethernet Network Integrated Fast NAS MAID Disk Archive SGI Sun IBM Fast Nas Fast NAS SAN Director Switch Single, integrated Storage Area Network for improved efficiency and RAS Volume mirroring, snapshots, LUN management Partition storage freely between shared NAS store and the cluster file System Further integration of MAID storage tier into shared storage hierarchy May 07 Cray Technical Review 25

CASA 2.0 Hardware (Potential) Servers, Clusters, and Workstations SSD 10G/1G Ethernet Network OST OST OST OST HSM/ Backup Server HSM/ Backup Server Fast NAS Fast NAS Fast NAS Fast NAS FC fabric HSM/Backup Tape Robot/ Archive Virtual Tape Archive May 07 Cray Technical Review 26

Questions? Comments? May 07 Cray Technical Review 27