IBM System x GPFS Storage Server



Similar documents
IBM System x GPFS Storage Server

IBM General Parallel File System (GPFS ) 3.5 File Placement Optimizer (FPO)

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

General Parallel File System (GPFS) Native RAID For 100,000-Disk Petascale Systems

IBM ELASTIC STORAGE SEAN LEE

Netapp HPC Solution for Lustre. Rich Fenton UK Solutions Architect

Data Protection Technologies: What comes after RAID? Vladimir Sapunenko, INFN-CNAF HEPiX Spring 2012 Workshop

HPC Advisory Council

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

SMB Direct for SQL Server and Private Cloud

How To Create A Multi Disk Raid

HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW

IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads

præsentation oktober 2011

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Software-defined Storage Architecture for Analytics Computing

File System & Device Drive. Overview of Mass Storage Structure. Moving head Disk Mechanism. HDD Pictures 11/13/2014. CS341: Operating System

Storage, Cloud, Web 2.0, Big Data Driving Growth

Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre

High-Availability and Scalable Cluster-in-a-Box HPC Storage Solution

DELL RAID PRIMER DELL PERC RAID CONTROLLERS. Joe H. Trickey III. Dell Storage RAID Product Marketing. John Seward. Dell Storage RAID Engineering

Business-centric Storage FUJITSU Hyperscale Storage System ETERNUS CD10000

Intel RAID SSD Cache Controller RCS25ZB040

Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 SMB Direct

PRIMERGY server-based High Performance Computing solutions

Guide to SATA Hard Disks Installation and RAID Configuration

Storage node capacity in RAID0 is equal to the sum total capacity of all disks in the storage node.

Highly-Available Distributed Storage. UF HPC Center Research Computing University of Florida

HadoopTM Analytics DDN

Hyper-V over SMB Remote File Storage support in Windows Server 8 Hyper-V. Jose Barreto Principal Program Manager Microsoft Corporation

Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers

Blueprints for Scalable IBM Spectrum Protect (TSM) Disk-based Backup Solutions

UCS M-Series Modular Servers

Understanding Microsoft Storage Spaces

Dell PowerVault MD Family. Modular storage. The Dell PowerVault MD storage family

Guide to SATA Hard Disks Installation and RAID Configuration

SATA II 4 Port PCI RAID Card RC217 User Manual

HPC Update: Engagement Model

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA

SCALABLE FILE SHARING AND DATA MANAGEMENT FOR INTERNET OF THINGS

Worry-free Storage. E-Series Simple SAN Storage

Enhancements of ETERNUS DX / SF

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007

Oracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

UNIFIED HYBRID STORAGE. Performance, Availability and Scale for Any SAN and NAS Workload in Your Environment

Connecting Flash in Cloud Storage

Deploying Flash- Accelerated Hadoop with InfiniFlash from SanDisk

NLSS: A Near-Line Storage System Design Based on the Combination of HDFS and ZFS

Guide to SATA Hard Disks Installation and RAID Configuration

Microsoft Hybrid Cloud IaaS Platforms

Introduction to Gluster. Versions 3.0.x

Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband

EMC VMAX3 FAMILY - VMAX 100K, 200K, 400K

The Pros and Cons of Erasure Coding & Replication vs. RAID in Next-Gen Storage Platforms. Abhijith Shenoy Engineer, Hedvig Inc.

Maxta Storage Platform Enterprise Storage Re-defined

Modernizing Hadoop Architecture for Superior Scalability, Efficiency & Productive Throughput. ddn.com

Nutanix Tech Note. Failure Analysis All Rights Reserved, Nutanix Corporation

Stovepipes to Clouds. Rick Reid Principal Engineer SGI Federal by SGI Federal. Published by The Aerospace Corporation with permission.

EMC SYMMETRIX VMAX 20K STORAGE SYSTEM

February, 2015 Bill Loewe

The last 18 months. AutoScale. IaaS. BizTalk Services Hyper-V Disaster Recovery Support. Multi-Factor Auth. Hyper-V Recovery.

OVERVIEW. CEP Cluster Server is Ideal For: First-time users who want to make applications highly available

SSDs and RAID: What s the right strategy. Paul Goodwin VP Product Development Avant Technology

Sun Constellation System: The Open Petascale Computing Architecture

NetApp High-Performance Computing Solution for Lustre: Solution Guide

Installing Hadoop over Ceph, Using High Performance Networking

Violin Memory Arrays With IBM System Storage SAN Volume Control

Improving Time to Results for Seismic Processing with Paradigm and DDN. ddn.com. DDN Whitepaper. James Coomer and Laurent Thiers

HPC data becomes Big Data. Peter Braam

Solution Brief July All-Flash Server-Side Storage for Oracle Real Application Clusters (RAC) on Oracle Linux

Flash Performance for Oracle RAC with PCIe Shared Storage A Revolutionary Oracle RAC Architecture

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing

Big Fast Data Hadoop acceleration with Flash. June 2013

How To Write An Article On An Hp Appsystem For Spera Hana

Sonexion GridRAID Characteristics

New Storage System Solutions

Hadoop on the Gordon Data Intensive Cluster

NVIDIA RAID Installation Guide

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute

Lessons learned from parallel file system operation

Chapter 10: Mass-Storage Systems

PARALLELS CLOUD STORAGE

Clusters: Mainstream Technology for CAE

Pivot3 Desktop Virtualization Appliances. vstac VDI Technology Overview

Moving Virtual Storage to the Cloud. Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage

Architecting a High Performance Storage System

Elasticsearch on Cisco Unified Computing System: Optimizing your UCS infrastructure for Elasticsearch s analytics software stack

owncloud Enterprise Edition on IBM Infrastructure

The safer, easier way to help you pass any IT exams. Exam : Storage Sales V2. Title : Version : Demo 1 / 5

3 Red Hat Enterprise Linux 6 Consolidation

Reliability and Fault Tolerance in Storage

Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar

IBM System Storage DS5020 Express

SFA10K-X & SFA10K-E. ddn.com. Storage Fusion Architecture TM. DDN Whitepaper

No matter what you need for Managed IT services, High-Performance Storage, you can count on us for low cost, fast and effective service.

Transcription:

IBM System x GPFS Storage Crispin Keable Technical Computing Architect 1

IBM Technical Computing comprehensive portfolio uniquely addresses supercomputing and mainstream client needs Technical Computing for Big Data Power Systems TM Engine for faster insights PureSystems TM Integrated expertise for improved economics Blue Gene Extremely fast, energy efficient supercomputer System x Redefining x86 System Storage System x idataplex Smarter storage solutions HPC Cloud Fast, dense, flexible GPFS Storage IBM Platform LSF Family IBM Platform HPC Big data storage IBM Platform Symphony Family IBM Platform Cluster Manager Intelligent Cluster GPFS Factory-integrated, interoperability-tested system with compute, storage, networking and cluster management 2

Perfect Storm of Synergetic Innovations Disruptive Integrated Storage Software: Declustered RAID with GPFS reduces overhead and speeds rebuilds by ~4-6x Performance: POWER, x86 cores more powerful than special-use controller chips GPFS Native RAID Storage Data Integrity, Reliability & Flexibility: End-to-end checksum, 2- & 3-fault tolerance, application-optimized RAID Integrated Hardware/Packaging: & Storage co-packaging improves density & efficiency High-Speed Interconnect: Clustering & storage traffic, including failover (PERCS/Power fabric, InfiniBand, or 10GE) Cost/Performance: Software-based controller reduces HW overhead & cost, and enables enhanced functionality. Big Data Converging with HPC Technology and Storage Convergence

High End POWER IBM GPFS Native RAID p775: High-Density Storage + Compute Based on Power 775 / PERCS Solution Basic Configuration: 32 Power7 32-core high bandwidth servers Configurable as GPFS Native RAID storage controllers, compute servers, I/O servers or spares Up to 5 Disk Enclosures per rack 384 Drives and 64 quad-lane SAS ports each Capacity: 1.1 PB/rack (900 GB SAS HDDs) Bandwidth: >150 GB/s per rack Read BW Compute Power: 18 TF + node sparing 1 Rack performs a 1TB Hadoop TeraSort in less than 3 minutes! Interconnect: IBM high-bw optical PERCS Multi-rack scalable, fully water-cooled 4

R D R D ol ol How does GNR work? Clients Clients FDR IB 10 GbE /Data s Custom Dedicated Disk Controllers 1 1 x3650 x3650 2 2 Migrate Migrate RAID RAID and and Disk Disk Management Management to to Commodity Commodity s! s! 1 1 GPFS GPFS Native Native RAID RAID 2 2 GPFS GPFS Native Native RAID RAID JBOD Disk Enclosures JBOD Disk Enclosures 5

A Scalable Building Block Approach to Storage Complete Storage Solution Data s, Disk (NL-SAS and SSD), Software, InfiniBand and Ethernet x3650 M4 6 Twin Tailed JBOD Disk Enclosure Model 24: Light and Fast 4 Enclosures, 20U 232 NL-SAS, 6 SSD 10 GB/Sec Performance based on IOR BM Model 26: HPC Workhorse! 6 Enclosures, 28U 348 NL-SAS, 6 SSD 12 GB/sec High-Density HPC Option 18 Enclosures 2-42U Standard Racks 1044 NL-SAS 18 SSD 6 36 GB/sec

GPFS Native RAID Feature Detail Declustered RAID Data and parity stripes are uniformly partitioned and distributed across a disk array. Arbitrary number of disks per array (unconstrained to an integral number of RAID stripe widths) 2-fault and 3-fault tolerance Reed-Solomon parity encoding 2 or 3-fault-tolerant: stripes = 8 data strips + 2 or 3 parity strips 3 or 4-way mirroring End-to-end checksum & dropped write detection Disk surface to GPFS user/client Detects and corrects off-track and lost/dropped disk writes Asynchronous error diagnosis while affected IOs continue If media error: verify and restore if possible If path problem: attempt alternate paths Supports live replacement of disks IO ops continue on for tracks whose disks have been removed during carrier service 7 7

Declustering Bringing parallel performance to disk maintenance Conventional RAID: Narrow data+parity arrays Rebuild can only use the IO capacity of 4 (surviving) disks 20 disks (5 disks per 4 conventional RAID arrays) 4x4 RAID stripes (data plus parity) Striping across all arrays, all file accesses are throttled by array 2 s rebuild overhead. Failed Disk Declustered RAID: Data+parity distributed over all disks Rebuild can use the IO capacity of all 19 (surviving) disks 20 disks in 1 Declustered RAID array 16 RAID stripes (data plus parity) Load on files accesses are reduced by 4.8x (=19/4) during array rebuild. Failed Disk 8 8

Declustered RAID Example 21 stripes (42 strips) 7 stripes per group (2 strips per stripe) 49 strips 9 3 1-fault-tolerant mirrored groups (RAID1) 3 groups 6 disks spare disk 7 spare strips 7 disks

Rebuild Overhead Reduction Example failed disk failed disk time time Rd Wr Rd-Wr Rebuild activity confined to just a few disks slow rebuild, disrupts user programs Rebuild activity spread across many disks, less disruption to user programs Rebuild overhead reduced by 3.5x 10

GPFS Native Raid Advantages Lower Cost! Software RAID No hardware storage controller 10-30% lower cost with higher performance Off-the-shelf SBODs Generic low-cost disk enclosures Standardized in-band SES management Standard Linux or AIX Generic high-volume servers Component of GPFS Extreme Data Integrity 2- and 3-fault-tolerant erasure codes 80% and 73% storage efficiency End-to-end checksum Protection against lost writes Fastest Rebuild times using Declustered RAID Industry Leading Performance Declustered RAID Reduced app load during rebuilds Up to 3x lower overhead to applications Aligned full-stripe writes disk limited Small writes backup-node NVRAM-log-write limited Faster than alternatives today and tomorrow! 11

12