Comparing Dynamic Disk Pools (DDP) with RAID-6 using IOR



Similar documents
Dynamic Disk Pools Technical Report

Dynamic Disk Pools Delivering Worry-Free Storage

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

WHITEPAPER: Understanding Pillar Axiom Data Protection Options

DELL RAID PRIMER DELL PERC RAID CONTROLLERS. Joe H. Trickey III. Dell Storage RAID Product Marketing. John Seward. Dell Storage RAID Engineering

Storage node capacity in RAID0 is equal to the sum total capacity of all disks in the storage node.

RAID Implementation for StorSimple Storage Management Appliance

This chapter explains how to update device drivers and apply hotfix.

Worry-free Storage. E-Series Simple SAN Storage

RAID 5 rebuild performance in ProLiant

RAID Level Descriptions. RAID 0 (Striping)

Sonexion GridRAID Characteristics

High Performance Tier Implementation Guideline

Netapp HPC Solution for Lustre. Rich Fenton UK Solutions Architect

HP Smart Array Controllers and basic RAID performance factors

IOmark- VDI. Nimbus Data Gemini Test Report: VDI a Test Report Date: 6, September

Dell SAS RAID Storage Manager. User s Guide. support.dell.com

Intel RAID Controllers

Intel RAID SSD Cache Controller RCS25ZB040

Comprehending the Tradeoffs between Deploying Oracle Database on RAID 5 and RAID 10 Storage Configurations. Database Solutions Engineering

Performance Impact on Exchange Latencies During EMC CLARiiON CX4 RAID Rebuild and Rebalancing Processes

Best Practices RAID Implementations for Snap Servers and JBOD Expansion

N /150/151/160 RAID Controller. N MegaRAID CacheCade. Feature Overview

RAID technology and IBM TotalStorage NAS products

TECHNOLOGY BRIEF. Compaq RAID on a Chip Technology EXECUTIVE SUMMARY CONTENTS

Picking the right number of targets per server for BeeGFS. Jan Heichler March 2015 v1.2

Practical issues in DIY RAID Recovery

IncidentMonitor Server Specification Datasheet

IOmark-VM. DotHill AssuredSAN Pro Test Report: VM a Test Report Date: 16, August

USER S GUIDE. MegaRAID SAS Software. June 2007 Version , Rev. B

NetApp Software. SANtricity Storage Manager Concepts for Version NetApp, Inc. 495 East Java Drive Sunnyvale, CA U.S.

Lecture 36: Chapter 6

Performance Report Modular RAID for PRIMERGY

User Guide - English. Embedded MegaRAID Software

Xanadu 130. Business Class Storage Solution. 8G FC Host Connectivity and 6G SAS Backplane. 2U 12-Bay 3.5 Form Factor

Intel RAID Software User s Guide:

Using RAID6 for Advanced Data Protection

SAN Conceptual and Design Basics

IBM ^ xseries ServeRAID Technology

HP reference configuration for entry-level SAS Grid Manager solutions

Evaluation Report: Supporting Microsoft Exchange on the Lenovo S3200 Hybrid Array

PowerVault MD1200/MD1220 Storage Solution Guide for Applications

The Effect of Priorities on LUN Management Operations

USER S GUIDE. MegaRAID SAS Software. April , Rev. D

Areas Covered. Chapter 1 Features (Overview/Note) Chapter 2 How to Use WebBIOS. Chapter 3 Installing Global Array Manager (GAM)

Improving Lustre OST Performance with ClusterStor GridRAID. John Fragalla Principal Architect High Performance Computing

Managing Storage Using RAID

Violin Memory Arrays With IBM System Storage SAN Volume Control

Evaluation Report: Accelerating SQL Server Database Performance with the Lenovo Storage S3200 SAN Array

EqualLogic PS Series Load Balancers and Tiering, a Look Under the Covers. Keith Swindell Dell Storage Product Planning Manager

Assessing RAID ADG vs. RAID 5 vs. RAID 1+0

MegaRAID SAS User Guide for Linux

HP ProLiant DL380p Gen mailbox 2GB mailbox resiliency Exchange 2010 storage solution

How to recover a failed Storage Spaces

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

Chapter 2 Array Configuration [SATA Setup Utility] This chapter explains array configurations using this array controller.

technology brief RAID Levels March 1997 Introduction Characteristics of RAID Levels

The IntelliMagic White Paper: Storage Performance Analysis for an IBM Storwize V7000

HP D3600 Disk Enclosure 4,000 Mailbox Resiliency Exchange 2013 Storage Solution

Dell Microsoft Business Intelligence and Data Warehousing Reference Configuration Performance Results Phase III

SSDs and RAID: What s the right strategy. Paul Goodwin VP Product Development Avant Technology

DELL TM PowerEdge TM T Mailbox Resiliency Exchange 2010 Storage Solution

Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance

Management Tools. Contents. Overview. MegaRAID Storage Manager. Supported Operating Systems MegaRAID CLI. Key Features

Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1

Blueprints for Scalable IBM Spectrum Protect (TSM) Disk-based Backup Solutions

FlexArray Virtualization

Virtualization Performance on SGI UV 2000 using Red Hat Enterprise Linux 6.3 KVM

Input / Ouput devices. I/O Chapter 8. Goals & Constraints. Measures of Performance. Anatomy of a Disk Drive. Introduction - 8.1

Q & A From Hitachi Data Systems WebTech Presentation:

RAID Basics Training Guide

Oracle Database Scalability in VMware ESX VMware ESX 3.5

S 2 -RAID: A New RAID Architecture for Fast Data Recovery

Intel RAID Software User s Guide:

Remote PC Guide Series - Volume 2b

Dell Exchange 2013 Reference Architecture for 500 to 20,000 Microsoft Users. 1 Overview. Reliable and affordable storage for your business

Technology Update White Paper. High Speed RAID 6. Powered by Custom ASIC Parity Chips

Nutanix Tech Note. Failure Analysis All Rights Reserved, Nutanix Corporation

Comparison of Hybrid Flash Storage System Performance

3PAR Fast RAID: High Performance Without Compromise

Architecting a High Performance Storage System

RAID HARDWARE. On board SATA RAID controller. RAID drive caddy (hot swappable) SATA RAID controller card. Anne Watson 1

DELL. Virtual Desktop Infrastructure Study END-TO-END COMPUTING. Dell Enterprise Solutions Engineering

As enterprise data requirements continue

Performance and scalability of a large OLTP workload

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

UK HQ RAID Chunk Size T F ISO 14001

Analysis of VDI Storage Performance During Bootstorm

EMC Virtual Infrastructure for SAP Enabled by EMC Symmetrix with Auto-provisioning Groups, Symmetrix Management Console, and VMware vcenter Converter

HP SN1000E 16 Gb Fibre Channel HBA Evaluation

Accelerating Server Storage Performance on Lenovo ThinkServer

Intel RAID Software User s Guide:

IOmark- VDI. HP HP ConvergedSystem 242- HC StoreVirtual Test Report: VDI- HC b Test Report Date: 27, April

Evaluation of Dell PowerEdge VRTX Shared PERC8 in Failover Scenario

MegaRAID Configuration Software

How To Understand And Understand The Power Of Aird 6 On Clariion

Dell EqualLogic Best Practices Series

QuickSpecs. HP Integrity Virtual Machines (Integrity VM) Overview. Currently shipping versions:

Xserve G5 Using the Hardware RAID PCI Card Instructions for using the software provided with the Hardware RAID PCI Card

Transcription:

Comparing Dynamic Disk Pools (DDP) with RAID-6 using IOR December, 2012 Peter McGonigal petermc@sgi.com Abstract Dynamic Disk Pools (DDP) offer an exciting new approach to traditional RAID sets by substantially improving rebuild times, limiting critical exposure during drive failures, reducing the performance penalty suffered during a rebuild, as well as significantly simplifying storage administration. This paper compares the performance using the IOR benchmark tool and benefits DDP can offer over traditional RAID-6. v1.6

Acknowledgement I would like to thank both Jerry Lohr and Scott Shaw for their help and advice that they provided me while doing this testing.

1.0 Introduction The introduction of release 10.83 of the SGI InfiniteStorage System Manager (ISSM) software provides a number of significant new features. This paper reports on one of the new features called Dynamic Disk Pools (DDP) and specifically how DDP performs in comparison to RAID-6. With RAID-6, data is striped across a set of physical drives and requires 2 parity data updates on different drives for every write. This dual parity is used to store and recover data in the event of a single or dual drive failure. In addition, further drives can be reserved as hot spares to act as stand-by replacements for failing drives. RAID-6 is designed to tolerate the failure of two drives. DDP uses an intelligent algorithm (utilizes a novel pseudo-random placement algorithm (CRUSH) 1 ) that defines which drives are used and distributes data, spare capacity, and protection information accordingly. When the disk pool is created it automatically includes all the storage needed to reconstruct and to re-allocate data (rebalances) if drive failures occur. Depending on the size of the pool (i.e. number if drives in pool), DDP reserves a number of reconstruction locations unknown as the Preservation Capacity. The Preservation Capacity provides rebuild locations for potential drive failures, unlike RAID-6 which uses dedicated stranded hot spares. Preservation Capacity is typically expressed as a drive count (e.g. a pool of 64 drives will have a preservation capacity of 4, this is sufficient storage has been reserved to rebuild up to 4 drives). A pool of drives can range from a minimum of 11 drives to the maximum of a fully configured storage array (e.g. 384 drives in the case of the SGI IS5500 Storage Array). The main benefits of using DDP include: Improved Data Protection o provides greater drive failure protection depending on size of pool than traditional RAID-6 o provides faster reconstruction time after drive failure than traditional RAID-6 Enhanced Performance Consistency o provides better performance than RAID-6 during drive failure and reconstruction Simple Storage Management o provides no RAID to manage o provides no hot-spares to allocate and manage o provides easier administration than RAID-6 1 Refer to http://www.ssrc.ucsc.edu/papers/weil-sc06.pdf for more details on CRUSH. 3

2.0 Test Environment Both the DDP and RAID-6 configurations were setup using the ISSM version 10.84.G8.21. The test environment used, including the hardware and storage setups, is described below. 2.1 Test Hardware The test environment consisted of a single SGI C2108-RP2 server and a SGI IS5500 60-bay Storage Array. A single SGI InfiniteStorage 5500 (IS5500) Storage Array base enclosure configured with: o Dual controllers o 60x Seagate ST9300553SS (300GB 6Gb SAS 15K) drives o Management Software version: 10.84.G8.21 o Controller firmware: 07.84.37.00 o Performance key enabled o Total un-configured capacity = 16TB Connected to a single SGI Rackable C2108-RP2 server configured with: o 2x Intel SandyBridge E5-2570 8-core processors o 384GB of memory o 4x dual port 8Gb FC HBAs o Software: SLES-11 SP2 SGI Foundation Software 2.6 SGI Performance Suite 1.4 SGI XVM 2.6 SGI MPI 1.4 IOR 2.10.3 2.2 DDP Configuration For the DDP test environment we configured the IS5500 Storage Array as follows: A single DDP configured with: o 60x Seagate ST9300553SS (300GB 6Gb SAS 15K) drives in the pool o Preservation Capacity = 3 drives o 8x volumes, each volume configured with 1230GB capacity o Mapped one FC 8Gb host path to one volume o Used XVM to stripe 8x volumes together (/dev/lxvm/ddpvol) 4

The following ISSM GUI window provides a summary of the DDP test configuration used. Note: (1) Used the following xvm command to create a DDP volume. I was unsure whether this is the best way of handling volumes in a DDP. xvm stripe -unit 2048 -volname DDPvol slice/lun* (2) Found setting up DDP more straightforward than setting up RAID-6 using ISSM. 2.3 RAID-6 Configuration For the RAID-6 test environment we configured the IS5500 Storage Array as follows: RAID-6 setup: o eight (8) RAID-6 (4+2) volume groups plus two hot spares o therefore using 50 drives (i.e. 48 + 2 hot spares) out of the 60x Seagate ST9300553SS (300GB 6Gb SAS 15K) drives available o One volume per volume group 5

o o Mapped one FC 8Gb host path to one volume Used XVM to stripe 8x volumes together (/dev/lxvm/r6vol) The following ISSM window provides a summary of the RAID-6 test configuration used. Note: (1) Used the following xvm command to create a RAID-6 volume: xvm stripe -unit 2048 -volname R6vol slice/lun* 2.4 Volume Initialization The time taken to initialize each of the eight (8) DDP volumes was on average 3 hours and 28 minutes, with the maximum time experienced on a volume being 4 hours and 15 minutes and the minimum being 3 hours and 10 minutes. With the eight volumes being initialized in parallel, the total time taken to complete initialization was 4 hours and 17 minutes. Initialization times will not hold up immediate use of storage since both DDP and RAID-6 volumes support Immediate Availability Format (IAF). The IAF feature provided by Santricity allows users to 6

access volumes during initialization. This enables setup of file-systems and applications during volume initialization, therefore providing users with zero wait time when initializing. Initialization does have some impact on performance; therefore full performance will not be reached until initialization has been completed. 2.5 IOR The IOR (v2.10.3) I/O benchmark tool was used for all performance testing. Testing conducted on both RAID-6 and DDP setups included: Sequential reads and writes Random reads and writes Sequential reads and write while manually failing a single drive and initiating reconstruction Each IOR run generated 16 I/O threads. Each I/O thread was pinned to one processor core. 7

3.0 DDP vs RAID6: Testing Sequential and Random READs and WRITEs 3.1 Sequential READs and WRITEs The testing conducted compares DDP and RAID-6 sequential I/O performance on the IS5500 Storage Array. Running IOR to generate sequential READs and WRITEs with I/O sizes ranging from 1KiB to 4MiB, in both in POSIX and MPI-IO modes The results of these tests are in graphs plotted below. 8

9

10

Note: (1) RAID-6 usually provides better sequential READ performance bandwidth than DDP. (2) DDP provides better sequential WRITE performance bandwidth than RAID-6. (3) For some reason sequential WRITE performance bandwidth is better than sequential READ performance bandwidth when using the MPI-IO API. 3.2 Random READs and WRITEs The testing conducted compares DDP and RAID-6 sequential I/O performance on the IS5500 Storage Array. Running IOR to generate random READs and WRITEs with I/O sizes ranging from 64KiB to 4MiB, in both in POSIX and MPI-IO modes Note: (1) Didn t conduct any random READs and WRITEs using small I/O sizes. Initial testing shows that doing random small I/O READs and WRITEs would take far too long, so decided to start testing from 64KiB I/O sizes. (2) Random WRITE performance bandwidth is nearly always better than random READ performance bandwidth (with the exception being 4MiB POSIX RAID6 random READs providing better performance bandwidth than random WRITEs). 11

12

13

4.0 DDP vs RAID6: Testing Performance during Single Drive Failure The testing also included: Forcing a single drive failure and reconstruction while running IOR to generate sequential READS and WRITES with I/O size of 4MB and recording both the time taken to reconstruct/rebalance the Storage Array and the (degraded) performance of the READs and WRITEs during reconstruction. 4.1 Single Drive Failure and Reconstruction using DDP The graphs show a slight drop in read and write performance during the period of the single drive failure and disk pool reconstruction/rebalancing. The default rebuild/rebalance priority settings were used. 14

15

The following ISSM GUI window below shows the events log that provides details from the beginning of the single drive failure through to the completion of the pool reconstruction. As you can see from the log below, the total time taken from the start of drive failure through to the completion of drive pool reconstruction was approximately 36 minutes (refer to red circled log entries). 16

The following ISSM GUI window below shows the DDP hardware drive layout during the single drive failure. It shows that the drive in drawer 3, slot 4 has failed. 4.2 Single Drive failure and reconstruction of Hot-Spare using RAID-6 The plot shows the slight drop in read and write performance during the period of the single drive failure and hot-spare update. 17

18

The following ISSM GUI window list the event log showing the starting date and time stamps of the beginning of the single drive failure through to the completion of the pool reconstruction. As you can see from the log below, the total time taken from drive failure through to completing of reconstruction was approximately 5 hours and 12 minutes (refer to red circled log entries). The following ISSM GUI window below shows the RAID-6 hardware drive layout during the single drive failure. It shows: The drive in drawer 3, slot 2 has failed. The drive in drawer 5, slot 10 is a hot spare and is currently active. The drive in drawer 5, slot 8 is another hot spare currently in stand-by. 19

Note: (1) That chart plotting RAID-6 performance while experiencing and single drive failure and reconstructing a hot-spare didn t cover the full period of reconstruction. This chart only plots up to 2:23:04am on the 30 th of November 2012. The reconstruction didn t complete until 3:01:24am on the 30 th of November 2012. (2) Reconstruction from a single drive failure was nearly nine (9) times faster for DDP compared to RAID-6 (i.e. 312 minutes for RAID-6, 36 minutes for DDP). 20

5.0 Summary In conclusion, the testing clearly shows that DDP is a winner when it comes to reducing the time to recover from a single drive failure (approximately 9 times faster). In regards to bandwidth while performing I/O: RAID-6 is generally slightly better than DDP when performing sequential READs. DDP is generally slightly better than RAID-6 when performing sequential WRITEs. RAID-6 is better under some circumstances when performing random READs and random WRITEs. DDP has slight performance degradation while experiencing a single drive failure and performing a reconstruction with default priority settings. RAID-6 appears to have minimal to no performance impact while experiencing a single drive failure and performing a hot-spare reconstruction with default priority settings. Further testing is required to see the impact on performance when handling single drive failures and hot-spare reconstructions with RAID-6 under other various workload scenarios. Further testing is required to see the impact on performance for both RAID-6 and DDP during dual disk failures and reconstructions. 21

Appendix The following provides the scripts used during testing to run the IOR benchmark. IOR script used to for sequential READs & WRITEs 22

IOR script used to for random READs & WRITEs 23

IOR script used during drive failure and reconstruction 24