SSDs and RAID: What s the right strategy. Paul Goodwin VP Product Development Avant Technology



Similar documents
High-Performance SSD-Based RAID Storage. Madhukar Gunjan Chakhaiyar Product Test Architect

Sistemas Operativos: Input/Output Disks

Disks and RAID. Profs. Bracy and Van Renesse. based on slides by Prof. Sirer

DELL RAID PRIMER DELL PERC RAID CONTROLLERS. Joe H. Trickey III. Dell Storage RAID Product Marketing. John Seward. Dell Storage RAID Engineering

RAID Overview: Identifying What RAID Levels Best Meet Customer Needs. Diamond Series RAID Storage Array

VERY IMPORTANT NOTE! - RAID

Using RAID6 for Advanced Data Protection

RAID Basics Training Guide

Lecture 36: Chapter 6

HP Smart Array Controllers and basic RAID performance factors

PIONEER RESEARCH & DEVELOPMENT GROUP

RAID. RAID 0 No redundancy ( AID?) Just stripe data over multiple disks But it does improve performance. Chapter 6 Storage and Other I/O Topics 29

An Introduction to RAID. Giovanni Stracquadanio

Reliability and Fault Tolerance in Storage

RAID HARDWARE. On board SATA RAID controller. RAID drive caddy (hot swappable) SATA RAID controller card. Anne Watson 1

Chapter Introduction. Storage and Other I/O Topics. p. 570( 頁 585) Fig I/O devices can be characterized by. I/O bus connections

How To Create A Multi Disk Raid

XtremIO DATA PROTECTION (XDP)

CS420: Operating Systems

RAID 5 rebuild performance in ProLiant

Definition of RAID Levels

RAID: Redundant Arrays of Independent Disks

technology brief RAID Levels March 1997 Introduction Characteristics of RAID Levels

Price/performance Modern Memory Hierarchy

Q & A From Hitachi Data Systems WebTech Presentation:

The Pitfalls of Deploying Solid-State Drive RAIDs

CS161: Operating Systems

NAND Flash Media Management Through RAIN

How to choose the right RAID for your Dedicated Server

Assessing RAID ADG vs. RAID 5 vs. RAID 1+0

RAID Levels and Components Explained Page 1 of 23

Storage node capacity in RAID0 is equal to the sum total capacity of all disks in the storage node.

Operating Systems. RAID Redundant Array of Independent Disks. Submitted by Ankur Niyogi 2003EE20367

Firebird and RAID. Choosing the right RAID configuration for Firebird. Paul Reeves IBPhoenix. mail:

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

GENERAL INFORMATION COPYRIGHT... 3 NOTICES... 3 XD5 PRECAUTIONS... 3 INTRODUCTION... 4 FEATURES... 4 SYSTEM REQUIREMENT... 4

Flash Memory Arrays Enabling the Virtualized Data Center. July 2010

PowerVault MD1200/MD1220 Storage Solution Guide for Applications

Storing Data: Disks and Files

RAID Level Descriptions. RAID 0 (Striping)

Configuring ThinkServer RAID 500 and RAID 700 Adapters. Lenovo ThinkServer

Flash 101. Violin Memory Switzerland. Violin Memory Inc. Proprietary 1

File System & Device Drive. Overview of Mass Storage Structure. Moving head Disk Mechanism. HDD Pictures 11/13/2014. CS341: Operating System

Chapter 6 External Memory. Dr. Mohamed H. Al-Meer

Outline. Database Management and Tuning. Overview. Hardware Tuning. Johann Gamper. Unit 12

Distribution One Server Requirements

RAID Implementation for StorSimple Storage Management Appliance

How To Use A Raid

Accelerating I/O- Intensive Applications in IT Infrastructure with Innodisk FlexiArray Flash Appliance. Alex Ho, Product Manager Innodisk Corporation

RAID Made Easy By Jon L. Jacobi, PCWorld

SSDs tend to be more rugged than hard drives with respect to shock and vibration because SSDs have no moving parts.

All-Flash Storage Solution for SAP HANA:

CS 6290 I/O and Storage. Milos Prvulovic

OBJECTIVE ANALYSIS WHITE PAPER MATCH FLASH. TO THE PROCESSOR Why Multithreading Requires Parallelized Flash ATCHING

Chapter 10: Mass-Storage Systems

Differential RAID: Rethinking RAID for SSD Reliability

Technologies Supporting Evolution of SSDs

TECHNOLOGY BRIEF. Compaq RAID on a Chip Technology EXECUTIVE SUMMARY CONTENTS

Performance Beyond PCI Express: Moving Storage to The Memory Bus A Technical Whitepaper

Disk Storage & Dependability

EMC XTREMIO EXECUTIVE OVERVIEW

Redundant Array of Independent Disks (RAID)

SSD Performance Tips: Avoid The Write Cliff

Intel RAID SSD Cache Controller RCS25ZB040

Automated Data-Aware Tiering

The What, Why and How of the Pure Storage Enterprise Flash Array

Striped Set, Advantages and Disadvantages of Using RAID

Dell PowerEdge RAID Controller Cards H700 and H800 Technical Guide

Input / Ouput devices. I/O Chapter 8. Goals & Constraints. Measures of Performance. Anatomy of a Disk Drive. Introduction - 8.1

SSD Server Hard Drives for IBM

Overview of I/O Performance and RAID in an RDBMS Environment. By: Edward Whalen Performance Tuning Corporation

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products

Systems Infrastructure for Data Science. Web Science Group Uni Freiburg WS 2014/15

Flash for Databases. September 22, 2015 Peter Zaitsev Percona

SOLID STATE DRIVES AND PARALLEL STORAGE

DELL SOLID STATE DISK (SSD) DRIVES

Deploying and Optimizing SQL Server for Virtual Machines

Impact of Stripe Unit Size on Performance and Endurance of SSD-Based RAID Arrays

What is RAID and how does it work?

Storage Technologies for Video Surveillance

The Pros and Cons of Erasure Coding & Replication vs. RAID in Next-Gen Storage Platforms. Abhijith Shenoy Engineer, Hedvig Inc.

VMware Virtual SAN 6.0 Design and Sizing Guide

PARALLELS CLOUD STORAGE

Solid State Drive Technology

The Technologies & Architectures. President, Demartek

Real-Time Storage Tiering for Real-World Workloads

Technology Update White Paper. High Speed RAID 6. Powered by Custom ASIC Parity Chips

IncidentMonitor Server Specification Datasheet

FUSION iocontrol HYBRID STORAGE ARCHITECTURE 1

The Shortcut Guide to Balancing Storage Costs and Performance with Hybrid Storage

NVIDIA RAID Installation Guide

SSDs tend to be more rugged than hard drives with respect to shock and vibration because SSDs have no moving parts.

UK HQ RAID Chunk Size T F ISO 14001

An Introduction to RAID 6 ULTAMUS TM RAID

Understanding endurance and performance characteristics of HP solid state drives

The Effect of Priorities on LUN Management Operations

Solid State Technology What s New?

Transcription:

SSDs and RAID: What s the right strategy Paul Goodwin VP Product Development Avant Technology

SSDs and RAID: What s the right strategy Flash Overview SSD Overview RAID overview Thoughts about Raid Strategies for SSDs

Flash Overview The smallest unit of NAND-based flash storage that can be written and subsequently read is a page. Today these pages are typically 4 KB in size. All bits in a blank page are set to 1s, and writing data to the page involves setting some of the bits within it to 0s. Individual bits within a page cannot be reset to 1s; rewriting the page to a different bit pattern requires an intermediate erase operation that resets all bits back to 1. Erasures are performed over large blocks (e.g., of 128 KB) spanning multiple pages. Blocks wear out as they are erased Exhibiting increasing BERs The PE cycle limit is the point where the BER becomes unmanageably high.

SSD Overview An SSD is comprised of a specialized processor with a I/O interface, typically SATA, on one side and a number of flash interfaces on the other.

The Function of the controller is to look like a Disk Drive. store and retrieve data to the flash devices connected to the back side of the controller. SSD Overview Additionally the controller has to manage the idiosyncrasies of the flash devices. In particular manage the PE cycle limitation of the flash devices. Wear leveling.

RAID Overview: Increasing performance of CPUs and memories will be squandered lf not matched by a similar performance increase in I/0. While the capacity of Single Large Expensive Disks (SLED) has grown rapidly, the performance improvement of SLED has been modest. Redundant Arrays of Inexpensive Disks (RAID), based on the magnetic disk technology developed for personal computers, offers an attractive alternative to SLED, promising improvements of an order of magnitude in performance, reliability, power consumption, and scalability. This paper introduces five levels of RAIDS - Patterson; A Case for Redundant Arrays of Inexpensive Disks (RAID) 1987

RAID Overview: Enter SSD Increasing performance of CPUs and memories will be squandered lf not matched by a similar performance increase in I/0. While the capacity of Single Large Expensive Disks (SLED) has grown rapidly, the performance improvement of SLED has been modest. Redundant Arrays of Inexpensive Disks (RAID), based on the magnetic disk technology developed for personal computers, offers an attractive alternative to SLED, promising improvements of an order of magnitude in performance, reliability, power consumption, and scalability. This paper introduces five levels of RAIDS Redundant Arrays Inexpensive Disks SSD are not really that really inexpensive SSD are not Disks New base to the acronym Redundant Array of Independent Drives

RAID Overview: Enter SSD Capacity HDs RAID Performance Increasing performance of CPUs and memories will be squandered lf not matched by a similar performance increase in I/0. While the capacity of Single Large Expensive Disks (SLED) has grown rapidly, the performance improvement of SLED has been modest. Redundant Arrays of Inexpensive Disks (RAID), based on the magnetic disk technology developed for personal computers, offers an attractive alternative to SLED, promising improvements of an order of magnitude in performance, reliability, power consumption, and scalability. This paper introduces five levels of RAIDS SSDs SLEDs

The advantages of RAID are not without cost Performance HD have only increased ~1.3X in speed over the past 10 years. SSDs can, on their own, approach the wire rate. Thru the RAID 0 component the SET can reach wire rate Reliability HDs have random, unpredictable failures SSDs have predictable and manageable failures RAID is good for random and unpredictable failures Power consumption HDs are power hungry SSDs are lower in power RAID has a power penalty for implementation Scalability HDs have a power and vibration problem when configuring large arrays SSDs scale well although are capacity challenged RAID has mechanical and power penalties RAID Overview: Increasing performance of CPUs and memories will be squandered lf not matched by a similar performance increase in I/0. While the capacity of Single Large Expensive Disks (SLED) has grown rapidly, the performance improvement of SLED has been modest. Redundant Arrays of Inexpensive Disks (RAID), based on the magnetic disk technology developed for personal computers, offers an attractive alternative to SLED, promising improvements of an order of magnitude in performance, reliability, power consumption, and scalability. This paper introduces five levels of RAIDS

Fairly large performance benefits RAID Overview: RAID 0 Not really RAID Data is split evenly across two or more disks with no parity or redundancy. For accesses that are larger than the stripe size the seek time of the array will be the same as that of a single drive. For accesses that are smaller than the stripe size allow the drives to seek independently. If the sectors accessed are spread evenly between the drives, the apparent seek time of the array will be half that of a single drive. The transfer speed of the array will be the transfer speed of all the disks added together, the speed of the RAID controller becomes the limiting factor. If you are going to have more than one drive you should set them up in a RAID 0 set.

RAID 0: Implementations RAID 0 may significantly increase the performance of an SSD Storage subsystem It s all a matter of where you place it An external RAID enclosure may not provide any performance benefit for SSDs SSDs running near SATA max won t benefit if connected via esata A Raid controller directly connected to the I/O will realize the benefits of RAID 0.

RAID 0: Implementations Storage subsystem on a card Density Form Factor IOPS Sequential or Random R/W RAID 40 GB to 4 TB PCIe, 8-lane Up to 240K Read/Write >2 GB/s 0, 1, 5, 6, 10, 50, 60

RAID Overview: RAID 1 Groups n + n disks as one virtual disk with the capacity of n disks. Data is replicated on the two disks. When a disk fails the data will be read from the failed mirror drive. read and write performance equal to a single drive Full Redundancy for protection of data. RAID 1 is more expensive in terms of disk space since twice the number of disks are used than required to store the data without redundancy.

Mirrored sets in a Striped configuration RAID Overview: RAID 10 Provides fault tolerance and improved performance. A1 B1 C1 A1 B1 C1 A2 B2 C2 A2 B2 C2 DISK 0 DISK 1 DISK 2 DISK 3

RAID Overview: RAID 5 provides data redundancy by using data striping in combination with parity across all physical drives in the set. Groups n disks as one large virtual disk with a capacity of (n-1) disks. Redundant information (parity) is alternately stored on all disks. When a disk fails, the virtual disk still works, but it is operating in a degraded state. The data is reconstructed from the surviving disks. Better read performance, but slower write performance. Redundancy for protection of data.

RAID 6 groups n disks as one large virtual disk with a capacity of (n-2) disks. RAID Overview: RAID 6 provides data redundancy by using data striping in combination with double distributed parity across all physical drives in the set. Parity information is alternately stored on all disks. The Raid 6 set remains functional with up to two disk failures. The data can be reconstructed from the surviving disks. Better read performance, but slower write performance. Increased redundancy for protection of data. Two disks per span are required for parity. RAID 6 is most expensive in terms of disk space.

RAID Overview: RAID 5 and RAID 6 RAID 5 and RAID 6 are expensive in terms of Flash operations Read the data block from drive Read the parity block from drive Compare the old data block with the write request. For each bit that has flipped in the data block, flip the corresponding bit in the parity block Write the updated data block Write the updated parity block

RAID Overview: RAID 50 Striping a RAID 5 drive set with a second RAID 5 Drive set. Groups n*s disks as one large virtual disk with a capacity of s*(n-1) disks, where s is the number of spans and n is the number of disks within each span. Redundant information (parity) is alternately stored on all disks of each RAID 5 span. Better read performance, but slower write performance. Requires proportionally as much parity information as standard RAID 5. Data is striped across all spans.

RAID Overview: RAID 60 Striping a RAID 6 drive set with a second RAID 6 drive set. Groups N disks as one large virtual disk with a capacity of S*(N-2) disks, where S is the number of sets and N is the number of disks within each span. Redundant information (parity) is alternately stored on all disks of each RAID 6 set. Better read performance, but slower write performance. Requires proportionally as much parity information as standard RAID 6. Data is striped across all spans.

Does RAID cause write amplification in SSDs? RAID and SSD: In Parity applications (RAID 5 and RAID 6) can writing a small block result in parity that affects multiple blocks? We are currently doing research on this to determine the effects of parity on SSD endurance. Some evidence is leaning towards supporting this. Some of this investigation is blind since we don t know the actual algorithms

RAID and SSD: SSDs don t eliminate the need for RAID Although less likely they can still have random failures For RAID using SSDs we prefer RAID 10 Because of the potential for RAID write amplification. Simple to implement RAID 10 is the fastest to rebuild Same cost as RAID 6 Faster that RAID 5 or RAID 6 RAID may help survive increased errors as the life cycle of a SSD is reached but swapping out drives is still necessary. Although they are likely to have been replaced by larger drives before that time is reached.

RAID and SSD: SSDs don t eliminate the need for RAID Although less likely they can still have random failures For RAID using SSDs we prefer RAID 10 Because of the potential for RAID write amplification. Simple to implement RAID 10 is the fastest to rebuild Same cost as RAID 6 Faster that RAID 5 or RAID 6 RAID may help survive increased errors as the life cycle of a SSD is reached but swapping out drives is still necessary. Although they are likely to have been replaced by larger drives before that time is reached.

RAID 10 vs. RAID 5 Tradeoff RAID 5 can t survive any 2 drive failures RAID 5's need to recalculate and redistribute parity data on a per-write basis. RAID 10 is faster to rebuild A1 B1 C1 A1 B1 C1 A2 B2 C2 DISK 0 DISK 1 DISK 2 A2 B2 C2 DISK 3

RAID 10 vs. RAID 6 Tradeoff RAID 6 can replace any 2 drive failures RAID 10 is faster to rebuild A1 B1 C1 A1 B1 C1 A2 B2 C2 A2 B2 C2 DISK 0 DISK 1 DISK 2 DISK 3

RAID 10 vs. RAID 50 Tradeoff RAID 50 only increases the speed of a RAID 5 based subsystem. SSDs are already faster. RAID 10 is faster to rebuild and doesn t suffer from any complexity A1 B1 C1 A1 B1 C1 A2 B2 C2 DISK 0 DISK 1 DISK 2 A2 B2 C2 DISK 3

RAID 10 vs. RAID 6 Tradeoff RAID 6 can replace any 2 drive failures RAID 10 is faster to rebuild A1 B1 C1 A1 B1 C1 A2 B2 C2 A2 B2 C2 DISK 0 DISK 1 DISK 2 DISK 3

RAID Strategies for SSDs: Next Steps More investigation A lot more data needs to be collected SSD specific RAID controllers Do we need new RAID controllers that work with SSDs? To be continued