Dynamic Disk Pools Technical Report



Similar documents
A Dell White Paper Dell Compellent

High Performance Tier Implementation Guideline

Dynamic Disk Pools Delivering Worry-Free Storage

Comparing Dynamic Disk Pools (DDP) with RAID-6 using IOR

Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays

Comprehending the Tradeoffs between Deploying Oracle Database on RAID 5 and RAID 10 Storage Configurations. Database Solutions Engineering

Getting Started With RAID

PowerVault MD1200/MD1220 Storage Solution Guide for Applications

StorTrends RAID Considerations

Dell PowerVault MD Series Storage Arrays: IP SAN Best Practices

DELL RAID PRIMER DELL PERC RAID CONTROLLERS. Joe H. Trickey III. Dell Storage RAID Product Marketing. John Seward. Dell Storage RAID Engineering

Dell Microsoft Business Intelligence and Data Warehousing Reference Configuration Performance Results Phase III

Dell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture. Dell Compellent Product Specialist Team

The Effect of Priorities on LUN Management Operations

Microsoft Exchange 2010 on Dell Systems. Simple Distributed Configurations

A Dell Technical White Paper Dell Compellent

Moving Beyond RAID DXi and Dynamic Disk Pools

Evaluation Report: Accelerating SQL Server Database Performance with the Lenovo Storage S3200 SAN Array

Flash-optimized Data Progression

RAID Basics Training Guide

RAID 5 rebuild performance in ProLiant

Reference Architecture for a Virtualized SharePoint 2010 Document Management Solution A Dell Technical White Paper

Best Practices RAID Implementations for Snap Servers and JBOD Expansion

Solid State Drive vs. Hard Disk Drive Price and Performance Study

Delivering Accelerated SQL Server Performance with OCZ s ZD-XL SQL Accelerator

High Performance SQL Server with Storage Center 6.4 All Flash Array

Quick SAS Cabling Guide

IP SAN Best Practices

Data Migration: Moving from Dell PowerVault MD3000i/MD3000 to MD3200i/MD3220i and MD3600i/MD3620i Series Storage Arrays

Netapp HPC Solution for Lustre. Rich Fenton UK Solutions Architect

Dell Virtualization Solution for Microsoft SQL Server 2012 using PowerEdge R820

Accelerating Server Storage Performance on Lenovo ThinkServer

Optimizing SQL Server Storage Performance with the PowerEdge R720

IBM ^ xseries ServeRAID Technology

IOmark- VDI. Nimbus Data Gemini Test Report: VDI a Test Report Date: 6, September

DELL s Oracle Database Advisor

Definition of RAID Levels

Microsoft SharePoint Server 2010

WebBIOS Configuration Utility Guide

DELL. Virtual Desktop Infrastructure Study END-TO-END COMPUTING. Dell Enterprise Solutions Engineering

The IntelliMagic White Paper: Storage Performance Analysis for an IBM Storwize V7000

Worry-free Storage. E-Series Simple SAN Storage

Transitioning to 6Gb/s SAS (Serial-Attached SCSI) A Dell White Paper

Dell Reference Configuration for Hortonworks Data Platform

Virtualizing SQL Server 2008 Using EMC VNX Series and Microsoft Windows Server 2008 R2 Hyper-V. Reference Architecture

Intel RAID SSD Cache Controller RCS25ZB040

RAID Implementation for StorSimple Storage Management Appliance

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

As enterprise data requirements continue

VERITAS Volume Manager. for Windows. Best Practices

PARALLELS CLOUD STORAGE

Storage node capacity in RAID0 is equal to the sum total capacity of all disks in the storage node.

Storage virtualization and the HP StorageWorks Enterprise Virtual Array

Accelerate Oracle Performance by Using SmartCache of T Series Unified Storage

Configuring ThinkServer RAID 500 and RAID 700 Adapters. Lenovo ThinkServer

RAID technology and IBM TotalStorage NAS products

WHITEPAPER: Understanding Pillar Axiom Data Protection Options

How To Understand And Understand The Power Of Aird 6 On Clariion

The introduction of storage presales tool-edesigner. Storage edesigner Project Team

Dell EqualLogic Best Practices Series. Dell EqualLogic PS Series Reference Architecture for Cisco Catalyst 3750X Two-Switch SAN Reference

How To Create A Multi Disk Raid

Contents. Overview. Drive Policy RAID 500 features. Disable BGI RAID 700 features. Management Tasks Choosing the RAID Level.

Intel RAID Software User s Guide:

Assessing RAID ADG vs. RAID 5 vs. RAID 1+0

IOmark-VM. DotHill AssuredSAN Pro Test Report: VM a Test Report Date: 16, August

Benefits of Intel Matrix Storage Technology

Lab Evaluation of NetApp Hybrid Array with Flash Pool Technology

Storing Data: Disks and Files

Sistemas Operativos: Input/Output Disks

Microsoft SharePoint Server 2010

USER S GUIDE. MegaRAID SAS Software. June 2007 Version , Rev. B

Dell PowerVault MD32xx Deployment Guide for VMware ESX4.1 Server

File System & Device Drive. Overview of Mass Storage Structure. Moving head Disk Mechanism. HDD Pictures 11/13/2014. CS341: Operating System

The Methodology Behind the Dell SQL Server Advisor Tool

Dell Exchange 2013 Reference Architecture for 500 to 20,000 Microsoft Users. 1 Overview. Reliable and affordable storage for your business

Comparison of Hybrid Flash Storage System Performance

Intel RAID Software User s Guide:

EqualLogic PS Series Architecture: Hybrid Array Load Balancer

Comparison of StorTrends and NetApp Implementations of SQL Services in a Virtualized Server Infrastructure

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014

Intel RAID Software User s Guide:

Configuring ThinkServer RAID 100 on the Lenovo TS430

RAID Levels and Components Explained Page 1 of 23

OPTIMIZING EXCHANGE SERVER IN A TIERED STORAGE ENVIRONMENT WHITE PAPER NOVEMBER 2006

Microsoft SharePoint Server 2010

MS Exchange Server Acceleration

SSDs and RAID: What s the right strategy. Paul Goodwin VP Product Development Avant Technology

EMC XTREMIO EXECUTIVE OVERVIEW

Best Practices for Optimizing Storage for Oracle Automatic Storage Management with Oracle FS1 Series Storage ORACLE WHITE PAPER JANUARY 2015

Leveraging EMC Fully Automated Storage Tiering (FAST) and FAST Cache for SQL Server Enterprise Deployments

Condusiv s V-locity Server Boosts Performance of SQL Server 2012 by 55%

Technology Update White Paper. High Speed RAID 6. Powered by Custom ASIC Parity Chips

N /150/151/160 RAID Controller. N MegaRAID CacheCade. Feature Overview

HP reference configuration for entry-level SAS Grid Manager solutions

Configuring ThinkServer RAID 100 on the TS140 and TS440

HP Smart Array Controllers and basic RAID performance factors

Demystifying Deduplication for Backup with the Dell DR4000

Performance Impact on Exchange Latencies During EMC CLARiiON CX4 RAID Rebuild and Rebalancing Processes

SQL Server Business Intelligence on HP ProLiant DL785 Server

New economies of storage with the Compellent Flash-optimized solutions. Flash at the price of disk for I/O-intensive applications

Transcription:

Dynamic Disk Pools Technical Report A Dell Technical White Paper Dell PowerVault MD3 Dense Series of Storage Arrays 9/5/2012

THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS AND TECHNICAL INACCURACIES. THE CONTENT IS PROVIDED AS IS, WITHOUT EXPRESS OR IMPLIED WARRANTIES OF ANY KIND. 2012 Dell Inc. All rights reserved. Reproduction of this material in any manner whatsoever without the express written permission of Dell Inc. is strictly forbidden. For more information, contact Dell. Dell, the DELL logo, and the DELL badge, PowerConnect, and PowerVault are trademarks of Dell Inc. Microsoft, and Windows are either trademarks or registered trademarks of Microsoft Corporation in the United States and/or other countries. Other trademarks and trade names may be used in this document to refer to either the entities claiming the marks and names or their products. Dell Inc. disclaims any proprietary interest in trademarks and trade names other than its own. September 5, 2012

Contents INTRODUCTION... 4 TECHNICAL OVERVIEW... 4 DATA AVAILABILITY... 6 MANAGEMENT... 8 RESTRICTIONS AND LIMITATIONS... 10 PERFORMANCE... 10 SEQUENTIAL PERFORMANCE... 10 RANDOM PERFORMANCE... 11 SUMMARY... 12 LIST OF FIGURES Figure 1: D-Piece and D-Stripe... 5 Figure 2: DDP Reconstruction... 6 Figure 4: Create Disk Pool... 9 Figure 5: Create Virtual Disk... 9 Figure 6: Traditional RAID vs. DDP Sequential Workload Performance... 11 Figure 7: Traditional RAID vs. DDP Random Workload Performance... 11 Figure 8: Traditional RAID vs. DDP Decision Tree... 13

Introduction With the release of PowerVault MD Dense enclosure and associated controller firmware, significant new capabilities have been added to the Dell PowerVault MD3 Dense storage systems, including a new option to create Dynamic Disk Pools (DDP) in addition to the previously available traditional RAID groups. Dynamic Disk Pools greatly simplify management, improve data protection, and provide enhanced performance consistency. The DDP feature dynamically distributes data, spare capacity, and protection information across a pool of disk drives. These pools may range in size from a minimum of 11 drives to potentially as large as of all of the drives in the PowerVault MD3 Dense storage system including the expansion units. In addition to creation of a single Dynamic Disk Pool, storage administrators may opt to create traditional disks in conjunction with Dynamic Disk Pools or even multiple Dynamic Disk Pools, providing an unprecendented level of flexiblity. This technical report will provide a technical overview of the DDP feature as well as providing best practice guidance for usage. Technical Overview In essence, DDP functions as effectively another RAID level offering in addition to the previously available RAID 0, 1, 10, 5, and 6 traditional RAID Disk Groups. DDP greatly simplifies storage administration since there is no need to manage Idle spares or RAID groups. The following discusses the intrinsic characteristics and architecture of the Dynamic Disk Pool feature. Dynamic Disk Pools are composed of several lower level elements. 1. D-Piece : The first of these is known as a D-Piece. A D-Piece consists of a contiguous 512MB section from a physical disk containing 4,096 128KB segments. Within a pool, 10 D-Pieces are selected using an intelligent optimization algorithm from selected drives within the pool. 2. D-Stripe : The ten associated D-Pieces are then considered a D-Stripe which is 4GB in size. Within the D-Stripe itself, the contents are similar to a RAID 6, 8+2 scenario whereby 8 of the underlying segments potentially contain user data, 1 segment contains parity (P) information calculated from the user data segments, and the final segment containing the Q value as defined by RAID 6.

From the virtual disk or LUN perspective, they are essentially created from an aggregation of multiple 4GB D-Stripes as required to satisfy the defined virtual disk size up to the maximum allowable within a Dynamic Disk Pool. Figure 1: D-Piece and D-Stripe In Figure 1 above, a simplified diagram is shown of a single Dynamic Disk Pool containing 12 disk drives. As shown above, there are ten D-Pieces organized into a D-Stripe located pseudo-randomly across ten of the disks within the pool and this continues similarly for each 4GB D-Stripe. It is important to note that while distribution of D-Pieces and D-Stripes is approximately equal across all disks, as seen in the above example it is possible that this will not completely be the case Once a storage administrator has completed the action of defining a Dynamic Disk Pool which largely consists of simply defining the number of desired drives in the pool, the D- Piece and D-Stripe structures are created, similar to how traditional RAID stripes are created during Virtual disk creation. After the Dynamic Disk Pool has been defined, a virtual disk may be created within the pool. This virtual disk will consist of some number of D-Stripes located across all the drives within the pool up to the defined value for the virtual disk capacity, such that Number of D-Stripes = Chosen capacity / 4GB. As an example, a 500GB virtual disk would consist of 125 D-Stripes. Allocation of D-Stripes for a given virtual disk is performed starting at the lowest available range of LBAs for a given D-Piece on a given disk drive. Multiple virtual disks may be defined within a Dynamic Disk Pool and as noted previously there may be multiple pools created within the supported storage system. Alternatively, the storage administrator could also choose to create traditional RAID Disk Groups in conjunction with Dynamic Disk Pools or any combination thereof.

As an example, with a 60-drive PowerVault MD3260 array with all drives being of the same capacity, the following combinations would all be supported: 1x 60 drive Dynamic Disk Pool 2x 30 drive Dynamic Disk Pools 1x RAID 5 (30 Disk Group) and 1x DDP (30 drive pool) Similarly to traditional RAID Disk Groups, Dynamic Disk Pools can be expanded by the addition of disk drives to the pool through the Dynamic Capacity Expansion (DCE) process and up to 12 disks may be added to a defined disk pool concurrently. When a DCE operation is initiated, a small percentage of the existing D-Pieces are effectively migrated to the new disks. Data Availability Another major benefit of a Dynamic Disk Pool is that, rather than using dedicated stranded hot spares the pool itself contains integrated preservation capacity to provide rebuild locations for potential drive failures. This simplifies management as it is no longer required to plan or manage individual hot spares as well as greatly improving both the time of rebuilds if they need to occur and enhancing the performance of the virtual disks themselves while under a rebuild as opposed to traditional hot spares. When a drive in a DDP fails, the D-Pieces from the failed drive are reconstructed using the same mechanisms as would be used by RAID 6 normally to potentially all other drives in the pool. During this process, it is ensured that no single drive contains two D-Pieces from the same D-Stripe. The individual D-Pieces are reconstructed at the lowest available LBA range on the selected disk drive. Figure 2: DDP Reconstruction

In Figure 2, disk drive 6 is shown to have failed. Subsequently, the D-Pieces that were previously resident on that disk are recreated across several other drives in the pool simultaneously. Since there are multiple disks participating in the effort, the overall performance impact of this situation is lessened as well as the length of time needed to complete the operation being reduced dramatically. In the event of multiple disk failures within a Dynamic Disk Pool, priority reconstruction is given to any D-Stripes which are missing two D-Pieces in order to minimize any data availability risk. After those critically affected D-Stripes are reconstructed, the remainder of the necessary data will continue to be reconstructed. From a controller resource allocation perspective, there are two user-modifiable reconstruction priorities within DDP: Degraded reconstruction priority is assigned for instances where only a single D- Piece needs to be rebuilt for affected D-Stripes. The default is high 1. Critical reconstruction priority is assigned for instances where a D-Stripe has two missing D-Pieces which need to rebuilt. The default is highest. It is recommended to use Low or Medium priority settings for NL-SAS drives, this will increase the drive reconstruction time but will also lessen the impact of I/O performance during rebuild. For very large disk pools with two simultaneous disk failures, only a relatively small number of D-Stripes are likely to encounter the critical situation where two D-Pieces need to be reconstructed. As discussed previously, these critical D-Pieces will be identified and reconstructed first at the highest priority thus returning the DDP to a degraded state from critical state very quickly such that a further drive failure could be tolerated. As an example, assume that a Dynamic Disk Pool consisting of 192 disk drives has been created and has a dual disk failure. In that event, it is likely that the critical D-Pieces would be reconstructed in less than a minute and, after that minute, a further disk failure could be tolerated. From a mathematical perspective given the same 192 drive pool, only 5.2% of D- Stripes would have a D-Piece on one drive in the pool and only 0.25% of D-Stripes would have two D-Pieces on those two particular drives, meaning that only 5.2GB of data would have to be reconstructed to exit the critical stage. A very large disk pool can continue to maintain multiple sequential failures without data loss up until there is no additional preservation capacity to continue the rebuilds. 1

After the reconstruction, the failed drive or drives may be subsequently replaced, although this is not specifically required. Fundamentally, this replacement of failed disk drives is treated in much the same way as a Dynamic Capacity Expansion (DCE) of the Dynamic Disk Pool. Failed drives may also be replaced prior to the DDP exiting from a critical or degraded state. Aside from the reduced time to move from a critical state to that of simply degraded, the general rebuild process for a Dynamic Disk Pool can be up to 10x faster than a traditional RAID Disk Group. In the above example, the rebuild time for several different configurations was determined empirically. The tests were conducted with 3TB 7200RPM NL-SAS drives on a Dell PowerVault MD3660f storage system. The RAID 6, 10+2 baseline with a single drive failure resulted in almost an 51 hour rebuild time, while a similar size DDP completed in less than ten hours. Management Management of Dynamic Disk Pools is substantially simplified versus that of traditional RAID Disk Groups. Due to the nature of a DDP, there is fundamentally only one decision required from a storage administrator- how many disks are preferred to be in the disk pool. As shown in the above,creation of a new disk pool is a simple matter of selecting the desired number of disk drives. Preservation capacity and many other attributes of the disk pool are predefined. The amount of preservation capacity to be utilized by a Dynamic Disk Pool is fixed depending on the number of drives that were created within the pool

Figure 3: Create Disk Pool After a DDP has been created, virtual disks may be created with the pool in much the same manner as was previously used for traditional RAID Disk Groups with the addition that in a pool thinly-provisioned virtual disks are also available. Figure 4: Create Virtual Disk

Restrictions and Limitations Dynamic Disk Pools do have some restrictions that are not in place for traditional RAID Disk Groups. The segment size for all virtual disks in a Dynamic Disk Pool is 128KB and this may not be modified. Therefore, the Dynamic Segment Sizing feature is not available. Dynamic Disk Pools do not support the Dynamic RAID Migration (DRM) feature and the only supported protection mechanism is that of RAID 6. The pre-read redundancy check feature is not available for use with virtual disks in a DDP. Tray and drawer loss protection is not considered as part of the disk pool creation. Dynamic Disk Pools cannot be migrated data intact to another PowerVault MD3 Dense Storage system. Maximum virtual disk size within a DDP is 64TB There is no support for SSDs within a DDP Performance Given the nature of D-Stripe distribution within a Dynamic Disk Pool, performance is largely going to depend on the frequency of I/O requests to specific disks within the pool. For environments where potential disk contention is not a limiting factor for performance but instead the effective bottleneck is at the controller level for items such as CPU resources and available host-side bandwidth, then very little difference would be expected between a Dynamic Disk Pool and a similarly configured traditional RAID Disk Group(s). Sequential Performance Within a Dynamic Disk Pool, D-Stripes are pseudo-randomly ditsributed across groupings of ten disk drives. This can have the effect of randomizing access patterns to the underying physical disks even in cases where the host read requests are sequential in nature and is particularly observable in disk pools where the spindle count is not sufficient to counter the potential for randomization. For a minimally-sized pool with a single virtual disk, the sequential read performance can be fairly similar to a traditional RAID 6 Group with a single virtual disk as the underlying architecture is similar. As with RAID Groups, if additional virtual disks exist within a DDP and concurrent I/O access is expected, sequential performance may suffer as a result of head movement with the physical disk.

IOPS Throughput (MB/s) Due in part to to caching mechanisms, sequential writes on a Dynamic Disk Pool are generally more similar to that of traditonal RAID Disk Groups. 2500 2000 1500 RAID 6 vs. DDP Sequential Performance Traditional RAID 6 Reads DDP Reads Traditional RAID 6 Writes DDP Writes 1000 500 0 512 1K 2K 4K 8K 16K 32K 64K 128K 256K 512K 1M 2M 4M Transfer Size Figure 5: Traditional RAID vs. DDP Sequential Workload Performance Random Performance Given the nature of Dynamic Disk Pools, random performance, both read and write, on a disk pool tends to be relatively similar in nature to that of traditional RAID Groups of comparable size. RAID 6 vs. DDP Random Performance 5000 4500 4000 3500 3000 2500 2000 1500 1000 500 0 512 1K 2K 4K 8K 16K 32K 64K 128K 256K 512K 1M 2M 4M Transfer Size Traditional RAID 6 Reads DDP Reads Traditional RAID 6 Writes DDP Writes Figure 6: Traditional RAID vs. DDP Random Workload Performance

In general, if an PowerVault storage system is expected to drive maximum performance for a sustained periods of time, then generically traditional RAID Disk Groups will outperform Dynamic Disk Pools for both sequential and random workflows. This is primarily due to the allocation of resources within the disk pool itself. For a traditional RAID group, the configuration may be such that a given I/O stream has unique access to the disk resources within that group. In contrast, all I/O streams within a Dynamic Disk Pool share access to the disks within the pool. This can create the potential for resource contention with the underlying physical media depending on the workload. Summary As disk capacities continue to increase, the rebuild times for disk failures within RAID Groups will likewise increase leaving storage systems potentially at risk for additional drive failures and prolonging the performance impact to applications during the rebuild process. Dynamic Disk Pools offer an exciting new approach to traditional RAID sets by substantially improving rebuild times, limiting critical exposure during dual drive failures, reducing the performance penalty suffered during a rebuild, as well as significantly simplifying storage administration. In addition, given that virtual disks within Dynamic Disk Pools can potentially span very large numbers of disk drives versus tradtional RAID 5 or RAID 6 Disk Groups, in environments with mixed or non-concurrent workflows there can be a tremendous advantage in terms of performance as all pool resources are available to all hosts Dell PowerVault MD3 Dense Storage arrays support both Dynamic Disk Pools and traditional RAID Disk Groups, the following table could be used as reference for comparison between the two technologies. Desired feature Ease of configuration and management Performance during drive failure Dynamic Disk Pool Traditional RAID Group Maximum performance for large, sequential workloads Maximum performance for small, random workloads Shortest drive reconstruction time

Solid State Drive support Least time required to add drives (Dynamic Capacity Expansion) Greater protection of data from multiple drive failures occurring over time Simplified administration of spare capacity Flexible and efficient capacity utilization Ability to fine-tune performance for specific workloads Table 3: DDP vs. Traditional RAID group Summary The following flow chart can be used as a guide to help with deciding whether a Dynamic Disk Pool or traditional RAID Disk Group may be the proper choice for a given environment. Rebuild times or Absolute Performance? Rebuild times Absolute Performance Are Thin Virtual Disks Required? YES NO Does the workload need uniform response times? NO YES Are all virtual disks accessed fairly uniformly or does the workload change dynamically? Varying Uniform Traditional RAID Dynamic Disk Pools Figure 7: Traditional RAID vs. DDP Decision Tree