1 RAID Basics Training Guide Discover a Higher Level of Performance RAID matters. Rely on Intel RAID.
2 Table of Contents 1. What is RAID? 2. RAID Levels RAID 0 RAID 1 RAID 5 RAID 6 RAID 10 RAID 0+1 RAID 1E RAID 50 RAID 60 JBOD 3. Management Console 4. Summary
3 What is RAID? RAID stands for a Redundant Array of Independent Drives (or disks). RAID gathers individual disk drives into a cohesive drive set. Often called RAID groups or Raid Arrays that can be operated in unison. RAID also offers other benefits. The first is that higher levels of data protection can be achieved as redundancy is accomplished through mirroring. Where the contents of one drive is duplicated on another drive. Also drive capacity is aggregated therefore larger storage volumes can be created. Another key benefit is performance, this is realized through disk striping in which the contents of a file are written to and read from across several drives in the RAID Group. Over the years RAID variations have been identified by their RAID level, there are multiple levels of RAID and at each level. There are specific costs, performance and fault tolerance benefits. The level of RAID itself is designated by a number and it s important to note that this numbering system is not intuitive and does not indicate which RAID level is better so it really goes back to understanding which RAID level will work best for your customer s environment. Let s look closer at some of these RAID levels.
4 RAID 0 RAID Level 0 is simply striping. Striping takes a series of drives, groups them so they are presented as a single device to the host and stripes the data across all these drives to improve performance. It certainly gives us a higher data transfer rate and is relatively low cost. The disadvantage here is that there is no redundancy or high availability. This does not mean that this RAID level is not used, in fact this RAID level is used with many applications in which performance and not data redundancy is important to your customer, but remember, if one drive fails the entire array fails because part of the data is missing with no way to recover it other than restoring from a back up. Disk striping enhances performance because multiple drives are accessed simultaneously, but remember disk striping does not provide data redundancy. It is recommended that you keep stripe sizes the same across RAID arrays. For example in a 3 disk system using only disk striping, Segment 1 is written to Disk 1, Segment 2 is written to Disk 2 and so on. Data transfer rates will be three times faster than a single disk or JBOD because no redundancy is required and all the reads and writes can be handled simultaneously from each disk.
5 RAID 1 With Mirroring used in RAID 1 data written to one disk is simultaneously written to another if one fails the contents of the other can be used to run the system and reconstruct the failed disk. The primary advantage of disk mirroring is that it provides 100% data redundancy but it is expensive because each drive in the system must be duplicated. Because the contents of the disk are completely written to a second, it does not matter if one of them fails. Both drives contain the same data at all times and either drive can act as the operational drive. This RAID 1 Illustration shows data being written to two drives at the same time creating an exact duplicate or mirrored data. If a drive fails the controller switches the mirror drive with no lapse in user acceptability. Extensions to RAID 1 One advantage of the LSI* MegaRAID RAID 1 algorithms utilized by Intel RAID is a technology called tier reads. During heavy read loads the controller will load balance the read request between both mirrored drives enhancing overall read performance. Traditional RAID 1 allows configuration of only two hard drives and RAID 10 configurations of up to 16 drives. With MegaRAID extensions to RAID 1, bandwidth-intensive applications can now benefit from much larger disk configurations. New enhancements allow up to 32 drives per RAID 1 volume improving overall system capacity and storage performance.
6 RAID 5 RAID 5 consists of block level striping and parity data distributed across all of RAID member disks. This gives more balanced access load across the drives. The array capacity is the sum of all the disks minus one with a minimum of three drives required. The parity information is used to recover data. If one drive fails so only one disk in an array is used to achieve data redundancy. This is the main reason this method is the most popular, also read performance in a RAID 5 configuration is virtually as good as RAID 0.
7 RAID 5 Cont. The disadvantage however is a relatively slow write cycle, 2 reads and 2 writes are required for each block written (one read and write for data block and another for the parity block). This RAID 5 illustration includes 6 physical disks where 5 data blocks are written on 5 physical disks, then parity data is written on the sixth disk. Parity is rotated and eventually written to every disk enabling the controller to recreate lost data onto a replacement disk without system interruption.
8 RAID 6 In a RAID 6 configuration, multiple parity operations are spread across the disk group. Which can survive the loss of two drives or the loss of a drive while another disk is being rebuilt. Of all of the standard RAID levels, RAID 6 provides the highest level of protection against drive failures. Read performance is similar to that of RAID 5, there is a performance penalty on write operations due to the over head association with the additional parity calculations. Performance is also further reduced during a drive rebuild. It is not well suited to tasks requiring a lot of writes or environments with few drives. Remember two complete disk failures in a single array is uncommon, an occasional bad blocks vary in frequency but the chance increases proportionally to capacity and number of disk drives. Some calculations show that arrays using the largest capacity disk drives are vulnerable to media errors and up to 1 in 4 rebuilds. In this illustration, Parity P+Q is rotated to every disk and written twice. The term P+Q indicates 2 algorithms are in use enabling the controller to recreate lost data from multiple disk failures on to replacement disks without system interruption. Disk spanning allows multiple physical disk arrays to function like one big drive. Spanning alone does not provide reliability or performance enhancements.
9 RAID 10 RAID 10 is the spanning of two or more RAID 1 mirrors. The advantages of RAID 10 are faster data access, like RAID 0 and single drive fault tolerance like RAID 1. RAID 10 still requires twice the number of disks like RAID 1. So it offers some performance improvements by striping but capacity is low since the mirror requires a duplicate set of drives.
10 RAID 10 Current Intel RAID controllers support up to 8 mirror groups in a RAID 10 configuration also note that spanned virtual disks must have the same stripe size and must be contiguous.
11 RAID 0+1 In RAID 0+1 data is striped across multiple drives and mirrored to a duplicate set of drives. RAID 0+1 is similar to RAID 10 with the exception that it cannot tolerate two simultaneous disk failures unless the second failed drive is from the same stripe as the first, that is once a single drive fails each of the disks in the other stripe is a single point of failure. Also, once the single failed drive is replaced all the disks in the array must participate in the rebuild. This Illustration of RAID 0+1 shows two groups of striped disks that are mirrors for redundancy.
12 RAID 1E RAID 1E is also a combination of mirroring with data striping. This RAID level stripes data and copies of the data across all the drives in the array. The first set of stripes are the data and the second set of stripes are mirrors of the first data stripe contained within the next logical drive. As with the standard RAID 1, the data is mirrored so the capacity of the logical drive is 50% of the total physical drive capacity of the array. RAID 1E requires a minimum of 3 drives. The following illustration is an example of a RAID 1E logical drive, each disk gets logically divided in half and mirrored data is written to the adjacent disk.
13 RAID 50 Another example of RAID spanning is RAID 50, like RAID 10 data is striped across multiple drive groups. However Raid 50 provides features of both RAID 0 and RAID 5. RAID 50 provides high throughput, redundancy and performance but required twice as many parity drives as a single RAID 5, configure RAID 50 by spanning two continuous RAID 5 virtual disks. As with RAID 10 the RAID 5 virtual disks must have the same stripe size. RAID 50 is commonly used in large disk groups. As the number of drives in a RAID set increases fault recovery time or the interval for rebuilding the RAID set increases. Instead of configuring one large RAID 5 array users can span multiple, smaller RAID 5 groups. The main advantage here is to reduce rebuild times thus reducing the likelihood of another disk failure while an array is in a degraded mode Also RAID 50 improves on the performance of RAID 5 particularly during writes. This level is recommended for applications that require high fault tolerance, capacity and random I/O activity performance. As you can see in this RAID 50 Illustration data is striped across multiple drive groups, data redundancy is achieved via rotated parity data.
14 RAID 60 RAID 60 combines data striping of RAID 0 with a distributed double parity of RAID 6, that is a RAID 0 array striped across RAID 6 elements. It typically requires at least 6 to 8 disks, RAID 60 has improved fault tolerance, as any 2 disk sets of each the RAID 6 sets can fail without data loss. Also drive failures or unrecoverable media errors occurring while a single disk is rebuilding in one RAID 6 set will not lead to data loss. Striping helps to increase capacity and performance without adding disks to each. RAID 6 set which would decrease data availability and could impact performance. RAID 60 improves upon the performance of RAID 6. Despite the fact that RAID 60 is slightly slower than RAID 50 in terms of writes, due to the added overhead of more parity calculations when data security is concerned this performance drop may be negligible.
15 RAID JBOD Concatenation or spanning of disks is not one of the numbered RAID levels but it is a popular method for combining multiple physical disk drives into a single virtual disk. It provides no data redundancy as the name implies disks are merely concatenated together so they appear to be a single large disk. This mode is sometimes called JBOD or just a bunch of disks. Performance is decreased because drives cannot be used concurrently. It is most commonly used when you have odd sized drives that need to be combined into a single virtual disk.
16 Management Console A RAID management console provides a simple way to manage and optimize storage application performance and data protection. It brings critical storage operations and reporting to the administrator s fingertips, allowing for easy deployment of storage functions. In addition to a Graphical User Interface, a Command Line Management tool also exists to provide additional flexibility, control and scripting capability. It enables administrators to easily deploy all critical system storage functions, including creating and managing virtual drives, adding a drive to a RAID virtual drive and on the fly RAID migration. Upgrading RAID levels is easy through a simple user interface. Depending on the console, it can upgrade to RAID 0, 1, 5, and 6 configurations and associated spans (10, 50 and 60), while allowing end-users to define specific properties for drive READs and WRITEs. Server operations are also allowed and can include creating a virtual drive, load configuration, firmware updating, silence alarm, and unlocking advanced software options, among other things.
17 Summary To summarize the different RAID levels discussed: RAID 0 is the fastest and most efficient but offers no fault tolerance. RAID 1 is ideal for highly fault tolerance environments but requires twice the storage. RAID 5 is the most cost efficient choice for server environments which are not write performance sensitive. RAID 10 is ideal for environments that require 100% redundancy and improved performance and capacity. RAID 50 is ideal for high capacity RAID 5 environments with additional reliability and performance. RAID 0+1 is optimal in systems requiring both fault tolerance and high performance but requires additional disk capacity investment. RAID 1E is a great choice for limited capacity environments such as small databases that need fault tolerance. RAID 6 is the choice for organizations that require high capacity along with high data redundancy where read performance is critical. RAID 60 is great for high capacity RAID 6 environments where additional data protection and performance is desired. Concatenation or JBOD mode is commonly used when combining odd sized drives into a single virtual disk. Intel RAID, or Redundant Array of Inexpensive (Independent) Disks, is a storage technology that combines multiple disk drive components into a logical unit, which helps with data redundancy and improves performance. RAID levels are the different ways data is distributed across the drives and depend on the level of redundancy and performance required. For more information visit: INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION WITH INTEL PRODUCTS. NO LICENSE, EXPRESS OR IMPLIED, BY ESTOPPEL OR OTHERWISE, TO ANY INTELLECTUAL PROPERTY RIGHTS IS GRANTED BY THIS DOCUMENT. EXCEPT AS PROVIDED IN INTEL S TERMS AND CONDITIONS OF SALE FOR SUCH PRODUCTS, INTEL ASSUMES NO LIABILITY WHATSOEVER AND INTEL DISCLAIMS ANY EXPRESS OR IMPLIED WARRANTY, RELATING TO SALE AND/OR USE OF INTEL PRODUCTS INCLUDING LIABILITY OR WARRANTIES RELATING TO FITNESS FOR A PARTICULAR PURPOSE, MERCHANTABILITY, OR INFRINGEMENT OF ANY PATENT, COPYRIGHT OR OTHER INTELLECTUAL PROPERTY RIGHT. Intel may make changes to specifications and product descriptions at any time, without notice. The information here is subject to change without notice. Do not finalize a design with this information. Intel, the Intel logo, Intel Inside, Xeon and Xeon Inside are trademarks of Intel Corporation in the U.S. and/or other countries. *Other names and brands may be claimed as the property of others. Copyright 2014 Intel Corporation. All rights reserved. 0414/SJ/EM/PDF Please Recycle
WHITE PAPER Intel Solid-State Drives Increase Productivity of Product Design and Simulation Intel Solid-State Drives Increase Productivity of Product Design and Simulation A study of how Intel Solid-State
Relational Database Management Systems in the Cloud: Microsoft SQL Server 2008 R2 Miles Ward July 2011 Page 1 of 22 Table of Contents Introduction... 3 Relational Databases on Amazon EC2... 3 AWS vs. Your
White Paper EMC VNXe HIGH AVAILABILITY Overview Abstract This white paper discusses the high availability (HA) features in the EMC VNXe system and how you can configure a VNXe system to achieve your goals
For Small and Medium Businesses Overview May 2008 Overview 2 Contents Page 3 Page 6 Page 9 Page 11 Introduction RAID storage Software versus hardware RAID Backup hardware and software Considerations for
Microsoft Corporation and HP Using Network Attached Storage for Reliable Backup and Recovery Microsoft Corporation Published: March 2010 Abstract Tape-based backup and restore technology has for decades
Best Practices for Virtualizing and Managing SQL Server v1.0 May 2013 Best Practices for Virtualizing and Managing SQL Server 2012 1 1 Copyright Information 2013 Microsoft Corporation. All rights reserved.
Best Practices Planning Abstract This white paper provides a set of proven practices for deploying EMC SourceOne Email Management. The information is intended as an enhancement to the information provided
HP B6200 Backup System Recommended Configuration Guidelines Introduction... 3 Purpose of this guide... 4 Executive summary... 4 Challenges in Enterprise Data Protection... 4 A summary of HP B6200 Backup
Best Practices for Deploying and Managing Linux with Red Hat Network Abstract This technical whitepaper provides a best practices overview for companies deploying and managing their open source environment
Dell NetVault Backup Technical Overview A technical overview of NetVault Backup, including its architecture, benefits, key components and licensing options. Written by Dell Software Introduction NetVault
Microsoft System Center 2012 R2 Why Microsoft? For Virtualizing & Managing SharePoint July 2014 v1.0 2014 Microsoft Corporation. All rights reserved. This document is provided as-is. Information and views
WHITE PAPER QUANTUM LATTUS: Next-Generation Object Storage for Big Data Archives CONTENTS Executive Summary....................................................................3 The Limits of Traditional
An Oracle White Paper June 2013 Oracle Real Application Clusters One Node Executive Overview... 1 Oracle RAC One Node 12c Overview... 2 Best In-Class Oracle Database Availability... 5 Better Oracle Database
White Paper EMC POWERPATH LOAD BALANCING AND FAILOVER Comparison with native MPIO operating system solutions Abstract EMC PowerPath and PowerPath/VE provide intelligent load balancing and failover. This
NetVault, NDMP and Network Attached Storage Simplicity and power for NAS Written by Adrian Moir, Dell Scott Hetrick, Dell Abstract This technical brief explains how Network Data Management Protocol (NDMP)
The Definitive Guide To tm Building Highly Scalable Enterprise File Serving Solutions Chris Wolf Chapter 5: Building High-Performance, Scalable, and Resilient Linux File-Serving Solutions...87 Challenges
NetXtreme Broadcom NetXtreme 57XX Introduction Functionality and Features Teaming Virtual LANs (VLANs) Manageability Installing the Hardware Installing the Driver Software Creating a Driver Disk Broadcom
High Availability and Scalability with Domino Clustering and Partitioning on AIX Marcelo R. Barrios, Ole Conradsen, Charles Haramoto, Didac Marin International Technical Support Organization http://www.redbooks.ibm.com
Introduction By leveraging the inherent benefits of a virtualization based platform, a Microsoft Exchange Server 2007 deployment on VMware Infrastructure 3 offers a variety of availability and recovery
Data Protection for Isilon Scale-Out NAS A Data Protection Best Practices Guide for Isilon IQ and OneFS By David Thomas, Solutions Architect An Isilon Systems Best Practices Guide May 2009 ISILON SYSTEMS
INTRODUCTION TO LINUX CLUSTERING DOCUMENT RELEASE 1.1 Copyright 2008 Jethro Carr This document may be freely distributed provided that it is not modified and that full credit is given to the original author.
An Oracle Technical White Paper May 2011 Oracle Optimized Solution for Enterprise Cloud Infrastructure Introduction... 1 Overview of the Oracle Optimized Solution for Enterprise Cloud Infrastructure...
HP StoreOnce Catalyst and HP Data Protector 7 Implementation and Best Practice Guide Release 1 Executive Summary This guide is intended to enable the reader to understand the basic technology of HP StoreOnce
STOR SOUTOS WT R Which R evel is Right for e? ontents ntroduction...1 R evel escriptions...1 R 0 ()...1 R 1 (irroring)...2 R 1 (Striped irror)...2 R 5 ( with parity)...2 R 5 (ot Space)...3 R 6 ( with dual