Surviving Two Disk Failures. Introducing Various RAID 6 Implementations
|
|
|
- Hester Gray
- 10 years ago
- Views:
Transcription
1 Surviving Two Failures Introducing Various RAID 6 Implementations
2 Notices The information in this document is subject to change without notice. While every effort has been made to ensure that all information in this document is accurate, Xyratex accepts no liability for any errors that may arise Xyratex (the trading name of Xyratex Technology Limited). Registered Office: Langstone Road, Havant, Hampshire, PO9 1SA, England. Registered number No part of this document may be transmitted or copied in any form, or by any means, for any purpose, without the written permission of Xyratex. Xyratex is a trademark of Xyratex Technology Limited. All other brand and product names are registered marks of their respective proprietors. Robert Maddock, Nigel Hart & Tom Kean Issue 3.0 October, 2007
3 Contents Introduction...2 Rationale A second disk drive failure A read error during rebuild...3 Repair error...3 What can be done?...3 Various design for multiple redundancy Triple mirroring RAID RAID 5& RAID RAID Reed-Solomon code...5 XOR Techniques...6 Rows and Columns...6 Horizontal and Diagonal parity...7 X-Code...9 ZZS (the Russian code)...10 Park (IBM)...10 Hybrid Approaches References About Xyratex
4 Introduction The following white paper discusses the reasons why RAID 6 has become more interesting to implement in Storage controllers in recent years. This is in part due to the adoption of high capacity (SATA) disk drives which has meant that the likelihood of failure during a rebuild operation has increased. The paper discusses the implications generically for RAID 6 algorithms against other techniques to protect against double drive failures. There are multiple RAID 6 algorithms and these are discussed at a high level. Each of these has some differences in their abilities, but generically some applications are more easily deployed onto RAID 6 arrays than others. Rationale The capacity of disk drives is continuously increasing, and the cost of storage is falling. So users are storing more and more data. But the reliability of disk drives does not seem to be improving (and moving to ATA drives may make it worse). So the number of data loss incidents in an installation is probably increasing. Customers of course want the number to fall, since dealing with such incidents is a time-consuming and expensive operation. The ideal would be a storage product that never loses data. Is this feasible? Many storage controllers today offer RAID arrays that can survive the failure of one disk drive without losing data. RAID-5 and RAID-1 (mirroring) are the most common examples. When a disk fails, it will be replaced and the contents rebuilt from the information on the remaining disks. There are two ways in which a further failure can lead to data loss. 1. A second disk drive failure If a second disk fails before the replace and rebuild is complete, the whole RAID array is lost. All the LUNs on this array will be lost, and if virtualisation is being used, sections of many LUNs may be lost. To reduce the chance of this happening, it is important to do the replace and rebuild as quickly as possible. Having a hot spare available allows the replacement to happen immediately, and rebuilding quickly reduces the exposure time, at the cost of a more severe performance reduction for normal operations. But as disk capacity increases, the disk data rate usually increases by a smaller amount, so the rebuild time increases, and with it the chance of data loss. Server class disk drives have a specified MTBF of 1 million hours or more. So an 8+P RAID-5 array will suffer a disk drive failure every 111,111 hours. With a replace and rebuild time of 10 hours, the chance of a second failure in an 8+P RAID-5 array, for example, is 80 in 1,000,000. So the chance of array loss is once every 1.4E9 hours. This sounds as though the chance of having an array failure is incredibly small, but it is doubtful whether the 1,000,000 hour MTBF figure is real. It isn t clear that any disk drives really reach this. Certainly some models of disk drive are worse than others, and if the factory builds a bad batch, they may well end up in a drawer 2
5 together built into a RAID array. If the figure is ten times worse, the incidence of array loss is 100 times worse. If the customer has 100 RAID arrays, the figure is 100 times worse again. Now we are at 1.4E5 hours, about 15 years. So over a ten-year life it is likely to happen. All this assumes the drive failures are independent events. Even worse, although difficult to quantify, are situations where some influence common to the drives of an array leads to failures. Stiction after a power failure was a notorious example in the past, but various environmental problems could cause linked failures. 2. A read error during rebuild To rebuild the contents of a failed disk drive in an 8+P RAID-5 array, we must read all the data on each of the 8 remaining disk drives. Server class drives specify an unrecoverable read error rate of about 1 in 1E14 bits. It isn t clear exactly what this means, and disk drive manufacturers provide little detail, but it suggests that reading eight drives of 300GB each, i.e. 8*300E9*10 = 2.4E13 bits, will have about one chance in 4 of a read error. If such a read error occurs, there is at least one block of the data on the array, which is lost. With or without virtualisation, this will be one block in one LUN. This will happen every 4 disk drive failures, i.e. every 4.4E5 hours. If the customer has 100 arrays it will happen every 4.4E3 hours every 6 months. This is with disk drives that meet their specification. If they are ten times worse, it happens 100 times more often every other day. Repair error Apart from data loss, there is one more argument that has been advanced for multiple redundancy. Apparently, it is quite common that when the repairman arrives to repair a RAID-5 array in which a disk has failed, he pulls out the wrong disk. Although this should not lead to long-term data loss, it certainly causes loss of access to data until the mistake is rectified. In an array with multiple redundancy, access would not be interrupted. What can be done? It seems unlikely that disk drive manufacturers will come up with orders of magnitude improvements in reliability, as long as we are using spinning platters and magnetics. So the only way to improve things substantially is to use arrays which can survive two (or more) failures without losing data. There are many different designs for arrays with multiple redundancy, with different advantages and disadvantages. The main trade-off is, as always, between cost and performance. The more disks that are used for a given amount of data, the higher the cost. With disks capable of a given number of operations per second, the number of disk operations required to implement each read or write to the controller determines performance. 3
6 (There are cost and performance trade-offs in the controller itself as well. If the algorithms used are complex, the controller is likely to be more expensive, both to develop and to manufacture, for a given level of performance.) Various design for multiple redundancy 1. Triple mirroring Three copies of the data are stored on three separate disks. It is expensive, since it uses three times as many disks as JBOD. But it is simple, and it performs quite well. Each read requires one disk read, from any of the three copies, and each write requires three disk writes: one to each copy. 2. RAID-51 Copies of the data are stored on two separate RAID-5 arrays, on separate disks of course. This is moderately expensive, since it uses more than twice as many disks as JBOD. The performance is quite good for reads, but modest for writes, since two RAID-5 writes are required. Each read requires one disk read, from either copy. Each write requires a read-modify-write to two disks on both RAID-5 arrays, although the reads only need to be done from one, making a total of six operations. As with RAID-5, long sequential writes can be optimised into full-stride writes. RAID-51 is particularly suitable when the two RAID-5 copies can be geographically separated, since complete failure of either site leaves one complete RAID-5, providing all the data with continuing redundancy. 3. RAID 5&0 A variant of RAID-51 is to store one RAID-5 copy of the data and one RAID-0 copy. This reduces the cost slightly (only one copy of parity) and improves the performance as well, since writing to the RAID-0 copy is more efficient. Writes now require read-modify writes to two disks on the RAID-5, and one disk write on the RAID RAID-6 The term RAID-6 here is used to describe any technique that stores strips of user data plus two derived redundancy strips, similar to the one parity strip in RAID-5. These redundancy strips are such that the data can be recovered after the loss of any two data or redundancy strips. The name RAID-6 should be used with caution, since it has been used for various different things (any salesman would see it as one better than RAID-5!). Another term is RAID-DP (double parity), but this is confusing too, since Network Appliance use this for their double parity technique, and HP use the even more confusing term RAID-5DP for the method used in their VA-7000 product. RAID-6 offers low cost, since only one copy of the data is stored, plus the redundancy strips. The performance is poor, although the performance per physical disk is no worse than RAID-51. Each read requires one disk read. Each write requires a read-modify-write on three disks. As with RAID-5, long sequential writes can be optimised. 4
7 The following table summarises the characteristics of these various methods and compares them to common methods with single or no redundancy. Protection against failure Total number of disks for N data disks Measure of cost Ratio of total:data (for N=8) ops per write Measures of performance Relative throughput 70R:30W (for N=8) Relative throughput with one disk failed JBOD none N 1 1W 1-1 RAID-5 any one disk N (R+W) RAID-1/10 any one disk 2N 2 2W RAID-6 any two disks N (R+W) RAID-51 any three disks 2N (R+W)+2W RAID-5&0 any two disks 2N (R+W)+1W Triple mirror any two disks 3N 3 3W Relative throughput per disk It is worth noting that the probability of three independent failures is very small, probably less likely than a geographical failure - fire, flood, earthquake, etc. - and so geographical separation is probably more useful than RAID-51 on one site. Such separation can be implemented with RAID 1/10, RAID-51, RAID 5&0, or triple mirroring. RAID-6 There are several ways of providing the double redundancy of RAID-6, and developing new ones has been an active field of research in recent years. These fall into two categories, those that use the standard XOR techniques as deployed in RAID 5 implementations and those that use redundant polynomial equations such as Reed-Solomon codes to recomputed the missing data. Reed-Solomon code There are well understood algorithms for taking N data bits and generating M more bits, such that the data can be calculated when several of the N+M bits are incorrect or have been lost. The development of these methods has taken place mainly in communications technology, and goes back more than fifty years. Such codes are also used in the error-correcting code (ECC) used with semiconductor memory. RAID-5 is of course a simple example of such a code, with N data bits and 1 parity bit, where the data can be retrieved after the loss of any one bit. For multiple redundancy, the most popular encoding seems to be the Reed-Solomon code, which can be used with any number of check disks. The calculation of the check fields must be done using data words of more than one bit from each data disk. The larger these words, the more data disks can be protected - the important equation here is 5
8 d = 2 b c Where d is the number of data disks, c the number of check disks, and b the number of bits in a word. From this it can be calculated that using bytes of data, up to 253 data disks can be protected against double disk loss by two check disks. So it will be seen that the method is more than sufficient for today s storage systems. The main difficulty with this technique is that calculating the check codes, and reconstructing data after failures, is quite complex. It involves polynomials and thus multiplication, and requires special hardware, or at least a signal processor, to do it at sufficient speed. With general purpose processors getting faster and faster, maybe this will change, but up to now it has been the case. The use of Reed-Solomon in communications means that hardware components are available for it. But since most existing storage controllers do not have this hardware, but do have exclusive-or hardware, there has been more emphasis in recent years on techniques that use exclusive-or. Whereas Reed-Solomon, like RAID-5, is one-dimensional, i.e. a vector of data words across the data disks generates the corresponding check words, all the exclusive-or techniques seem to need some kind of two-dimensional design. XOR Techniques The following are a selection of IP references to solutions to double drive parity/r6 using XOR techniques. Most RAID controller contain dedicated XOR hardware and therefore their designs are extendable to adopt one or other of these techniques. Rows and Columns Perhaps the simplest of these is rows and columns. In this, an array of M x N data disks is used, with M parity disks in one dimension and N parity disks in the other N Parity 1x N Parity 2x N Parity 3x M1 M2 M3 M4... MN Parity Mx Parity x1 Parity x2 Parity x3 Parity x4... Parity xn In the diagram, each cell represents one disk. There are M x N data disks. The M disks in the rightmost column contain the parity of the data disks in the same row; the N disks in the lowest row contain the parity of the data disks in the same column. When data is written, we must do read-modify-write on the data, the horizontal parity disk and the vertical parity disk. If a disk fails, if there is only one failed disk in the same column the data can be reconstructed using 6
9 the disks in that column, just like RAID-5. Similarly if there is only one failed disk in the same row. So the data can always be reconstructed after two failures, and very often after more. The disadvantage of the technique is, of course, that for the check disk overhead to be small, there have to be a lot of disks, so the method is really only suitable for very large storage systems where there are always many disks. Note, however, that the reconstruction does not involve all the disks, so the performance after failure is no worse than RAID-5 with N or M disks. It is possible to perform full-stride writes, but they do have to be very long because of the number of disks. The method as shown here puts data and parity on separate disks, with the possible uneven load that this entails (as with RAID-4). It might be possible to design a rotating RAID-5-like version of this, but it is a lot more complicated, and would need the extra disk in the corner. Horizontal and Diagonal parity As an alternative to such a large number of disks, it is possible to use multiple stripes on one array of disks as the way to get the second dimension. This is done in the RAID-DP method used by Network Appliance. Data 1 Data 2 Data 3 Data 4 Horizontal Parity Diagonal Parity Data1 Data2 Data3 Data4 Parity DParity1 Data Data1 Data2 Data3 Parity4 DParity2 Data4 Data Data1 Data2 Parity3 DParity3 Data3 Data4 Data Data1 Parity2 DParity4 The diagram shows a 4+2P array, but any number of data disks can be used. The data disks and the horizontal parity disk are a conventional RAID-4: the parity is the exclusive-or of the data strips in the same row. The diagonal parity strips are the exclusive-or of the other strips marked with the same number. Note that the horizontal parity does take part, and that there are some strips that do not contribute to a diagonal parity. Surprisingly, it is possible to reconstruct everything after any two-disk failures. Read operations are read directly from the disks where the data is stored. Short write operations require readmodify-write operations to the data, the horizontal parity, and usually the diagonal parity. But most writes also require read-modify-write to the diagonal parity to which the horizontal parity contributes. On average, a short writes requires 3.5 read-modify-write operations, more than the minimum three used by some other designs, so short writes are not efficient. Full stride writes work well, but the data length has to be N*N strips. In the Network Appliance implementation, the parity is not distributed, so the two parity disks do not contribute to read performance. However, the design is easily extensible to allow rotation of successive blocks of N*(N+2) strips. The algorithm to recover from the loss of a single drive in the array is the normal RAID 5 degraded array reconstruction, using the horizontal parity. 7
10 The algorithm to recover from a double drive failure involves a number of stages of using alternating Q and P stripe reconstructions. These in various sequences allow full reconstruction of the original data. The following worked example is a simple case to show the sequence. Note that the sequence of operations becomes larger as the width of the array becomes wider. Taking the previous layout of relationship between :- Data 1 Data 2 Data 3 Data 4 Horizontal Parity Diagonal Parity Data1 Data2 Data3 Data4 Parity DParity1 Data Data1 Data2 Data3 Parity4 DParity2 Data4 Data Data1 Data2 Parity3 DParity3 Data3 Data4 Data Data1 Parity2 DParity4 Should drives 1 and 2 fail, the situation would be :- Data 1 Data 2 Data 3 Data 4 Horizontal Parity Diagonal Parity Data1 Data2 Data3 Data4 Parity DParity1 Data Data1 Data2 Data3 Parity4 DParity2 Data4 Data Data1 Data2 Parity3 DParity3 Data3 Data4 Data Data1 Parity2 DParity4 Note that there will be a single strip of data that can be re computed :- Data 1 Data 2 Data 3 Data 4 Horizontal Parity Diagonal Parity Data1 Data2 Data3 Data4 Parity DParity1 Data Data1 Data2 Data3 Parity4 DParity2 Data4 Data Data1 Data2 Parity3 DParity3 Data3 Data4 Data Data1 Parity2 DParity4 From this a standard RAID 5 can recomputed the row :- Data 1 Data 2 Data 3 Data 4 Horizontal Parity Diagonal Parity Data1 Data2 Data3 Data4 Parity DParity1 Data Data1 Data2 Data3 Parity4 DParity2 Data4 Data Data1 Data2 Parity3 DParity3 Data3 Data4 Data Data1 Parity2 DParity4 8
11 Another Q drive relationship can be used to compute an additional strip :- Data 1 Data 2 Data 3 Data 4 Horizontal Parity Diagonal Parity Data1 Data2 Data3 Data4 Parity DParity1 Data Data1 Data2 Data3 Parity4 DParity2 Data4 Data Data1 Data2 Parity3 DParity3 Data3 Data4 Data Data1 Parity2 DParity4 And so the process goes on until fully reconstructed. X-Code A more balanced design is called X-Code (ref 4), for reasons which will become clear Data11 Data22 Data33 Data44 Data55 Data52 Data13 Data24 Data35 Data41 Data43 Data54 Data15 Data21 Data32 Parity3X Parity4X Parity5X Parity1X Parity2X ParityX4 ParityX5 ParityX1 ParityX2 ParityX3 The diagram shows the layout of an X-Code array of 5 disks. Each box in the diagram represents a strip of data or parity on a disk. ParityXn is the exclusive-or of the three data strips DataXn, and ParitynX of the three data strips DatanX. The groups of data strips and the parity strips generated from them run in both diagonal directions, hence X-Code. The layout of 5x5 strips shown in the diagram is then repeated across the whole of the 5 disks. With this encoding, data is read directly from the disks where it is stored, as in RAID-5. On short writes, the data, and the two corresponding parities, must be read, modified and rewritten. It is possible to write in full stride mode, but the length of data required is 5x3 strips. If a disk fails, it is possible to reconstruct each lost data strip by calculating exclusive-or of the other strips in either of the parity diagonals though the strip. If two disks are lost, it is still always possible to find a diagonal with only one missing strip, and reconstruct that one. Then there will be another diagonal with only one missing strip, and so on. In practice it is probably easiest to reconstruct a whole 5x5 section at a time. Both the calculation of the parities and the reconstruction are done entirely with exclusive-or operations. The data and parity are distributed, so that the load on the disks is even. 9
12 ZZS (the Russian code) In 1983, three Russians, Zaitsev, Zinov ev and Semakov (ZZS) published a method which can be applied to disk storage. The paper is in Russian (ref 5), but is described in (ref 6). It is similar to X-Code, but uses fewer stripes on each disk in the repeat pattern Data0 Parity0 Parity1 Parity2 Parity3 Parity4 Parity5 Data1 Data2 Data3 Data4 Data5 Data6 Data7 Data8 Data9 Data10 Data11 Data12 Data13 Data14 The number of disks is a prime number (p), and the number of stripes in the repeat pattern is (p-1)/2. Each data strip contributes to two parity strips on other disks, and the parity is computed as the exclusive-or of the data strips. Which strips are related is not a straightforward geometric pattern, as in X-Code. It is the result of an algorithm, but in an implementation it might be easier to use a table-driven technique. Read operations proceed as usual by reading the data directly from the disks where it is stored. Short write operations require three read-modify-write operations, and full-stride write operations are possible - the length required is about half that with X-Code. If a disk fails, each data strip on it can be reconstructed by exclusive-or of the remaining data and parity strips in either of the parity sets to which it belongs. If two disks fail, it is always possible to find a parity set with only one missing strip, and reconstruct that one first, etc. The parity is distributed, although the load on the disks is not completely even because of the one disk without parity. This method still has the disadvantage of requiring a prime number of disks, although it is possible to use it with one less than a prime number, by omitting 1, and instead using an imaginary 1 with all zeroes recorded on it. Park (IBM) Park, working at IBM Yorktown, discovered that methods very similar to ZZS could be used for arrays of various numbers of disks. He did this simply by exhaustive computer search, as described in his patent (ref 7), and discovered suitable data-parity patterns for all array sizes from 3 to 38 disks, with the exception of 8 and 9. (Lihao Xu subsequently found a method for 8 and 9 disks). Since the methods were discovered by search, rather than by algorithm, an implementation would have to be table-driven. As with ZZS, the repeat pattern is (N-1)/2 for an odd number of disks, and N/2 for an even number. Park always has one parity strip on each disk - he does not seem to have discovered that one disk in an odd number can contain all data. 10
13 Parity0 Parity1 Parity2 Parity3 Parity4 Parity5 Data0 Data1 Data2 Data3 Data4 Data5 Data6 Data7 Data8 Data9 Data10 Data11 The operation of Park s arrays is similar to ZZS arrays, and they have similar characteristics, with the further advantage of allowing any number of disks. Hybrid Approaches RAID 6 P+Q is a hybrid approach of the combination of XOR and polynomial codes. The following table demonstrates the terms that compute the 2 paritiy fields P&Q. Stripe D(0,0) D(0,1) D (0,2) D (0,3) D (0,4) P (0 ) Q (0) 1 D(1,0) D(1,1) D (1,2) D (1,3) P (1 ) Q (1) D (1,4) 2 D(2,0) D(2,1) D (2,2) P (2 ) Q (2) D (2,3) D (2,4) 3 D(3,0) D(3,1) P (3 ) Q (3) D (3,2) D (3,3) D (3,4) 4 D(4,0) P (4 ) Q (4) D(4,1) D (4,2) D (4,3) D (4,4) 5 P (5 ) Q (5) D(5,0) D(5,1) D (5,2) D (5,3) D (5,4) 6 Q (6) D (6,4) D(6,0) D(6,1) D (6,2) D (6,3) P (6)) P is computed as a standard XOR of D0-4 in this case Q is a polynomial based on Q(j) = (k0 D(j,0)) (k1 D(j,1)) (k4 D(j,4)). This has the characteristic that it can compute 1 term if 2 terms are lost. Reconstruction in the case of a double drive failure is a process of recreation of 1 of the D values from the Q term and then a standard XOR to compute the final term. As in this example, the stripes are rotated to balance read IO s over all of the disks in the array. 11
14 References L. Xu and J. Bruck, X-Code: MDS Array Codes with Optimal Encoding, IEEE Trans. on Information Theory, 45(1), , Jan., G. V Zaitsev, V.A. Zinov ev and N.V. Semakov, Minimum-Check-Density Codes for Correcting Bytes of Errors, Erasures,or Defects, Problems of Information Transmission 19(3), M. Blaum, R. M. Roth, On Lowest-Density MDS Codes, IEEE Trans. on Information Theory, 45(1), 46-59, Jan pdf 7. US patent 5,862,158, Baylor, Corbett and Park (IBM), US patent 5,485,474, Rabin (Harvard) RAID 6 P+Q, About Xyratex Xyratex is a leading provider of enterprise class data storage subsystems and storage process technology. The company designs and manufactures enabling technology that provides OEM and disk drive manufacturer customers with data storage products to support high-performance storage and data communication networks. Xyratex has over 20 years of experience in research and development relating to disk drives, storage systems and high-speed communication protocols. Founded in 1994 in a management buy-out from IBM, and with its headquarters in the UK, Xyratex has an established global base with R&D and operational facilities in Europe, the United States and South East Asia. 12
15
16 UK HQ Langstone Road Havant Hampshire PO9 1SA United Kingdom T +44(0) F +44(0) ISO 14001: 2004 Cert No. EMS Xyratex (The trading name of Xyratex Technology Limited). Registered in England & Wales. Company no: Registered Office: Langstone Road, Havant, Hampshire PO9 1SA, England. The information given in this brochure is for marketing purposes and is not intended to be a specification nor to provide the basis for a warranty. The products and their details are subject to change. For a detailed specification or if you need to meet a specific requirement please contact Xyratex.
RAID Basics Training Guide
RAID Basics Training Guide Discover a Higher Level of Performance RAID matters. Rely on Intel RAID. Table of Contents 1. What is RAID? 2. RAID Levels RAID 0 RAID 1 RAID 5 RAID 6 RAID 10 RAID 0+1 RAID 1E
Definition of RAID Levels
RAID The basic idea of RAID (Redundant Array of Independent Disks) is to combine multiple inexpensive disk drives into an array of disk drives to obtain performance, capacity and reliability that exceeds
RAID-DP: NetApp Implementation of Double- Parity RAID for Data Protection
Technical Report RAID-DP: NetApp Implementation of Double- Parity RAID for Data Protection Jay White & Chris Lueth, NetApp May 2010 TR-3298 ABSTRACT This document provides an in-depth overview of the NetApp
Using RAID6 for Advanced Data Protection
Using RAI6 for Advanced ata Protection 2006 Infortrend Corporation. All rights reserved. Table of Contents The Challenge of Fault Tolerance... 3 A Compelling Technology: RAI6... 3 Parity... 4 Why Use RAI6...
Why disk arrays? CPUs speeds increase faster than disks. - Time won t really help workloads where disk in bottleneck
1/19 Why disk arrays? CPUs speeds increase faster than disks - Time won t really help workloads where disk in bottleneck Some applications (audio/video) require big files Disk arrays - make one logical
How To Write A Disk Array
200 Chapter 7 (This observation is reinforced and elaborated in Exercises 7.5 and 7.6, and the reader is urged to work through them.) 7.2 RAID Disks are potential bottlenecks for system performance and
DELL RAID PRIMER DELL PERC RAID CONTROLLERS. Joe H. Trickey III. Dell Storage RAID Product Marketing. John Seward. Dell Storage RAID Engineering
DELL RAID PRIMER DELL PERC RAID CONTROLLERS Joe H. Trickey III Dell Storage RAID Product Marketing John Seward Dell Storage RAID Engineering http://www.dell.com/content/topics/topic.aspx/global/products/pvaul/top
RAID-DP : NETWORK APPLIANCE IMPLEMENTATION OF RAID DOUBLE PARITY FOR DATA PROTECTION A HIGH-SPEED IMPLEMENTATION OF RAID 6
RAID-DP : NETWORK APPLIANCE IMPLEMENTATION OF RAID DOUBLE PARITY FOR DATA PROTECTION A HIGH-SPEED IMPLEMENTATION OF RAID 6 Chris Lueth, Network Appliance, Inc. TR-3298 [12/2006] ABSTRACT To date, RAID
VERY IMPORTANT NOTE! - RAID
Disk drives are an integral part of any computing system. Disk drives are usually where the operating system and all of an enterprise or individual s data are stored. They are also one of the weakest links
Why disk arrays? CPUs improving faster than disks
Why disk arrays? CPUs improving faster than disks - disks will increasingly be bottleneck New applications (audio/video) require big files (motivation for XFS) Disk arrays - make one logical disk out of
CS420: Operating Systems
NK YORK COLLEGE OF PENNSYLVANIA HG OK 2 RAID YORK COLLEGE OF PENNSYLVAN James Moscola Department of Physical Sciences York College of Pennsylvania Based on Operating System Concepts, 9th Edition by Silberschatz,
PIONEER RESEARCH & DEVELOPMENT GROUP
SURVEY ON RAID Aishwarya Airen 1, Aarsh Pandit 2, Anshul Sogani 3 1,2,3 A.I.T.R, Indore. Abstract RAID stands for Redundant Array of Independent Disk that is a concept which provides an efficient way for
An Introduction to RAID 6 ULTAMUS TM RAID
An Introduction to RAID 6 ULTAMUS TM RAID The highly attractive cost per GB of SATA storage capacity is causing RAID products based on the technology to grow in popularity. SATA RAID is now being used
UK HQ RAID Chunk Size T F www.xyratex.com ISO 14001
RAID Chunk Size Notices The information in this document is subject to change without notice. While every effort has been made to ensure that all information in this document is accurate, Xyratex accepts
Fault Tolerance & Reliability CDA 5140. Chapter 3 RAID & Sample Commercial FT Systems
Fault Tolerance & Reliability CDA 5140 Chapter 3 RAID & Sample Commercial FT Systems - basic concept in these, as with codes, is redundancy to allow system to continue operation even if some components
Disk Array Data Organizations and RAID
Guest Lecture for 15-440 Disk Array Data Organizations and RAID October 2010, Greg Ganger 1 Plan for today Why have multiple disks? Storage capacity, performance capacity, reliability Load distribution
Reliability and Fault Tolerance in Storage
Reliability and Fault Tolerance in Storage Dalit Naor/ Dima Sotnikov IBM Haifa Research Storage Systems 1 Advanced Topics on Storage Systems - Spring 2014, Tel-Aviv University http://www.eng.tau.ac.il/semcom
Technology Update White Paper. High Speed RAID 6. Powered by Custom ASIC Parity Chips
Technology Update White Paper High Speed RAID 6 Powered by Custom ASIC Parity Chips High Speed RAID 6 Powered by Custom ASIC Parity Chips Why High Speed RAID 6? Winchester Systems has developed High Speed
RAID 5 rebuild performance in ProLiant
RAID 5 rebuild performance in ProLiant technology brief Abstract... 2 Overview of the RAID 5 rebuild process... 2 Estimating the mean-time-to-failure (MTTF)... 3 Factors affecting RAID 5 array rebuild
Data Storage - II: Efficient Usage & Errors
Data Storage - II: Efficient Usage & Errors Week 10, Spring 2005 Updated by M. Naci Akkøk, 27.02.2004, 03.03.2005 based upon slides by Pål Halvorsen, 12.3.2002. Contains slides from: Hector Garcia-Molina
Dependable Systems. 9. Redundant arrays of. Prof. Dr. Miroslaw Malek. Wintersemester 2004/05 www.informatik.hu-berlin.de/rok/zs
Dependable Systems 9. Redundant arrays of inexpensive disks (RAID) Prof. Dr. Miroslaw Malek Wintersemester 2004/05 www.informatik.hu-berlin.de/rok/zs Redundant Arrays of Inexpensive Disks (RAID) RAID is
Best Practices RAID Implementations for Snap Servers and JBOD Expansion
STORAGE SOLUTIONS WHITE PAPER Best Practices RAID Implementations for Snap Servers and JBOD Expansion Contents Introduction...1 Planning for the End Result...1 Availability Considerations...1 Drive Reliability...2
IBM ^ xseries ServeRAID Technology
IBM ^ xseries ServeRAID Technology Reliability through RAID technology Executive Summary: t long ago, business-critical computing on industry-standard platforms was unheard of. Proprietary systems were
How To Create A Multi Disk Raid
Click on the diagram to see RAID 0 in action RAID Level 0 requires a minimum of 2 drives to implement RAID 0 implements a striped disk array, the data is broken down into blocks and each block is written
an analysis of RAID 5DP
an analysis of RAID 5DP a qualitative and quantitative comparison of RAID levels and data protection hp white paper for information about the va 7000 series and periodic updates to this white paper see
Introduction. What is RAID? The Array and RAID Controller Concept. Click here to print this article. Re-Printed From SLCentral
Click here to print this article. Re-Printed From SLCentral RAID: An In-Depth Guide To RAID Technology Author: Tom Solinap Date Posted: January 24th, 2001 URL: http://www.slcentral.com/articles/01/1/raid
HP Smart Array Controllers and basic RAID performance factors
Technical white paper HP Smart Array Controllers and basic RAID performance factors Technology brief Table of contents Abstract 2 Benefits of drive arrays 2 Factors that affect performance 2 HP Smart Array
technology brief RAID Levels March 1997 Introduction Characteristics of RAID Levels
technology brief RAID Levels March 1997 Introduction RAID is an acronym for Redundant Array of Independent Disks (originally Redundant Array of Inexpensive Disks) coined in a 1987 University of California
RAID Implementation for StorSimple Storage Management Appliance
RAID Implementation for StorSimple Storage Management Appliance Alpa Kohli June, 2012 KB-00008 Document Revision 1 StorSimple knowledge base articles are intended to provide customers with the information
Benefits of Intel Matrix Storage Technology
Benefits of Intel Matrix Storage Technology White Paper December 2005 Document Number: 310855-001 INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION WITH INTEL PRODUCTS. NO LICENSE, EXPRESS OR IMPLIED,
NetApp RAID-DP : Dual-Parity RAID 6 Protection Without Compromise
White Paper NetApp RAID-DP : Dual-Parity RAID 6 Protection Without Compromise Paul Feresten, NetApp April 2010 WP-7005-0410 EXECUTIVE SUMMARY As disk drive densities continue to increase, and companies
RAID HARDWARE. On board SATA RAID controller. RAID drive caddy (hot swappable) SATA RAID controller card. Anne Watson 1
RAID HARDWARE On board SATA RAID controller SATA RAID controller card RAID drive caddy (hot swappable) Anne Watson 1 RAID The word redundant means an unnecessary repetition. The word array means a lineup.
SSDs and RAID: What s the right strategy. Paul Goodwin VP Product Development Avant Technology
SSDs and RAID: What s the right strategy Paul Goodwin VP Product Development Avant Technology SSDs and RAID: What s the right strategy Flash Overview SSD Overview RAID overview Thoughts about Raid Strategies
RAID Technology Overview
RAID Technology Overview HP Smart Array RAID Controllers HP Part Number: J6369-90050 Published: September 2007 Edition: 1 Copyright 2007 Hewlett-Packard Development Company L.P. Legal Notices Copyright
Read this before starting!
Points missed: Student's Name: Total score: /100 points East Tennessee State University Department of Computer and Information Sciences CSCI 4717 Computer Architecture TEST 2 for Fall Semester, 2006 Section
Overview of I/O Performance and RAID in an RDBMS Environment. By: Edward Whalen Performance Tuning Corporation
Overview of I/O Performance and RAID in an RDBMS Environment By: Edward Whalen Performance Tuning Corporation Abstract This paper covers the fundamentals of I/O topics and an overview of RAID levels commonly
1 Storage Devices Summary
Chapter 1 Storage Devices Summary Dependability is vital Suitable measures Latency how long to the first bit arrives Bandwidth/throughput how fast does stuff come through after the latency period Obvious
RAID 6 with HP Advanced Data Guarding technology:
RAID 6 with HP Advanced Data Guarding technology: a cost-effective, fault-tolerant solution technology brief Abstract... 2 Introduction... 2 Functions and limitations of RAID schemes... 3 Fault tolerance
The Benefit of Migrating from 4Gb to 8Gb Fibre Channel
The Benefit of Migrating from 4Gb to 8Gb Fibre Channel Notices The information in this document is subject to change without notice. While every effort has been made to ensure that all information in this
How To Understand And Understand The Power Of Aird 6 On Clariion
A Detailed Review Abstract This white paper discusses the EMC CLARiiON RAID 6 implementation available in FLARE 26 and later, including an overview of RAID 6 and the CLARiiON-specific implementation, when
RAID Levels and Components Explained Page 1 of 23
RAID Levels and Components Explained Page 1 of 23 What's RAID? The purpose of this document is to explain the many forms or RAID systems, and why they are useful, and their disadvantages. RAID - Redundant
RAID. Storage-centric computing, cloud computing. Benefits:
RAID Storage-centric computing, cloud computing. Benefits: Improved reliability (via error correcting code, redundancy). Improved performance (via redundancy). Independent disks. RAID Level 0 Provides
Comprehending the Tradeoffs between Deploying Oracle Database on RAID 5 and RAID 10 Storage Configurations. Database Solutions Engineering
Comprehending the Tradeoffs between Deploying Oracle Database on RAID 5 and RAID 10 Storage Configurations A Dell Technical White Paper Database Solutions Engineering By Sudhansu Sekhar and Raghunatha
CS 61C: Great Ideas in Computer Architecture. Dependability: Parity, RAID, ECC
CS 61C: Great Ideas in Computer Architecture Dependability: Parity, RAID, ECC Instructor: Justin Hsia 8/08/2013 Summer 2013 Lecture #27 1 Review of Last Lecture MapReduce Data Level Parallelism Framework
How To Improve Performance On A Single Chip Computer
: Redundant Arrays of Inexpensive Disks this discussion is based on the paper:» A Case for Redundant Arrays of Inexpensive Disks (),» David A Patterson, Garth Gibson, and Randy H Katz,» In Proceedings
Getting Started With RAID
Dell Systems Getting Started With RAID www.dell.com support.dell.com Notes, Notices, and Cautions NOTE: A NOTE indicates important information that helps you make better use of your computer. NOTICE: A
RAID Made Easy By Jon L. Jacobi, PCWorld
9916 Brooklet Drive Houston, Texas 77099 Phone 832-327-0316 www.safinatechnolgies.com RAID Made Easy By Jon L. Jacobi, PCWorld What is RAID, why do you need it, and what are all those mode numbers that
Q & A From Hitachi Data Systems WebTech Presentation:
Q & A From Hitachi Data Systems WebTech Presentation: RAID Concepts 1. Is the chunk size the same for all Hitachi Data Systems storage systems, i.e., Adaptable Modular Systems, Network Storage Controller,
Data Corruption In Storage Stack - Review
Theoretical Aspects of Storage Systems Autumn 2009 Chapter 2: Double Disk Failures André Brinkmann Data Corruption in the Storage Stack What are Latent Sector Errors What is Silent Data Corruption Checksum
RAID Overview: Identifying What RAID Levels Best Meet Customer Needs. Diamond Series RAID Storage Array
ATTO Technology, Inc. Corporate Headquarters 155 Crosspoint Parkway Amherst, NY 14068 Phone: 716-691-1999 Fax: 716-691-9353 www.attotech.com [email protected] RAID Overview: Identifying What RAID Levels
Input / Ouput devices. I/O Chapter 8. Goals & Constraints. Measures of Performance. Anatomy of a Disk Drive. Introduction - 8.1
Introduction - 8.1 I/O Chapter 8 Disk Storage and Dependability 8.2 Buses and other connectors 8.4 I/O performance measures 8.6 Input / Ouput devices keyboard, mouse, printer, game controllers, hard drive,
Storing Data: Disks and Files
Storing Data: Disks and Files (From Chapter 9 of textbook) Storing and Retrieving Data Database Management Systems need to: Store large volumes of data Store data reliably (so that data is not lost!) Retrieve
Technical White paper RAID Protection and Drive Failure Fast Recovery
Technical White paper RAID Protection and Drive Failure Fast Recovery RAID protection is a key part of all ETERNUS Storage Array products. Choices of the level chosen to meet customer application requirements
Lecture 36: Chapter 6
Lecture 36: Chapter 6 Today s topic RAID 1 RAID Redundant Array of Inexpensive (Independent) Disks Use multiple smaller disks (c.f. one large disk) Parallelism improves performance Plus extra disk(s) for
Distributed Storage Networks and Computer Forensics
Distributed Storage Networks 5 Raid-6 Encoding Technical Faculty Winter Semester 2011/12 RAID Redundant Array of Independent Disks Patterson, Gibson, Katz, A Case for Redundant Array of Inexpensive Disks,
Trends in NAND Flash Memory Error Correction
Trends in NAND Flash Memory Error Correction June 2009 Eric Deal Cyclic Design Introduction NAND Flash is a popular storage media because of its ruggedness, power efficiency, and storage capacity. Flash
Algorithms and Methods for Distributed Storage Networks 5 Raid-6 Encoding Christian Schindelhauer
Algorithms and Methods for Distributed Storage Networks 5 Raid-6 Encoding Institut für Informatik Wintersemester 2007/08 RAID Redundant Array of Independent Disks Patterson, Gibson, Katz, A Case for Redundant
Operating Systems. RAID Redundant Array of Independent Disks. Submitted by Ankur Niyogi 2003EE20367
Operating Systems RAID Redundant Array of Independent Disks Submitted by Ankur Niyogi 2003EE20367 YOUR DATA IS LOST@#!! Do we have backups of all our data???? - The stuff we cannot afford to lose?? How
RAID Level Descriptions. RAID 0 (Striping)
RAID Level Descriptions RAID 0 (Striping) Offers low cost and maximum performance, but offers no fault tolerance; a single disk failure results in TOTAL data loss. Businesses use RAID 0 mainly for tasks
HARD DRIVE CHARACTERISTICS REFRESHER
The read/write head of a hard drive only detects changes in the magnetic polarity of the material passing beneath it, not the direction of the polarity. Writes are performed by sending current either one
Price/performance Modern Memory Hierarchy
Lecture 21: Storage Administration Take QUIZ 15 over P&H 6.1-4, 6.8-9 before 11:59pm today Project: Cache Simulator, Due April 29, 2010 NEW OFFICE HOUR TIME: Tuesday 1-2, McKinley Last Time Exam discussion
Practical issues in DIY RAID Recovery
www.freeraidrecovery.com Practical issues in DIY RAID Recovery Based on years of technical support experience 2012 www.freeraidrecovery.com This guide is provided to supplement our ReclaiMe Free RAID Recovery
How To Encrypt Data With A Power Of N On A K Disk
Towards High Security and Fault Tolerant Dispersed Storage System with Optimized Information Dispersal Algorithm I Hrishikesh Lahkar, II Manjunath C R I,II Jain University, School of Engineering and Technology,
OPTIMIZING VIRTUAL TAPE PERFORMANCE: IMPROVING EFFICIENCY WITH DISK STORAGE SYSTEMS
W H I T E P A P E R OPTIMIZING VIRTUAL TAPE PERFORMANCE: IMPROVING EFFICIENCY WITH DISK STORAGE SYSTEMS By: David J. Cuddihy Principal Engineer Embedded Software Group June, 2007 155 CrossPoint Parkway
File System & Device Drive. Overview of Mass Storage Structure. Moving head Disk Mechanism. HDD Pictures 11/13/2014. CS341: Operating System
CS341: Operating System Lect 36: 1 st Nov 2014 Dr. A. Sahu Dept of Comp. Sc. & Engg. Indian Institute of Technology Guwahati File System & Device Drive Mass Storage Disk Structure Disk Arm Scheduling RAID
TECHNOLOGY BRIEF. Compaq RAID on a Chip Technology EXECUTIVE SUMMARY CONTENTS
TECHNOLOGY BRIEF August 1999 Compaq Computer Corporation Prepared by ISSD Technology Communications CONTENTS Executive Summary 1 Introduction 3 Subsystem Technology 3 Processor 3 SCSI Chip4 PCI Bridge
XtremIO DATA PROTECTION (XDP)
White Paper XtremIO DATA PROTECTION (XDP) Flash-Specific Data Protection, Provided by XtremIO (Ver..0) Abstract This white paper introduces the and discusses its benefits and advantages over RAID, with
How to recover a failed Storage Spaces
www.storage-spaces-recovery.com How to recover a failed Storage Spaces ReclaiMe Storage Spaces Recovery User Manual 2013 www.storage-spaces-recovery.com Contents Overview... 4 Storage Spaces concepts and
RAID for the 21st Century. A White Paper Prepared for Panasas October 2007
A White Paper Prepared for Panasas October 2007 Table of Contents RAID in the 21 st Century...1 RAID 5 and RAID 6...1 Penalties Associated with RAID 5 and RAID 6...1 How the Vendors Compensate...2 EMA
RAID. Contents. Definition and Use of the Different RAID Levels. The different RAID levels: Definition Cost / Efficiency Reliability Performance
RAID Definition and Use of the Different RAID Levels Contents The different RAID levels: Definition Cost / Efficiency Reliability Performance Further High Availability Aspects Performance Optimization
Chapter 6. 6.1 Introduction. Storage and Other I/O Topics. p. 570( 頁 585) Fig. 6.1. I/O devices can be characterized by. I/O bus connections
Chapter 6 Storage and Other I/O Topics 6.1 Introduction I/O devices can be characterized by Behavior: input, output, storage Partner: human or machine Data rate: bytes/sec, transfers/sec I/O bus connections
Intel RAID Software User s Guide:
Intel RAID Software User s Guide: Intel Embedded Server RAID Technology II Intel Integrated Server RAID Intel RAID Controllers using the Intel RAID Software Stack 3 Revision 8.0 August, 2008 Intel Order
How To Make A Backup System More Efficient
Identifying the Hidden Risk of Data De-duplication: How the HYDRAstor Solution Proactively Solves the Problem October, 2006 Introduction Data de-duplication has recently gained significant industry attention,
Enterprise-class versus Desktopclass
Enterprise-class versus Desktopclass Hard Drives April, 2008 Enterprise Platforms and Services Division - Marketing Revision History Date Revision Number April, 2008 1.0 Initial Release Modifications Disclaimers
Data Protection Technologies: What comes after RAID? Vladimir Sapunenko, INFN-CNAF HEPiX Spring 2012 Workshop
Data Protection Technologies: What comes after RAID? Vladimir Sapunenko, INFN-CNAF HEPiX Spring 2012 Workshop Arguments to be discussed Scaling storage for clouds Is RAID dead? Erasure coding as RAID replacement
Disk Storage & Dependability
Disk Storage & Dependability Computer Organization Architectures for Embedded Computing Wednesday 19 November 14 Many slides adapted from: Computer Organization and Design, Patterson & Hennessy 4th Edition,
Filing Systems. Filing Systems
Filing Systems At the outset we identified long-term storage as desirable characteristic of an OS. EG: On-line storage for an MIS. Convenience of not having to re-write programs. Sharing of data in an
Enova X-Wall LX Frequently Asked Questions
Enova X-Wall LX Frequently Asked Questions Q: What is X-Wall LX? A: X-Wall LX is the third generation of Enova real-time hard drive cryptographic gateway ASIC (Application Specific Integrated Circuit)
Block1. Block2. Block3. Block3 Striping
Introduction to RI Team members: 電 機 一 94901150 王 麒 鈞, 電 機 一 94901151 吳 炫 逸, 電 機 一 94901154 孫 維 隆. Motivation Gosh, my hard disk is broken again, and my computer can t boot normally. I even have no chance
California Software Labs
WHITE PAPERS FEBRUARY 2006 California Software Labs CSWL INC R e a l i z e Y o u r I d e a s Redundant Array of Inexpensive Disks (RAID) Redundant Array of Inexpensive Disks (RAID) aids development of
Introduction Disks RAID Tertiary storage. Mass Storage. CMSC 412, University of Maryland. Guest lecturer: David Hovemeyer.
Guest lecturer: David Hovemeyer November 15, 2004 The memory hierarchy Red = Level Access time Capacity Features Registers nanoseconds 100s of bytes fixed Cache nanoseconds 1-2 MB fixed RAM nanoseconds
What Types of ECC Should Be Used on Flash Memory?
What Types of ECC Should Be Used on Flash Memory? Application 1. Abstract NOR Flash normally does not need ECC (Error-Correcting Code). On the other hand, NAND requires ECC to ensure data integrity. NAND
Comparing Dynamic Disk Pools (DDP) with RAID-6 using IOR
Comparing Dynamic Disk Pools (DDP) with RAID-6 using IOR December, 2012 Peter McGonigal [email protected] Abstract Dynamic Disk Pools (DDP) offer an exciting new approach to traditional RAID sets by substantially
Hard Disk Drives and RAID
Hard Disk Drives and RAID Janaka Harambearachchi (Engineer/Systems Development) INTERFACES FOR HDD A computer interfaces is what allows a computer to send and retrieve information for storage devices such
Solving Data Loss in Massive Storage Systems Jason Resch Cleversafe
Solving Data Loss in Massive Storage Systems Jason Resch Cleversafe 2010 Storage Developer Conference. Insert Your Company Name. All Rights Reserved. 1 In the beginning There was replication Long before
NAND Flash Media Management Through RAIN
Flash Media Management Through RAIN Scott Shadley, Senior roduct Marketing Manager Micron Technology, Inc. Technical Marketing Brief What Is RAIN, and Why Do SSDs Need It? This brief compares redundant
Database Management Systems
4411 Database Management Systems Acknowledgements and copyrights: these slides are a result of combination of notes and slides with contributions from: Michael Kiffer, Arthur Bernstein, Philip Lewis, Anestis
RAID technology and IBM TotalStorage NAS products
IBM TotalStorage Network Attached Storage October 2001 RAID technology and IBM TotalStorage NAS products By Janet Anglin and Chris Durham Storage Networking Architecture, SSG Page No.1 Contents 2 RAID
Today s Papers. RAID Basics (Two optional papers) Array Reliability. EECS 262a Advanced Topics in Computer Systems Lecture 4
EECS 262a Advanced Topics in Computer Systems Lecture 4 Filesystems (Con t) September 15 th, 2014 John Kubiatowicz Electrical Engineering and Computer Sciences University of California, Berkeley Today
Chapter 6 External Memory. Dr. Mohamed H. Al-Meer
Chapter 6 External Memory Dr. Mohamed H. Al-Meer 6.1 Magnetic Disks Types of External Memory Magnetic Disks RAID Removable Optical CD ROM CD Recordable CD-R CD Re writable CD-RW DVD Magnetic Tape 2 Introduction
