CS 61C: Great Ideas in Computer Architecture. Dependability: Parity, RAID, ECC
|
|
|
- Maryann Dean
- 10 years ago
- Views:
Transcription
1 CS 61C: Great Ideas in Computer Architecture Dependability: Parity, RAID, ECC Instructor: Justin Hsia 8/08/2013 Summer 2013 Lecture #27 1
2 Review of Last Lecture MapReduce Data Level Parallelism Framework to divide up data to be processed in parallel Mapper outputs intermediate (key, value) pairs Optional Combiner in between for better load balancing Reducer combines intermediate values with same key Handles worker failure and does redundant execution 8/08/2013 Summer 2013 Lecture #27 2
3 Agenda Dependability Administrivia RAID Error Correcting Codes 8/08/2013 Summer 2013 Lecture #27 3
4 Six Great Ideas in Computer Architecture 1 Layers of Representation/Interpretation 2 Technology Trends 3 Principle of Locality/Memory Hierarchy 4 Parallelism 5 Performance Measurement & Improvement 6 Dependability via Redundancy 8/08/2013 Summer 2013 Lecture #27 4
5 Great Idea #6: Dependability via Redundancy Redundancy so that a failing piece doesn t make the whole system fail 1+1=2 2 of 3 agree 1+1=2 1+1=2 1+1=1 FAIL! 8/08/2013 Summer 2013 Lecture #27 5
6 Great Idea #6: Dependability via Redundancy Applies to everything from datacenters to memory Redundant datacenters so that can lose 1 datacenter but Internet service stays online Redundant routes so can lose nodes but Internet doesn t fail Redundant disks so that can lose 1 disk but not lose data (Redundant Arrays of Independent Disks/RAID) Redundant memory bits of so that can lose 1 bit but no data (Error Correcting Code/ECC Memory) 8/08/2013 Summer 2013 Lecture #27 6
7 Dependability Restoration Service accomplishment Service delivered as specified Failure Fault: failure of a component May or may not lead to system failure Applies to any part of the system Service interruption Deviation from specified service 8/08/2013 Summer 2013 Lecture #27 7
8 Dependability Measures Reliability: Mean Time To Failure (MTTF) Service interruption: Mean Time To Repair (MTTR) Mean Time Between Failures (MTBF) MTBF = MTTR + MTTF Availability = MTTF / (MTTF + MTTR) = MTTF / MTBF Improving Availability Increase MTTF: more reliable HW/SW + fault tolerance Reduce MTTR: improved tools and processes for diagnosis and repair 8/08/2013 Summer 2013 Lecture #27 8
9 Reliability Measures 1) MTTF, MTBF measured in hours/failure eg average MTTF is 100,000 hr/failure 2) Annualized Failure Rate (AFR) Average rate of failures per year (%) AFR Disks MTTF 8760 hr yr 1 Disks Total disk failures/yr 8760 hr/yr MTTF 8/08/2013 Summer 2013 Lecture #27 9
10 Availability Measures Availability = MTTF / (MTTF + MTTR) usually written as a percentage (%) Want high availability, so categorize by number of 9s of availability per year 1 nine: 90% => 36 days of repair/year 2 nines: 99% => 36 days of repair/year 3 nines: 999% => 526 min of repair/year 4 nines: 9999% => 53 min of repair/year 5 nines: 99999% => 5 min of repair/year 8/08/2013 Summer 2013 Lecture #27 10
11 Dependability Example 1000 disks with MTTF = 100,000 hr and MTTR = 100 hr MTBF = MTTR + MTTF = 100,100 hr Availability = MTTF/MTBF = = 999% 3 nines of availability! AFR = 8760/MTTF = = 876% Faster repair to get 4 nines of availability? MTTF = MTTR MTTR = hr 8/08/2013 Summer 2013 Lecture #27 11
12 Dependability Design Principle No single points of failure Chain is only as strong as its weakest link Dependability Corollary of Amdahl s Law Doesn t matter how dependable you make one portion of system Dependability limited by part you do not improve 8/08/2013 Summer 2013 Lecture #27 12
13 Question: There s a hardware glitch in our system that makes the Mean Time To Failure (MTTF) decrease Are the following statements TRUE or FALSE? 1) Our system s Availability will increase 2) Our system s Annualized Failure Rate (AFR) will increase (A) (B) (C) (D) 1 F 2 F F T T F 13
14 Agenda Dependability Administrivia RAID Error Correcting Codes 8/08/2013 Summer 2013 Lecture #27 14
15 Administrivia Project 3 (individual) due Sun 8/11 Final Review Tue 8/13, 7 10pm in 10 Evans Final Fri 8/16, 9am 12pm, 155 Dwinelle 2 nd half material + self modifying MIPS MIPS Green Sheet provided again Two two sided handwritten cheat sheets Can re use your midterm cheat sheet! Dead Week WTh 8/14 15 Optional Lab 13 (EC) due anytime next week 8/08/2013 Summer 2013 Lecture #27 15
16 Agenda Dependability Administrivia RAID Error Correcting Codes 8/08/2013 Summer 2013 Lecture #27 16
17 Arrays of Small Disks Can smaller disks be used to close the gap in performance between disks and CPUs? Conventional: 4 disk types Disk Array: 1 disk type 35 Low End High End 35 8/08/2013 Summer 2013 Lecture #27 17
18 Replace Large Disks with Large Number of Small Disks! (Data from 1988 disks) Capacity Volume Power Data Rate I/O Rate MTTF Cost IBM 3390K 20 GBytes 97 cu ft 3 KW 15 MB/s 600 I/Os/s 250 KHrs $250K IBM 35" MBytes 01 cu ft 11 W 15 MB/s 55 I/Os/s 50 KHrs $2K x72 23 GBytes 11 cu ft 1 KW 120 MB/s 3900 IOs/s??? ~700 Hrs Hrs $150K 9X 3X 8X 6X Disk Arrays have potential for large data and I/O rates, high MB/ft 3, high MB/KW, but what about reliability? 8/08/2013 Summer 2013 Lecture #27 18
19 RAID: Redundant Arrays of Inexpensive Disks Files are striped across multiple disks Concurrent disk accesses improve throughput Redundancy yields high data availability Service still provided to user, even if some components (disks) fail Contents reconstructed from data redundantly stored in the array Capacity penalty to store redundant info Bandwidth penalty to update redundant info 8/08/2013 Summer 2013 Lecture #27 19
20 RAID 0: Data Striping logical record striped physical records Stripe data across all disks Generally faster accesses (access disks in parallel) No redundancy (really AID ) Bit striping shown here, can do in larger chunks /08/2013 Summer 2013 Lecture #27 20
21 RAID 1: Disk Mirroring recovery group Each disk is fully duplicated onto its mirror Very high availability can be achieved Bandwidth sacrifice on write: Logical write = two physical writes Logical read = one physical read Most expensive solution: 100% capacity overhead 8/08/2013 Summer 2013 Lecture #27 21
22 Parity Bit Describes whether a group of bits contains an even or odd number of 1 s Define 1 = odd and 0 = even Can use XOR to compute parity bit! Adding the parity bit to a group will always result in an even number of 1 s ( even parity ) 100 Parity: 1, 101 Parity: 0 If we know number of 1 s must be even, can we figure out what a single missing bit should be? 10?11 missing bit is 1 8/08/2013 Summer 2013 Lecture #27 22
23 RAID 3: Parity Disk Logical data is bytestriped across disks Parity disk P contains parity bytes of other disks If any one disk fails, can use other disks to recover data! We have to know which disk failed X Y Z Must update Parity data on EVERY write Logical write = min 2 to max N physical reads and writes parity new = data old data new parity old P 0 2 P 3 5 P 6 8 8/08/2013 Summer 2013 Lecture #27 23 D 0 D 3 D 6 D 2 D 4 D 7 D 3 D 5 D 8 P
24 Updating the Parity Data Examine small write in RAID 3 (1 byte) 1 logical write = 2 physical reads + 2 physical writes Same concept applies for later RAIDs, too What if writing halfword (2 B)? Word (4 B)? D 0 D 0 P P new XOR data D 0 + D 0 D 1 D 2 D 3 P old data (1 Read) 1 only if bit changed XOR (3 Write) P D 0 D 1 D 2 D 3 + P old parity (2 Read) flip if changed (4 Write) 8/08/2013 Summer 2013 Lecture #27 24
25 RAID 4: Higher I/O Rate Logical data is now block striped across disks Parity disk P contains all parity blocks of other disks Because blocks are large, can handle small reads in parallel Must be blocks in different disks Still must update Parity data on EVERY write Logical write = min 2 to max N physical reads and writes Performs poorly on small writes 8/08/2013 Summer 2013 Lecture #27 25
26 RAID 4: Higher I/O Rate Insides of 5 disks D0 D1 D2 D3 P D4 D5 D6 D7 P D8 D9 D10 D11 P Increasing Logical Disk Address Example: small read D0 & D5, large write D12 D15 D12 D13 D14 D15 D16 D17 D18 D19 D20 D21 D22 D23 P 8/08/2013 Summer 2013 Lecture #27 26 Disk Columns P P Stripe
27 Inspiration for RAID 5 When writing to a disk, need to update Parity Small writes are bottlenecked by Parity Disk: Write to D0, D5 both also write to P disk D0 D1 D2 D3 P D4 D5 D6 D7 P 8/08/2013 Summer 2013 Lecture #27 27
28 RAID 5: Interleaved Parity Independent writes possible because of interleaved parity D0 D1 D2 D3 P D4 D5 D6 P D7 D8 D9 P D10 D11 D12 P D13 D14 D15 Increasing Logical Disk Addresses Example: write to D0, D5 uses disks 0, 1, 3, 4 P D16 D17 D18 D19 D20 D21 D22 D23 P Disk Columns 8/08/2013 Summer 2013 Lecture #27 28
29 Agenda Dependability Administrivia RAID Error Correcting Codes 8/08/2013 Summer 2013 Lecture #27 29
30 Error Detection/Correction Codes Memory systems generate errors (accidentally flipped bits) DRAMs store very little charge per bit Soft errors occur occasionally when cells are struck by alpha particles or other environmental upsets Hard errors occur when chips permanently fail Problem gets worse as memories get denser and larger 8/08/2013 Summer 2013 Lecture #27 30
31 Error Detection/Correction Codes Protect against errors with EDC/ECC Extra bits are added to each M bit data chunk to produce an N bit code word Extra bits are a function of the data Each data word value is mapped to a valid code word Certain errors change valid code words to invalid ones (ie can tell something is wrong) 8/08/2013 Summer 2013 Lecture #27 31
32 Detecting/Correcting Code Concept Space of all possible bit patterns: Error changes bit pattern to an invalid code word 2 N patterns, but only 2 M are valid code words Detection: fails code word validity check Correction: can map to nearest valid code word 8/08/2013 Summer 2013 Lecture #27 32
33 Hamming Distance Hamming distance = # of bit changes to get from one code word to another p = , q = , Hdist(p,q) = 2 p = , q = , Hdist(p,q) =? 3 Richard Hamming ( ) Turing Award Winner If all code words are valid, then min Hdist between valid code words is 1 Change one bit, at another valid code word 8/08/2013 Summer 2013 Lecture #27 33
34 3 Bit Visualization Aid Want to be able to see Hamming distances Show code words as nodes, Hdist of 1 as edges For 3 bits, show each bit in a different dimension: Bit 1 Bit 2 Bit 0 8/08/2013 Summer 2013 Lecture #27 34
35 Minimum Hamming Distance 2 Half the available code words are valid Let 000 be valid If 1 bit error, is code word still valid? No! So can detect If 1 bit error, know which code word we came from? No! Equidistant, so cannot correct 8/08/2013 Summer 2013 Lecture #27 35
36 Minimum Hamming Distance 3 Nearest 000 (one 1) Only a quarter of the available code words are valid Let 000 be valid Nearest 111 (one 0) How many bit errors can we detect? Two! Takes 3 errors to reach another valid code word If 1 bit error, know which code word we came from? Yes! 8/08/2013 Summer 2013 Lecture #27 36
37 Parity: Simple Error Detection Coding Add parity bit when writing block of data: b 7 b 6 b 5 b 4 b 3 b 2 b 1 b 0 p Check parity on block read: Error if odd number of 1s Valid otherwise b 7 b 6 b 5 b 4 b 3 b 2 b 1 b 0 p + + error Minimum Hamming distance of parity code is 2 Parity of code word = 1 indicates an error occurred: 2 bit errors not detected (nor any even #of errors) Detects an odd # of errors 8/08/2013 Summer 2013 Lecture #27 37
38 Parity Examples 1) Data ones, even parity now Write to memory to keep parity even 2) Data ones, odd parity now Write to memory: to make parity even 3) Read from memory ones even parity, so no error 4) Read from memory ones odd parity, so error What if error in parity bit? Can detect! 8/08/2013 Summer 2013 Lecture #27 38
39 Get To Know Your Instructor 8/08/2013 Summer 2013 Lecture #27 39
40 Agenda Dependability Administrivia RAID Error Correcting Codes (Cont) 8/08/2013 Summer 2013 Lecture #27 40
41 How to Correct 1 bit Error? Recall: Minimum distance for correction? Three Richard Hamming came up with a mapping to allow Error Correction at min distance of 3 Called Hamming ECC for Error Correction Code 8/08/2013 Summer 2013 Lecture #27 41
42 Hamming ECC (1/2) Use extra parity bits to allow the position identification of a single error Interleave parity bits within bits of data to form code word Note: Number bits starting at 1 from the left 1) Use all bit positions in the code word that are powers of 2 for parity bits (1, 2, 4, 8, 16, ) 2) All other bit positions are for the data bits (3, 5, 6, 7, 9, 10, ) 8/08/2013 Summer 2013 Lecture #27 42
43 Hamming ECC (2/2) 3) Set each parity bit to create even parity for a group of the bits in the code word The position of each parity bit determines the group of bits that it checks Parity bit p checks every bit whose position number in binary has a 1 in the bit position corresponding to p Bit 1 ( ) checks bits 1,3,5,7, (XXX1 2 ) Bit 2 ( ) checks bits 2,3,6,7, (XX1X 2 ) Bit 4 ( ) checks bits 4 7, 12 15, (X1XX 2 ) Bit 8 ( ) checks bits 8 15, 24 31, (1XXX 2 ) 8/08/2013 Summer 2013 Lecture #27 43
44 Hamming ECC Example (1/3) A byte of data: Create the code word, leaving spaces for the parity bits: _ 1 _ _ _ /08/2013 Summer 2013 Lecture #27 44
45 Hamming ECC Example (2/3) Calculate the parity bits: Parity bit 1 group (1, 3, 5, 7, 9, 11):? _ 1 _ _ Parity bit 2 group (2, 3, 6, 7, 10, 11): 0? 1 _ _ Parity bit 4 group (4, 5, 6, 7, 12): 0 1 1? _ Parity bit 8 group (8, 9, 10, 11, 12): ? /08/2013 Summer 2013 Lecture #27 45
46 Hamming ECC Example (3/3) Valid code word: Recover original data: Suppose we see instead fix the error! Check each parity group Parity bits 2 and 8 are incorrect As 2+8=10, bit position 10 is the bad bit, so flip it! Corrected value: /08/2013 Summer 2013 Lecture #27 46
47 Hamming ECC Cost Space overhead in single error correction code Form p + d bit code word, where p = # parity bits and d = # data bits Want the p parity bits to indicate either no error or 1 bit error in one of the p + d places Need 2 p p + d + 1, thus p log 2 (p + d + 1) For large d, p approaches log 2 (d) Example: d = 8 p = log 2 (p+8+1) p = 4 d = 16 p = 5; d = 32 p = 6; d = 64 p = 7 8/08/2013 Summer 2013 Lecture #27 47
48 Hamming Single Error Correction, Double Error Detection (SEC/DED) Adding extra parity bit covering the entire SEC code word provides double error detection as well! p 1 p 2 d 1 p 3 d 2 d 3 d 4 p 4 Let H be the position of the incorrect bit we would find from checking p 1, p 2,and p 3 (0 means no error) and let P be parity of complete code word H=0 P=0, no error H 0 P=1, correctable single error (p 4 =1 odd # errors) H 0 P=0, double error detected (p 4 =0 even # errors) H=0 P=1, an error occurred in p 4 bit, not in rest of word 8/08/2013 Summer 2013 Lecture #27 48
49 SEC/DED: Hamming Distance 4 1 bit error (one 0) Nearest bit error (one 1) Nearest bit error (two 0 s, two 1 s) halfway between 8/08/2013 Summer 2013 Lecture #27 49
50 Modern Use of RAID and ECC (1/3) Typical modern code words in DRAM memory systems: 64 bit data blocks (8 B) with 72 bit codes (9 B) d = 64 p = 7, +1 for DED What happened to RAID 2? Bit striping with extra disks just for ECC parity bits Very expensive computationally and in terms of physical writes ECC implemented in all current HDDs, so RAID 2 became obsolete (redundant, ironically) 8/08/2013 Summer 2013 Lecture #27 50
51 Modern Use of RAID and ECC (2/3) RAID 6: Recovering from two disk failures! RAID 5 with an extra disk s amount of parity blocks (also interleaved) Extra parity computation more complicated than Double Error Detection (not covered here) When useful? Operator replaces wrong disk during a failure Disk bandwidth is growing more slowly than disk capacity, so MTTR a disk in a RAID system is increasing (increases the chances of a 2 nd failure during repair) 8/08/2013 Summer 2013 Lecture #27 51
52 Modern Use of RAID and ECC (3/3) Common failure mode is bursts of bit errors, not just 1 or 2 Network transmissions, disks, distributed storage Contiguous sequence of bits in which first, last, or any number of intermediate bits are in error Caused by impulse noise or by fading signal strength; effect is greater at higher data rates Other tools: cyclic redundancy check, Reed Solomon, other linear codes 8/08/2013 Summer 2013 Lecture #27 52
53 Summary Great Idea: Dependability via Redundancy Reliability: MTTF & Annual Failure Rate Availability: % uptime = MTTF/MTBF RAID: Redundant Arrays of Inexpensive Disks Improve I/O rate while ensuring dependability Memory Errors: Hamming distance 2: Parity for Single Error Detect Hamming distance 3: Single Error Correction Code + encode bit position of error Hamming distance 4: SEC/Double Error Detection 8/08/2013 Summer 2013 Lecture #27 53
Price/performance Modern Memory Hierarchy
Lecture 21: Storage Administration Take QUIZ 15 over P&H 6.1-4, 6.8-9 before 11:59pm today Project: Cache Simulator, Due April 29, 2010 NEW OFFICE HOUR TIME: Tuesday 1-2, McKinley Last Time Exam discussion
Disk Storage & Dependability
Disk Storage & Dependability Computer Organization Architectures for Embedded Computing Wednesday 19 November 14 Many slides adapted from: Computer Organization and Design, Patterson & Hennessy 4th Edition,
Operating Systems. RAID Redundant Array of Independent Disks. Submitted by Ankur Niyogi 2003EE20367
Operating Systems RAID Redundant Array of Independent Disks Submitted by Ankur Niyogi 2003EE20367 YOUR DATA IS LOST@#!! Do we have backups of all our data???? - The stuff we cannot afford to lose?? How
How To Improve Performance On A Single Chip Computer
: Redundant Arrays of Inexpensive Disks this discussion is based on the paper:» A Case for Redundant Arrays of Inexpensive Disks (),» David A Patterson, Garth Gibson, and Randy H Katz,» In Proceedings
Disks and RAID. Profs. Bracy and Van Renesse. based on slides by Prof. Sirer
Disks and RAID Profs. Bracy and Van Renesse based on slides by Prof. Sirer 50 Years Old! 13th September 1956 The IBM RAMAC 350 Stored less than 5 MByte Reading from a Disk Must specify: cylinder # (distance
Definition of RAID Levels
RAID The basic idea of RAID (Redundant Array of Independent Disks) is to combine multiple inexpensive disk drives into an array of disk drives to obtain performance, capacity and reliability that exceeds
RAID. RAID 0 No redundancy ( AID?) Just stripe data over multiple disks But it does improve performance. Chapter 6 Storage and Other I/O Topics 29
RAID Redundant Array of Inexpensive (Independent) Disks Use multiple smaller disks (c.f. one large disk) Parallelism improves performance Plus extra disk(s) for redundant data storage Provides fault tolerant
Lecture 36: Chapter 6
Lecture 36: Chapter 6 Today s topic RAID 1 RAID Redundant Array of Inexpensive (Independent) Disks Use multiple smaller disks (c.f. one large disk) Parallelism improves performance Plus extra disk(s) for
CS 6290 I/O and Storage. Milos Prvulovic
CS 6290 I/O and Storage Milos Prvulovic Storage Systems I/O performance (bandwidth, latency) Bandwidth improving, but not as fast as CPU Latency improving very slowly Consequently, by Amdahl s Law: fraction
Reliability and Fault Tolerance in Storage
Reliability and Fault Tolerance in Storage Dalit Naor/ Dima Sotnikov IBM Haifa Research Storage Systems 1 Advanced Topics on Storage Systems - Spring 2014, Tel-Aviv University http://www.eng.tau.ac.il/semcom
Input / Ouput devices. I/O Chapter 8. Goals & Constraints. Measures of Performance. Anatomy of a Disk Drive. Introduction - 8.1
Introduction - 8.1 I/O Chapter 8 Disk Storage and Dependability 8.2 Buses and other connectors 8.4 I/O performance measures 8.6 Input / Ouput devices keyboard, mouse, printer, game controllers, hard drive,
Fault Tolerance & Reliability CDA 5140. Chapter 3 RAID & Sample Commercial FT Systems
Fault Tolerance & Reliability CDA 5140 Chapter 3 RAID & Sample Commercial FT Systems - basic concept in these, as with codes, is redundancy to allow system to continue operation even if some components
1 Storage Devices Summary
Chapter 1 Storage Devices Summary Dependability is vital Suitable measures Latency how long to the first bit arrives Bandwidth/throughput how fast does stuff come through after the latency period Obvious
Why disk arrays? CPUs speeds increase faster than disks. - Time won t really help workloads where disk in bottleneck
1/19 Why disk arrays? CPUs speeds increase faster than disks - Time won t really help workloads where disk in bottleneck Some applications (audio/video) require big files Disk arrays - make one logical
PIONEER RESEARCH & DEVELOPMENT GROUP
SURVEY ON RAID Aishwarya Airen 1, Aarsh Pandit 2, Anshul Sogani 3 1,2,3 A.I.T.R, Indore. Abstract RAID stands for Redundant Array of Independent Disk that is a concept which provides an efficient way for
How To Write A Disk Array
200 Chapter 7 (This observation is reinforced and elaborated in Exercises 7.5 and 7.6, and the reader is urged to work through them.) 7.2 RAID Disks are potential bottlenecks for system performance and
Why disk arrays? CPUs improving faster than disks
Why disk arrays? CPUs improving faster than disks - disks will increasingly be bottleneck New applications (audio/video) require big files (motivation for XFS) Disk arrays - make one logical disk out of
CS161: Operating Systems
CS161: Operating Systems Matt Welsh [email protected] Lecture 18: RAID April 19, 2007 2007 Matt Welsh Harvard University 1 RAID Redundant Arrays of Inexpensive Disks Invented in 1986-1987 by David Patterson
Disk Array Data Organizations and RAID
Guest Lecture for 15-440 Disk Array Data Organizations and RAID October 2010, Greg Ganger 1 Plan for today Why have multiple disks? Storage capacity, performance capacity, reliability Load distribution
Data Storage - II: Efficient Usage & Errors
Data Storage - II: Efficient Usage & Errors Week 10, Spring 2005 Updated by M. Naci Akkøk, 27.02.2004, 03.03.2005 based upon slides by Pål Halvorsen, 12.3.2002. Contains slides from: Hector Garcia-Molina
Storing Data: Disks and Files
Storing Data: Disks and Files (From Chapter 9 of textbook) Storing and Retrieving Data Database Management Systems need to: Store large volumes of data Store data reliably (so that data is not lost!) Retrieve
RAID Basics Training Guide
RAID Basics Training Guide Discover a Higher Level of Performance RAID matters. Rely on Intel RAID. Table of Contents 1. What is RAID? 2. RAID Levels RAID 0 RAID 1 RAID 5 RAID 6 RAID 10 RAID 0+1 RAID 1E
technology brief RAID Levels March 1997 Introduction Characteristics of RAID Levels
technology brief RAID Levels March 1997 Introduction RAID is an acronym for Redundant Array of Independent Disks (originally Redundant Array of Inexpensive Disks) coined in a 1987 University of California
Chapter 6. 6.1 Introduction. Storage and Other I/O Topics. p. 570( 頁 585) Fig. 6.1. I/O devices can be characterized by. I/O bus connections
Chapter 6 Storage and Other I/O Topics 6.1 Introduction I/O devices can be characterized by Behavior: input, output, storage Partner: human or machine Data rate: bytes/sec, transfers/sec I/O bus connections
How To Create A Multi Disk Raid
Click on the diagram to see RAID 0 in action RAID Level 0 requires a minimum of 2 drives to implement RAID 0 implements a striped disk array, the data is broken down into blocks and each block is written
Solving Data Loss in Massive Storage Systems Jason Resch Cleversafe
Solving Data Loss in Massive Storage Systems Jason Resch Cleversafe 2010 Storage Developer Conference. Insert Your Company Name. All Rights Reserved. 1 In the beginning There was replication Long before
CS420: Operating Systems
NK YORK COLLEGE OF PENNSYLVANIA HG OK 2 RAID YORK COLLEGE OF PENNSYLVAN James Moscola Department of Physical Sciences York College of Pennsylvania Based on Operating System Concepts, 9th Edition by Silberschatz,
Dependable Systems. 9. Redundant arrays of. Prof. Dr. Miroslaw Malek. Wintersemester 2004/05 www.informatik.hu-berlin.de/rok/zs
Dependable Systems 9. Redundant arrays of inexpensive disks (RAID) Prof. Dr. Miroslaw Malek Wintersemester 2004/05 www.informatik.hu-berlin.de/rok/zs Redundant Arrays of Inexpensive Disks (RAID) RAID is
Read this before starting!
Points missed: Student's Name: Total score: /100 points East Tennessee State University Department of Computer and Information Sciences CSCI 4717 Computer Architecture TEST 2 for Fall Semester, 2006 Section
COMP 7970 Storage Systems
COMP 797 Storage Systems Dr. Xiao Qin Department of Computer Science and Software Engineering Auburn University http://www.eng.auburn.edu/~xqin [email protected] COMP 797, Auburn University Slide 3b- Problems
Filing Systems. Filing Systems
Filing Systems At the outset we identified long-term storage as desirable characteristic of an OS. EG: On-line storage for an MIS. Convenience of not having to re-write programs. Sharing of data in an
RAID Overview 91.520
RAID Overview 91.520 1 The Motivation for RAID Computing speeds double every 3 years Disk speeds can t keep up Data needs higher MTBF than any component in system IO Performance and Availability Issues!
RAID Levels and Components Explained Page 1 of 23
RAID Levels and Components Explained Page 1 of 23 What's RAID? The purpose of this document is to explain the many forms or RAID systems, and why they are useful, and their disadvantages. RAID - Redundant
Chapter 6 External Memory. Dr. Mohamed H. Al-Meer
Chapter 6 External Memory Dr. Mohamed H. Al-Meer 6.1 Magnetic Disks Types of External Memory Magnetic Disks RAID Removable Optical CD ROM CD Recordable CD-R CD Re writable CD-RW DVD Magnetic Tape 2 Introduction
Computer Systems Structure Main Memory Organization
Computer Systems Structure Main Memory Organization Peripherals Computer Central Processing Unit Main Memory Computer Systems Interconnection Communication lines Input Output Ward 1 Ward 2 Storage/Memory
RAID. Storage-centric computing, cloud computing. Benefits:
RAID Storage-centric computing, cloud computing. Benefits: Improved reliability (via error correcting code, redundancy). Improved performance (via redundancy). Independent disks. RAID Level 0 Provides
HARD DRIVE CHARACTERISTICS REFRESHER
The read/write head of a hard drive only detects changes in the magnetic polarity of the material passing beneath it, not the direction of the polarity. Writes are performed by sending current either one
Sistemas Operativos: Input/Output Disks
Sistemas Operativos: Input/Output Disks Pedro F. Souto ([email protected]) April 28, 2012 Topics Magnetic Disks RAID Solid State Disks Topics Magnetic Disks RAID Solid State Disks Magnetic Disk Construction
File System Design and Implementation
Transactions and Reliability Sarah Diesburg Operating Systems CS 3430 Motivation File systems have lots of metadata: Free blocks, directories, file headers, indirect blocks Metadata is heavily cached for
Introduction. What is RAID? The Array and RAID Controller Concept. Click here to print this article. Re-Printed From SLCentral
Click here to print this article. Re-Printed From SLCentral RAID: An In-Depth Guide To RAID Technology Author: Tom Solinap Date Posted: January 24th, 2001 URL: http://www.slcentral.com/articles/01/1/raid
Summer Student Project Report
Summer Student Project Report Dimitris Kalimeris National and Kapodistrian University of Athens June September 2014 Abstract This report will outline two projects that were done as part of a three months
RAID HARDWARE. On board SATA RAID controller. RAID drive caddy (hot swappable) SATA RAID controller card. Anne Watson 1
RAID HARDWARE On board SATA RAID controller SATA RAID controller card RAID drive caddy (hot swappable) Anne Watson 1 RAID The word redundant means an unnecessary repetition. The word array means a lineup.
What is RAID and how does it work?
What is RAID and how does it work? What is RAID? RAID is the acronym for either redundant array of inexpensive disks or redundant array of independent disks. When first conceived at UC Berkley the former
Using RAID6 for Advanced Data Protection
Using RAI6 for Advanced ata Protection 2006 Infortrend Corporation. All rights reserved. Table of Contents The Challenge of Fault Tolerance... 3 A Compelling Technology: RAI6... 3 Parity... 4 Why Use RAI6...
RAID 5 rebuild performance in ProLiant
RAID 5 rebuild performance in ProLiant technology brief Abstract... 2 Overview of the RAID 5 rebuild process... 2 Estimating the mean-time-to-failure (MTTF)... 3 Factors affecting RAID 5 array rebuild
COSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters
COSC 6374 Parallel I/O (I) I/O basics Fall 2012 Concept of a clusters Processor 1 local disks Compute node message passing network administrative network Memory Processor 2 Network card 1 Network card
RAID Made Easy By Jon L. Jacobi, PCWorld
9916 Brooklet Drive Houston, Texas 77099 Phone 832-327-0316 www.safinatechnolgies.com RAID Made Easy By Jon L. Jacobi, PCWorld What is RAID, why do you need it, and what are all those mode numbers that
RAID: Redundant Arrays of Independent Disks
RAID: Redundant Arrays of Independent Disks Dependable Systems Dr.-Ing. Jan Richling Kommunikations- und Betriebssysteme TU Berlin Winter 2012/2013 RAID: Introduction Redundant array of inexpensive disks
RAID Level Descriptions. RAID 0 (Striping)
RAID Level Descriptions RAID 0 (Striping) Offers low cost and maximum performance, but offers no fault tolerance; a single disk failure results in TOTAL data loss. Businesses use RAID 0 mainly for tasks
CS 153 Design of Operating Systems Spring 2015
CS 153 Design of Operating Systems Spring 2015 Lecture 22: File system optimizations Physical Disk Structure Disk components Platters Surfaces Tracks Arm Track Sector Surface Sectors Cylinders Arm Heads
RAID Technology Overview
RAID Technology Overview HP Smart Array RAID Controllers HP Part Number: J6369-90050 Published: September 2007 Edition: 1 Copyright 2007 Hewlett-Packard Development Company L.P. Legal Notices Copyright
CS 464/564 Introduction to Database Management System Instructor: Abdullah Mueen
CS 464/564 Introduction to Database Management System Instructor: Abdullah Mueen LECTURE 14: DATA STORAGE AND REPRESENTATION Data Storage Memory Hierarchy Disks Fields, Records, Blocks Variable-length
RAID Implementation for StorSimple Storage Management Appliance
RAID Implementation for StorSimple Storage Management Appliance Alpa Kohli June, 2012 KB-00008 Document Revision 1 StorSimple knowledge base articles are intended to provide customers with the information
An Introduction to RAID. Giovanni Stracquadanio [email protected] www.dmi.unict.it/~stracquadanio
An Introduction to RAID Giovanni Stracquadanio [email protected] www.dmi.unict.it/~stracquadanio Outline A definition of RAID An ensemble of RAIDs JBOD RAID 0...5 Configuring and testing a Linux
Chapter 9: Peripheral Devices: Magnetic Disks
Chapter 9: Peripheral Devices: Magnetic Disks Basic Disk Operation Performance Parameters and History of Improvement Example disks RAID (Redundant Arrays of Inexpensive Disks) Improving Reliability Improving
IBM ^ xseries ServeRAID Technology
IBM ^ xseries ServeRAID Technology Reliability through RAID technology Executive Summary: t long ago, business-critical computing on industry-standard platforms was unheard of. Proprietary systems were
Striped Set, Advantages and Disadvantages of Using RAID
Algorithms and Methods for Distributed Storage Networks 4: Volume Manager and RAID Institut für Informatik Wintersemester 2007/08 RAID Redundant Array of Independent Disks Patterson, Gibson, Katz, A Case
Lecture 23: Multiprocessors
Lecture 23: Multiprocessors Today s topics: RAID Multiprocessor taxonomy Snooping-based cache coherence protocol 1 RAID 0 and RAID 1 RAID 0 has no additional redundancy (misnomer) it uses an array of disks
Guide to SATA Hard Disks Installation and RAID Configuration
Guide to SATA Hard Disks Installation and RAID Configuration 1. Guide to SATA Hard Disks Installation...2 1.1 Serial ATA (SATA) Hard Disks Installation...2 2. Guide to RAID Configurations...3 2.1 Introduction
RAID. Tiffany Yu-Han Chen. # The performance of different RAID levels # read/write/reliability (fault-tolerant)/overhead
RAID # The performance of different RAID levels # read/write/reliability (fault-tolerant)/overhead Tiffany Yu-Han Chen (These slides modified from Hao-Hua Chu National Taiwan University) RAID 0 - Striping
Silent data corruption in SATA arrays: A solution
Silent data corruption in SATA arrays: A solution Josh Eddy August 2008 Abstract Recent large academic studies have identified the surprising frequency of silent read failures that are not identified or
RAID Overview: Identifying What RAID Levels Best Meet Customer Needs. Diamond Series RAID Storage Array
ATTO Technology, Inc. Corporate Headquarters 155 Crosspoint Parkway Amherst, NY 14068 Phone: 716-691-1999 Fax: 716-691-9353 www.attotech.com [email protected] RAID Overview: Identifying What RAID Levels
Outline. Database Management and Tuning. Overview. Hardware Tuning. Johann Gamper. Unit 12
Outline Database Management and Tuning Hardware Tuning Johann Gamper 1 Free University of Bozen-Bolzano Faculty of Computer Science IDSE Unit 12 2 3 Conclusion Acknowledgements: The slides are provided
Reliability of Systems and Components
Reliability of Systems and Components GNS Systems GmbH Am Gaußberg 2 38114 Braunschweig / Germany Fri 13 Apr 2012, Revision 0.1.1 Please send comments about this document to: [email protected]
Hard Disk Drives and RAID
Hard Disk Drives and RAID Janaka Harambearachchi (Engineer/Systems Development) INTERFACES FOR HDD A computer interfaces is what allows a computer to send and retrieve information for storage devices such
What is RAID? data reliability with performance
What is RAID? RAID is the use of multiple disks and data distribution techniques to get better Resilience and/or Performance RAID stands for: Redundant Array of Inexpensive / Independent Disks RAID can
Achieving High Availability
Achieving High Availability What You Need to Know before You Start James Bottomley SteelEye Technology 21 January 2004 1 What Is Availability? Availability measures the ability of a given service to operate.
RAID. Contents. Definition and Use of the Different RAID Levels. The different RAID levels: Definition Cost / Efficiency Reliability Performance
RAID Definition and Use of the Different RAID Levels Contents The different RAID levels: Definition Cost / Efficiency Reliability Performance Further High Availability Aspects Performance Optimization
Input/output (I/O) I/O devices. Performance aspects. CS/COE1541: Intro. to Computer Architecture. Input/output subsystem.
Input/output (I/O) CS/COE1541: Intro. to Computer Architecture Input/output subsystem Sangyeun Cho Computer Science Department I/O connects User (human) and CPU (or a program running on it) Environment
Block1. Block2. Block3. Block3 Striping
Introduction to RI Team members: 電 機 一 94901150 王 麒 鈞, 電 機 一 94901151 吳 炫 逸, 電 機 一 94901154 孫 維 隆. Motivation Gosh, my hard disk is broken again, and my computer can t boot normally. I even have no chance
Homework # 2. Solutions. 4.1 What are the differences among sequential access, direct access, and random access?
ECE337 / CS341, Fall 2005 Introduction to Computer Architecture and Organization Instructor: Victor Manuel Murray Herrera Date assigned: 09/19/05, 05:00 PM Due back: 09/30/05, 8:00 AM Homework # 2 Solutions
COS 318: Operating Systems. Storage Devices. Kai Li Computer Science Department Princeton University. (http://www.cs.princeton.edu/courses/cos318/)
COS 318: Operating Systems Storage Devices Kai Li Computer Science Department Princeton University (http://www.cs.princeton.edu/courses/cos318/) Today s Topics Magnetic disks Magnetic disk performance
System Availability and Data Protection of Infortrend s ESVA Storage Solution
System Availability and Data Protection of Infortrend s ESVA Storage Solution White paper Abstract This white paper analyzes system availability and data protection on Infortrend s ESVA storage systems.
IBM System x GPFS Storage Server
IBM System x GPFS Storage Server Schöne Aussicht en für HPC Speicher ZKI-Arbeitskreis Paderborn, 15.03.2013 Karsten Kutzer Client Technical Architect Technical Computing IBM Systems & Technology Group
Programming NAND devices
Technical Guide Programming NAND devices Kelly Hirsch, Director of Advanced Technology, Data I/O Corporation Recent Design Trends In the past, embedded system designs have used NAND devices for storing
RAID technology and IBM TotalStorage NAS products
IBM TotalStorage Network Attached Storage October 2001 RAID technology and IBM TotalStorage NAS products By Janet Anglin and Chris Durham Storage Networking Architecture, SSG Page No.1 Contents 2 RAID
File System & Device Drive. Overview of Mass Storage Structure. Moving head Disk Mechanism. HDD Pictures 11/13/2014. CS341: Operating System
CS341: Operating System Lect 36: 1 st Nov 2014 Dr. A. Sahu Dept of Comp. Sc. & Engg. Indian Institute of Technology Guwahati File System & Device Drive Mass Storage Disk Structure Disk Arm Scheduling RAID
William Stallings Computer Organization and Architecture 7 th Edition. Chapter 6 External Memory
William Stallings Computer Organization and Architecture 7 th Edition Chapter 6 External Memory Types of External Memory Magnetic Disk RAID Removable Optical CD-ROM CD-Recordable (CD-R) CD-R/W DVD Magnetic
IncidentMonitor Server Specification Datasheet
IncidentMonitor Server Specification Datasheet Prepared by Monitor 24-7 Inc October 1, 2015 Contact details: [email protected] North America: +1 416 410.2716 / +1 866 364.2757 Europe: +31 088 008.4600
Guide to SATA Hard Disks Installation and RAID Configuration
Guide to SATA Hard Disks Installation and RAID Configuration 1. Guide to SATA Hard Disks Installation... 2 1.1 Serial ATA (SATA) Hard Disks Installation... 2 2. Guide to RAID Configurations... 3 2.1 Introduction
Distributed Storage Networks and Computer Forensics
Distributed Storage Networks 5 Raid-6 Encoding Technical Faculty Winter Semester 2011/12 RAID Redundant Array of Independent Disks Patterson, Gibson, Katz, A Case for Redundant Array of Inexpensive Disks,
Database Management Systems
4411 Database Management Systems Acknowledgements and copyrights: these slides are a result of combination of notes and slides with contributions from: Michael Kiffer, Arthur Bernstein, Philip Lewis, Anestis
Algorithms and Methods for Distributed Storage Networks 5 Raid-6 Encoding Christian Schindelhauer
Algorithms and Methods for Distributed Storage Networks 5 Raid-6 Encoding Institut für Informatik Wintersemester 2007/08 RAID Redundant Array of Independent Disks Patterson, Gibson, Katz, A Case for Redundant
Data Corruption In Storage Stack - Review
Theoretical Aspects of Storage Systems Autumn 2009 Chapter 2: Double Disk Failures André Brinkmann Data Corruption in the Storage Stack What are Latent Sector Errors What is Silent Data Corruption Checksum
An Introduction to RAID 6 ULTAMUS TM RAID
An Introduction to RAID 6 ULTAMUS TM RAID The highly attractive cost per GB of SATA storage capacity is causing RAID products based on the technology to grow in popularity. SATA RAID is now being used
Chapter 13 Disk Storage, Basic File Structures, and Hashing.
Chapter 13 Disk Storage, Basic File Structures, and Hashing. Copyright 2004 Pearson Education, Inc. Chapter Outline Disk Storage Devices Files of Records Operations on Files Unordered Files Ordered Files
Q & A From Hitachi Data Systems WebTech Presentation:
Q & A From Hitachi Data Systems WebTech Presentation: RAID Concepts 1. Is the chunk size the same for all Hitachi Data Systems storage systems, i.e., Adaptable Modular Systems, Network Storage Controller,
How To Understand And Understand The Power Of Aird 6 On Clariion
A Detailed Review Abstract This white paper discusses the EMC CLARiiON RAID 6 implementation available in FLARE 26 and later, including an overview of RAID 6 and the CLARiiON-specific implementation, when
Lecture 13: I/O: RAID, I/O Benchmarks, UNIX File System Performance Professor David A. Patterson Computer Science 252 Fall 1996
Lecture 3: I/O: RAID, I/O Benchmarks, UNIX File System Performance Professor David A. Patterson Computer Science 252 Fall 996 DAP.F96 Summary: A Little Queuing Theory Queue System server Proc IOC Device
California Software Labs
WHITE PAPERS FEBRUARY 2006 California Software Labs CSWL INC R e a l i z e Y o u r I d e a s Redundant Array of Inexpensive Disks (RAID) Redundant Array of Inexpensive Disks (RAID) aids development of
