COMP 7970 Storage Systems



Similar documents
CS 6290 I/O and Storage. Milos Prvulovic

Input / Ouput devices. I/O Chapter 8. Goals & Constraints. Measures of Performance. Anatomy of a Disk Drive. Introduction - 8.1

Price/performance Modern Memory Hierarchy

Fault Tolerance & Reliability CDA Chapter 3 RAID & Sample Commercial FT Systems

D.1 Introduction D-2 D.2 Advanced Topics in Disk Storage D-2 D.3 Definition and Examples of Real Faults and Failures D-10 D.

How To Improve Performance On A Single Chip Computer

Lecture 36: Chapter 6

File System & Device Drive. Overview of Mass Storage Structure. Moving head Disk Mechanism. HDD Pictures 11/13/2014. CS341: Operating System

Hard Disk Drives and RAID

Distribution One Server Requirements

CS 61C: Great Ideas in Computer Architecture. Dependability: Parity, RAID, ECC

IncidentMonitor Server Specification Datasheet

Disk Storage & Dependability

How To Create A Multi Disk Raid

1 Storage Devices Summary

technology brief RAID Levels March 1997 Introduction Characteristics of RAID Levels

How To Write A Disk Array

Data Storage - II: Efficient Usage & Errors

RAID. RAID 0 No redundancy ( AID?) Just stripe data over multiple disks But it does improve performance. Chapter 6 Storage and Other I/O Topics 29

Using Multipathing Technology to Achieve a High Availability Solution

Dependable Systems. 9. Redundant arrays of. Prof. Dr. Miroslaw Malek. Wintersemester 2004/05

RAID 5 rebuild performance in ProLiant

QuickSpecs. HP Smart Array 5312 Controller. Overview

Input/output (I/O) I/O devices. Performance aspects. CS/COE1541: Intro. to Computer Architecture. Input/output subsystem.

Disk Array Data Organizations and RAID

Chapter Introduction. Storage and Other I/O Topics. p. 570( 頁 585) Fig I/O devices can be characterized by. I/O bus connections

Overview of I/O Performance and RAID in an RDBMS Environment. By: Edward Whalen Performance Tuning Corporation

Definition of RAID Levels

RAID Technology Overview

PIONEER RESEARCH & DEVELOPMENT GROUP

HP Smart Array 5i Plus Controller and Battery Backed Write Cache (BBWC) Enabler

Enhancing IBM Netfinity Server Reliability

Operating Systems. RAID Redundant Array of Independent Disks. Submitted by Ankur Niyogi 2003EE20367

RAID HARDWARE. On board SATA RAID controller. RAID drive caddy (hot swappable) SATA RAID controller card. Anne Watson 1

SAN Conceptual and Design Basics

NEC Corporation of America Intro to High Availability / Fault Tolerant Solutions

CS161: Operating Systems

HP recommended configuration for Microsoft Exchange Server 2010: HP LeftHand P4000 SAN

IBM ^ xseries ServeRAID Technology

Models Smart Array 6402A/128 Controller 3X-KZPEC-BF Smart Array 6404A/256 two 2 channel Controllers

Data Storage and Backup. Sanjay Goel School of Business University at Albany, SUNY

Chapter 10: Mass-Storage Systems

Today s Papers. RAID Basics (Two optional papers) Array Reliability. EECS 262a Advanced Topics in Computer Systems Lecture 4

Reliability and Fault Tolerance in Storage

RAID Overview: Identifying What RAID Levels Best Meet Customer Needs. Diamond Series RAID Storage Array

Solving Data Loss in Massive Storage Systems Jason Resch Cleversafe

Hardware Performance Optimization and Tuning. Presenter: Tom Arakelian Assistant: Guy Ingalls

EMC Backup and Recovery for Microsoft SQL Server 2008 Enabled by EMC Celerra Unified Storage

Traditionally, a typical SAN topology uses fibre channel switch wiring while a typical NAS topology uses TCP/IP protocol over common networking

HA / DR Jargon Buster High Availability / Disaster Recovery

COS 318: Operating Systems. Storage Devices. Kai Li Computer Science Department Princeton University. (

Assessing RAID ADG vs. RAID 5 vs. RAID 1+0

COSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters

RAID Overview

Data Corruption In Storage Stack - Review

Network Design. Yiannos Mylonas

Enterprise Edition. Hardware Requirements

Why disk arrays? CPUs speeds increase faster than disks. - Time won t really help workloads where disk in bottleneck

Chapter 6 External Memory. Dr. Mohamed H. Al-Meer

Distributed RAID Architectures for Cluster I/O Computing. Kai Hwang

Distributed Storage Networks and Computer Forensics

Distributed File System Performance. Milind Saraph / Rich Sudlow Office of Information Technologies University of Notre Dame

Algorithms and Methods for Distributed Storage Networks 5 Raid-6 Encoding Christian Schindelhauer

Intel RAID Controllers

How Does the ECASD Network Work?

HARVARD RESEARCH GROUP, Inc.

RAID Level Descriptions. RAID 0 (Striping)

Introduction. What is RAID? The Array and RAID Controller Concept. Click here to print this article. Re-Printed From SLCentral

BrightStor ARCserve Backup for Windows

ORACLE DATABASE HIGH AVAILABILITY STRATEGY, ARCHITECTURE AND SOLUTIONS

Deploying a File Server Lesson 2

Outline. Database Management and Tuning. Overview. Hardware Tuning. Johann Gamper. Unit 12

Seradex White Paper. Focus on these points for optimizing the performance of a Seradex ERP SQL database:

OVERVIEW. CEP Cluster Server is Ideal For: First-time users who want to make applications highly available

Filing Systems. Filing Systems

MANAGING DISK STORAGE

Professional and Enterprise Edition. Hardware Requirements

Microsoft Exchange 2010 on Dell Systems. Simple Distributed Configurations

RAID Implementation for StorSimple Storage Management Appliance

COMPARISON OF VMware VSHPERE HA/FT vs stratus

Using RAID6 for Advanced Data Protection

COS 318: Operating Systems. Storage Devices. Kai Li and Andy Bavier Computer Science Department Princeton University

Evaluation of Dell PowerEdge VRTX Shared PERC8 in Failover Scenario

Server and Storage Virtualization with IP Storage. David Dale, NetApp

Module 6. RAID and Expansion Devices

New Advanced RAID Level for Today's Larger Storage Capacities: Advanced Data Guarding

Lecture 13: I/O: RAID, I/O Benchmarks, UNIX File System Performance Professor David A. Patterson Computer Science 252 Fall 1996

Stretching A Wolfpack Cluster Of Servers For Disaster Tolerance. Dick Wilkins Program Manager Hewlett-Packard Co. Redmond, WA dick_wilkins@hp.

Fault Tolerant Servers: The Choice for Continuous Availability on Microsoft Windows Server Platform

Availability and Disaster Recovery: Basic Principles

CSE-E5430 Scalable Cloud Computing P Lecture 5

LSI SAS inside 60% of servers. 21 million LSI SAS & MegaRAID solutions shipped over last 3 years. 9 out of 10 top server vendors use MegaRAID

VTrak SATA RAID Storage System

Transcription:

COMP 797 Storage Systems Dr. Xiao Qin Department of Computer Science and Software Engineering Auburn University http://www.eng.auburn.edu/~xqin xqin@auburn.edu COMP 797, Auburn University Slide 3b-

Problems of Disk Arrays Small Writes RAID-5: Small Write Algorithm Logical Write = 2 Physical Reads + 2 Physical Writes D' D D D2 D3 P new data old data (. Read) old (2. Read) parity + XOR + XOR (3. Write) (4. Write) D' D D2 D3 P' COMP 797, Auburn University Slide 3b-2

RAID 6: P + Q Redundancy block block block 2 block 3 P(-3) block 4 block 5 block 6 P(4-6) Q(9 2 5...) block 7 block 8 P(7-9) Q(3 4...) block 9 block P(-2) Q(2 6 3...) block block 2 P(2-5) Q( 5 8...) block 3 block 4 block 5 Q( 4 7...) An extension to RAID 5 but with two-dimensional parity. Each row has P parity and each row has Q parity. (Reed-Solomon Codes) Has an extremely high data fault tolerance and can sustain multiple simultaneous drive failures Rarely implemented More information, please see the paper: A tutorial on Reed-Solomon Coding for Fault Tolerance in RAID-like Systems COMP 797, Auburn University Slide 3b-3

Berkeley History: RAID-I RAID-I (989) Consisted of a Sun 4/28 workstation with 28 MB of DRAM, four dualstring SCSI controllers, 28 5.25-inch SCSI disks and specialized disk striping software Today RAID is $24 billion dollar industry, 8% nonpc disks sold in RAIDs COMP 797, Auburn University Slide 3b-4

Summary: RAID Techniques: Goal was performance, popularity due to reliability of storage Disk Mirroring, Shadowing (RAID ) Each disk is fully duplicated onto its "shadow" Logical write = two physical writes % capacity overhead Parity Data Bandwidth Array (RAID 3) Parity computed horizontally Logically a single high data bw disk High I/O Rate Parity Array (RAID 5) Interleaved parity blocks Independent reads and writes Logical write = 2 reads + 2 writes COMP 797, Auburn University Slide 3b-5

Definitions Examples on why precise definitions so important for reliability Is a programming mistake a fault, error, or failure? Are we talking about the time it was designed or the time the program is run? If the running program doesn t exercise the mistake, is it still a fault/error/failure? If an alpha particle hits a DRAM memory cell, is it a fault/error/failure if it doesn t change the value? Is it a fault/error/failure if the memory doesn t access the changed bit? Did a fault/error/failure still occur if the memory had error correction and delivered the corrected value to the CPU? COMP 797, Auburn University Slide 3b-6

IFIP Standard terminology Computer system dependability: quality of delivered service such that reliance can be placed on service Service is observed actual behavior as perceived by other system(s) interacting with this system s users Each module has ideal specified behavior, where service specification is agreed description of expected behavior A system failure occurs when the actual behavior deviates from the specified behavior failure occurred because an error, a defect in module The cause of an error is a fault When a fault occurs it creates a latent error, which becomes effective when it is activated When error actually affects the delivered service, a failure occurs (time from error to failure is error latency) COMP 797, Auburn University Slide 3b-7

Fault v. (Latent) Error v. Failure An error is manifestation in the system of a fault, a failure is manifestation on the service of an error Is If an alpha particle hits a DRAM memory cell, is it a fault/error/failure if it doesn t change the value? Is it a fault/error/failure if the memory doesn t access the changed bit? Did a fault/error/failure still occur if the memory had error correction and delivered the corrected value to the CPU? An alpha particle hitting a DRAM can be a fault if it changes the memory, it creates an error error remains latent until effected memory word is read if the effected word error affects the delivered service, a failure occurs COMP 797, Auburn University Slide 3b-8

Fault Categories. Hardware faults: Devices that fail, such alpha particle hitting a memory cell 2. Design faults: Faults in software (usually) and hardware design (occasionally) 3. Operation faults: Mistakes by operations and maintenance personnel 4. Environmental faults: Fire, flood, earthquake, power failure, and sabotage Also by duration:. Transient faults exist for limited time and not recurring 2. Intermittent faults cause a system to oscillate between faulty and faultfree operation 3. Permanent faults do not correct themselves over time COMP 797, Auburn University Slide 3b-9

Fault Tolerance vs Disaster Tolerance Fault-Tolerance (or more properly, Error- Tolerance): mask local faults (prevent errors from becoming failures) RAID disks Uninterruptible Power Supplies Cluster Failover Disaster Tolerance: masks site errors (prevent site errors from causing service failures) Protects against fire, flood, sabotage,.. Redundant system and service at remote site. Use design diversity From Jim Gray s Talk at UC Berkeley on Fault Tolerance " /9/ COMP 797, Auburn University Slide 3b-

Case Studies - Tandem Trends Reported MTTF by Component 45 4 35 3 25 2 5 Mean Time to System Failure (years) by Cause environment maintenance hardware operations 5 total 985 987 989 COMP 797, Auburn University software 985 987 99 SOFTWARE 2 53 33 Years HARDWARE 29 9 3 Years MAINTENANCE 45 62 49 Years OPERATIONS 99 7 36 Years ENVIRONMENT 42 24 346 Years SYSTEM 8 2 2 Years Problem: Systematic Under-reporting From Jim Gray s Talk at UC Berkeley on Fault Tolerance " /9/ Slide 3b-

% 8% 6% 4% 2% % 5% 5% 2% Is Maintenance the Key? Rule of Thumb: Maintenance X HW so over 5 year product life, ~ 95% of cost is maintenance Cause of System Crashes 5% 8% 2% 53% 8% % 985 993 2 VAX crashes 85, 93 [Murp95]; extrap. to Sys. Man.: N crashes/problem, SysAdmin action Actions: set params bad, bad config, bad app install HW/OS 7% in 85 to 28% in 93. In, %? COMP 797, Auburn University 69% 5% 5% (est.) Other: app, power, network failure System management: actions + N/problem Operating System failure Hardware failure Slide 3b-2

HW Failures in Real Systems: Tertiary Disks A cluster of 2 PCs in seven 7-foot high, 9-inch wide racks with 368 8.4 GB, 72 RPM, 3.5-inch IBM disks. The PCs are P6-2MHz with 96 MB of DRAM each. They run FreeBSD 3. and the hosts are connected via switched Mbit/second Ethernet Component Total in System Total Failed % Failed SCSI Controller 44 2.3% SCSI Cable 39 2.6% SCSI Disk 368 7.9% IDE Disk 24 6 25.% Disk Enclosure -Backplane 46 3 28.3% Disk Enclosure - Power Supply 92 3 3.3% Ethernet Controller 2 5.% Ethernet Switch 2 5.% Ethernet Cable 42 2.3% CPU/Motherboard 2 % COMP 797, Auburn University Slide 3b-3

Does Hardware Fail Fast? 4 of 384 Disks that failed in Tertiary Disk Messages in system log for failed disk No. log msgs Duration (hours) Hardware Failure (Peripheral device write fault [for] Field Replaceable Unit) Not Ready (Diagnostic failure: ASCQ = Component ID [of] Field Replaceable Unit) Recovered Error (Failure Prediction Threshold Exceeded [for] Field Replaceable Unit) Recovered Error (Failure Prediction Threshold Exceeded [for] Field Replaceable Unit) 763 86 46 9 33 5 43 7 COMP 797, Auburn University Slide 3b-4

High Availability System Classes Goal: Build Class 6 Systems System Type Unmanaged Managed Well Managed Fault Tolerant High-Availability Very-High-Availability Ultra-Availability Unavailable (min/year) 5, 5, 5 5 5.5.5 Availability 9.% 99.% 99.9% 99.99% 99.999% 99.9999% 99.99999% Availability Class 2 3 4 5 6 7 UnAvailability = MTTR/MTBF can cut it in ½ by cutting MTTR or MTBF From Jim Gray s Talk at UC Berkeley on Fault Tolerance " /9/ COMP 797, Auburn University Slide 3b-5

How Realistic is "5 Nines"? HP claims HP-9 server HW and HP-UX OS can deliver 99.999% availability guarantee in certain pre-defined, pre-tested customer environments Application faults? Operator faults? Environmental faults? Collocation sites (lots of computers in building on Internet) have network outage per year (~ day) power failure per year (~ day) Microsoft Network unavailable recently for a day due to problem in Domain Name Server: if only outage per year, 99.7% or 2 Nines COMP 797, Auburn University Slide 3b-6