Reliability-Aware Energy Management for Hybrid Storage Systems



Similar documents
Energy aware RAID Configuration for Large Storage Systems

Lab Evaluation of NetApp Hybrid Array with Flash Pool Technology

Understanding the Economics of Flash Storage

Understanding Flash SSD Performance

Price/performance Modern Memory Hierarchy

Storage Class Memory and the data center of the future

Write Off-Loading: Practical Power Management for Enterprise Storage

The Case for Massive Arrays of Idle Disks (MAID)

Silverton Consulting, Inc. StorInt Briefing

The Data Placement Challenge

Optimizing enterprise data storage capacity and performance to reduce your data footprint

Power-Aware Autonomous Distributed Storage Systems for Internet Hosting Service Platforms

2009 Oracle Corporation 1

EMC XtremSF: Delivering Next Generation Performance for Oracle Database

Analyzing Big Data with Splunk A Cost Effective Storage Architecture and Solution

Firebird and RAID. Choosing the right RAID configuration for Firebird. Paul Reeves IBPhoenix. mail:

The Technologies & Architectures. President, Demartek

The Use of Flash in Large-Scale Storage Systems.

Xanadu 130. Business Class Storage Solution. 8G FC Host Connectivity and 6G SAS Backplane. 2U 12-Bay 3.5 Form Factor

Accelerating Server Storage Performance on Lenovo ThinkServer

SLIDE 1 Previous Next Exit

NetApp FAS Hybrid Array Flash Efficiency. Silverton Consulting, Inc. StorInt Briefing

Oracle Aware Flash: Maximizing Performance and Availability for your Database

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

HP Smart Array Controllers and basic RAID performance factors

Comparison of Hybrid Flash Storage System Performance

COS 318: Operating Systems. Storage Devices. Kai Li and Andy Bavier Computer Science Department Princeton University

WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE

Input / Ouput devices. I/O Chapter 8. Goals & Constraints. Measures of Performance. Anatomy of a Disk Drive. Introduction - 8.1

RAID. RAID 0 No redundancy ( AID?) Just stripe data over multiple disks But it does improve performance. Chapter 6 Storage and Other I/O Topics 29

All-Flash Storage Solution for SAP HANA:

LSI MegaRAID CacheCade Performance Evaluation in a Web Server Environment

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

Clash of the Titans. I/O System Performance. mag. Sergej Rožman; Abakus plus d.o.o.

Solid State Drive Architecture

Intel RAID Controllers

The IntelliMagic White Paper: Storage Performance Analysis for an IBM Storwize V7000

Energy Efficient Storage Management Cooperated with Large Data Intensive Applications

COS 318: Operating Systems. Storage Devices. Kai Li Computer Science Department Princeton University. (

Reducing Storage Energy Consumption by up to 75%

Application-Focused Flash Acceleration

Evaluation Report: Supporting Microsoft Exchange on the Lenovo S3200 Hybrid Array

An Overview of Flash Storage for Databases

How To Scale Myroster With Flash Memory From Hgst On A Flash Flash Flash Memory On A Slave Server

1 Storage Devices Summary

White paper. QNAP Turbo NAS with SSD Cache

SSD Performance Tips: Avoid The Write Cliff

SSDs and RAID: What s the right strategy. Paul Goodwin VP Product Development Avant Technology

Technologies Supporting Evolution of SSDs

Database Hardware Selection Guidelines

Energy Efficient Management Scheme for Heterogeneous Secondary Storage System in Mobile Computers

Increasing Storage Performance

Performance of vsphere Flash Read Cache in VMware vsphere 5.5

OPTIMIZING EXCHANGE SERVER IN A TIERED STORAGE ENVIRONMENT WHITE PAPER NOVEMBER 2006

Flash Memory Technology in Enterprise Storage

Flash 101. Violin Memory Switzerland. Violin Memory Inc. Proprietary 1

Difference between Enterprise SATA HDDs and Desktop HDDs. Difference between Enterprise Class HDD & Desktop HDD

Accelerating MS SQL Server 2012

HYBRID STORAGE WITH FASTier ACCELERATION TECHNOLOGY

Chapter Introduction. Storage and Other I/O Topics. p. 570( 頁 585) Fig I/O devices can be characterized by. I/O bus connections

Getting the Most Out of Flash Storage

VNX HYBRID FLASH BEST PRACTICES FOR PERFORMANCE

Improve Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database

A POWER AND PERFORMANCE MEASUREMENT FRAMEWORK FOR SERVER- CLASS STORAGE

File System & Device Drive. Overview of Mass Storage Structure. Moving head Disk Mechanism. HDD Pictures 11/13/2014. CS341: Operating System

HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW

Performance Report Modular RAID for PRIMERGY

Enhancements of ETERNUS DX / SF

June Blade.org 2009 ALL RIGHTS RESERVED

How Virtualization Breaks Traditional Storage

Disk Storage & Dependability

SSDs: Practical Ways to Accelerate Virtual Servers

3PAR Fast RAID: High Performance Without Compromise

PARALLELS CLOUD STORAGE

Flash Memory Arrays Enabling the Virtualized Data Center. July 2010

Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance

Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays

RAID Storage System of Standalone NVR

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database

Flash Storage Roles & Opportunities. L.A. Hoffman/Ed Delgado CIO & Senior Storage Engineer Goodwin Procter L.L.P.

Q & A From Hitachi Data Systems WebTech Presentation:

Flash Performance in Storage Systems. Bill Moore Chief Engineer, Storage Systems Sun Microsystems

Enterprise-class versus Desktopclass

DELL TM PowerEdge TM T Mailbox Resiliency Exchange 2010 Storage Solution

Boosting Database Batch workloads using Flash Memory SSDs

SSDs: Practical Ways to Accelerate Virtual Servers

NEXSAN NST STORAGE FOR THE VIRTUAL DESKTOP

EqualLogic PS Series Load Balancers and Tiering, a Look Under the Covers. Keith Swindell Dell Storage Product Planning Manager

Virtuoso and Database Scalability

Reducing Energy Consumption of Disk Storage Using Power-Aware Cache Management

High Performance SQL Server with Storage Center 6.4 All Flash Array

Leveraging EMC Fully Automated Storage Tiering (FAST) and FAST Cache for SQL Server Enterprise Deployments

IOmark- VDI. Nimbus Data Gemini Test Report: VDI a Test Report Date: 6, September

RAID Level Descriptions. RAID 0 (Striping)

Accelerate Oracle Performance by Using SmartCache of T Series Unified Storage

Choosing Storage Systems

Reduce Latency and Increase Application Performance Up to 44x with Adaptec maxcache 3.0 SSD Read and Write Caching Solutions

89 Fifth Avenue, 7th Floor. New York, NY White Paper. HP 3PAR Adaptive Flash Cache: A Competitive Comparison

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014

DMS Performance Tuning Guide for SQL Server

Transcription:

MSST Research Track, May 2011 Reliability-Aware Energy Management for Hybrid Storage Systems Wes Felter, Anthony Hylick, John Carter IBM Research - Austin

Energy Saving using Hybrid Storage with Flash Caching Storage Controller Server Server Server I/O requests DRAM cache Flash cache RAID SATA Spindown Control 85% energy savings Goal: Demonstrate significant disk energy savings for storage systems Constraint: Maintain performance and reliability Target: Medium-duty workloads can tolerate infrequent multi-second spin up delays, e.g., email, web, and file servers How we do it: Use flash SSD as a secondary cache behind DRAM Exploit (or create) opportunities to spin down idle disks Use token bucket to limit disk spinup wear Why this saves energy: Replaces high-energy disks (e.g., SAS/FC) with low-energy disks (e.g., SATA) and SSDs Spins down disks that are idle because I/O requests are serviced by the flash cache Page 2

Disk spindown background Disks are not very energy-proportional idle uses nearly the same power as active Significant energy savings requires spindown Spinup takes time and consumes significant energy Breakeven time is critical Plenty of work in this area Extending battery life in laptops Spindown timeout of 2x breakeven time shown to be competitive Workload-adaptive timeouts Servers MAID, power-aware RAID and caching Most prior work treats disk reliability naively Page 3

Disk Energy Management Disks starting to support multiple idle states: Idle_A: Everything on Idle_B: A + some electronics off Idle_C: B + Lower RPM, park head Standby: C + Spindle motor off Trade off power and response time Most savings comes from Standby Mode Power Recovery Time Breakeven Time Max Rate Idle_A 5.8 W 0 s 0 s 1 s Idle_B 4.5 W 0 s 1 s 4 min Idle_C 3.5 W 0.4 s 2.3 s 10 min Standby 0.3 W 6 s 15.4 s 15 min From Western Digital RE2-GP and Seagate Constellation 3.5 SATA disks Observation: Caching can increase idle intervals enable more spindown Non-linear relationship between I/O rate and power consumption Constraint: Each state has a reliability limit Page 4

Managing Reliability with Token Bucket Spindown Disks are rated for a limited number of lifetime spin-ups Number varies depending on technology (e.g., SAS vs SATA) Typical conservative default spindown policy: fixed timeout = lifetime / # of spin-ups Reliability dictates spindown frequency Energy break-even point: 15 seconds (measured) Reliability constraint: one spindown per 15 minutes (lifetime average) Spindowns are a precious resource do not waste opportunities Fixed timeout policy wastes spindown opportunities during long idle/active phases (e.g., 10 hours of idle time overnight 40 unused spindown opportunities) Key Idea: Use token bucket (from networking) to jointly manage energy & reliability Add one spindown token to bucket as often as reliability allows (e.g., 15 mins) Energy management policy can only spin down disk if token is available Allows more aggressive spindown (e.g., after 1 idle minute) Separate token bucket for each idle state Workloa d Disk Lifetime Page 5 Add token every 15 mins Accumulated spindown opportunities Remove token before spinning down Can be combined with any spindown policy. proj_1 proj_2 prxy_1 usr_1 src1_1 4 years 14 years 1 year 2 years 5 years

Experimental Evaluation Used five MSR block I/O traces proj_1, proj_2, prxy_1, usr_1, src1_1 Two sets of experiments: Simulation Hardware testbed Baseline Configuration 8 450 GB 3.5 SAS disks, RAID-6 (2.7 TB) Hybrid Storage Configuration 8 750 GB 3.5 SATA disks, RAID-6 (4.5 TB) 2 100 GB (128 GB raw) SandForce SF-1500 SSD cache (mirrored) Approximately equal-cost configurations Note: SATA gives extra capacity (unused in our experiments) Page 6

TRAIDe Simulator Trace-driven RAID array energy Simulator Block trace-driven storage array software simulator RAID-5 and RAID-6 Energy-aware LRU flash caching Several disk spindown policies Outputs the time and energy spent in each power state (reading, writing, seeking, spindown, idle, etc.) per disk Based upon prior research that accurately generates disk energy models from performance characteristics Minimal disk profiling required Seek time taken from disk data sheets Does not model detailed timing for each request Simulator output validated to be within 5% of measured energy Page 7

Policies Studied SAS: Conventional configuration w/ SAS disks, no flash, no spindown SATA: SATA disks, no flash cache, conservative fixedtimeout spindown Write-back caching (WC): SATA + write-back mirrored energy-aware flash cache (Zhu et al.) Token Bucket (TB): WC + competitive spindown algorithm moderated by token bucket our contribution Disk spins down when it is idle for twice the breakeven time and a token is available Safe Oracle (SO): WC + reliability-aware oracle spindown Disk spins down during the longest 672 intervals (avg. one spindown per 15 mins for one week) Disk exactly meets its reliability target Lowest possible energy while maintaining reliability Page 8

Time spent in each power state Simulation Page 9

Energy consumption simulation Page 10

The image cannot be displayed. Your computer may not have enough memory to open the image, or the image may have been corrupted. Restart your computer, and then open the file again. If the red x still appears, you may have to delete the image and then insert it again. IBM Research - Austin Token accumulation over time Tokens accumulate over weekend Bursty I/O consumes tokens Some workloads never consume all their tokens One week Page 11

Performance measured experimentally Hybrid SATA+Flash SAS Heaviest hour of proj2 Measured on real hardware x86 server Linux storage stack Custom flash cache Equal-cost comparison: -8 15K SAS disks vs -8 7200 SATA disks + 100 GB SSD flash cache SATA capacity >> SAS System with caching is as fast or faster than without (note log scale!) proj2 representative of all runs (essentially identically-shaped CDF plots) Page 12

Related Work Making hard disks more energy efficient DRPM (Gurumurthi 2003) Intra-Disk Parallelism (Sankar 2008) Disk Spin-down Techniques Laptops (Wilkes 1992, Douglis 1995) Massive Array of Idle Disks (MAID) (Colerelli 2002) Popular Data Concentration (Pinheiro 2004) PARAID (Weddle 2007) Write Off-Loading (Narayanan 2008) Flash Caching SieveStore (Pritchett 2010) FlashCache (Kgil 2006) Energy-Aware Caching Power-Aware Cache Management (Zhu 2004) NVCache (Bisson 2006) Augmenting RAID with SSD (Lee 2008) C-Burst (Chen 2008) Disk Reliability Failure trends in a large disk drive population (Pinheiro 2007) Page 13

Conclusions 85% energy savings possible with spindown and hybrid storage Disk energy management must be reliability-aware Reliability management and energy management are separable concerns Token bucket reliability management is near-optimal Intermediate power states provide little benefit Page 14

Thank you! Source: If applicable, describe source origin Page 15 Questions?

SAS vs. SATA disks SAS : High RPM (10-15K), lower latency, lower capacity, higher power, higher MTBF*, fewer spinups, higher cost SATA : Low RPM (5-7K), higher latency, higher capacity, lower power, lower MTBF, more spinups, lower cost Conventional wisdom: Only SAS drives can meet enterprise workload demands E.g. Sub-10 ms latency Flash changes the situation Sub-ms flash latency can offset slower SATA disks Page 16

Reliability-Aware (Safe) vs. Unsafe Oracle Page 17

Why time in each state Include hit rates here Workloa d Cache Read Hit Rate (%) proj_1 39 proj_2 52 prxy_1 65 usr_1 67 src1_1 85 Page 18

Performance (Simulated) Page 19 -- IBM Confidential--