Linux VM Infrastructure for memory power management



Similar documents
Linux on POWER for Green Data Center

Overview of the Current Approaches to Enhance the Linux Scheduler. Preeti U. Murthy IBM Linux Technology Center

An Oracle Technical White Paper June Oracle VM Windows Paravirtual (PV) Drivers 2.0: New Features

SmartCloud Monitoring - Capacity Planning ROI Case Study

Autodesk 3ds Max 2010 Boot Camp FAQ

Enery Efficient Dynamic Memory Bank and NV Swap Device Management

8Gb Fibre Channel Adapter of Choice in Microsoft Hyper-V Environments

Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1

COLO: COarse-grain LOck-stepping Virtual Machine for Non-stop Service

The Benefits of POWER7+ and PowerVM over Intel and an x86 Hypervisor

DELL. Virtual Desktop Infrastructure Study END-TO-END COMPUTING. Dell Enterprise Solutions Engineering

KVM & Memory Management Updates

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

Deployment Options for Microsoft Hyper-V Server

IBM Tivoli Storage Manager for Virtual Environments

SUSE Linux Enterprise 10 SP2: Virtualization Technology Support

Oracle Database Scalability in VMware ESX VMware ESX 3.5

Memory Sizing for Server Virtualization. White Paper Intel Information Technology Computer Manufacturing Server Virtualization

Deploying Citrix MetaFrame on IBM eserver BladeCenter with FAStT Storage Planning / Implementation

Using VMware VMotion with Oracle Database and EMC CLARiiON Storage Systems

WHITE PAPER Optimizing Virtual Platform Disk Performance

Performance tuning Xen

Cloud Computing through Virtualization and HPC technologies

IOS110. Virtualization 5/27/2014 1

MS Exchange Server Acceleration

Containers: Challenges with the memory resource controller and its performance

solution brief September 2011 Can You Effectively Plan For The Migration And Management of Systems And Applications on Vblock Platforms?

Virtualization Performance Analysis November 2010 Effect of SR-IOV Support in Red Hat KVM on Network Performance in Virtualized Environments

How to Run the MQX RTOS on Various RAM Memories for i.mx 6SoloX

can you effectively plan for the migration and management of systems and applications on Vblock Platforms?

IBM Security SiteProtector System Migration Utility Guide

Capacity Estimation for Linux Workloads

Steps to Migrating to a Private Cloud

z/os V1R11 Communications Server system management and monitoring

PARALLELS SERVER 4 BARE METAL README

SAS deployment on IBM Power servers with IBM PowerVM dedicated-donating LPARs

IBM ^ xseries ServeRAID Technology

Cloud Server. Parallels. An Introduction to Operating System Virtualization and Parallels Cloud Server. White Paper.

Energy-aware Memory Management through Database Buffer Control

Communications Server for Linux

Virtualizing a Virtual Machine

Hybrid Virtualization The Next Generation of XenLinux

CA Nimsoft Monitor. Probe Guide for Active Directory Server. ad_server v1.4 series

Datasheet Fujitsu Cloud Infrastructure Management Software V1

Platform LSF Version 9 Release 1.2. Migrating on Windows SC

PARALLELS SERVER BARE METAL 5.0 README

Windows Server 2008 R2 Hyper-V Live Migration

Red Hat enterprise virtualization 3.0 feature comparison

CA Capacity Manager. Product overview. agility made possible

Local and Remote Memory: Memory in a Linux/NUMA System

White Paper Amy Chong Yew Ee Online Sales Account Manager APAC Online Sales Center Intel Corporation. BOM Cost Reduction by Removing S3 State

Virtualization. Dr. Yingwu Zhu

Intel X38 Express Chipset Memory Technology and Configuration Guide

Redbooks Paper. Local versus Remote Database Access: A Performance Test. Victor Chao Leticia Cruz Nin Lei

An Oracle White Paper November Oracle Real Application Clusters One Node: The Always On Single-Instance Database

The Evolving Role of Flash in Memory Subsystems. Greg Komoto Intel Corporation Flash Memory Group

Autodesk Inventor on the Macintosh

N /150/151/160 RAID Controller. N MegaRAID CacheCade. Feature Overview

Chapter 12 File Management

Chapter 12 File Management. Roadmap

Introduction to Windows Server 2016 Nested Virtualization

HP ProLiant BL660c Gen9 and Microsoft SQL Server 2014 technical brief

Performance with the Oracle Database Cloud

How To Write To A Linux Memory Map On A Microsoft Zseries (Amd64) On A Linux (Amd32) (

Database Instance Caging: A Simple Approach to Server Consolidation. An Oracle White Paper September 2009

Hyper-V R2: What's New?

Practical Performance Understanding the Performance of Your Application

Parallels Desktop 4 for Windows and Linux Read Me

Enabling Technologies for Distributed Computing

IBM Maximo Asset Management Essentials

Intel Virtualization Technology (VT) in Converged Application Platforms

ORACLE OPS CENTER: VIRTUALIZATION MANAGEMENT PACK

Agenda. Context. System Power Management Issues. Power Capping Overview. Power capping participants. Recommendations

CA ARCserve Replication and High Availability Deployment Options for Hyper-V

Basics in Energy Information (& Communication) Systems Virtualization / Virtual Machines

Best Practices for Monitoring a Vmware Environment. Gary Powell Senior Consultant IBM SWG Tivoli

How To Manage Energy At An Energy Efficient Cost

Energy Management in a Cloud Computing Environment

Tivoli Endpoint Manager for Security and Compliance Analytics. Setup Guide

SAS Business Analytics. Base SAS for SAS 9.2

Data Deduplication: An Essential Component of your Data Protection Strategy

Dell Virtualization Solution for Microsoft SQL Server 2012 using PowerEdge R820

Using process address space on the GPU

WebSphere Commerce V7.0

Measuring Cache and Memory Latency and CPU to Memory Bandwidth

Best Practices for Increasing Ceph Performance with SSD

COSBench: A benchmark Tool for Cloud Object Storage Services. Jiangang.Duan@intel.com

Answering the Requirements of Flash-Based SSDs in the Virtualized Data Center

Dialogic PowerMedia XMS

RED HAT ENTERPRISE VIRTUALIZATION FOR SERVERS: COMPETITIVE FEATURES

Virtualization Performance on SGI UV 2000 using Red Hat Enterprise Linux 6.3 KVM

Using esxtop to Troubleshoot Performance Problems

Top 5 Reasons to choose Microsoft Windows Server 2008 R2 SP1 Hyper-V over VMware vsphere 5

An Oracle White Paper August Oracle VM 3: Server Pool Deployment Planning Considerations for Scalability and Availability

Cloud Operating Systems for Servers

An Oracle White Paper September Advanced Java Diagnostics and Monitoring Without Performance Overhead

Intel Media SDK Library Distribution and Dispatching Process

Transcription:

Linux VM Infrastructure for memory power management Ankita Garg Vaidyanathan Srinivasan IBM Linux Technology Center

Agenda - Saving Power Motivation - Why Save Power Benefits How can it be achieved Role of Linux Virtual Manager Implementation & challenges

Motivation - Why save memory power Density and capacity of memory has increased: o Leading to increased memory power consumption o Modern workload stack and virtualization demands more memory Advances in OS cooperative runtime CPU power management has reduced cpu power consumption power in the system is increasing power portion is higher in idle system CPU Mem IO OS cooperative memory power savings can help reduce system level power consumption Mem Power Higher idle power in pre-charge MEM MEM MEM MEM CPU Power CPU CPU Lower idle power due to low power states

Runtime low power modes in memory DDR3 memory allows different runtime low power states o contents are preserved o Possible performance reduction Idle memory is mostly in pre-charge state. Moving to other low power states help save power Read/Write Pre-charge High power Short duration Low power Long duration Self-refresh Power save modes (content preserved) Power-down DDR3 Power States

Runtime partial power off Partial Array Self Refresh is a method to self refresh parts of system memory - o When system is active, all regions of memory are refreshed o When system is inactive (suspend state), memory is placed in self refresh to save content o Unimportant and free memory need not be self refreshed while in suspend mode, leading to longer battery life Self-refresh power Partial array self refresh Self-refresh enabled bank Self-refresh disabled bank Self-refresh area Rank Content not preserved but power saved on free memory Power off unused memory Unused parts of memory hierarchy can be turned off (contents not preserved) even in an active system Controller 0 Controller 1 CPU

What this means to Linux OS Type-A (performance loss) o Less memory access consumes less power banks that are not accessed for some time (10s micro seconds) can be transitioned to low power state Performance impact on first access to get the memory out to active state o power savings happen at certain granularity Consolidating access around this granularity saves power Type-B (content loss) o Parts of memory can be turned off losing contents Chunks of free memory or unimportant data is aligned with the power off granularity, then save power during runtime as well as suspend Granularity varies for each of the above feature Data Pages region Type-A Low power state Content preserved Low power state Content lost Type-B

The benefits - Longer battery runtime Saves 2.7% Actively used data pages Type-A: in self refresh un-referenced data pages 1GB Controller 0 1GB Controller 1 2-6% Power savings demonstrated in Samsung EXYNOS board with 2GB RAM Further cooperation with DEVFREQ can improve savings o bandwidth control Saves 6.3% Actively used data pages CPU Type-B: Power off unused memory No data pages 1GB 1GB Controller 0 Controller 1 CPU

Opportunity for Power Savings is over-provisioned There are blocks that are not referenced Potential Savings

The Goal accesses Organize/shape reference accesses Region 0 Region 1 Time Time Standby Power Down Less power Self-refresh Least power Consolidate memory reference for Type-A savings Consolidate allocations for Type-B savings

Role of Virtual Manager Organize memory into distinct memory regions Align memory operations across these regions Allocate pages so that we fill one region at a time Maintain per-region allocated/free memory statistics Free pages with an intention to completely free a region Track page-type (dirty, read-only, etc), to estimate the cost of reclaim vs migrate vs spilling over to next region Design OS callbacks to track first page allocation and last page deallocation Mechanism to group regions Present VM organization

Implementation Schemes regions o Data structure between nodes and zones, that aligns with memory hardware power management granularity

Pros & Cons of Regions Advantages o A region aligns with hardware address boundary o Sequential allocation within region o Targeted reclaim within region o Statistics gathered on a per-region granularity o Grouping of regions Disadvantages o Zones are fragmented o Dereference overhead o Modification of core VM structures (memory layout change)

Region Experimental Data Allocation of memory from one regions at a time

Region Overhead Estimating overhead of memory region -- zone fragmentation Lmbench: File & VM latencies vary from 0-3% variation Kernbench on 2 GB RAM with two memory regions 2 GB RAM Kernbench -j 4 Samsung Exynos board with 2GB RAM 2 GB RAM Kernbench -j 4 Regions 0 (1GB) Regions 1 (1GB) 2% Performance variation

Alternate Implementation Schemes Track statistics using buddy allocator o No address boundary information, so cumbersome to traverse and find pages belonging to the same region o Can control access to region o Beneficial for PASR compaction technique or Lumpy reclaim o Free-up region sized chunks of memory o Move allocated/free pages around Ballooning driver -- create artifical memory pressure o Free unused pages to aid PASR Contiguous Allocator (CMA) Fake NUMA nodes

Implementation Challenges interleaving vs power savings Accurately capturing rate of memory references Low overhead tracking of region usage statistics is difficult Kernel memory allocations are non-movable Evaluating the threshold for reclaim vs migration Determining the amount of memory required to sustain certain performance level (unlike CPU) Virtualization Guest VM does not have complete view of memory

Reference Ankita's memory regions (LWN) Summary of requirements Kernel Summit 2011 proposal and discussions o Compaction (Mel Gorman) Devfreq, DVFS for devices General PASR Reference Pattern Instrumentation Power Management http://www.energystar.gov/ia/products/downloads/gpapadopoulos_keynote.pdf http://www.crucial.com/pdf/crucial_energy_efficient_memory.pdf

Legal statements Copyright International Business Machines Corporation 2011. Permission to redistribute in accordance with Linux Plumbers Conference submission guidelines is granted; all other rights reserved. This work represents the view of the authors and does not necessarily represent the view of IBM. IBM, IBM logo, ibm.com are trademarks of International Business Machines Corporation in the United States, other countries, or both. Intel is a trademark or registered trademark of Intel Corporation or its subsidiaries in the United States and other countries. Linux is a registered trademark of Linus Torvalds in the United States, other countries, or both. Other company, product, and service names may be trademarks or service marks of others. References in this publication to IBM products or services do not imply that IBM intends to make them available in all countries in which IBM operates. INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of express or implied warranties in certain transactions, therefore, this statement may not apply to you. This information could include technical inaccuracies or typographical errors. Changes are periodically made to the information herein; these changes will be incorporated in new editions of the publication. IBM may make improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time without notice.