Capacity Estimation for Linux Workloads



Similar documents
IOS110. Virtualization 5/27/2014 1

Introduction 1 Performance on Hosted Server 1. Benchmarks 2. System Requirements 7 Load Balancing 7

BridgeWays Management Pack for VMware ESX

Managing Capacity Using VMware vcenter CapacityIQ TECHNICAL WHITE PAPER

Enabling Technologies for Distributed Computing

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

White Paper. Recording Server Virtualization

Introduction to the NI Real-Time Hypervisor

Virtualization Technologies and Blackboard: The Future of Blackboard Software on Multi-Core Technologies

Performance Comparison of Fujitsu PRIMERGY and PRIMEPOWER Servers

Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging

Enabling Technologies for Distributed and Cloud Computing

Cloud Computing through Virtualization and HPC technologies

Rackspace Cloud Databases and Container-based Virtualization

Thomas Fahrig Senior Developer Hypervisor Team. Hypervisor Architecture Terminology Goals Basics Details

An Oracle White Paper November Oracle Real Application Clusters One Node: The Always On Single-Instance Database

Microkernels, virtualization, exokernels. Tutorial 1 CSC469

Performance And Scalability In Oracle9i And SQL Server 2000

Parallels Virtuozzo Containers vs. VMware Virtual Infrastructure:

Understanding Linux on z/vm Steal Time

Deputy Secretary for Information Technology Date Issued: November 20, 2009 Date Revised: December 20, Revision History Description:

Delivering Quality in Software Performance and Scalability Testing

Intro to Virtualization

Cloud Based Application Architectures using Smart Computing

Virtual Machines.

Multiprocessor Scheduling and Scheduling in Linux Kernel 2.6

Oracle Database Scalability in VMware ESX VMware ESX 3.5

Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1

Performance and scalability of a large OLTP workload

Agenda. Capacity Planning practical view CPU Capacity Planning LPAR2RRD LPAR2RRD. Discussion. Premium features Future

OPTIMIZING SERVER VIRTUALIZATION

Hard Partitioning and Virtualization with Oracle Virtual Machine. An approach toward cost saving with Oracle Database licenses

Windows Server 2008 R2 Hyper-V Live Migration

Operating Systems OBJECTIVES 7.1 DEFINITION. Chapter 7. Note:

Understanding the Performance of an X User Environment

Why Relative Share Does Not Work

Virtualization. Dr. Yingwu Zhu

Systemverwaltung 2009 AIX / LPAR

Developing a dynamic, real-time IT infrastructure with Red Hat integrated virtualization

Mark Bennett. Search and the Virtual Machine

TPC-W * : Benchmarking An Ecommerce Solution By Wayne D. Smith, Intel Corporation Revision 1.2

Monitoring Databases on VMware

BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB

nanohub.org An Overview of Virtualization Techniques

Kernel Optimizations for KVM. Rik van Riel Senior Software Engineer, Red Hat June

Server Virtualization with VMWare

Solution Guide Parallels Virtualization for Linux

Integrated Application and Data Protection. NEC ExpressCluster White Paper

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

Best Practices for Monitoring Databases on VMware. Dean Richards Senior DBA, Confio Software

All-Flash Arrays Weren t Built for Dynamic Environments. Here s Why... This whitepaper is based on content originally posted at

Microsoft Exchange Server 2007 and Hyper-V high availability configuration on HP ProLiant BL680c G5 server blades

White Paper Perceived Performance Tuning a system for what really matters

Virtuoso and Database Scalability

Hyper-V R2: What's New?

IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud

Why Computers Are Getting Slower (and what we can do about it) Rik van Riel Sr. Software Engineer, Red Hat

Virtualization 101 for Power Systems

Eloquence Training What s new in Eloquence B.08.00

PERFORMANCE ANALYSIS OF KERNEL-BASED VIRTUAL MACHINE

IBM Software Group. Lotus Domino 6.5 Server Enablement

System Models for Distributed and Cloud Computing

Zadara Storage Cloud A

Intel DPDK Boosts Server Appliance Performance White Paper

Final Report. Cluster Scheduling. Submitted by: Priti Lohani

Database Instance Caging: A Simple Approach to Server Consolidation. An Oracle White Paper September 2009

Operating Systems, 6 th ed. Test Bank Chapter 7

COS 318: Operating Systems. Virtual Machine Monitors

Basics in Energy Information (& Communication) Systems Virtualization / Virtual Machines

Performance with the Oracle Database Cloud

Virtualization and Cloud Management Using Capacity Planning

Muse Server Sizing. 18 June Document Version Muse

Virtualization of Oracle Evolves to Best Practice for Production Systems

COLO: COarse-grain LOck-stepping Virtual Machine for Non-stop Service

Cloud Management: Knowing is Half The Battle

Linux VM Infrastructure for memory power management

How To Virtualize A Storage Area Network (San) With Virtualization

Crystal Reports Server 2008

Avoiding Performance Bottlenecks in Hyper-V

Chapter 1 - Web Server Management and Cluster Topology

Virtual Machine Monitors. Dr. Marc E. Fiuczynski Research Scholar Princeton University

AlphaTrust PRONTO - Hardware Requirements

Technology Insight Series

Distributed and Cloud Computing

Performance and Scale in Cloud Computing

ELEC 377. Operating Systems. Week 1 Class 3

VIRTUALIZATION, The next step for online services

An Oracle White Paper August Oracle VM 3: Server Pool Deployment Planning Considerations for Scalability and Availability

Architecting for the next generation of Big Data Hortonworks HDP 2.0 on Red Hat Enterprise Linux 6 with OpenJDK 7

Full and Para Virtualization

Virtualization NIIT WHITE PAPER SEPTEMBER 2011

HRG Assessment: Stratus everrun Enterprise

Balancing CPU, Storage

Virtualization management tools

Evaluation Report: Supporting Microsoft Exchange on the Lenovo S3200 Hybrid Array

Transcription:

Capacity Estimation for Linux Workloads Session L985 David Boyes Sine Nomine Associates 1

Agenda General Capacity Planning Issues Virtual Machine History and Value Unique Capacity Issues in Virtual Machines Empirical Impact and Observations Measurements, Benchmarks, and Methods, Oh My! Case Study of a Real Deployment References 2

Caveats Capacity planning is at best an educated guess about what might happen if something happens, plus an error term Statistics are voodoo Work in progress Your results will vary Etc, etc, etc 3

General Capacity Planning Issues Relative vs Absolute Measuring Workload Workload Composition Utilization White Space 4

Relative Vs Absolute Balance of: Processor power (effective work per instruction) Internal bandwidth (data rate between cache and RAM) Scheduler efficiency (processor utilization typical to this system type) Bandwidth Processors 1 0.5 0 Scheduler A B C Perfect World: If equal stress in test, relative capacity is geometric mean of measurements 5

Measuring Workload If workload shifts away from balance, relative capacity is always affected. Server workload tends to be skew-rich CPU intense applications tend to favor more/faster CPUs (skewless) Mixed workloads/non uniform usage patters favor higher internal bandwidth and better scheduling (skew/spiky) 6

Workload Summary Relative capacity of machines differs by workload: There is no single solution Right Tool, Right Job 7

Utilization Typical server farm is deliberately underutilized most of the time (peak load allowance, or white space ) White space is 1 minus the utilization. Sources of White Space Spikes Headroom Redundancy Fragmentation Partitioning Skew 8

Utilization Be careful when you talk about utilization (peak vs average). Utilization figures are a statistical sample, thus subject to manipulation Short time samples tend to poorly represent the utilization of a machine Fine granularity impacts figures (Heisenburg) Corollary: utilization often depends on workload combined workload smoothing often can optimize utilization over time 9

Virtual Machines: History and Value 10 Not a new idea OS/8 for the PDP-8 IBM VM family (1964) Commonplace in large systems deployments for more than 3 full decades Simple concept: allow single physical machine to emulate a dedicated system for each individual user Each emulated system receives a portion of the physical resources of the underlying hardware The emulated system does NOT have to match the physical configuration of the system Emulated systems are independent of each other.

Graphical View of Virtual Machine Environment Virtual Machine Virtual Machine Virtual Machine Virtual Machine White Space Hypervisor CPU1 CPU2 CPU3 CPU4 System Physical Resources Sy s tem RAM CPUs/RAM/Disk/Tape/etc 11

Unique Capacity Issues I/O Overlapping CPU and FPU Overlapping Memory Overlap Virtual Device Emulation Overhead 12

I/O Overlap Virtual machines allow overlapping usage of I/O devices based on individual virtual machines being statistically unlikely to be performing simultaneous I/O. This overlap can affect performance positively in most cases in most applications, the lions share of cycles are wasted waiting on external I/O to complete. 13

CPU/FPU Overlap CPU overlap occurs when a virtual machine suspends to complete a blocking operation such as I/O and cannot use it s full timeslice, releasing it to be available for other virtual machines to use as needed. Unused space allows additional overlap cycles to be utilized according to demand 14

Memory Overlap This is not particularly different from any other virtual memory value proposition, however, the scale is somewhat larger entire systems instead of processes within a single system. Storage overload/overlap factors are generally much higher (60-80%) in virtual machines 15

Device Emulation Overlap Permitting devices to be sub-allocated (simulating full disks) allows significant cost savings Consider normal workstation deployment: Every machine has: Operating system files all the same in most environments, so why not share? 16

Empirical Impact for System Sizing Standard sizing models for Unix and traditional large systems grossly overestimate the requirement for CPU and memory. Most server farm machines operate at less than 10% nominal utilization Most server farms have less than 20% of systems active at any specific interval There is no valid direct comparison for disparate architectures the numbers don t necessarily mean the same for different systems Meaningless/Misleading Indicator of Processor Speed (MIPS) Overall instruction throughput for effective work differs on RISC/CISC 17

Empirical Observations Smaller is Better For most virtual machine hypervisors, the contents of the virtual machine is a opaque black box. Optimizing for minimum size impermeable boxes allows better control of hypervisor resources Smaller scheduling units Less storage/disk paging fragmentation Less paging/swapping latency Maximum benefit for block paging algorithms Best opportunity for maximum I/O and CPU overlap 18

Empirical Observations Scaling via Horizontal Additions Corollary of Smaller is Better Applications best suited to virtual machine environment scale by addition of multiple virtual instances Statistical sampling of workload spreads impact of load across multiple virtual CPU engines Minimizes deadlock opportunities Increases requirement for synchronization Does awful things to CPU cache consistency in hypervisor Augments demand for sophisticated system/configuration management tools 19

Empirical Observations Hypervisor impact biases application selection toward I/O intensive applications Self-overlapping effect of I/O operations as unit transactions at the hypervisor level Allows dispatch of next VM at I/O request block boundary and more effective fair share algorithms Rapid context switch capability imperative for efficient hypervisor management Intel and Sparc not strong in this area without careful thought Strong benefit for S/390 architecture 20

Empirical Observations Virtual machines demand LESS hardware resources for most server-class tasks due to overlap. Many applications spend up to 60% of CPU time in I/O wait/spin locks Context switch/dispatch/cpu pipeline effect can demonstrate 35-45% overlap for CPU Corollary: You don t need 900 Mhz systems to deliver equivalent performance apples/pears comparison Corresponding impact on effective transaction throughput more transactions with 35-45% smaller systems Reduced hypervisor overhead 21

What About Benchmarks? Benchmarks are by definition arbitrary, and are usually misleading across disparate hardware. 22

What Can We Rely On? Workload Analysis Overlap Analysis Error Estimation Terms In capacity planning, the answer is always it depends. With virtual systems and Linux, it depends even more. -- Bill Bitner, IBM 23

Analysis Framework Instrumentation Application Classification Data Model Overlap Analysis Error Estimation 24

Instrumentation Internal Instrumentation vmstat/iostat over multiple runs sar Best for detail of individual task implementation Hypervisor Instrumentation Surrounds virtual systems Gives overall summary of usage for system 25

Application Structure Application Classes I/O Intensive CPU/FP Intensive Memory Intensive Network Intensive Composite Distribution To some degree, all applications are composite, but a consistent judgement call generally gives better results 26

Overlap Analysis Transaction Borders Preemption Timeslice Multitasking/Multiprogramming Level Threading 27

Error Estimation Terms While we can t cover this in detail here, the Merrill book in the reference section does a very detailed job of providing error estimation processes. Major areas: Measurement overhead Workload estimation errors over time Imprecise transaction definitions 28

A More Comprehensive Example Global Insurance Company Windows File and Print Application Replacing 180 800 Mhz 1-way NT4 servers Driving 850 network printers Consolidate to IBM mainframe and VM 29

Analysis Framework File/print primarily I/O intensive, little CPU involved Much of existing capacity dedicated to redundancy (~40% for failure & clustering requirements) Strong transactional model (file update/print job) 30

Overlap Analysis CPU utilization rarely over 10% on any individual server Overlay of all server processes under 10% nominal, peak 62% during cluster failover; normal idle percentage of 60-70% of theoretical capacity Network and disk I/O rarely idle 31

Error Estimation Terms CPU measurement introduces +/- 2.5-3% error Disk usage sampling imposes +/- 5-7% error Inefficiency of MS performance tools: priceless 32

Sizing Possibilities Simplistic 1800 MIPS 12 (11.25) way mainframe Next generation only growth path Some operational savings, but not compelling Sized 900 MIPS 6 way mainframe 1+ million saved Room to grow with single asset Capacity on demand 33

References Linux for S/390: ISP/ASP Solutions RedBook, IBM Generalized Capacity Planning Models for Large Systems, David Merrill 34

Contact Info David Boyes Sine Nomine Associates +1 703 723 6673 dboyes@sinenomine.net dboyes@de.sinenomine.net 35