Small is Better: Avoiding Latency Traps in Virtualized DataCenters

Similar documents
Small is Better: Avoiding Latency Traps in Virtualized Data Centers

Prioritizing Soft Real-Time Network Traffic in Virtualized Hosts Based on Xen

Assessing the Performance of Virtualization Technologies for NFV: a Preliminary Benchmarking

How To Stop A Malicious Process From Running On A Hypervisor

Virtual Switching Without a Hypervisor for a More Secure Cloud

Optimizing Converged Cisco Networks (ONT)

Intel Ethernet Switch Load Balancing System Design Using Advanced Features in Intel Ethernet Switch Family

Real-time Virtual NIC on KVM for Real-Time Network with OpenFlow

Virtualization and Cloud Computing. The Threat of Covert Channels. Related Work. Zhenyu Wu, Zhang Xu, and Haining Wang 1

CIVSched: A Communication-aware Inter-VM Scheduling Technique for Decreased Network Latency between Co-located VMs

Networking for Caribbean Development

IOS110. Virtualization 5/27/2014 1

Performance Evaluation of Linux Bridge

Friends, not Foes Synthesizing Existing Transport Strategies for Data Center Networks

Workload-Aware Provisioning in Public Clouds

Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study

Network Functions Virtualization on top of Xen

Network performance in virtual infrastructures

Impact of Virtualization on Network Performance The TCP Case

Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build

Load Balancing in Data Center Networks

Cloud Computing CS

Storage I/O Control: Proportional Allocation of Shared Storage Resources

Masters Project Proposal

AS cloud computing becomes widespread, there are

Nutanix Tech Note. VMware vsphere Networking on Nutanix

InfiniBand Software and Protocols Enable Seamless Off-the-shelf Applications Deployment

Achieving a High-Performance Virtual Network Infrastructure with PLUMgrid IO Visor & Mellanox ConnectX -3 Pro

Enabling Technologies for Distributed and Cloud Computing

On the Performance Isolation Across Virtual Network Adapters in Xen

Network Performance Comparison of Multiple Virtual Machines

Global Headquarters: 5 Speen Street Framingham, MA USA P F

Full and Para Virtualization

Feature Comparison. Windows Server 2008 R2 Hyper-V and Windows Server 2012 Hyper-V

Best Practices for Monitoring Databases on VMware. Dean Richards Senior DBA, Confio Software

Nested Virtualization

QoS & Traffic Management

VON/K: A Fast Virtual Overlay Network Embedded in KVM Hypervisor for High Performance Computing

This topic lists the key mechanisms use to implement QoS in an IP network.

CON9577 Performance Optimizations for Cloud Infrastructure as a Service

Optimizing Network Virtualization in Xen

Rackspace Cloud Databases and Container-based Virtualization

Optimizing Network Virtualization in Xen

Chapter 5 Cloud Resource Virtualization

Leveraging NIC Technology to Improve Network Performance in VMware vsphere

High Performance OpenStack Cloud. Eli Karpilovski Cloud Advisory Council Chairman

Comparison of the Three CPU Schedulers in Xen

Decentralized Task-Aware Scheduling for Data Center Networks

Multi-core Programming System Overview

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

CS312 Solutions #6. March 13, 2015

How Router Technology Shapes Inter-Cloud Computing Service Architecture for The Future Internet

VMWARE WHITE PAPER 1

Virtual Computing and VMWare. Module 4

Real- Time Mul,- Core Virtual Machine Scheduling in Xen

Enabling Technologies for Distributed Computing

Virtualization for Cloud Computing

Big Data in the Background: Maximizing Productivity while Minimizing Virtual Machine Interference

Deploying Extremely Latency-Sensitive Applications in VMware vsphere 5.5

Simplify VMware vsphere* 4 Networking with Intel Ethernet 10 Gigabit Server Adapters

Cloud security CS642: Computer Security Professor Ristenpart h9p:// rist at cs dot wisc dot edu University of Wisconsin CS 642

Quality of Service (QoS) for Enterprise Networks. Learn How to Configure QoS on Cisco Routers. Share:

Chapter 14 Virtual Machines

Building Docker Cloud Services with Virtuozzo

Application Performance in the Cloud, and its Relationship to QoS

基 於 SDN 與 可 程 式 化 硬 體 架 構 之 雲 端 網 路 系 統 交 換 器

Quality of Service. Traditional Nonconverged Network. Traditional data traffic characteristics:

IxChariot Virtualization Performance Test Plan

New Data Center architecture

Quality of Service (QoS)) in IP networks

Fibre Channel Over and Under

Windows Server 2008 R2 Hyper-V Live Migration

Windows Server 2012 R2 Hyper-V: Designing for the Real World

Virtualizing a Virtual Machine

Virtualization: TCP/IP Performance Management in a Virtualized Environment Orlando Share Session 9308

QoS Queuing on Cisco Nexus 1000V Class-Based Weighted Fair Queuing for Virtualized Data Centers and Cloud Environments

TRILL Large Layer 2 Network Solution

Enhancing Hypervisor and Cloud Solutions Using Embedded Linux Iisko Lappalainen MontaVista

Configuring QoS. Understanding QoS CHAPTER

Windows Server Performance Monitoring

COLO: COarse-grain LOck-stepping Virtual Machine for Non-stop Service

Open vswitch and the Intelligent Edge

UCS Network Utilization Monitoring: Configuration and Best Practice

Hyper-V R2: What's New?

Virtualization: Concepts, Applications, and Performance Modeling

EVALUATING NETWORK BUFFER SIZE REQUIREMENTS

Software-Defined Networking Architecture Framework for Multi-Tenant Enterprise Cloud Environments

Presentation of Diagnosing performance overheads in the Xen virtual machine environment

WAN Performance Analysis A Study on the Impact of Windows 7

Optimizing TCP Forwarding

Optimizing Data Center Networks for Cloud Computing

Performance Management in a Virtual Environment. Eric Siebert Author and vexpert. whitepaper

Thomas Fahrig Senior Developer Hypervisor Team. Hypervisor Architecture Terminology Goals Basics Details

SERVICE SCHEDULE PULSANT ENTERPRISE CLOUD SERVICES

HRG Assessment: Stratus everrun Enterprise

Securing your Virtual Datacenter. Part 1: Preventing, Mitigating Privilege Escalation

Performance evaluation of Linux Bridge and OVS in Xen

MANAGING NETWORK COMPONENTS USING SNMP

Cloud Computing and the Internet. Conferenza GARR 2010

Migration and Building of Data Centers in IBM SoftLayer with the RackWare Management Module

Transcription:

Small is Better: Avoiding Latency Traps in Virtualized DataCenters SOCC 2013 Yunjing Xu, Michael Bailey, Brian Noble, Farnam Jahanian University of Michigan 1

Outline Introduction Related Work Source of latency Design and Implementation Evaluation Conclusion 2

Introduction Public clouds have become a popular platform for building Internet-scale applications Applications built with public clouds are often highly sensitive to response time Large data centers have become the cornerstone of modern, Internet-scale Web applications Unlike dedicated data centers,a public cloud relies on virtualization to both hide the details of the underlying host infrastructure as well as support multi-tenancy 3

Related Work Kernel-centric modify the operating system (OS) kernel to deploy new TCP congestion control algorithms. DCTCP, HULL Application-centric applications must be modified to tag the packets they generate with scheduling hints. D 3, D 2 TCP, DeTail, PDQ, and pfabric Operator-centric require operators to change their application deployment. Bobtail 4

Host-centric It does not require or trust guest cooperation, and it only modifies the host infrastructure controlled by cloud providers. The existing host-centric method: EyeQ it mainly focuses on bandwidth sharing in the cloud require feedback loops between hypervisors need explicit bandwidth headroom to reduce latency only solves one of the three latency problems addressed by our solution. 5

Sources of latency VM scheduling delay the server VM cannot process the packets until scheduled by the hypervisor Host network queueing delay the response packets first go through the host network stack, which processes I/O requests on behalf of all guest VMs Switch queueing delay response packets on the wire may experience switch queueing delay on congested links 6

7 Sources of latency

EC2 and the Xen hypervisor These studies find that virtualization and multi-tenancy are keys to EC2 s performance variation Xen runs on bare metal hardware to manage guest VMs A privileged guest VM called dom0 is used to fulfill I/O requests for non-privileged guest VMs 8

9 EC2 measurements --VM Scheduling delay and Swich queueing delay

10 EC2 measurements --VM Scheduling delay and Swich queueing delay

Testbed experiments --Host network queueing delay Kernel network stack NIC transmission queue Byte Queue Limits (BQL) A new device driver interface CoDel If queued packets have already spent too much time in the queue, the upper layer of the network stack is notified to slow down 11

Testbed experiments --Host network queueing delay physical machine B1 C1 physical machine Cause congestion switch ping A1 A2 12 physical machine

Testbed experiments --Host network queueing delay Congestion Free: A1 and B1 are left idle. Congestion Enabled: A1 sends bulk traffic to B1 without BQL or CoDel. Congestion Managed: A1 sends bulk traffic to B1 with BQL and CoDel enabled. 13

Testbed experiments --Host network queueing delay While BQL and CoDel can significantly reduce the latency, the result is still four to six times as large when compared to the baseline 14

Summary Switch queueing delay increases network tail latency by over 10 times; together with VM scheduling delay, it becomes more than 20 times as bad Host network queueing delay also worsens the FCT tail by four to six times. 15

Design and Implementation Principle I: Not trusting guest VMs Principle II: Shortest remaining time first Principle III: No undue harm to throughput 16

VM scheduling delay Credit Scheduler is currently Xen s default VM scheduler Each guest VM has at least one VCPU Credits are redistributed in 30ms intervals VCPU has three states OVER UNDER BOOST 17

VM scheduling delay The BOOST mechanism only prioritizes VMs over others in UNDER or OVER states BOOSTed VMs can not preempt each other, and they are round-robin scheduled We deploy a more aggressive VM scheduling policy to allow BOOSTed VMs to preempt each other Xen has a rate limit mechanism that maintains overall system throughput by preventing preemption when the running VM has run for less than 1ms in its default setting. 18

Host network queueing delay From Table2, when BQL and CoDel are both enabled, the queue delay is four to six times The root cause is requests often too large and hard to be preempted. Our solution is to break large jobs into smaller ones to allow CoDel to conduct fine-grained packet scheduling. 19

Host network queueing delay guest VM dom0 frontend backend NIC A packet sent out by guest VMs first goes to the frontend, which copies the packet to the backend Xen s backend announce to the guest VMs that hardware segmentation offload is not supported, then guests have to segment the packets before copying them 20

Switch queueing delay Letting switches favor small flows when scheduling packets Define a flow as the collection of any IP packets Define flow size as the instant size of a message the flow In reality, it is also difficult to define message boundaries. Thus, we measure flow by rate as an approximation of the message semantic We classify flows into two classes, small and large 21

Switch queueing delay First, we build a monitoring and tagging module in the host that sets priority on outgoing packets. Small flows are tagged as high-priority Second, we need switches that support basic priority queueing 22

Putting it all together We change a single line in the credit scheduler of Xen 4.2.1 to enable the new scheduling policy We also modify the CoDel kernel module in Linux 3.6.6 with about 20 lines to segment large packets in the host We augment the Xen s network backend with about 200 lines of changes to do flow monitoring and tagging 23

Evaluation The testbed consists of five four-core physical machines running Linux 3.6.6 and Xen 4.2.1. They are connected to a Cisco Catalyst 2970 switch 24

Evaluation 25 A1 serves small responses to E1 A2 sends bulk traffic to B1 (Host network queueing delay) A3 A4 run CPU-bound tasks (VM scheduler delay) C1 D1 respond to E2 s queries for large flows and congest E s access link (Switch queueing delay)

Evaluation Our solution achieves about 40% reduction in mean latency Over 56% for the 99th percentile, and almost 90% for the 99.9th percentile. 26

VM scheduling delay A1 A3 switch E1 Use E1 to query A1 for small responses Keep A3 running a CPU-bound workload 27

VM scheduling delay Our new policy reduces latency at the 99.9th percentile by 95% at the cost of 3.8% reduction in CPU throughput 28

Host network queueing delay B1 A1 A2 switch E1 A1 to ping E1 once every 10ms for round-trip time A2 to saturate B1 s access link with iperf 29

Host network queueing delay Our solution can yield an additional 50% improvement at both the body and tail of the distribution 30

Switch queueing delay B1 C1 D1 A1 switch E1 E2 E1 queries A1 and B1 in parallel for small flows E2 queries C1 and D1 for large flows to congest the access link 31

Switch queueing delay When QoS support on the switch is enabled to recognize our tags, all small flows enjoy a low latency with an order of magnitude improvement at both the 99th and 99.9 th percentiles On the other hand, the average throughput loss for large flows is less than 3% 32

Conclusion We design a host-centric solution that extends the classic shortest remaining time first scheduling policy from the virtualization layer, through the host network stack, to the network switches 33