Data Center Networking with Multipath TCP



Similar documents
Computer Networks COSC 6377

Multipath TCP in Data Centres (work in progress)

Multipath TCP design, and application to data centers. Damon Wischik, Mark Handley, Costin Raiciu, Christopher Pluntke

Data Center Networking with Multipath TCP

Improving datacenter performance and robustness with multipath TCP

Data Center Netwokring with Multipath TCP

Improving Datacenter Performance and Robustness with Multipath TCP

MMPTCP: A Novel Transport Protocol for Data Centre Networks

Multipath TCP in Practice (Work in Progress) Mark Handley Damon Wischik Costin Raiciu Alan Ford

An Overview of Multipath TCP

On the Impact of Packet Spraying in Data Center Networks

MAPS: Adaptive Path Selection for Multipath Transport Protocols in the Internet

Mul$path Networking OpenFlow and MPTCP Friend or Foe?

Data Center Network Topologies: FatTree

Load Balancing Mechanisms in Data Center Networks

Advanced Computer Networks. Datacenter Network Fabric

Data Center Load Balancing Kristian Hartikainen

Data Center Network Topologies: VL2 (Virtual Layer 2)

Load Balancing in Data Center Networks

OpenFlow based Load Balancing for Fat-Tree Networks with Multipath Support

Operating Systems. Cloud Computing and Data Centers

Congestion Control of Multipath TCP: Problems and Solutions

BigPi: Sharing Link Pools in Cloud Networks

Boosting mobility performance with Multi-Path TCP

TinyFlow: Breaking Elephants Down Into Mice in Data Center Networks

On implementation of DCTCP on three tier and fat tree data center network topologies

Per-packet Load-balanced, Low-Latency Routing for Clos-based Data Center Networks

DARD: Distributed Adaptive Routing for Datacenter Networks

Exploring Mobile/WiFi Handover with Multipath TCP

DiFS: Distributed Flow Scheduling for Adaptive Routing in Hierarchical Data Center Networks

Paolo Costa

Xiaoqiao Meng, Vasileios Pappas, Li Zhang IBM T.J. Watson Research Center Presented by: Payman Khani

International Journal of Emerging Technology in Computer Science & Electronics (IJETCSE) ISSN: Volume 8 Issue 1 APRIL 2014.

Lecture 7: Data Center Networks"

Data Center Switch Fabric Competitive Analysis

Avoiding Network Polarization and Increasing Visibility in Cloud Networks Using Broadcom Smart- Hash Technology

A Reliability Analysis of Datacenter Topologies

Scaling 10Gb/s Clustering at Wire-Speed

[Yawen comments]: The first author Ruoyan Liu is a visting student coming from our collaborator, A/Prof. Huaxi Gu s research group, Xidian

Cisco s Massively Scalable Data Center

TRILL Large Layer 2 Network Solution

T. S. Eugene Ng Rice University

High-Speed TCP Performance Characterization under Various Operating Systems

TCP loss sensitivity analysis ADAM KRAJEWSKI, IT-CS-CE

How To Analyse Cloud Data Centre Performance

Using Random Neural Network for Load Balancing in Data Centers

Outline. VL2: A Scalable and Flexible Data Center Network. Problem. Introduction 11/26/2012

SDN and Data Center Networks

Experiences with MPTCP in an intercontinental OpenFlow network

Powerful Duo: MapR Big Data Analytics with Cisco ACI Network Switches

Low-rate TCP-targeted Denial of Service Attack Defense

Optimization models for congestion control with multipath routing in TCP/IP networks

Data Center Networks

Protagonist International Journal of Management And Technology (PIJMT) Online ISSN Vol 2 No 3 (May-2015) Active Queue Management

Optimization of Communication Systems Lecture 6: Internet TCP Congestion Control

Local Area Networks transmission system private speedy and secure kilometres shared transmission medium hardware & software

Traffic Engineering for Multiple Spanning Tree Protocol in Large Data Centers

CROSS LAYER BASED MULTIPATH ROUTING FOR LOAD BALANCING

The Software Defined Hybrid Packet Optical Datacenter Network SDN AT LIGHT SPEED TM CALIENT Technologies

Longer is Better? Exploiting Path Diversity in Data Centre Networks

Question: 3 When using Application Intelligence, Server Time may be defined as.

ICTCP: Incast Congestion Control for TCP in Data Center Networks

How the Port Density of a Data Center LAN Switch Impacts Scalability and Total Cost of Ownership

Data Center Network Topologies

Advanced Computer Networks. Scheduling

Non-blocking Switching in the Cloud Computing Era

Energy Optimizations for Data Center Network: Formulation and its Solution

A Cooperative Game Based Allocation for Sharing Data Center Networks

TRILL for Service Provider Data Center and IXP. Francois Tallet, Cisco Systems

Intel Ethernet Switch Converged Enhanced Ethernet (CEE) and Datacenter Bridging (DCB) Using Intel Ethernet Switch Family Switches

A Talari Networks White Paper. Turbo Charging WAN Optimization with WAN Virtualization. A Talari White Paper

On Uplink Measurement of Cellular Networks for Mobile Multi-Path TCP

Adaptive Routing for Layer-2 Load Balancing in Data Center Networks

Dahu: Commodity Switches for Direct Connect Data Center Networks

TCP and UDP Performance for Internet over Optical Packet-Switched Networks

MultiWAN: WAN Aggregation for Developing Regions

Binder: A System to Aggregate Multiple Internet Gateways in Community Networks

On the effect of forwarding table size on SDN network utilization

DARD: Distributed Adaptive Routing for Datacenter Networks

Experiences with MPTCP in an intercontinental multipathed OpenFlow network

Lecture 15: Congestion Control. CSE 123: Computer Networks Stefan Savage

Storage I/O Control: Proportional Allocation of Shared Storage Resources

Windows Server Performance Monitoring

PortLand:! A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric

CONGA: Distributed Congestion-Aware Load Balancing for Datacenters

Transcription:

Data Center Networking with Multipath TCP Costin Raiciu, Christopher Pluntke, Sebastien Barre, Adam Greenhalgh, Damon Wischik, Mark Handley Hotnets 2010 報 告 者 : 莊 延 安

Outline Introduction Analysis Conclusion

Introduction How can we ensure that no matter the traffic pattern, the load is distributed between the many possible parallel paths as evenly as possible?

Analysis Single path TCP with RLB The current wisdom seems to be to use randomised load balancing (RLB) to randomly choose a path for each flow from among the possible parallel paths. MPTCP To utilize multiple parallel paths for each TCP connection

Analysis Fat tree VL2 BCube

Analysis Fig. shows the total throughput of all flows when we use a random permutation matrix where each host sends flat out (as determined by the TCP response function) to a single other host. The benefits of MPTCP are clear; as additional subflows are added, the overall throughput increases.

Analysis Fig. shows the minimum number of subflows that can achieve 90% utilization for each size of network. The result is encouraging: beyond a certain size, the number of subflows needed does not increase significantly with network size

Analysis Fig. shows the throughput of the lowest speed flow (as a percentage of what should be achievable) and Jain s fairness index for the three topologies. Multipath always improves fairness, even for the VL2 topology which performed relatively well if we only examine throughput.

Analysis The results in Fig.(a) show that MPTCP does not increase network load, as measured by either mean or max loss rate. In contrast, independent congestion control for each subflow causes the network to become increasingly congested as we add more subflows. Fig. (b) shows the number of retransmit timeouts in this experiment. Unsurprisingly, independent TCP suffers many timeouts, which may impact application performance.

Conclusion MPTCP is a simple solution that can effectively utilize the dense parallel network topologies proposed for modern data centers, significantly improving throughput and fairness compared to singlepath TCP.

Design, Implementation and Evaluation of Congestion Control for Multipath TCP Damon Wischik, Costin Raiciu, Adam Greenhalgh, Mark Handley University College London Proc. Usenix NSDI 2011 報 告 者 : 黃 聖 荃

Outline Introduction Related Work Problems Solution Experimental Results Conclusions

Introduction Multipath TCP allows a single data stream to be split across multiple paths. This has obvious benefits for reliability, the connection can persist when a path fails.

Introduction (cont.) Multipath TCP also raises questions, some obvious and some subtle, about how network capacity should be shared efficiently and fairly between competing flows.

Related Work

Regular Single TCP Congestion Control Each ACK, increase the congestion window w by 1/w, resulting in an increase of one packet per RTT Each loss, decrease w by w/2.

Equally-Weighted TCP Main idea : Split traffic evenly The multipath flow gets the same throughput as a regular TCP at the bottleneck link.

Equally-Weighted TCP(CONT.) Each subflow gets window size proportional to a^2, and a = 1/ n, where n is the number of paths. For each ACK on path r, increase window wr by a/ wr For each loss on path r, decrease window wr by wr /2

Coupled Main idea : A multipath flow should shift all its traffic onto the least-congested path.

Coupled(cont.) Wtotal is the total window size across all subflows. For each ACK on path r, increase window wr by 1/wtotal. For each loss on path r, decrease window wr by wtotal /2

Coupled(cont.) Consider that all paths have the same loss rate p Wtotal = (where the approximation is good if p is small)

Problems Adapting to load changes

Problems(cont.) RTT mismatch

Solution Main idea : 1. A multipath flow should give a connection at least as much throughput as it would get with single-path TCP on the best of its paths. 2. A multipath flow should take no more capacity on any path or collection of paths than if it was a single path TCP flow using the best of those paths.

Solution(cont.) Let be the equilibrium window obtained by multipath TCP on path r Let be the equilibrium window that would be obtained by a single-path TCP experiencing path r s loss rate.

Solution(cont.)

Solution(cont.) Final algorithm : MPTCP Each ACK on subflow r, increase the window wr by min(ɑ/wtotal, 1/wr). Each loss on subflow r, decrease the window wr by wr/2.

Experimental Results TP1 is a random permutation where each host opens a flow to a single destination chosen uniformly at random, such that each host has a single incoming flow. TP2 each host opens 12 flows to 12 destinations TP3 is a sparse traffic pattern: 30% of the hosts open one flow to a single destination chosen uniformly at random.

Experimental Results(cont.)

Experimental Results(cont.)

Conclusions MPTCP selects efficient paths and balances congestion. MPTCP performs well across wide range traffic patterns.

Improving Datacenter Performance and Robustness with Multipath TCP Costin Raiciu, Sebastien Barre, Christopher Pluntke, Adam Greenhalgh, Damon Wischik, Mark Handley SIGCOMM 2011 報 告 者 : 陳 孟 緯 1

Benefit of MPTCP in Data Centers Better aggregate throughput: more paths and load-balancing, reduce number of underutilized and idle links Better robustness : once a link or switch fail, with MPTCP, others can continue without pausing 2

Examples of benefits Throughput Traffic pattern: permutation matrix & every host sends to one other host chosen at random Left&right histogram: increase network size slightly reduce overall performance Grey& black histogram: packet-level & flow level are in agreement about performance benefit& number of subflows 3

Examples of benefits(cont) Throughput Reason :single-path performs poorly Fairness: Most MPTCP flows get at least 90% of the available capacity 35

Analysis What impact to performance of TCP vs MPTCP Influence of topology Number of subflows Oversubscribed topologies 36

Influence of Topology Performance: BCube is similar to FatTree VL2- gain smaller- most gain Cause Bcube same problem with collision with FatTree does VL2: capacity of links is huge 37

Number of subflows On FatTree & permutation traffic matrix 8 subflows 90%throughput indicate how many of subflows are required the variance settles down above eight subflows 38

Oversubscribed Topologies Low-load: most flows never share a link with another flow saturate between TCP & MPTCP High load: core congested & MPTCP provides significant improvements in throughput 39

What is an optimal datacenter topology for multipath transport? 40

Evolving topologies with MPTCP Consider 2 topologies: Perfect Switch : FatTree and VL2 both try to emulate a single huge non-blocking switch, giving an upper bound on what any network core might provide using single links to the hosts. Dual-homed FatTree (DHFT): traditional,fattree with 5 switch port per host; one is to the host and 4 for connect 4 layer of switches. DHFT, remove one port per host from the core and connect to second interface on switch ports 41

Fat Tree Topology Upper Pod Switch ToR Switch Servers 42

Dual Homed Fat Tree Topology Upper Pod Switch ToR Switch Servers 43

Evolving topologies with MPTCP (const) Analysis: Effects of Locality Dark grey region shows throughputs that are feasible as locality changes Light grey region show the possible throughput if one of the two interfaces on each host is used Effects of Load Light-to-medium load: twice performance High load: bottleneck 44

Evolving topologies with MPTCP (const) Discussion DHFT cost the same as a Fat Tree DHFT improves robustness DHFT is not optimal by any measure but it shows that we can create topologies with better performance if we assume MPTCP is the transport protocol With MPTCP as transport, a wider range of topologies are cost-effective create a topology with 2Gb/s full bisection bandwidth, we could use a k-port Fat Tree with k3/8 dual-homed hosts 45

Conclusion One flow, one path thinking has constrained datacenter design Collisions, limited utilization MPTCP could improve data center performance by performing very short timescale distributed load balancing Provide benefits for bottlenecked problem 46

Question Consider with EWTCP and MPTCP, which of the following answers are true? (A) EWTCP splits its traffic evenly. (B) MPTCP splits its traffic evenly. (C) EWTCP tends to shift its traffic onto the leastcongested path. (D) MPTCP tends to shift its traffic onto the leastcongested path.

Answer (A) EWTCP splits its traffic evenly. (B) MPTCP splits its traffic evenly. (C) EWTCP tends to shift its traffic onto the leastcongested path. (D) MPTCP tends to shift its traffic onto the leastcongested path.