Experiences with MPTCP in an intercontinental multipathed OpenFlow network



Similar documents
Experiences with MPTCP in an intercontinental OpenFlow network

Multipathing with MPTCP and OpenFlow

Mul$path Networking OpenFlow and MPTCP Friend or Foe?

Lightpath Planning and Monitoring

D1.2 Network Load Balancing

Network and Capacity Planning in SURFnet6

TRILL Large Layer 2 Network Solution

How To Build A Lightpath Network For Multiple Lightpath Projects

Ethernet OAM enabled OpenFlow Controller

TRILL for Service Provider Data Center and IXP. Francois Tallet, Cisco Systems

An Overview of Multipath TCP

Leveraging Advanced Load Sharing for Scaling Capacity to 100 Gbps and Beyond

Network Virtualization for Large-Scale Data Centers

IP Traffic Engineering over OMP technique

OpenFlow & Software Defined Networking

Computer Networks. Definition of LAN. Connection of Network. Key Points of LAN. Lecture 06 Connecting Networks

Performance of Software Switching

Evaluation of 40 Gigabit Ethernet technology for data servers

CHAPTER 6. VOICE COMMUNICATION OVER HYBRID MANETs

Boosting mobility performance with Multi-Path TCP

Computer Networking Networks

CROSS LAYER BASED MULTIPATH ROUTING FOR LOAD BALANCING

Virtual Private LAN Service

Data Center Convergence. Ahmad Zamer, Brocade

LAN Switching and VLANs

What is VLAN Routing?

TRILL for Data Center Networks

Based on Computer Networking, 4 th Edition by Kurose and Ross

QoS Switching. Two Related Areas to Cover (1) Switched IP Forwarding (2) 802.1Q (Virtual LANs) and 802.1p (GARP/Priorities)

Ethernet. Ethernet Frame Structure. Ethernet Frame Structure (more) Ethernet: uses CSMA/CD

PERFORMANCE OF MOBILE AD HOC NETWORKING ROUTING PROTOCOLS IN REALISTIC SCENARIOS

Performance Analysis of IPv4 v/s IPv6 in Virtual Environment Using UBUNTU

Introduction to MPIO, MCS, Trunking, and LACP

DEMYSTIFYING ROUTING SERVICES IN SOFTWAREDEFINED NETWORKING

Alkit Reflex RTP reflector/mixer

Hybrid Optical and Packet Infrastructure (HOPI) Project

Introduction to Network Security Lab 1 - Wireshark

Optimizing Enterprise Network Bandwidth For Security Applications. Improving Performance Using Antaira s Management Features

Measurement of IP Transport Parameters for IP Telephony

IP Networking. Overview. Networks Impact Daily Life. IP Networking - Part 1. How Networks Impact Daily Life. How Networks Impact Daily Life

Ethernet-based Software Defined Network (SDN) Cloud Computing Research Center for Mobile Applications (CCMA), ITRI 雲 端 運 算 行 動 應 用 研 究 中 心

CCNA R&S: Introduction to Networks. Chapter 5: Ethernet

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago

Performance Evaluation of Linux Bridge

PART II. OPS-based metro area networks

SDN and Data Center Networks

Application Note Gigabit Ethernet Port Modes

ADVANCED NETWORK CONFIGURATION GUIDE

Voice Over IP. MultiFlow IP Phone # 3071 Subnet # Subnet Mask IP address Telephone.

Applications. Network Application Performance Analysis. Laboratory. Objective. Overview

OVERLAYING VIRTUALIZED LAYER 2 NETWORKS OVER LAYER 3 NETWORKS

Section 1: Network monitoring based on flow measurement techniques

IT 3202 Internet Working (New)

100Gigabit and Beyond: Increasing Capacity in IP/MPLS Networks Today Rahul Vir Product Line Manager Foundry Networks

Performance Evaluation of VoIP Services using Different CODECs over a UMTS Network

Objectives of Lecture. Network Architecture. Protocols. Contents

10/100/1000Mbps Ethernet MAC with Protocol Acceleration MAC-NET Core with Avalon Interface

Transport and Network Layer

REDUCING PACKET OVERHEAD IN MOBILE IPV6

Software Defined Cloud Networking

ENABLING INNOVATION THROUGH NETWORK VIRTUALIZATION (AND INTEGRATION OF COMPUTE AND STORAGE)

2. What is the maximum value of each octet in an IP address? A. 28 B. 255 C. 256 D. None of the above

Exploring Mobile/WiFi Handover with Multipath TCP

Transport layer issues in ad hoc wireless networks Dmitrij Lagutin,

The Network Hypervisor

Networking Topology For Your System

IXP Member connection Best Practice. Kittinan Sriprasert BKNIX

EXPERIMENTAL STUDY FOR QUALITY OF SERVICE IN VOICE OVER IP

Introduction to IP v6

Multi-site Datacenter Network Infrastructures

MikroTik RouterOS Workshop Load Balancing Best Practice. Warsaw MUM Europe 2012

Performance Evaluation of AODV, OLSR Routing Protocol in VOIP Over Ad Hoc

Intel Ethernet Switch Converged Enhanced Ethernet (CEE) and Datacenter Bridging (DCB) Using Intel Ethernet Switch Family Switches

UPPER LAYER SWITCHING

Basic Networking Concepts. 1. Introduction 2. Protocols 3. Protocol Layers 4. Network Interconnection/Internet

Ethernet Fabric Requirements for FCoE in the Data Center

Cloud Infrastructure Planning. Chapter Six

Network congestion control using NetFlow

Network performance in virtual infrastructures

You will work in groups of two on the labs. It is OK to talk to others and help each other in the lab.

Bandwidth Aggregation, Teaming and Bonding

Top-Down Network Design

Disjoint Path Algorithm for Load Balancing in MPLS network

UK Interconnect White Paper

Influence of Load Balancing on Quality of Real Time Data Transmission*

Comparison of RIP, EIGRP, OSPF, IGRP Routing Protocols in Wireless Local Area Network (WLAN) By Using OPNET Simulator Tool - A Practical Approach

Avaya VENA Fabric Connect

Implementation of Virtual Local Area Network using network simulator

Data Center Networking Designing Today s Data Center

What is TCP/IP. Introduction to TCP/IP &216</ &216</ z Transmission Control Protocol/Internet Protocol

Accelerating Network Virtualization Overlays with QLogic Intelligent Ethernet Adapters

Development of the FITELnet-G20 Metro Edge Router

LAN Switching Computer Networking. Switched Network Advantages. Hubs (more) Hubs. Bridges/Switches, , PPP. Interconnecting LANs

High-Speed TCP Performance Characterization under Various Operating Systems

Transcription:

Experiences with MP in an intercontinental multipathed network Ronald van der Pol, Michael Bredel, Artur Barczyk SURFnet Radboudkwartier 273 3511 CK Utrecht, The Netherlands Email: Ronald.vanderPol@SURFnet.nl California Institute of Technology, c/o CERN 1211 Geneva, Switzerland Email: michael.bredel@cern.ch, Artur.Barczyk@cern.ch KEYWORDS MP,, Floodlight, multipath, big data ABSTRACT In this paper we present our experiences with multipath (MP) in an intercontinental multipathed testbed. We believe that multipathing will become an important network concept in the next few years. One of the major reasons is that data sets in e-science are increasing exponentially in size. To transfer these huge data sets we need to make efficient use of all available network capacity. This means using multiple paths simultaneously wherever possible. To get experience with this concept we set up a testbed in which we transferred data between two servers. The testbed consisted of es and multiple link disjoint paths between the servers. An application provisioned multiple paths between the servers and MP was used on the servers to simultaneously send traffic across all those paths. Figure 1 shows the testbed that was Fig. 1. SC12 Multipath Network Topology used for the SC12 demonstration in Salt Lake City in November 2012.

Multipathing can be done at L3 with Equal Cost Multipath (ECMP) routing or at L2 with protocols like TRILL (IETF RFC 5556) [1] or IEEE 802.1aq (Shortest Path Bridging - P802.1aq-2012 [2]). In all these cases load balancing across the paths is done based on flows by calculating a hash (based on e.g. Ethernet addresses, IP addresses and /UDP port numbers) of the packets. Each packet of such a flow follows the same path through the network, which prevents out of order delivery within a flow. When the traffic has many different flows the traffic will be evenly spread across the various paths. But when there are only a few flows, which is typically the case in large data e-science applications, this is not the case. Another disadvantage of hashing is that usually all links get the same percentage of the hash values and therefore all the paths need to have the same capacity. Multipath is a new approach towards efficient load balancing. Instead of letting the network do the load balancing by using hashes and ECMP, MP is doing the load balancing in the end nodes as part of the process. Multipath (MP) is described in RFC 6182 [3] and the Extensions for Multipath Operation with Multiple Addresses internet draft [4]. MP is an active working group in the IETF. Figure 2 shows how MP works. In a MP enabled kernel application IP application MP IP IP IP Fig. 2. Traditional vs Multipath the component is split in a MP component and components for each interface. The MP component receives a byte stream from the application (MP uses an unmodified socket API and semantics northbound, so applications do not need to be adapted). The MP component splits the byte stream into multiple segments which are handed to the components. Each behaves as a normal flow to the network. MP can handle paths of different bandwidth because there is a congestion control mechanism across the s. This congestion control mechanism takes care that traffic on a congested path is moved to a path with less congestion. So it adapts the load balancing according to the load of other traffic on the network. The MP component implements three functions. It takes care of path management by detecting and using multiple paths to a destination. Packet scheduling splits the byte stream received from the application in multiple segments and transmits these segments on one of the available s. These segments are numbered, so that the receiving MP component can put the segments in the correct order and reconstruct the original byte stream. Finally there is congestion control across the s. This function spreads the load over the s. When a becomes congested, traffic is moved to a that is less congested. This function also takes care of retransmissions on another when one of the s fails. In 2012 two demonstrations were given, one in October in Chicago during the GLIF meeting and one in November in Salt Lake City during SC12. During the GLIF meeting we showed streaming from Geneva to Chicago over multiple 10GE paths. On our servers we used the Linux MP implementation of the IP networking lab of the Université de Louvain in Belgium [7]. Both servers had two 10GE NICs each. On these physical interfaces we configured two MAC VLAN virtual interfaces so that we could give each virtual interface its own MAC address. In our testbed the various paths through the network are set up by provisioning forwarding entries on the es. Each of the four virtual interfaces was mapped to its own path through the network and each path had it own IP subnet assigned to it. The forwarding entries matched on destination MAC and IP address. There was no routing in our testbed, so four s (paths) could be used by MP. During the GLIF demonstration we were able to reach an aggregated end-to-end throughput of around 3 Gbit/s. Later analysis showed that we used an MP segment size of only 1400 bytes, which was probably one of the reasons for the relatively low throughput. During SC12, we showed OLiMPS and MP by streaming from Geneva to Salt Lake City as part of the SCInet Research Sandbox. Figure 1 shows the topology used during SC12. Figure 3 shows the live monitoring website at SC12. We have also streamed between Geneva and Amsterdam and by using four paths we were able to reach an aggregated end-to-end

Fig. 3. Live monitoring website at SC12 throughput of around 13 Gbit/s. Each of the four s is mapped to one of the four paths and has its own colour in the graphs (see figure 4). When the stream was started all four s were used. After a while only the red and green s server server Fig. 4. Four paths between Geneva and Amsterdam were used. These are the only two link disjoint paths between the servers, so it makes sense that MP would eventually use only these paths. Figure 5 shows how initially all flows are used and after some time only the two link disjoint paths. Figure 6 show the usage after some time. In the first half of 2013 we will continue this work in several areas: MP throughput measurements between Geneva, Amsterdam and Chicago MP throughput on links of different bandwidth (e.g. 1GE and 10GE) Continue the development of the OLiMPS application MP and multipathing experiments on the Géant testbed Demo at Joint Techs in Hawaii Demo at TNC2013 The results of these activities will be presented at TNC2013. ACKNOWLEDGEMENTS The work of SURFnet is made possible by the GigaPort program of the Dutch Government and the Géant3 project of the European Commission. Caltech research in this area is supported by the DOE Office of Advanced Scientific Computing

Fig. 5. Streaming between Geneva and Amsterdam, initial phase Fig. 6. Streaming between Geneva and Amsterdam, stable phase Research (OASCR). REFERENCES [1] J. Touch and R. Perlman, Transparent Interconnection of Lots of Links (TRILL): Problem and Applicability Statement, RFC 5556, May 2009 [2] IEEE Approved Draft Standard for Local and Metropolitan Area Networks: Bridges and Virtual Bridged Local Area Networks - Amendment 9: Shortest Path Bridging, IEEE 802.1aq-2012, January 2012 [3] A. Ford, C. Raicu, M. Handley, S. Barre and J. Iyengar, Architectural Guidelines for Multipath Development, RFC 6182, March 2011 [4] A. Ford, C. Raiciu, M. Handley, O. Bonaventure, Extensions for Multipath Operation with Multiple Addresses, draft-ietf-mptcp-multiaddressed-08, November 2012 [5] Floodlight Controller http://floodlight.openflowhub.org/ [6] IEEE Standard for Local and Metropolitan Area Networks - Station and Media Access Control Connectivity Discovery, IEEE 802.1AB-2009, ISBN 978-0-7381-6038-2, September 2009 [7] Multipath - Linux Kernel implementation http://mptcp.info.ucl.ac.be/ VITAE Ronald van der Pol is a network researcher at SURFnet. He has been working in the field of Education and Research Networks for more than twenty years. His former employers include the VU University in Amsterdam, NLnet Labs and SARA. His current interests are in new network technologies and how these can be applied to next generation networking. In recent years he worked on monitoring and management of optical transport networks, carrier Ethernet, end-to-end performance of demanding applications and. He holds masters degrees in both Physics and Computer Science and is co-author of several IPv6 RFCs. Michael Bredel is a network research engineer at Caltech based at CERN. He received a Diploma of Electrical Engineering from the Technische Universität Darmstadt, Germany and a Ph.D. from the Leibniz Universität Hannover, Germany. His current

research activities are related to network measurements, network management, software defined networks and. His particular interests are new technologies that can improve the movement of Big Data, like for the Large Hadron Collider. Artur Barczyk is working since 2007 in the Caltech HEP Networking group as a senior research engineer. He is the technical team lead in US LHCNet, the transatlantic, mission-oriented network for LHC data. Artur holds a PhD in Particle Physics.