Architecting Low Latency Cloud Networks

Similar documents
Switching Architectures for Cloud Network Designs

Cloud Networking Services

Impact of Virtualization on Cloud Networking Arista Networks Whitepaper

Block based, file-based, combination. Component based, solution based

The Impact of Virtualization on Cloud Networking Arista Networks Whitepaper

Bioscience. Introduction. The Importance of the Network. Network Switching Requirements. Arista Technical Guide

Hadoop Cluster Applications

Introduction. Need for ever-increasing storage scalability. Arista and Panasas provide a unique Cloud Storage solution

The Future of Cloud Networking. Idris T. Vasi

Scaling 10Gb/s Clustering at Wire-Speed

Cloud Networking: A Novel Network Approach for Cloud Computing Models CQ1 2009

Achieving Data Center Networking Efficiency Breaking the Old Rules & Dispelling the Myths

Cisco s Massively Scalable Data Center

Juniper Networks QFabric: Scaling for the Modern Data Center

Data Center Networking Designing Today s Data Center

Intel Ethernet Switch Converged Enhanced Ethernet (CEE) and Datacenter Bridging (DCB) Using Intel Ethernet Switch Family Switches

High-Performance Automated Trading Network Architectures

Cloud Networking: Design Patterns for Cloud-Centric Application Environments. January 2009

Chapter 1 Reading Organizer

Optimizing Data Center Networks for Cloud Computing

Fibre Channel Overview of the Technology. Early History and Fibre Channel Standards Development

TRILL for Service Provider Data Center and IXP. Francois Tallet, Cisco Systems

Large Scale Clustering with Voltaire InfiniBand HyperScale Technology

LAYER3 HELPS BUILD NEXT GENERATION, HIGH-SPEED, LOW LATENCY, DATA CENTER SOLUTION FOR A LEADING FINANCIAL INSTITUTION IN AFRICA.

IEEE Congestion Management Presentation for IEEE Congestion Management Study Group

Cloud Networking: A Network Approach that Meets the Requirements of Cloud Computing CQ2 2011

Scalable Approaches for Multitenant Cloud Data Centers

RoCE vs. iwarp Competitive Analysis

Data Center Switch Fabric Competitive Analysis

ioscale: The Holy Grail for Hyperscale

Building Virtualization-Optimized Data Center Networks

IP ETHERNET STORAGE CHALLENGES

VMware Virtual SAN 6.2 Network Design Guide

Networking in the Hadoop Cluster

InfiniBand Clustering

Non-blocking Switching in the Cloud Computing Era

Building a Scalable Big Data Infrastructure for Dynamic Workflows

Low Latency 10 GbE Switching for Data Center, Cluster and Storage Interconnect

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

全 新 企 業 網 路 儲 存 應 用 THE STORAGE NETWORK MATTERS FOR EMC IP STORAGE PLATFORMS

Extreme Networks: Building Cloud-Scale Networks Using Open Fabric Architectures A SOLUTION WHITE PAPER

Enabling High performance Big Data platform with RDMA

VMDC 3.0 Design Overview

High Performance Computing (HPC)

Data Center Convergence. Ahmad Zamer, Brocade

ARISTA WHITE PAPER Why Big Data Needs Big Buffer Switches

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

Ultra Low Latency Data Center Switches and iwarp Network Interface Cards

SDN CENTRALIZED NETWORK COMMAND AND CONTROL

Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging

Testing Network Virtualization For Data Center and Cloud VERYX TECHNOLOGIES

SummitStack in the Data Center

Fabrics that Fit Matching the Network to Today s Data Center Traffic Conditions

Next Steps Toward 10 Gigabit Ethernet Top-of-Rack Networking

Interconnect Analysis: 10GigE and InfiniBand in High Performance Computing

Flattening the Data Center Architecture

3G Converged-NICs A Platform for Server I/O to Converged Networks

Meeting the Five Key Needs of Next-Generation Cloud Computing Networks with 10 GbE

THE BIG DATA REVOLUTION

Advanced Computer Networks. Datacenter Network Fabric

BUILDING A SCALABLE BIG DATA INFRASTRUCTURE FOR DYNAMIC WORKFLOWS

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA

10Gb Ethernet: The Foundation for Low-Latency, Real-Time Financial Services Applications and Other, Latency-Sensitive Applications

The Next Generation Trading Infrastructure What You Will Learn. The State of High-Performance Trading Architecture

DEDICATED NETWORKS FOR IP STORAGE

Latency Considerations for 10GBase-T PHYs

Ultra-Low Latency, High Density 48 port Switch and Adapter Testing

White Paper Solarflare High-Performance Computing (HPC) Applications

Cut I/O Power and Cost while Boosting Blade Server Performance

InfiniBand -- Industry Standard Data Center Fabric is Ready for Prime Time

Performance Evaluation of the RDMA over Ethernet (RoCE) Standard in Enterprise Data Centers Infrastructure. Abstract:

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

Brocade Solution for EMC VSPEX Server Virtualization

The virtualization of SAP environments to accommodate standardization and easier management is gaining momentum in data centers.

Virtual PortChannels: Building Networks without Spanning Tree Protocol

Scaling Cloud-Native Virtualized Network Services with Flash Memory

STATE OF THE ART OF DATA CENTRE NETWORK TECHNOLOGIES CASE: COMPARISON BETWEEN ETHERNET FABRIC SOLUTIONS

CENTER I S Y O U R D ATA

Joint ITU-T/IEEE Workshop on Carrier-class Ethernet

Ethernet Fabrics: An Architecture for Cloud Networking

Data Center Infrastructure of the future. Alexei Agueev, Systems Engineer

Arista and Leviton Technology in the Data Center

Fibre Channel over Ethernet in the Data Center: An Introduction

Cloud Infrastructure Planning. Chapter Six

Introduction to Cloud Design Four Design Principals For IaaS

Building a Scalable Storage with InfiniBand

Solution Brief Network Design Considerations to Enable the Benefits of Flash Storage

Cloud Computing and the Internet. Conferenza GARR 2010

A Whitepaper on. Building Data Centers with Dell MXL Blade Switch

Lustre Networking BY PETER J. BRAAM

Top Ten Reasons for Deploying Oracle Virtual Networking in Your Data Center

Mellanox Academy Online Training (E-learning)

Unified Fabric: Cisco's Innovation for Data Center Networks

WAN Performance Analysis A Study on the Impact of Windows 7

Why the Cloud Titans Drive DC Strategy and how you benefit! Mark Carnighan. November 2015

Data Center Architecture Overview

ARISTA WHITE PAPER Deploying IP Storage Infrastructures

Interoperability Testing and iwarp Performance. Whitepaper

Direct Scale-out Flash Storage: Data Path Evolution for the Flash Storage Era

Transcription:

Architecting Low Latency Cloud Networks Introduction: Application Response Time is Critical in Cloud Environments As data centers transition to next generation virtualized & elastic cloud architectures, high performance and resilient cloud networking has become a requirement for delivering always-on, always-available applications. Over the last decade, response time requirements have been rigorously tightened (see Figure 1), in some cases by as much as two to three orders of magnitude. In addition to the financial, education/research institutions and Web 2.0 firms, there is a growing list of customer verticals where fast response time is a top-of-mind business requirement, including richmedia companies (e.g. those involved in video post-production, digital animation and broadcasting), oil & gas producers, and health care. Figure1:ApplicationResponseTimeTrend As response times gravitate towards fractions of a second, both network and compute latencies need to be reduced in parallel. In effect an architectural approach is needed so that response times are consistently and measurably enhanced across a wide set of cloud applications. Unlike prior efforts that often deploy niche and sometimes even proprietary technologies, today it is critical that low latency infrastructure fabric be built using standards-based and widely used technologies for broadbased cloud deployments. Also, because latency optimization is implemented at the infrastructure level, it needs to be future proof to meet latency demands 3 to 5 years out. Architecting Low-latency Cloud Networks A key attribute of latency-sensitive workloads is that they are built using distributed compute architecture. Each workload transaction spawns a large number of interactions between compute nodes, in some cases across thousands of machines and data stores, as depicted in Figure 2. For example: In algorithmic trading, machines receive market data feeds from other machines, calculate trading decisions and place trades with another set of machines. In HPC, thousands of commodity servers are clustered to behave as a super computer, connected to a clustered file system (e.g. Lustre) to crunch complex mathematical formulae and simulations. 1

Architecting Low Latency Cloud Networks In web based cloud applications, workloads operate on thousands of dynamically clustered servers and implement multi-stage processing mechanisms to sift through petabytes of data. Example applications include those implemented by Web 2.0 and Internet Search companies, as well as emerging open-source initiatives such as Hadoop which provides scalable and reliable framework for distributed computing (MapReduce), distributed file system and distributed database. Figure2:Networkbasedmachine to machineinteractionsofasingleworkload:atransaction To boost system-wide response time, both compute and network latencies must be minimized. Typically, commodity servers are used for computing where degrees of freedom for latency optimization are limited. One essentially selects a server with denser and faster multi-core CPUs, larger and faster memories/caches, and speedier peripheral interconnects. The other side of latency coin is the cloud network where fabric-wide optimization is called for to minimize transport latencies. In particular, network latencies need to be optimized across four key axes: Reduce latency of each network node Reduce number of network nodes needed to traverse from one stage to another Eliminate network congestion Reduce transport protocol latency The Arista Advantage: Arista Networks has adopted an architectural approach to building low latency cloud networking. Its wire-rate, non-blocking and ultra low latency 1Gb/10GbE switches, along with extensible operating system (EOS), enable resilient cloud networks for transporting data, multi-media, storage and compute traffic. Architecting cloud networks based on Arista s 7xxx series switch fabric has the following advantages: Reduce network latencies across all four axes (see details below); Future-proof the network infrastructure for latency demands of today and tomorrow; Implement standard Ethernet, which is widely deployed and operationally well understood. Network Node Latency: For ultra low latency workloads, such as those in financials and HPC, shaving off every microsecond is important. Deploying the optimal 1/10Gb Ethernet switch can improve latency by 10+ microseconds for 1500B Ethernet frames (even more if the 9KB jumbo frames are used). Also, the cut-through switch architecture can reduce additional tens of microseconds of over older store-and-forward designs. Arista provides dense 10Gb switches with cut-through architecture that yields switching latencies in the range of 0.6-1.2 microsecond (see Figure 3). 2

Architecting Low Latency Cloud Networks (a) Legacy Store-and-forward Design Intra-Rack Latency: 8us Inter-Rack Latency: 36us (b) Arista Cut-through Design Intra-Rack Latency: 0.6us Inter-Rack Latency: 2.4us Figure3:LatencybenefitsofArista'scut throughswitchesovertraditionalstore and forwardswitches Number of Network Hops: Typical data center architecture consists of a three-tier architecture, with server access, distribution and core switch layers. Often packets need to traverse all the way to the core layer when communicating across server racks. In Arista s two-tier topology with leaf and spine switches (see Figure 4), latency is significantly reduced as packets traverse fewer hops. Figure4:Arista's2 tier(leafandspine)topologyreduceslatencyovertraditional3 tier(access,distribution,core) topology 3

Architecting Low Latency Cloud Networks Network Congestion: In most applications, IP/Ethernet packets are transported using the TCP protocol. To guarantee reliability and maintain throughput, TCP state machine uses windowing to adjust to dynamic network congestion. During congestion periods, TCP uses smaller windows, thus decreasing throughput and increasing overall transmission latency. Traditional data center architecture is oversubscribed, thus leading to congestions & dropped packets and hence large TCP latencies. Arista s 7100 switches are wire-rate switches (line rate for every 10Gb Ethernet port), thus enabling non-oversubscribed cloud network architectures with 1:1 bisectional bandwidth (as shown in Figure 4). Large, congestion-free, fabric can be architected that allows TCP to operate using large window sizes and hence at high throughput and low latency. Hundreds or thousands of machine-tomachine TCP interactions of application workloads, be they web, file/storage, multi-media or database, benefits immensely in terms of TCP throughput and latency. Transport Protocol Latency: Certain low latency environments (e.g. HPC, applications leveraging Lustre file system algorithmic trading), it is desired to bypass TCP altogether and instead leverage direct machine-to-machine communication using message-passing (MPI) or remote direct memory access (RDMA). Infiniband and other proprietary interconnects have been used because of their ultra low latency, high throughput and lossless characteristics. Customers have longed for standardsbased Ethernet to provide such capabilities as well as price-performance advantage so that they can benefit from the cost-efficiencies of managing a single Ethernet fabric across the public or private cloud. Arista s 7100 series of 10Gb Ethernet switches, with line rate and cut-through low latency performance, enables a single low latency cloud network. Arista s architecture also supports current and emerging Ethernet standards in IEEE and IETF on Layer-2/3 multi-pathing and congestion control to span large congestion free domains. Large Layer-2 domains, along with Arista s low latency, wirerate and congestion-free cloud network, enable rapid migration of application workloads anywhere in the cloud. Low-latency Application Example: Capital Markets Algorithmic trading and market data feeds are the most celebrated use cases of ultra low latency designs where it is said that 1-millisecond advantage in trading applications can be worth $100 million a year to a major brokerage firm 1. Shaving off milliseconds is so critical that brokerage firms co-locate their trading systems right on the exchange floor for direct access to market data feeds, thus eliminating distance-based latency barriers. Modern designs strive for one millisecond response time for trading transactions, with the eye towards deep sub-millisecond designs in few years. Quest for speed has often led to deployment of exotic interconnects, such as Infiniband that are operationally difficult to deploy & manage at cloud scale. Arista s line-rate Ethernet switch fabric, along with eco system partners such as Solace Systems, provides a complete algorithmic trading solution with 31-microsecond latency, as shown in Figure 5(b). 1 BusinessatLightSpeed byrichardmartin,informationweek,april2007 (http://www.informationweek.com/news/infrastructure/showarticle.jhtml?articleid=199200297) 4

Architecting Low Latency Cloud Networks Figure5(a):Marketdatafeedandalgorithmictrading5(b):Arista+SolaceSystemsprovidesultralowlatency algorithmictradingsolution Summary The drive for scalable and efficient cloud networks has led to 10Gb Ethernet as the interface of choice for simultaneously transporting data, storage and compute information over a common Ethernet cloud. Multiple classes of applications, including those in capital markets, high performance computing and web-centric computing, rely on latency-optimized architectures for ultra fast response times. After all, it is faster response times that ensure users stickiness to a service, enhance scalability of HPC clusters and increase trading profits by speeding up algorithmic trade execution during price fluctuations. The Arista Networks 7XX series is designed solve these real word cloudcomputing challenges for latency and scale. InformationinthisdocumentisprovidedinconnectionwithAristaNetworksproducts.Formoreinformation,visit usathttp://www.aristanetworks.com,orcontactusatsales@aristanetworks.com 5