Deploying 10/40G InfiniBand Applications over the WAN



Similar documents
Cisco SFS 7000P InfiniBand Server Switch

Long-Haul System Family. Highest Levels of RDMA Scalability, Simplified Distance Networks Manageability, Maximum System Productivity

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

Ethernet: THE Converged Network Ethernet Alliance Demonstration as SC 09

Storage at a Distance; Using RoCE as a WAN Transport

Introduction to Infiniband. Hussein N. Harake, Performance U! Winter School

Primary Data Center. Remote Data Center Plans (COOP), Business Continuity (BC), Disaster Recovery (DR), and data

White Paper Solarflare High-Performance Computing (HPC) Applications

100 Gigabit Ethernet is Here!

High Speed I/O Server Computing with InfiniBand

Flexible SDN Transport Networks With Optical Circuit Switching

3G Converged-NICs A Platform for Server I/O to Converged Networks

SummitStack in the Data Center

High Performance Data-Transfers in Grid Environment using GridFTP over InfiniBand

Data Center Fabric Convergence for Cloud Computing (the Debate of Ethernet vs. Fibre Channel is Over)

Government Transport Networks: Minimize Lifetime Costs

InfiniBand Software and Protocols Enable Seamless Off-the-shelf Applications Deployment

Please purchase PDF Split-Merge on to remove this watermark.

SMB Direct for SQL Server and Private Cloud

Data Sheet. V-Net Link 700 C Series Link Load Balancer. V-NetLink:Link Load Balancing Solution from VIAEDGE

Virtualized Converged Data Centers & Cloud how these trends are effecting Optical Networks

Performance of RDMA-Capable Storage Protocols on Wide-Area Network

How To Build A Network For Storage Area Network (San)

Campus Network Design Science DMZ

A Tour of the Linux OpenFabrics Stack

Fibre Channel over Ethernet in the Data Center: An Introduction

Brocade Solution for EMC VSPEX Server Virtualization

Optical Networks for Next Generation Disaster Recovery Networking Solutions with WDM Systems Cloud Computing and Security

High Speed Ethernet. Dr. Sanjay P. Ahuja, Ph.D. Professor School of Computing, UNF

Value Proposition for Data Centers

10 Gigabit Ethernet: Scaling across LAN, MAN, WAN

DEVELOPING A STRATEGIC APPROACH TO DATA CENTER NETWORKING

TP-LINK. 24-Port Gigabit L2 Managed Switch with 4 SFP Slots. Overview. Datasheet TL-SG

Lustre Networking BY PETER J. BRAAM

Low Latency 10 GbE Switching for Data Center, Cluster and Storage Interconnect

Frequently Asked Questions

Cisco SFS 7000D Series InfiniBand Server Switches

Towards an Open Data Center with an Interoperable Network Volume 5: WAN and Ultra Low Latency Applications Last update: May 2012

Cisco Bandwidth Quality Manager 3.1

Using High Availability Technologies Lesson 12

Block based, file-based, combination. Component based, solution based

TP-LINK. 24-Port 10/100Mbps + 4-Port Gigabit L2 Managed Switch. Overview. Datasheet TL-SL5428E.

Building Enterprise-Class Storage Using 40GbE

Server Consolidation and Remote Disaster Recovery: The Path to Lower TCO and Higher Reliability

Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 SMB Direct

FIBRE CHANNEL OVER ETHERNET

TP-LINK 24-Port Gigabit L2 Managed Switch with 4 SFP Slots

EVALUATING NETWORK BUFFER SIZE REQUIREMENTS

Mellanox Academy Online Training (E-learning)

TP-LINK L2 Managed Switch

INTRODUCTION TO MEDIA CONVERSION

NET ACCESS VOICE PRIVATE CLOUD

Best Practice and Deployment of the Network for iscsi, NAS and DAS in the Data Center

NetTESTER Embedded 'Always-On' Network Testing & In-Service Performance Assurance

10G CWDM Conversion Technology

RDMA over Ethernet - A Preliminary Study

Extending SANs Over TCP/IP by Richard Froom & Erum Frahim

WanVelocity. WAN Optimization & Acceleration

CTS2134 Introduction to Networking. Module 07: Wide Area Networks

Windows TCP Chimney: Network Protocol Offload for Optimal Application Scalability and Manageability

How To Get More Bandwidth From Your Business Network

Recession-Proof Consulting Services with CWDM Network Design

Accelerate Private Clouds with an Optimized Network

Four Ways High-Speed Data Transfer Can Transform Oil and Gas WHITE PAPER

The Software Defined Hybrid Packet Optical Datacenter Network SDN AT LIGHT SPEED TM CALIENT Technologies

Advancing Applications Performance With InfiniBand

TP-LINK. 24-Port 10/100Mbps + 4-Port Gigabit L2 Managed Switch. Overview. Datasheet TL-SL

Central Office Testing of Network Services

IEEE Congestion Management Presentation for IEEE Congestion Management Study Group

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Multi-protocol Label Switching

A SENSIBLE GUIDE TO LATENCY MANAGEMENT

The Impact Of The WAN On Disaster Recovery Capabilities A commissioned study conducted by Forrester Consulting on behalf of F5 Networks

ethernet alliance Data Center Bridging Version 1.0 November 2008 Authors: Steve Garrison, Force10 Networks Val Oliva, Foundry Networks

The Hybrid Enterprise. Enhance network performance and build your hybrid WAN

APV9650. Application Delivery Controller

Oracle Exalogic Elastic Cloud: Datacenter Network Integration

PRODUCTS & TECHNOLOGY

Overview of Requirements and Applications for 40 Gigabit and 100 Gigabit Ethernet

The Future of Storage Area Network

Arista 7060X and 7260X series: Q&A

CloudEngine Series Data Center Switches. Cloud Fabric Data Center Network Solution

TP-LINK. Gigabit L2 Managed Switch. Overview. Datasheet TL-SG3216 / TL-SG

The Next Generation Network:

Oracle Big Data Appliance: Datacenter Network Integration

Net Optics and Cisco NAM

Preparing Your IP Network for High Definition Video Conferencing

SummitStack in the Data Center

Smart Solutions for Network IP Migration

nappliance Network Virtualization Gateways

Business Services. Is Ethernet the Right Choice for Your Network? Learn More: Call us at

How To Use A Network Instrument Ntap

Highly-Available Distributed Storage. UF HPC Center Research Computing University of Florida

Issue December SAN Interworking. Contents

High Availability Server Clustering Solutions

Network Simulation Traffic, Paths and Impairment

Ethernet Link SGI-4844F

Resilient Metropolitan Area Networks

Chapter 13 Selected Storage Systems and Interface

iscsi Top Ten Top Ten reasons to use Emulex OneConnect iscsi adapters

Transcription:

Deploying 10/40G InfiniBand Applications over the WAN Eric Dube (eric@baymicrosystems.com) Senior Product Manager of Systems November 2011

Overview About Bay Founded in 2000 to provide high performance networking solutions Silicon Engineering & Headquarters: San Jose, CA Systems Engineering & Business Development: Germantown, MD; MA Corporate Focus Development of complex integrated circuits that are applied to high performance packet processing and optical transport applications in support of our systems Systems that deliver high performance protocol agnostic encryption adaptation, protocol inter-working, and WAN acceleration for government agencies and commercial enterprises 2

Wide Area Networking Challenges Wide Area Networks can often be difficult to deploy for many popular compute and storage applications Common issues include: Ability to maintain link utilization over extended distances Providing congestion control and avoiding packet loss Fair link sharing of bandwidth resources across multiple, concurrent applications TCP/IP packet acknowledgement delays exponentially grows as distance increase causing performance degradation for most applications 3

Benefits of using InfiniBand / RDMA for Wide Area Network Applications Improved performance Enables application and storage acceleration through RDMA Increased efficiency Provides maximum link utilization over the WAN with fair sharing of resources between applications Minimal latency Adds very little latency to the overall WAN connection for latency-sensitive applications Cost savings Expands existing WAN link capacity and offloads CPU for application processing (saving on both hardware processing and network bandwidth) Seamless implementation Transparent application interoperability with existing and new applications and storage solutions 4

Challenges of Extending InfiniBand Globally Campus, Metro, or Wide Area Network (from 1 to 1000 s of kilometers) 0100110 The need to extend InfiniBand between data centers is essential for providing disaster recovery, multi-site backups, and real-time data access solutions. While InfiniBand s crediting mechanism is an excellent and reliable way to provide flow control, existing InfiniBand LAN hardware doesn t provide enough port buffering for deployment beyond a single site. A reduction in sustained bandwidth starts occurring at 500-600 meters or less (depending on the data rate) due to inadequate port buffering if the number of virtual lanes aren t reduced. Even with the minimum amount of virtual lanes configured, not enough packets can be kept in-flight on the wire due to the port buffer credit starvation that occurs over extended distances such as greater than 4 kilometers 5

Welcome to IBEx WAN Acceleration Solutions Intelligent Bandwidth Exchange The IBEx product family enables wide area networking acceleration using RDMA over InfiniBand for compute and storage applications to any point on the globe (up to 15,000 km and beyond.) Highlights: Improves link utilization with 80-99% efficiency Supports 4X InfiniBand QDR connectivity today with future software upgradability to FDR10 for up to 40Gbps data rates Provides lossless communication and true QoS capabilities for workflows Flexible 10/40G WAN connectivity options over SONET/SDH, ITU-T G.709 OTN, and Ethernet 6

IBEx InfiniBand Support The IBEx InfiniBand product family supports: All native InfiniBand protocols IPoIB, SDP, RDS, MPI, udapl, iser, SRP, IB Verbs Layer, etc. Support RDMA data transfers over the WAN for applications Connectivity for all InfiniBand data rates SDR, DDR, QDR, FDR10* Standard QSFP InfiniBand Interface Accepts both active optical and passive copper cabling in addition to optical transceivers Operates as a typical InfiniBand switch device Appears as a 2-Port Switch in the InfiniBand fabric True 10/40G WAN-side data rates for extending native InfiniBand Provides 10/40G actual data rate throughput for InfiniBand extension with QDR and FDR10* connectivity * Future support for FDR10 data rates through IBEx system software upgrade 7

IBEx Platform Typical Data Center Deployment Diagrams DWDM Metro / Wide Area Network Deployment Storage 10/40G Wavelength Customer Premise Carrier Network 1-15,000+ km Servers InfiniBand LAN Switching Infrastructure IBEx Extension Platform DWDM Optical Transport Platform Optical Amplification Metro / Wide Area Network Applications Point-to-Point Campus / Metro Area Network Deployment Storage 10/40G Dark Fiber Customer Premise Carrier / Local Fiber Network 1-250 km optional Servers InfiniBand LAN Switching Infrastructure IBEx Extension Platform Optical Amplification optional Campus / Metro Area Network Applications 8

Need for Distributed InfiniBand Applications and Multi-site Deployments Global File Systems & Storage High performance/high volume data sharing and storage virtualization between sites High Performance Computing (HPC) Clustered applications and cloud computing Post-processing and visualization Financial Services Clustered Databases and Warehouses Distributed Healthcare Applications Disaster recovery solution for low latency trading and market data feed applications Multi-site failover and data mirroring Real-time local access and information sharing High resolution patient imaging sharing between offices Content Distribution Global distribution for thousands of HD videos over a single connection 9

Network Protocol Efficiencies Network Protocols: TCP/IP Typically software protocol stack implemented TCP subject to significant saw-tooth performance effects upon any loss with slow ramp to nominal utilization Conversion to UDP, using TCP spoofing techniques, helps performance but looses all notion of congestion management, reliable transport, and in-order delivery TCP/IP utilization significantly degrades with multi-session and any notion of congestion due to its reactive congestion control RDMA over InfiniBand (IB) RDMA (Remote Direct Memory Access) is hardware transfer initiated by the software application from local memory, across the network, to the remote server or mass storage system InfiniBand is lossless with end-to-end flow pro-active flow control and reliable, in order, detection and delivery With InfiniBand extension, InfiniBand can run on nearly any optical or traffic engineered network utilizing up to 90%+ efficiency of the available bandwidth 10

Large Data JCTD Protocol Performance Comparison Typical RDMA/IB Performance Typical RDMA/IB Performance Typical TCP/IP/ETHERNET Performance TESTS ON 1 Gbps CIRCUIT (~8000 miles) [ ~13,000 fiber miles] TESTS ON 8 Gbps CIRCUIT (~1200 miles) [~2000 fiber miles] RDMA over IB provides very efficient use of available bandwidth with near linear scaling RDMA/IB performance 80% TCP/IP performance 40% RDMA/IB CPU usage estimated 4x less InfiniBand connection is lossless with nearly perfect fair sharing of bandwidth across multiple, concurrent data flows * Slide content and performance data obtained from Large Data JCTD Public Presentation 11

Orange / ESnet / Bay Microsystems 40G IB Extension over SONET OC-768 370 km Fiber Loop AOFA ESnet ANI Testbed BNL IBEx G40 Applications, Servers, & Storage Infinera Optical Transport Platform NY Long Island Metro Area Network SONET OC-768 Service Infinera Optical Transport Platform IBEx G40 Applications, Servers, & Storage Application testing performed in July/August 2011 at Brookhaven National Laboratory as part of the ESnet ANI Testbed project Obtained 96% efficiency of useable bandwidth through concurrent streaming of RDMA applications Utilized a SONET OC-768 (40G) WAN circuit spanning 370 km from Upton, NY to Long Island and back 12

Orange / ESnet / Bay Microsystems ANI Testbed Performance Data MB/second 9000 8000 7000 6000 5000 4000 3000 2000 1000 Bidirectional Maximum Bandwidth (RC) Message Size 64 128 256 512 1024 2048 4096 8192 16384 32768 65536 131072 262144 524288 0 1 2 4 8 16 32 64 128 256 512 1024 Transmit Queue Depth 13

SC11: Orange / ESnet / Bay Microsystems World s First Long Distance 40G RDMA over InfiniBand Data Transfer Demonstration Seattle ~7000 Fiber Miles Loop! Salt Lake City Chicago Native 4X InfiniBand QDR is extended over 40G Ethernet / 100G MPLS network circuit provided by ESnet 14

SC11: Booth Demonstrations Remote Visualization Demonstration Visualization accessing remote data, leaving the dataset intact at the remote node Uncompressed Parallel HD Video Streaming over distance Transfer parallel, independent streams consuming file wire bandwidth High Performance Big Data File Transfers Demonstrate high bandwidth transfers over long-haul wide area networks 15

IBEx M40 4X InfiniBand QDR / FDR10 Extension / 40G WAN Acceleration Platform Management Ethernet 2 x 1G Ethernet 4X IB QDR 1 x 40G Ethernet 2 x 10G Ethernet Serial Console IBEx M40 Main Features 40G InfiniBand extension platform providing connectivity for: 4X InfiniBand QDR / FDR10 (up to 40Gbps) [1 x QSFP], 10G Ethernet [2 x SFP+], and 1G Ethernet [2 x SFP] Provides 40G WAN extension over: 40G Ethernet (40GBase-SR4/LR4), IPv4/IPv6, or dark fiber Enhanced internal port buffering and flow control capabilities enabling global InfiniBand extension at full line rate from 1-15,000km Easy To Use Secure Graphical User [HTTPS] and Command Line Interface [SSH] Compact, low power (<150 watts), 1U 19-inch rack mountable chassis Redundant, hot-swappable (dual-input) power supplies and fans 16

IBEx G40 4X InfiniBand QDR / FDR10 Extension / 40G WAN Acceleration Platform Management Ethernet 40G WAN (SONET OC-768/SDH STM-256, ITU-T G.709 OTU3, Dark Fiber) 4X IB QDR 2 x 1G Ethernet 2 x 10G Ethernet Serial Console IBEx G40 Main Features 40G InfiniBand extension platform providing connectivity for: 4X InfiniBand QDR / FDR10 (up to 40Gbps) [1 x QSFP], 10G Ethernet [2 x SFP+], and 1G Ethernet [2 x SFP] Provides 40G WAN extension over: SONET OC-768/SDH STM-256, ITU-T G.709 OTU3, or dark fiber Enhanced internal port buffering and flow control capabilities enabling global InfiniBand extension at full line rate from 1-15,000km Easy To Use Secure Graphical User [HTTPS] and Command Line Interface [SSH] Compact, low power (<150 watts), 1U 19-inch rack mountable chassis Redundant, hot-swappable (dual-input) power supplies and fans 17

IBEx M10/G10/M20/G20 4X InfiniBand QDR Extension / 10G WAN Acceleration Platforms Single 10G (actual data rate) InfiniBand over the WAN via SONET OC-192/SDH STM-64, ITU-T G.709 OTU2, or 10G Ethernet Dual 10G (actual data rate) InfiniBand over the WAN for site-to-side link redundancy or multi-site connectivity configurations 18

For more information please contact: Eric Dube Senior Product Manager of Systems Bay Microsystems, Inc. Phone: (301) 944-8149 Email: eric@baymicrosystems.com http://www.baymicrosystems.com 19

40G WAN (LC Fiber) 1G Ethernet LAN (SFP Transceiver) 4X InfiniBand QDR (QSFP Port 1) 10G Ethernet LAN (2 x SFP+ Transceiver) 1G Ethernet LAN (SFP Transceiver) 4X InfiniBand QDR (QSFP Port 1) 10G Ethernet LAN (2 x SFP+ Transceiver) IBEx G40 Platform Connectivity Diagram SONET OC-768/SDH STM-256, ITU-T G.709 OTU3, WDM, or Dark Fiber 4X InfiniBand QDR and 1/10G Ethernet LAN connections are encapsulated over the 40G WAN link Management Ethernet (RJ45) Serial Console (RJ45) 20