Storage at a Distance; Using RoCE as a WAN Transport
|
|
|
- Lionel Spencer
- 9 years ago
- Views:
Transcription
1 Storage at a Distance; Using RoCE as a WAN Transport Paul Grun Chief Scientist, System Fabric Works, Inc. (503) [email protected] Why Storage at a Distance the Storage Cloud Following close on the heels of Cloud Computing, the notion of the Storage Cloud is a powerful concept now finding significant traction in the enterprise. In the last generation, SANs within the data center served to virtualize storage and in so doing disaggregated storage from the compute platforms hosting application processing. This yielded important benefits in the way data is managed and in the way in which applications access and share data. Among those benefits is the ability to flexibly relocate application processing as needed to satisfy demands for performance or load balancing. In the same way, federating information storage across multiple data centers into a storage cloud brings unprecedented flexibility and reliability to the enterprise. These benefits are driving progress in the realization of practical private storage clouds. The Challenge of Storage over the WAN Creating a storage cloud, by definition, requires an ability to keep data synchronized even though it is stored in two or more distinct geographies. This is a delicate balancing act that requires attacking three key issues: 1. The efficient transfer of large data blocks or files over long distances, 2. Caching technologies that can help to overcome some of the distance delays, and 3. Synchronization and coordination among the storage sites The second and third items on the list are being addressed by a number of storage technology vendors such as EMC. But the issue of efficient high-speed transfer remains elusive. Users have become accustomed to accessing locally stored data at nearly electronic data rates. The problem with so-called cloud computing is that the delays in accessing storage at a distance can be orders of magnitude beyond what is experienced when accessing local storage. Some of this is due to speed of light delays and cannot be reduced, but a huge amount of it arises in the low level details of transmitting large blocks of data over long distance links using traditional networking technologies. If we can successfully address these issues, we can bring storage access delays down to a realm where they can be managed using caching and other advanced data management techniques. In fact, that should be our goal: When designing a long distance interconnect for storage, we should strive to raise the performance level of remote storage to the point where the performance delta between remote storage and local storage can be effectively concealed through caching techniques. So what are the challenges that inhibit high-speed storage access over the WAN? We have already mentioned speed of light which can be expected to contribute delays. Even though this number cannot be reduced, there are other significant contributors that can and should be addressed. Current methods for synchronizing or moving data between remote sites depend on a file copy application like FTP, which in turn depends on a reliable transport such as TCP. As part of its reliability 1
2 mechanism, TCP employs a sliding window protocol ( congestion window ) to detect failures or collisions within the network. These failures or collisions result in dropped packets and a subsequent loss of reliability and they can impact the performance of the network. The congestion window determines the number of bytes that the sender can transmit before it must stop and wait for an acknowledgement from the receiver. Once an acknowledgement is received, the sender can again begin transmitting until the congestion window is again full. A key determinant of the performance of any windowing protocol like TCP is the bandwidth-delay product of the interconnect. As its name implies, the bandwidth-delay product is calculated as the product of the bandwidth of the wire multiplied by the length of the wire measured in units of time; thus the bandwidth-delay product is a measure of the amount of data that can be stored in transit on the wire. There is an obvious connection between the bandwidth-delay product of the wire and the TCP congestion window; if the maximum congestion window supported by the transport protocol (TCP) is much smaller than the bandwidth-delay product for the wire connection, the transport will not be able to keep the wire continuously full; it will have to stop transmitting periodically while it waits for bytes to propagate down the length of the wire and for acknowledgments to be returned. This results in large gaps in time where the wire is idle as the transmitter waits for acknowledgements to be returned from the far end. For LANs, the bandwidth-delay product of a reasonable wire length is on the order of magnitude of a typical TCP congestion window size. For WANs, however, the delay component of the product becomes increasingly significant. For example, the delay associated with an 8600 mile WAN represents about 163mS of roundtrip delay. For a 100Mb/s link, this represents a bandwidth-delay product of about 200KB; well within the size of a typical congestion window. As the wire speed increases to 1Gb/s, the bandwidth-delay product increases to 20MB and at 10Gb/s it increases again to 200MB. Given that typical TCP implementations use a default window size of 4MB, it is clear that the bandwidthdelay product swamps the available congestion window. Thus, as the bandwidth-delay product of the wire increases, either due to increasing length or increasing bandwidth, so too must the congestion window size. Naturally, there is a tradeoff involved in increasing the congestion window; a larger window size means that more bytes are in-flight at any given point in time and thus at risk of loss due to congestion dropping or error. And, as the congestion window size increases, so too does the overhead associated with recovering from a lost byte. At some point it becomes impractical to increase the window size further. There is anecdotal and experimental evidence to suggest that TCP is satisfactory over the WAN at 1Gb/s speeds, and decreasingly less so at 10Gb/s speeds. Extrapolating from the empirical data suggests that significant problems will arise at the emerging generation of 10Gb/s and grow worse as 40Gb/s and 100 Gb/s links emerge during The DOE s ESnet, for example, which currently runs at 10Gb/s speeds, is expected to graduate to 40 and 100 Gb/s speeds during RDMA over the WAN Remote Direct Memory Access (RDMA) is a technology that is well established in the HPC space and is rapidly gaining traction in the enterprise space. An easy way to think of RDMA is as a message passing service; it enables applications to exchange messages with each other directly without requiring the involvement of the operating system. While the acronym RDMA originally represented a particular memory access method and programming style, as the technology has emerged in the marketplace the simple acronym has come to represent much more than that. The nuance arises because of the R in RDMA, which refers to remote, meaning outside the local server, and thus implies a networking technology. It turns out that efficiently implementing the RDMA access method over a network depends on a network transport mechanism which is suitable for transporting memory buffers (i.e. messages) 2
3 from one server to another. It is for this reason that the leading implementation of RDMA in the marketplace does not use the TCP transport which is considered to be inefficient for message based transfers. The best-known implementation of RDMA is the InfiniBand Architecture, which is widely recognized in HPC and within data centers for its extremely low end-to-end latencies and its ability to reduce the CPU and memory bandwidth burdens that are common to standard network protocols. These performance characteristics are due largely to the remote memory access method, which is key to avoiding both buffer copies and CPU involvement in the endnodes. But what of the RDMA transport? While well known within the data center, InfiniBand is less well known for its performance over the WAN. However, over the past several years the DoD has been engaged in a demonstration project dubbed Large Data which, among other major breakthroughs, has demonstrated great success in applying RDMA over the WAN for interconnecting widely distributed parallel storage systems. The Large Data system consists of multiple geographically distributed nodes; each node comprises a complete data center including servers for application processing and a large scale parallel file system and accompanying storage. At the heart of each of the nodes is a low latency InfiniBand 40Gb/s network. The geographically distributed nodes are connected to one another via 10Gb/s WAN links to create a fully connected mesh. By executing the RDMA protocol over the WAN links instead of the better known TCP, the RDMA networks at the heart of each of the nodes are effectively merged into a large, high performance RDMA network. Because of this interconnection of the nodes, a user of the system, such as an intelligence analyst, is presented with a view of not only the data stored at his local node, but of all the data stored anywhere on the entire large data system. In effect, the Large Data system is a complete cloud architecture with an effective physical diameter on the order of 10,000 kilometers. Experimental results clearly show the system s ability to transport storage data over the WAN at sustained rates very close to available wire speed just under 10Gb/s. By comparison, when interconnected via conventional TCP/IP protocols running over the same 10Gb/s WAN links, the sustained data rates for more than one flow dropped precipitously to just a fraction of the available WAN wire speed. So what is it about InfiniBand that accounts for its remarkable performance over the WAN? As described above, the bandwidth-delay product associated with the network, and its relationship to the transport s sliding congestion window, are the key determinants in achieving maximum sustained throughput over the WAN. The InfiniBand transport, like TCP, is based on a sliding congestion window. But unlike TCP, IB s congestion window is based on outstanding packets instead of outstanding bytes. With a packet sequence numbering range of more than 8 million packets, and each packet ranging in size from 256 bytes to 4K bytes, it is easy for the InfiniBand transport to completely fill even the longest wire long before its congestion window is filled. Is there a way to capture the features of the InfiniBand transport that make it suitable for the WAN without the requirement for a wholesale upgrade to an InfiniBand network? The answer is yes. RDMA Implementations RDMA over Ethernet RoCE As described so far, wringing the best possible bandwidth utilization from 10Gb/s and faster WAN links is facilitated through the unique features of the InfiniBand transport protocol; in particular, its message orientation coupled with its use of a packet-based sliding congestion window. Up until now, the only 3
4 route to accessing the InfiniBand transport has been to deploy a complete IB network, including the use of WAN gateway adapters to allow the extension of the RDMA protocols over the WAN. Recently, the InfiniBand Trade Association announced the release of a new specification, called RoCE (RDMA over Converged Ethernet). This new specification defines a mechanism for layering InfiniBand s efficient message-based transport directly over an Ethernet layer 2 fabric, replacing the InfiniBand network, link and phy layers. The notion of RoCE was proposed at the OpenFabric Alliance s 2009 Sonoma workshop and publicly discussed by the InfiniBand Trade Association at the October 2009 T11 meeting using the working title, IBXoE. The specification was ratified by the IBTA and released in April 2010 and can be found in the downloads area of the IBTA website. Non-members of the IBTA can register and freely download and use the specification at no cost. The diagram below is a notional representation of the RoCE architecture illustrating how IB s RDMA service is cleanly married with Ethernet s layer 2 link and phy layers. RoCE maps an efficient RDMA transport directly onto a layer 2 Ethernet fabric This specification for the first time permits the use of the proven RDMA transport that is at the heart of the InfiniBand Architecture directly on top of layer 2 Ethernet. This major advancement means that an IT user now has access to the performance and latency advantages associated with the best in RDMA technology, while preserving existing Ethernet cables and switches. Since RoCE conforms to InfiniBand s verbs definitions, it can be easily deployed using the Open Fabrics Enterprise Distribution (OFED) software which is freely available in open source from the Open Fabrics Alliance. There are presently two different, but wholly interoperable implementations of the RoCE specification available. Mellanox Connect-X HCA fully supports RoCE with a hardware-based transport engine. And System Fabric Works has made available through the OFA a software implementation of the transport engine, known as soft RoCE. This implementation, which interoperates with the Mellanox hardware version, is used in conjunction with a standard 10Gb/s Ethernet NIC. Implementing RoCE in the Storage Cloud A storage cloud can be implemented as two or more storage sites, usually combined with application processing capability and connected together with wide area links. Each node is built around the usual 4
5 hierarchy of Ethernet switches, with one or more of the Ethernet switch ports exiting to the WAN. The WAN port is used to keep the cloud nodes synchronized. The figure below describes the basic concept of a two node storage cloud. By taking advantage of RoCE s transport characteristics in the WAN, it should be possible to dramatically increase throughput across the WAN links compared to FTP/TCP and in so doing reduce the performance delta between local storage and remote storage. Ideally, the delta can be reduced to the point that caching techniques can be effective in hiding performance differences. As a distinct side benefit, the use of RoCE as the WAN transport also improves the utilization of the relatively expensive WAN links, thus maximizing the return on investment in the WAN equipment and links. Running the RoCE protocol over the WAN can be effective for a range of different WAN choices, for example SONET or Ethernet WAN PHY are likely choices. In the SONET case, a WAN gateway device is required in order to provide the interface between the asynchronous Ethernet data center network and the synchronous SONET network. For Ethernet WAN PHY, a standard Ethernet switch could support an integrated WAN PHY port potentially eliminating the need for a relatively expensive WAN gateway device. The diagram below illustrates one storage cloud node. Within the node, application and storage server(s) are interconnected via RoCE running over conventional Ethernet switches. As described, the set of nodes comprising the cloud benefits from the advantages of using RoCE as the WAN transport. But more than that, each individual node within the cloud also benefits from the performance advantages that RDMA brings to each server; for example, RDMA s ability to deliver data directly to a user s virtual address space dramatically reduces latency, eliminates buffer copies, reduces the demand on memory bandwidth and reduces CPU utilization considerably. These advantages are the traditional realm of RDMA technology and are independent of the use of RoCE as a WAN transport. Note the illustration of two different types of 10GbE NICs in the diagram below. The heavily loaded storage servers are connected via hard RoCE NICs. These NICs implement the complete RoCE RDMA 5
6 protocol in hardware and provide maximum performance, lowest latency and lowest CPU utilization. The application servers, on the other hand, are assumed to be more lightly loaded and can be connected via soft RoCE. Soft RoCE consists of the RDMA protocol implemented in software coupled together with a conventional 10GbE NIC. Although not required by the RoCE specification, one option is the use of the newly emerging Data Center Bridging variation of Ethernet. This variation provides certain features, such as link level flow control and congestion control mechanisms that are intended to reduce the instance of dropped packets in the Ethernet fabric. The relatively lightly loaded application servers take advantage of embedded 10GbE LOM NICs by deploying soft RoCE. This allows application servers to participate in the RDMA protocol while minimizing the physical changes required in the servers, for example standard blade servers. Heavily loaded storage servers are supported by hard RoCE. The RDMA protocol is implemented in hardware. Summary The emergence of high-speed networks extended over long geographical distances demands a re-look at networking transport protocols. Although many approaches to fine tuning TCP s congestion window mechanism have been proposed (High Speed TCP, H-TCP, BIC and others) each of them represents a fairly sophisticated tuning of the TCP protocol. Although these improvements can yield some performance gains, the improvements are incremental. The opportunity for large scale advances in the well-known TCP transport appears to be greatly receding. Something new is required to transport data over long distances. A by-product of the emergence of practical and effective RDMA technologies is the appearance of a transport protocol which is capable of tolerating high bandwidth-delay product environments. The InfiniBand Trade Association has recently released a specification for RoCE, which, for the first time, 6
7 combines the RDMA transport protocol with Ethernet. This development opens the possibility of applying RDMA technologies, including in the WAN, in environments where it is not practical or desirable to deploy a non-ethernet fabric. The ability to transfer large data blocks or files over long distances at reasonable bandwidths and latencies will propel the trend toward greater data center flexibility. It can do this by making it possible to federate storage over long distances, effectively decoupling the data storage location from the application execution location. In addition, with the advent of fast long distance links, the barriers to migrating entire virtual machines among data centers are substantially lowered. Again, this is another important step in improving the flexibility with which IT resources can be allocated. About System Fabric Works System Fabric Works ( is an Austin Texas based engineering company providing consulting, professional engineering services and products to assist our client in maximizing the value of his investment in applications, hardware and systems. By focusing on the application s needs for data movement and data storage, we are able to assist our client by recommending specific I/O and storage technologies, providing engineering services including software development and assisting with system integration and deployment. We also supply and support a highly customizable range of high performance, low cost storage and networking products. We are worldwide experts in the development and application of RDMA technologies and are widely recognized as leading experts in developing and deploying advanced software such as the Open Fabrics Enterprise Distribution (OFED) software stacks from the Open Fabrics Alliance. 7
Performance Evaluation of the RDMA over Ethernet (RoCE) Standard in Enterprise Data Centers Infrastructure. Abstract:
Performance Evaluation of the RDMA over Ethernet (RoCE) Standard in Enterprise Data Centers Infrastructure Motti Beck Director, Marketing [email protected] Michael Kagan Chief Technology Officer [email protected]
Solving I/O Bottlenecks to Enable Superior Cloud Efficiency
WHITE PAPER Solving I/O Bottlenecks to Enable Superior Cloud Efficiency Overview...1 Mellanox I/O Virtualization Features and Benefits...2 Summary...6 Overview We already have 8 or even 16 cores on one
Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014
Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet Anand Rangaswamy September 2014 Storage Developer Conference Mellanox Overview Ticker: MLNX Leading provider of high-throughput,
3G Converged-NICs A Platform for Server I/O to Converged Networks
White Paper 3G Converged-NICs A Platform for Server I/O to Converged Networks This document helps those responsible for connecting servers to networks achieve network convergence by providing an overview
Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks
WHITE PAPER July 2014 Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks Contents Executive Summary...2 Background...3 InfiniteGraph...3 High Performance
Block based, file-based, combination. Component based, solution based
The Wide Spread Role of 10-Gigabit Ethernet in Storage This paper provides an overview of SAN and NAS storage solutions, highlights the ubiquitous role of 10 Gigabit Ethernet in these solutions, and illustrates
Choosing the Best Network Interface Card for Cloud Mellanox ConnectX -3 Pro EN vs. Intel XL710
COMPETITIVE BRIEF April 5 Choosing the Best Network Interface Card for Cloud Mellanox ConnectX -3 Pro EN vs. Intel XL7 Introduction: How to Choose a Network Interface Card... Comparison: Mellanox ConnectX
RoCE vs. iwarp Competitive Analysis
WHITE PAPER August 21 RoCE vs. iwarp Competitive Analysis Executive Summary...1 RoCE s Advantages over iwarp...1 Performance and Benchmark Examples...3 Best Performance for Virtualization...4 Summary...
Mellanox Academy Online Training (E-learning)
Mellanox Academy Online Training (E-learning) 2013-2014 30 P age Mellanox offers a variety of training methods and learning solutions for instructor-led training classes and remote online learning (e-learning),
From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller
White Paper From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller The focus of this paper is on the emergence of the converged network interface controller
SMB Direct for SQL Server and Private Cloud
SMB Direct for SQL Server and Private Cloud Increased Performance, Higher Scalability and Extreme Resiliency June, 2014 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server
State of the Art Cloud Infrastructure
State of the Art Cloud Infrastructure Motti Beck, Director Enterprise Market Development WHD Global I April 2014 Next Generation Data Centers Require Fast, Smart Interconnect Software Defined Networks
InfiniBand Software and Protocols Enable Seamless Off-the-shelf Applications Deployment
December 2007 InfiniBand Software and Protocols Enable Seamless Off-the-shelf Deployment 1.0 Introduction InfiniBand architecture defines a high-bandwidth, low-latency clustering interconnect that is used
Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 SMB Direct
Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 Direct Increased Performance, Scaling and Resiliency July 2012 Motti Beck, Director, Enterprise Market Development [email protected]
Enabling High performance Big Data platform with RDMA
Enabling High performance Big Data platform with RDMA Tong Liu HPC Advisory Council Oct 7 th, 2014 Shortcomings of Hadoop Administration tooling Performance Reliability SQL support Backup and recovery
Cloud-ready network architecture
IBM Systems and Technology Thought Leadership White Paper May 2011 Cloud-ready network architecture 2 Cloud-ready network architecture Contents 3 High bandwidth with low latency 4 Converged communications
I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology
I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology Reduce I/O cost and power by 40 50% Reduce I/O real estate needs in blade servers through consolidation Maintain
Advancing Applications Performance With InfiniBand
Advancing Applications Performance With InfiniBand Pak Lui, Application Performance Manager September 12, 2013 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server and
Cisco Application Networking for IBM WebSphere
Cisco Application Networking for IBM WebSphere Faster Downloads and Site Navigation, Less Bandwidth and Server Processing, and Greater Availability for Global Deployments What You Will Learn To address
Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks. An Oracle White Paper April 2003
Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks An Oracle White Paper April 2003 Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building
Converging Data Center Applications onto a Single 10Gb/s Ethernet Network
Converging Data Center Applications onto a Single 10Gb/s Ethernet Network Explanation of Ethernet Alliance Demonstration at SC10 Contributing Companies: Amphenol, Broadcom, Brocade, CommScope, Cisco, Dell,
Accelerate Private Clouds with an Optimized Network
Accelerate Private Clouds with an Optimized Network An Allstream White Paper 1 Table of contents The importance of WAN 1 WAN challenges for Private Clouds 1 WAN Optimization methods 2 Benefits of WAN Optimization
High Speed I/O Server Computing with InfiniBand
High Speed I/O Server Computing with InfiniBand José Luís Gonçalves Dep. Informática, Universidade do Minho 4710-057 Braga, Portugal [email protected] Abstract: High-speed server computing heavily relies on
Can High-Performance Interconnects Benefit Memcached and Hadoop?
Can High-Performance Interconnects Benefit Memcached and Hadoop? D. K. Panda and Sayantan Sur Network-Based Computing Laboratory Department of Computer Science and Engineering The Ohio State University,
10 Gigabit Ethernet: Scaling across LAN, MAN, WAN
Arasan Chip Systems Inc. White Paper 10 Gigabit Ethernet: Scaling across LAN, MAN, WAN By Dennis McCarty March 2011 Overview Ethernet is one of the few protocols that has increased its bandwidth, while
Brocade Solution for EMC VSPEX Server Virtualization
Reference Architecture Brocade Solution Blueprint Brocade Solution for EMC VSPEX Server Virtualization Microsoft Hyper-V for 50 & 100 Virtual Machines Enabled by Microsoft Hyper-V, Brocade ICX series switch,
Latency Considerations for 10GBase-T PHYs
Latency Considerations for PHYs Shimon Muller Sun Microsystems, Inc. March 16, 2004 Orlando, FL Outline Introduction Issues and non-issues PHY Latency in The Big Picture Observations Summary and Recommendations
Accelerating High-Speed Networking with Intel I/O Acceleration Technology
White Paper Intel I/O Acceleration Technology Accelerating High-Speed Networking with Intel I/O Acceleration Technology The emergence of multi-gigabit Ethernet allows data centers to adapt to the increasing
Windows TCP Chimney: Network Protocol Offload for Optimal Application Scalability and Manageability
White Paper Windows TCP Chimney: Network Protocol Offload for Optimal Application Scalability and Manageability The new TCP Chimney Offload Architecture from Microsoft enables offload of the TCP protocol
Deploying 10/40G InfiniBand Applications over the WAN
Deploying 10/40G InfiniBand Applications over the WAN Eric Dube ([email protected]) Senior Product Manager of Systems November 2011 Overview About Bay Founded in 2000 to provide high performance
SMB Advanced Networking for Fault Tolerance and Performance. Jose Barreto Principal Program Managers Microsoft Corporation
SMB Advanced Networking for Fault Tolerance and Performance Jose Barreto Principal Program Managers Microsoft Corporation Agenda SMB Remote File Storage for Server Apps SMB Direct (SMB over RDMA) SMB Multichannel
Building a Scalable Storage with InfiniBand
WHITE PAPER Building a Scalable Storage with InfiniBand The Problem...1 Traditional Solutions and their Inherent Problems...2 InfiniBand as a Key Advantage...3 VSA Enables Solutions from a Core Technology...5
InfiniBand in the Enterprise Data Center
InfiniBand in the Enterprise Data Center InfiniBand offers a compelling value proposition to IT managers who value data center agility and lowest total cost of ownership Mellanox Technologies Inc. 2900
networks Live & On-Demand Video Delivery without Interruption Wireless optimization the unsolved mystery WHITE PAPER
Live & On-Demand Video Delivery without Interruption Wireless optimization the unsolved mystery - Improving the way the world connects - WHITE PAPER Live On-Demand Video Streaming without Interruption
TCP Offload Engines. As network interconnect speeds advance to Gigabit. Introduction to
Introduction to TCP Offload Engines By implementing a TCP Offload Engine (TOE) in high-speed computing environments, administrators can help relieve network bottlenecks and improve application performance.
Frequently Asked Questions
Frequently Asked Questions 1. Q: What is the Network Data Tunnel? A: Network Data Tunnel (NDT) is a software-based solution that accelerates data transfer in point-to-point or point-to-multipoint network
Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA
WHITE PAPER April 2014 Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA Executive Summary...1 Background...2 File Systems Architecture...2 Network Architecture...3 IBM BigInsights...5
Choosing the Best Network Interface Card Mellanox ConnectX -3 Pro EN vs. Intel X520
COMPETITIVE BRIEF August 2014 Choosing the Best Network Interface Card Mellanox ConnectX -3 Pro EN vs. Intel X520 Introduction: How to Choose a Network Interface Card...1 Comparison: Mellanox ConnectX
Low Latency 10 GbE Switching for Data Center, Cluster and Storage Interconnect
White PAPER Low Latency 10 GbE Switching for Data Center, Cluster and Storage Interconnect Introduction: High Performance Data Centers As the data center continues to evolve to meet rapidly escalating
Cisco Application Networking for Citrix Presentation Server
Cisco Application Networking for Citrix Presentation Server Faster Site Navigation, Less Bandwidth and Server Processing, and Greater Availability for Global Deployments What You Will Learn To address
A Tour of the Linux OpenFabrics Stack
A Tour of the OpenFabrics Stack Johann George, QLogic June 2006 1 Overview Beyond Sockets Provides a common interface that allows applications to take advantage of the RDMA (Remote Direct Memory Access),
New Features in SANsymphony -V10 Storage Virtualization Software
New Features in SANsymphony -V10 Storage Virtualization Software Updated: May 28, 2014 Contents Introduction... 1 Virtual SAN Configurations (Pooling Direct-attached Storage on hosts)... 1 Scalability
High Performance Data-Transfers in Grid Environment using GridFTP over InfiniBand
High Performance Data-Transfers in Grid Environment using GridFTP over InfiniBand Hari Subramoni *, Ping Lai *, Raj Kettimuthu **, Dhabaleswar. K. (DK) Panda * * Computer Science and Engineering Department
Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck
Sockets vs. RDMA Interface over 1-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Pavan Balaji Hemal V. Shah D. K. Panda Network Based Computing Lab Computer Science and Engineering
Cisco Application Networking for BEA WebLogic
Cisco Application Networking for BEA WebLogic Faster Downloads and Site Navigation, Less Bandwidth and Server Processing, and Greater Availability for Global Deployments What You Will Learn To address
White Paper Solarflare High-Performance Computing (HPC) Applications
Solarflare High-Performance Computing (HPC) Applications 10G Ethernet: Now Ready for Low-Latency HPC Applications Solarflare extends the benefits of its low-latency, high-bandwidth 10GbE server adapters
Building Enterprise-Class Storage Using 40GbE
Building Enterprise-Class Storage Using 40GbE Unified Storage Hardware Solution using T5 Executive Summary This white paper focuses on providing benchmarking results that highlight the Chelsio T5 performance
Intel Data Direct I/O Technology (Intel DDIO): A Primer >
Intel Data Direct I/O Technology (Intel DDIO): A Primer > Technical Brief February 2012 Revision 1.0 Legal Statements INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION WITH INTEL PRODUCTS. NO LICENSE,
ECLIPSE Performance Benchmarks and Profiling. January 2009
ECLIPSE Performance Benchmarks and Profiling January 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox, Schlumberger HPC Advisory Council Cluster
RDMA over Ethernet - A Preliminary Study
RDMA over Ethernet - A Preliminary Study Hari Subramoni, Miao Luo, Ping Lai and Dhabaleswar. K. Panda Computer Science & Engineering Department The Ohio State University Outline Introduction Problem Statement
LS DYNA Performance Benchmarks and Profiling. January 2009
LS DYNA Performance Benchmarks and Profiling January 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox HPC Advisory Council Cluster Center The
Computer Network. Interconnected collection of autonomous computers that are able to exchange information
Introduction Computer Network. Interconnected collection of autonomous computers that are able to exchange information No master/slave relationship between the computers in the network Data Communications.
Chapter 1 Reading Organizer
Chapter 1 Reading Organizer After completion of this chapter, you should be able to: Describe convergence of data, voice and video in the context of switched networks Describe a switched network in a small
Introduction to Cloud Design Four Design Principals For IaaS
WHITE PAPER Introduction to Cloud Design Four Design Principals For IaaS What is a Cloud...1 Why Mellanox for the Cloud...2 Design Considerations in Building an IaaS Cloud...2 Summary...4 What is a Cloud
Simplifying Big Data Deployments in Cloud Environments with Mellanox Interconnects and QualiSystems Orchestration Solutions
Simplifying Big Data Deployments in Cloud Environments with Mellanox Interconnects and QualiSystems Orchestration Solutions 64% of organizations were investing or planning to invest on Big Data technology
TCP Servers: Offloading TCP Processing in Internet Servers. Design, Implementation, and Performance
TCP Servers: Offloading TCP Processing in Internet Servers. Design, Implementation, and Performance M. Rangarajan, A. Bohra, K. Banerjee, E.V. Carrera, R. Bianchini, L. Iftode, W. Zwaenepoel. Presented
Ethernet: THE Converged Network Ethernet Alliance Demonstration as SC 09
Ethernet: THE Converged Network Ethernet Alliance Demonstration as SC 09 Authors: Amphenol, Cisco, Dell, Fulcrum Microsystems, Intel, Ixia, JDSU, Mellanox, NetApp, Panduit, QLogic, Spirent, Tyco Electronics,
10GBASE T for Broad 10_Gigabit Adoption in the Data Center
10GBASE T for Broad 10_Gigabit Adoption in the Data Center Contributors Carl G. Hansen, Intel Carrie Higbie, Siemon Yinglin (Frank) Yang, Commscope, Inc 1 Table of Contents 10Gigabit Ethernet: Drivers
PCI Express Overview. And, by the way, they need to do it in less time.
PCI Express Overview Introduction This paper is intended to introduce design engineers, system architects and business managers to the PCI Express protocol and how this interconnect technology fits into
- An Essential Building Block for Stable and Reliable Compute Clusters
Ferdinand Geier ParTec Cluster Competence Center GmbH, V. 1.4, March 2005 Cluster Middleware - An Essential Building Block for Stable and Reliable Compute Clusters Contents: Compute Clusters a Real Alternative
High Throughput File Servers with SMB Direct, Using the 3 Flavors of RDMA network adapters
High Throughput File Servers with SMB Direct, Using the 3 Flavors of network adapters Jose Barreto Principal Program Manager Microsoft Corporation Abstract In Windows Server 2012, we introduce the SMB
Applications. Network Application Performance Analysis. Laboratory. Objective. Overview
Laboratory 12 Applications Network Application Performance Analysis Objective The objective of this lab is to analyze the performance of an Internet application protocol and its relation to the underlying
Long-Haul System Family. Highest Levels of RDMA Scalability, Simplified Distance Networks Manageability, Maximum System Productivity
Long-Haul System Family Highest Levels of RDMA Scalability, Simplified Distance Networks Manageability, Maximum System Productivity Mellanox continues its leadership by providing RDMA Long-Haul Systems
The proliferation of the raw processing
TECHNOLOGY CONNECTED Advances with System Area Network Speeds Data Transfer between Servers with A new network switch technology is targeted to answer the phenomenal demands on intercommunication transfer
Performance Evaluation of Linux Bridge
Performance Evaluation of Linux Bridge James T. Yu School of Computer Science, Telecommunications, and Information System (CTI) DePaul University ABSTRACT This paper studies a unique network feature, Ethernet
Connecting the Clouds
Connecting the Clouds Mellanox Connected Clouds Mellanox s Ethernet and InfiniBand interconnects enable and enhance worldleading cloud infrastructures around the globe. Utilizing Mellanox s fast server
Overview of Requirements and Applications for 40 Gigabit and 100 Gigabit Ethernet
Overview of Requirements and Applications for 40 Gigabit and 100 Gigabit Ethernet Version 1.1 June 2010 Authors: Mark Nowell, Cisco Vijay Vusirikala, Infinera Robert Hays, Intel 1. This work represents
Global Headquarters: 5 Speen Street Framingham, MA 01701 USA P.508.872.8200 F.508.935.4015 www.idc.com
Global Headquarters: 5 Speen Street Framingham, MA 01701 USA P.508.872.8200 F.508.935.4015 www.idc.com W H I T E P A P E R O r a c l e V i r t u a l N e t w o r k i n g D e l i v e r i n g F a b r i c
High Performance OpenStack Cloud. Eli Karpilovski Cloud Advisory Council Chairman
High Performance OpenStack Cloud Eli Karpilovski Cloud Advisory Council Chairman Cloud Advisory Council Our Mission Development of next generation cloud architecture Providing open specification for cloud
Intel Ethernet Switch Load Balancing System Design Using Advanced Features in Intel Ethernet Switch Family
Intel Ethernet Switch Load Balancing System Design Using Advanced Features in Intel Ethernet Switch Family White Paper June, 2008 Legal INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION WITH INTEL
Campus Network Design Science DMZ
Campus Network Design Science DMZ Dale Smith Network Startup Resource Center [email protected] The information in this document comes largely from work done by ESnet, the USA Energy Sciences Network see
Lecture 4: Introduction to Computer Network Design
Lecture 4: Introduction to Computer Design Prof. Shervin Shirmohammadi SITE, University of Ottawa Prof. Shervin Shirmohammadi CEG 4185 4-1 Computer s Prof. Shervin Shirmohammadi CEG 4185 4-2 1 Background
Enabling the Use of Data
Enabling the Use of Data Michael Kagan, CTO June 1, 2015 - Technion Computer Engineering Conference Safe Harbor Statement These slides and the accompanying oral presentation contain forward-looking statements
WAN Performance Analysis A Study on the Impact of Windows 7
A Talari Networks White Paper WAN Performance Analysis A Study on the Impact of Windows 7 Test results demonstrating WAN performance changes due to upgrading to Windows 7 and the network architecture and
HyperQ Remote Office White Paper
HyperQ Remote Office White Paper Parsec Labs, LLC. 7101 Northland Circle North, Suite 105 Brooklyn Park, MN 55428 USA 1-763-219-8811 www.parseclabs.com [email protected] [email protected] Introduction
InfiniBand Switch System Family. Highest Levels of Scalability, Simplified Network Manageability, Maximum System Productivity
InfiniBand Switch System Family Highest Levels of Scalability, Simplified Network Manageability, Maximum System Productivity Mellanox continues its leadership by providing InfiniBand SDN Switch Systems
IEEE Congestion Management Presentation for IEEE Congestion Management Study Group
IEEE Congestion Management Presentation for IEEE Congestion Management Study Group Contributors Jeff Lynch IBM Gopal Hegde -- Intel 2 Outline Problem Statement Types of Traffic & Typical Usage Models Traffic
Question: 3 When using Application Intelligence, Server Time may be defined as.
1 Network General - 1T6-521 Application Performance Analysis and Troubleshooting Question: 1 One component in an application turn is. A. Server response time B. Network process time C. Application response
Quantum StorNext. Product Brief: Distributed LAN Client
Quantum StorNext Product Brief: Distributed LAN Client NOTICE This product brief may contain proprietary information protected by copyright. Information in this product brief is subject to change without
Upgrading Data Center Network Architecture to 10 Gigabit Ethernet
Intel IT IT Best Practices Data Centers January 2011 Upgrading Data Center Network Architecture to 10 Gigabit Ethernet Executive Overview Upgrading our network architecture will optimize our data center
Replacing SAN with High Performance Windows Share over a Converged Network
WHITE PAPER November 2015 Replacing SAN with High Performance Windows Share over a Converged Network Abstract...1 Introduction...1 Early FC SAN (Storage Area Network)...1 FC vs. Ethernet...1 Changing SAN
Introduction to InfiniBand for End Users Industry-Standard Value and Performance for High Performance Computing and the Enterprise
Introduction to InfiniBand for End Users Industry-Standard Value and Performance for High Performance Computing and the Enterprise Paul Grun InfiniBand Trade Association INTRO TO INFINIBAND FOR END USERS
OFA Training Program. Writing Application Programs for RDMA using OFA Software. Author: Rupert Dance Date: 11/15/2011. www.openfabrics.
OFA Training Program Writing Application Programs for RDMA using OFA Software Author: Rupert Dance Date: 11/15/2011 www.openfabrics.org 1 Agenda OFA Training Program Program Goals Instructors Programming
Mike Canney Principal Network Analyst getpackets.com
Mike Canney Principal Network Analyst getpackets.com 1 My contact info contact Mike Canney, Principal Network Analyst, getpackets.com [email protected] 319.389.1137 2 Capture Strategies capture Capture
RAMCloud and the Low- Latency Datacenter. John Ousterhout Stanford University
RAMCloud and the Low- Latency Datacenter John Ousterhout Stanford University Most important driver for innovation in computer systems: Rise of the datacenter Phase 1: large scale Phase 2: low latency Introduction
Storage, Cloud, Web 2.0, Big Data Driving Growth
Storage, Cloud, Web 2.0, Big Data Driving Growth Kevin Deierling Vice President of Marketing October 25, 2013 Delivering the Highest ROI Across all Markets HPC Web 2.0 DB/Enterprise Cloud Financial Services
Cloud Computing and the Internet. Conferenza GARR 2010
Cloud Computing and the Internet Conferenza GARR 2010 Cloud Computing The current buzzword ;-) Your computing is in the cloud! Provide computing as a utility Similar to Electricity, Water, Phone service,
Deploying in a Distributed Environment
Deploying in a Distributed Environment Distributed enterprise networks have many remote locations, ranging from dozens to thousands of small offices. Typically, between 5 and 50 employees work at each
Network Attached Storage. Jinfeng Yang Oct/19/2015
Network Attached Storage Jinfeng Yang Oct/19/2015 Outline Part A 1. What is the Network Attached Storage (NAS)? 2. What are the applications of NAS? 3. The benefits of NAS. 4. NAS s performance (Reliability
IP SAN Best Practices
IP SAN Best Practices A Dell Technical White Paper PowerVault MD3200i Storage Arrays THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS AND TECHNICAL INACCURACIES.
Deploying Silver Peak VXOA with EMC Isilon SyncIQ. February 2012. www.silver-peak.com
Deploying Silver Peak VXOA with EMC Isilon SyncIQ February 2012 www.silver-peak.com Table of Contents Table of Contents Overview... 3 Solution Components... 3 EMC Isilon...3 Isilon SyncIQ... 3 Silver Peak
Deploying Riverbed wide-area data services in a LeftHand iscsi SAN Remote Disaster Recovery Solution
Wide-area data services (WDS) Accelerating Remote Disaster Recovery Reduce Replication Windows and transfer times leveraging your existing WAN Deploying Riverbed wide-area data services in a LeftHand iscsi
Computer Organization & Architecture Lecture #19
Computer Organization & Architecture Lecture #19 Input/Output The computer system s I/O architecture is its interface to the outside world. This architecture is designed to provide a systematic means of
Video Streaming Without Interruption
Video Streaming Without Interruption Adaptive bitrate and content delivery networks: Are they enough to achieve high quality, uninterrupted Internet video streaming? WHITE PAPER Abstract The increasing
Application Note. Windows 2000/XP TCP Tuning for High Bandwidth Networks. mguard smart mguard PCI mguard blade
Application Note Windows 2000/XP TCP Tuning for High Bandwidth Networks mguard smart mguard PCI mguard blade mguard industrial mguard delta Innominate Security Technologies AG Albert-Einstein-Str. 14 12489
