A Low-Latency Solution for High- Frequency Trading from IBM and Mellanox
|
|
|
- Charity Richard
- 10 years ago
- Views:
Transcription
1 Effective low-cost High-Frequency Trading solutions from IBM and Mellanox May 2011 A Low-Latency Solution for High- Frequency Trading from IBM and Mellanox Vinit Jain IBM Systems and Technology Group Falke Bruinsma IBM Software Group
2 Page 2 Executive Overview In the world of High-Frequency Trading (HFT), opportunities exist only fleetingly and therefore trading solutions must run at the lowest latency to be competitive. Low-latency 10 Gigabit Ethernet has become the interconnect of choice for HFT solutions. IBM and Mellanox have demonstrated a solution that performs at high throughput rates and low latency to facilitate High- Frequency Trading solutions. This solution uses the IBM BNT /40Gb Ethernet Switch, coupled with IBM's WebSphere MQ Low Latency Messaging (WMQLLM) software using the Mellanox ConnectX Gigabit Ethernet adapter with RoCE (RDMA over Converged Ethernet) drivers. This solution delivers a powerful combination of networking hardware and messaging software that meets the latency and throughput requirements of high-frequency trading environments. This has been demonstrated through independently audited benchmarks published using the STAC-M2 Benchmark test. In addition to the STAC-M2 Benchmark test, further testing was performed by IBM to explore the latency and throughput performance that this solution delivers. The results show that the average latency of this solution does not exceed 5 µsec, with standard deviation remaining below 2 µsec. These results are obtained using a typical message size of 128 bytes and at message rates of 1 million messages / second. High-Frequency Trading HFT has gained a strong foothold in financial markets, driven by several factors, including advances in information technology that have been conducive to its growth. Unlike traditional traders who hold their positions long term, high-frequency traders hold their positions for shorter durations, which can be as little as a few seconds. They typically end their day with few to no positions carried over to the next business day. HFT uses strategies such as statistical arbitrage and liquidity detection. Statistical arbitrage relies on the principle that stocks in a pair or pool of correlated stocks that diverge from their statistically expected behavior will converge and profit is achieved by taking positions based on this expectation. Liquidity detection is the strategy by which small trades are sent out to detect large orders that are not visible and then taking positions based on the expectation that the large orders will move the market. These strategies require programs to analyze massive amounts of market data using complex algorithms to exploit opportunities that exist for as little as a fraction of a second. The execution of these strategies requires information technology that can compute complex algorithms, and exchange messages at extremely low latencies even at very high rates to handle volume spikes without impacting system performance. In fact, it is at exactly those times when the volume spikes that HFT systems may take advantage of delays in other system. Therefore IT organizations in the financial services industry face tremendous pressures to optimize the transaction lifecycle. There is a critical need for the underlying messaging infrastructures to deliver extremely low latency and very high message throughputs. Solution IT organizations are expected to deliver solutions offering low latency with high throughput without using specialized technology, to avoid high cost for capital and skills. The trend has been to favor solutions that use commodity hardware and software components. As a result, lowlatency 10 Gigabit Ethernet has become the interconnect of choice.
3 Page 3 IBM and Mellanox have demonstrated an ultralow-latency messaging solution that performs at high throughput rates with reliability. This solution stack, shown in Figure 1, addresses the requirements of the financial industry and delivers the solution using commodity hardware and software components. IBM WMQLLM 2.6 Red Hat Enterprise Linux 5.5 (64-bit) 2 x Intel Xeon (X5570) 4-core 2.93GHz processors IBM System x3550 M2, model 7946-E2U Mellanox MNPH29C-XTR ConnectX-2 EN with RoCE IBM BNT G /40Gb Ethernet Switch Figure 1. Solution Stack Components This solution uses the IBM BNT Gigabit Ethernet switch, coupled with IBM's WMQLLM software using the Mellanox ConnectX-2 10 Gigabit Ethernet adapter with RoCE drivers. This solution delivers a powerful combination of networking hardware and messaging software that meets the latency and throughput requirements of high-frequency trading environments. This has been demonstrated through independently audited benchmarks published using the STAC-M2 Benchmark test. In this paper we present additional test results demonstrating the latency characteristics of this solution. IBM BNT RackSwitch G /40 GbE Switch The IBM BNT RackSwitch G8264 is high-performance switch designed to meet the demanding requirements of high-frequency trading systems. It provides line-rate, high-bandwidth switching, filtering, and traffic forwarding, without delaying data. This switch offers up to 64 10GbE ports and up to 4 40GbE ports, 1.2 Terabits per second of non-blocking bidirectional throughput in a 1U footprint. In addition to a rich set of Layer-2 and Layer-3 connectivity, the G8264 supports the newest protocols, including Data Center Bridging / Converged Enhanced Ethernet (DCB/CEE) for support of Fibre Channel over Ethernet (FCoE). Redundant power and fans, along with numerous high-availability features, enable the RackSwitch G8264 to always be available for businesscritical traffic. The single-chip design and the default cut-through mode are key to enabling extremely low deterministic latency and jitter. Large data-center-grade buffers enable congestion free operation. Furthermore, the G8264 delivers best-of-breed performance and function, including Layer-3 with a standard 40GbE interconnect into the core, rather than taking the approach of a proprietary core interconnection, as chosen by some Ethernet vendors. IBM Websphere MQ Low Latency Messaging WebSphere MQ Low Latency Messaging is a transport fabric product engineered for the rigorous latency and throughput requirements typical of today s financial trading environments. The product is daemonless and provides peer-to-peer transport for one-to-one, one-to-many and many-to-many data exchange. It also exploits the IP multicast infrastructure to enable scalable resource conservation and timely information distribution.
4 Page 4 Designed to dramatically improve throughput and reduce latency while maximizing system reliability, WMQLLM can help high-frequency trading organizations enhance the responsiveness of their existing trade infrastructure while developing new solutions for emerging business opportunities. Several factors contribute to the high performance enabled by WMQLLM. For example, a unique method of message packetization enables delay-free, high-speed data delivery. Unique batching technology dynamically optimizes packetization for reliable delivery and lowest latency, based on throughput, message sizes, receiver, and system feedback. In addition, very compact packet headers leave more network bandwidth for application data. WMQLLM supports high performance interconnects, such as 10 Gigabit Ethernet and InfiniBand to enable higher throughput with lower latency, reduced latency variability, and low CPU utilization. Mellanox ConnectX-2 10GbE Server Adapter with RoCE Mellanox ConnectX-2 EN Ethernet Network Interface Cards (NICs) deliver low latency, high throughput and low CPU utilization leveraging the RoCE standard. RoCE is based on the IBTA RoCE specifications, and utilizes the Open Fabrics Enterprise Distribution (OFED) verbs interface as the software interface between application layer and ConnectX-2 EN hardware. RoCE takes advantage of transport services support of various modes of communication, such as reliable connected services and datagram services. RoCE uses well-defined verb operations, including kernel bypass, send/receive semantics, RDMA read/write, user-level multicast, user-level I/O access, zero copy and atomic operations. The ConnectX-2 EN adapters with RoCE are widely used in financial services for removing I/O bottlenecks, lowering latency and jitter, and increasing message rates for high-frequency trading, market data distribution and real-time risk management. Performance Testing STAC-M2 Benchmark Test The STAC-M2 Benchmark specifications test the ability of a solution to handle real-time market data in a variety of configurations found in typical trading environments. The specifications are defined by end-user IT executives within the financial industry with input from vendors of highperformance messaging solutions. The STAC-M2 Benchmarks provide key performance metrics such as latency, throughput, power efficiency, and CPU/memory consumption under several scenarios, including both undisturbed flow and exception conditions like slow consumers. STAC Report Highlights are made available at the STAC website ( for systems that have been independently audited by STAC. The full STAC Report is available in the STAC Vault to STAC s premium end-user subscribers. IBM and Mellanox completed an audited STAC-M2 Benchmark using the solution described in this paper. The latency results below are from three of the 10 test sequences that were carried out. The tests we describe the results for here are: BASELINE Each of five consumers has a unique watchlist, with one consumer per motherboard. BASELINE emulates applications such as a smart order router that partitions symbols across servers. OVERLAP This is similar to the BASELINE test except for some overlap in the consumer watchlists. OVERLAP emulates deployments such as multiple black-box applications that are independent of one another. FLEXIBLE Emulates a general enterprise deployment, where costs matter more than latency. FLEXIBLE requires 15 consumer applications, with no restriction on the number of consumers per motherboard and with some overlap in the consumer watchlists.
5 Page 5 STAC-M2 Latency IBM BNT G /40 GbE and Mellanox ConnectX-2 with RoCE (SUT ID LLM110421) Test Description (Spec ID) SupplyToReceive Latency (Hybrid) at base rate in the 1:5 setup with no watchlist overlap (STAC.M2.v1.0.BASELINE.LAT1) SupplyToReceive Latency (Hybrid) at base rate in the 1:5 setup with some watchlist overlap (STAC.M2.v1.0.OVERLAP.LAT1) SupplyToReceive Latency (Hybrid) at base rate in the setup with flexible Consumer resources (STAC.M2.v1.0.FLEXIBLE.LAT1) Mean (µsec) Med (µsec) 99P (µsec) STDV (µsec) Table 1. STAC-M2 Latency Results This solution exhibited the best mean latencies ever published for the foregoing benchmarks in a STAC-M2 Report. Single-Hop Latency Test In addition to the STAC-M2 Benchmark test, further testing was performed by IBM to explore the performance such a solution can deliver in terms of latency. As shown in Figure 2, the test is a reflector test and the setup consists of two machines, A and B, connected through an IBM BNT G8264 RackSwitch. System A Sender / Receiver WMQLLM Transmitter WMQLLM Receiver G GbE WMQLLM Transmitter / Receiver System B Reflector Figure 2. Reflector Test Layout On System A, the sender sends packets at the rate being tested to the reflector on System B. The reflector receives every packet, and only forwards to the receiver on System A those packets that have a time stamp for latency measurement. The receiver on System A extracts the time stamps from the reflected packet and uses it to measure round-trip time. The single-hop
6 Page 6 latency is calculated as half of the round trip time. The standard deviation is calculated using the round trip time. All latency tests ran for 5 minutes. Approximately 300,000 latency samples were recorded for each 5-minute test. From these 300,000 samples latency statistics were calculated. Two test parameters were used to vary the workload for this testing: message size and message rate. Table 2 shows the results for each message rate and size test combination. LLM Latency using IBM BNT G /40 GbE and Mellanox ConnectX-2 with RoCE Msg Rate [msgs/sec] Msg Size [bytes] Average [µsec] Single Hop Median [µsec] 99P [µsec] RTT Std Dev [µsec] 10, , ,000, , , ,000, , , ,000, Table 2. Single-hop latency Key Takeaway The average latency of this solution remains in the extremely low range of µsec with standard deviation remaining less than 2 µsec, even as message sizes grow large and at very high rates. Conclusions These results clearly show that a messaging solution stack created using IBM s WMQ Low Latency Messaging, IBM BNT G GbE switch, and Mellanox s ConnectX-2 EN with RoCE adapters, delivers the technology that High-Frequency Trading applications need. This result has been audited by STAC and conforms to the STAC M2 benchmark. Defined by trading firms, the STAC M2 benchmark represents typical requirements of a trading system infrastructure. It measures latency and throughput in both optimal and failure scenarios. The results of the STAC audit and subsequent tests by IBM confirm that the components described in this paper produce the lowest latency solution while scaling to very high message rates.
7 Page 7 References High Performance Business Computing in Financial Institutions, Sue Gouws Korn, CFA, Christopher G. Willard, Ph.D, Addison Snell, February 2011 High-frequency trading, Deutsche Bank Research, February 7, 2011 STAC Report Highlights: IBM WMQ LLM with IBM x3550 servers, Mellanox ConnectX-2 EN and IBM BNT G /40Gb Ethernet Switch, May 2011 Develop high-volume, low-latency finance solutions with IBM WebSphere MQ Low Latency Messaging, Financial industry solutions White paper, IBM, October 2009
8 Page 8 For More Information IBM System Networking IBM System x Servers IBM Systems Director Service and Support Manager IBM System x and BladeCenter Power Configurator IBM Standalone Solutions Configuration Tool IBM Configuration and Options Guide IBM ServerProven Program Technical Support Other Technical Support Resources Legal Information IBM Corporation 2011 IBM Systems and Technology Group Dept. USA 3039 Cornwallis Road Research Triangle Park, NC Produced in the USA May 2011 All rights reserved. For a copy of applicable product warranties, write to: Warranty Information, P.O. Box 12195, RTP, NC 27709, Attn: Dept. JDJA/B203. IBM makes no representation or warranty regarding third-party products or services including those designated as ServerProven or ClusterProven. Telephone support may be subject to additional charges. For onsite labor, IBM will attempt to diagnose and resolve the problem remotely before sending a technician. IBM, the IBM logo, ibm.com, BNT, ClusterProven, RackSwitch, ServerProven, and WebSphere, are trademarks of IBM Corporation in the United States and/or other countries. If these and other IBM trademarked terms are marked on their first occurrence in this information with a trademark symbol ( or ), these symbols indicate U.S. registered or common law trademarks owned by IBM at the time this information was published. Such trademarks may also be registered or common law trademarks in other countries. For a list of additional IBM trademarks, please see InfiniBand is a trademark of InfiniBand Trade Association. Intel, the Intel logo, and Xeon are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United States and other countries. Linux is a registered trademark of Linus Torvalds. Mellanox and ConnectX-2 are trademarks or registered trademarks of Mellanox. STAC and all STAC names are trademarks or registered trademarks of the Securities Technology Analysis Center, LLC. Other company, product and service names may be trademarks or service marks of others IBM reserves the right to change specifications or other product information without notice. References in this publication to IBM products or services do not imply that IBM intends to make them available in all countries in which IBM operates. IBM PROVIDES THIS PUBLICATION AS IS WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. Some jurisdictions do not allow disclaimer of express or implied warranties in certain transactions; therefore, this statement may not apply to you. This publication may contain links to third party sites that are not under the control of or maintained by IBM. Access to any such third party site is at the user's own risk and IBM is not responsible for the accuracy or reliability of any information, data, opinions, advice or statements made on these sites. IBM provides these links merely as a convenience and the inclusion of such links does not imply an endorsement. Information in this presentation concerning non-ibm products was obtained from the suppliers of these products, published announcement material or other publicly available sources. IBM has not tested these products and cannot confirm the accuracy of performance, compatibility or any other claims related to non-ibm products. Questions on the capabilities of non-ibm products should be addressed to the suppliers of those products. MB, GB and TB = 1,000,000, 1,000,000,000 and 1,000,000,000,000 bytes, respectively, when referring to storage capacity. Accessible capacity is less; up to 3GB is used in service partition. Actual storage capacity will vary based upon many factors and may be less than stated. Performance is in Internal Throughput Rate (ITR) ratio based on measurements and projections using standard IBM benchmarks in a controlled environment. The actual throughput that any user will experience will depend on considerations such as the amount of multiprogramming in the user s job stream, the I/O configuration, the storage configuration and the workload processed. Therefore, no assurance can be given that an individual user will achieve throughput improvements equivalent to the performance ratios stated here. Maximum internal hard disk and memory capacities may require the replacement of any standard hard drives and/or memory and the population of all hard disk bays and memory slots with the largest currently supported drives available. When referring to variable speed CD-ROMs, CD-Rs, CD-RWs and DVDs, actual playback speed will vary and is often less than the maximum possible. QCW03013-USEN-00
IBM BladeCenter H with Cisco VFrame Software A Comparison with HP Virtual Connect
IBM BladeCenter H with Cisco VFrame Software A Comparison with HP Connect Executive Overview This white paper describes how Cisco VFrame Server Fabric ization Software works with IBM BladeCenter H to provide
Comparison of NAND Flash Technologies Used in Solid- State Storage
An explanation and comparison of SLC and MLC NAND technologies August 2010 Comparison of NAND Flash Technologies Used in Solid- State Storage By Shaluka Perera IBM Systems and Technology Group Bill Bornstein
Performance Evaluation of the RDMA over Ethernet (RoCE) Standard in Enterprise Data Centers Infrastructure. Abstract:
Performance Evaluation of the RDMA over Ethernet (RoCE) Standard in Enterprise Data Centers Infrastructure Motti Beck Director, Marketing [email protected] Michael Kagan Chief Technology Officer [email protected]
Virtualization Performance Analysis November 2010 Effect of SR-IOV Support in Red Hat KVM on Network Performance in Virtualized Environments
Virtualization Performance Analysis November 2010 Effect of SR-IOV Support in Red Hat KVM on Network Performance in Steve Worley System x Performance Analysis and Benchmarking IBM Systems and Technology
WAN Optimization For Software Defined Networks
Solution Brief IBM Systems and Technology Group WAN Optimization For Software Defined Networks Silver Peak Velocity Replication Acceleration Appliance and IBM Software Defined Network for Virtual Environments
Advanced application delivery over software defined networks
Solution Brief IBM Systems and Technology Group Advanced application delivery over software defined networks Citrix NetScaler Application Delivery Controller with IBM Software Defined Network for Virtual
Running Microsoft SQL Server 2012 on a Scale-Out File Server Cluster via SMB Direct Connection Solution Utilizing IBM System x Servers
Highly Available Scale-Out File Server on IBM System x3650 M4 November 2012 Running Microsoft SQL Server 2012 on a Scale-Out File Server Cluster via SMB Direct Connection Solution Utilizing IBM System
The MAX5 Advantage: Clients Benefit running Microsoft SQL Server Data Warehouse (Workloads) on IBM BladeCenter HX5 with IBM MAX5.
Performance benefit of MAX5 for databases The MAX5 Advantage: Clients Benefit running Microsoft SQL Server Data Warehouse (Workloads) on IBM BladeCenter HX5 with IBM MAX5 Vinay Kulkarni Kent Swalin IBM
Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering
Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays Red Hat Performance Engineering Version 1.0 August 2013 1801 Varsity Drive Raleigh NC
RoCE vs. iwarp Competitive Analysis
WHITE PAPER August 21 RoCE vs. iwarp Competitive Analysis Executive Summary...1 RoCE s Advantages over iwarp...1 Performance and Benchmark Examples...3 Best Performance for Virtualization...4 Summary...
Oracle Database Scalability in VMware ESX VMware ESX 3.5
Performance Study Oracle Database Scalability in VMware ESX VMware ESX 3.5 Database applications running on individual physical servers represent a large consolidation opportunity. However enterprises
Performance and scalability of a large OLTP workload
Performance and scalability of a large OLTP workload ii Performance and scalability of a large OLTP workload Contents Performance and scalability of a large OLTP workload with DB2 9 for System z on Linux..............
z/os V1R11 Communications Server system management and monitoring
IBM Software Group Enterprise Networking Solutions z/os V1R11 Communications Server z/os V1R11 Communications Server system management and monitoring z/os Communications Server Development, Raleigh, North
Introduction to PCI Express Positioning Information
Introduction to PCI Express Positioning Information Main PCI Express is the latest development in PCI to support adapters and devices. The technology is aimed at multiple market segments, meaning that
Evaluation Report: Emulex OCe14102 10GbE and OCe14401 40GbE Adapter Comparison with Intel X710 10GbE and XL710 40GbE Adapters
Evaluation Report: Emulex OCe14102 10GbE and OCe14401 40GbE Adapter Comparison with Intel X710 10GbE and XL710 40GbE Adapters Evaluation report prepared under contract with Emulex Executive Summary As
10Gb Ethernet: The Foundation for Low-Latency, Real-Time Financial Services Applications and Other, Latency-Sensitive Applications
10Gb Ethernet: The Foundation for Low-Latency, Real-Time Financial Services Applications and Other, Latency-Sensitive Applications Testing conducted by Solarflare and Arista Networks reveals single-digit
Choosing the Best Network Interface Card for Cloud Mellanox ConnectX -3 Pro EN vs. Intel XL710
COMPETITIVE BRIEF April 5 Choosing the Best Network Interface Card for Cloud Mellanox ConnectX -3 Pro EN vs. Intel XL7 Introduction: How to Choose a Network Interface Card... Comparison: Mellanox ConnectX
Mellanox Academy Online Training (E-learning)
Mellanox Academy Online Training (E-learning) 2013-2014 30 P age Mellanox offers a variety of training methods and learning solutions for instructor-led training classes and remote online learning (e-learning),
Accelerating High-Speed Networking with Intel I/O Acceleration Technology
White Paper Intel I/O Acceleration Technology Accelerating High-Speed Networking with Intel I/O Acceleration Technology The emergence of multi-gigabit Ethernet allows data centers to adapt to the increasing
InfiniBand Software and Protocols Enable Seamless Off-the-shelf Applications Deployment
December 2007 InfiniBand Software and Protocols Enable Seamless Off-the-shelf Deployment 1.0 Introduction InfiniBand architecture defines a high-bandwidth, low-latency clustering interconnect that is used
10G Ethernet: The Foundation for Low-Latency, Real-Time Financial Services Applications and Other, Future Cloud Applications
10G Ethernet: The Foundation for Low-Latency, Real-Time Financial Services Applications and Other, Future Cloud Applications Testing conducted by Solarflare Communications and Arista Networks shows that
Intel Ethernet Switch Load Balancing System Design Using Advanced Features in Intel Ethernet Switch Family
Intel Ethernet Switch Load Balancing System Design Using Advanced Features in Intel Ethernet Switch Family White Paper June, 2008 Legal INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION WITH INTEL
Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks
WHITE PAPER July 2014 Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks Contents Executive Summary...2 Background...3 InfiniteGraph...3 High Performance
From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller
White Paper From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller The focus of this paper is on the emergence of the converged network interface controller
Solving I/O Bottlenecks to Enable Superior Cloud Efficiency
WHITE PAPER Solving I/O Bottlenecks to Enable Superior Cloud Efficiency Overview...1 Mellanox I/O Virtualization Features and Benefits...2 Summary...6 Overview We already have 8 or even 16 cores on one
Intel Data Direct I/O Technology (Intel DDIO): A Primer >
Intel Data Direct I/O Technology (Intel DDIO): A Primer > Technical Brief February 2012 Revision 1.0 Legal Statements INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION WITH INTEL PRODUCTS. NO LICENSE,
Storage at a Distance; Using RoCE as a WAN Transport
Storage at a Distance; Using RoCE as a WAN Transport Paul Grun Chief Scientist, System Fabric Works, Inc. (503) 620-8757 [email protected] Why Storage at a Distance the Storage Cloud Following
Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging
Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging In some markets and scenarios where competitive advantage is all about speed, speed is measured in micro- and even nano-seconds.
QLogic 16Gb Gen 5 Fibre Channel in IBM System x Deployments
QLogic 16Gb Gen 5 Fibre Channel in IBM System x Deployments Increase Virtualization Density and Eliminate I/O Bottlenecks with QLogic High-Speed Interconnects Key Findings Support for increased workloads,
EMC Unified Storage for Microsoft SQL Server 2008
EMC Unified Storage for Microsoft SQL Server 2008 Enabled by EMC CLARiiON and EMC FAST Cache Reference Copyright 2010 EMC Corporation. All rights reserved. Published October, 2010 EMC believes the information
IBM System x and BladeCenter for Oracle s JD Edwards EnterpriseOne
IBM System x and BladeCenter for Oracle s JD Edwards EnterpriseOne Help lower your total cost of ownership while adding flexibility to help meet unforeseen requirements Corning Data Services, Inc. Experience
Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build 164009
Performance Study Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build 164009 Introduction With more and more mission critical networking intensive workloads being virtualized
Cisco for SAP HANA Scale-Out Solution on Cisco UCS with NetApp Storage
Cisco for SAP HANA Scale-Out Solution Solution Brief December 2014 With Intelligent Intel Xeon Processors Highlights Scale SAP HANA on Demand Scale-out capabilities, combined with high-performance NetApp
White Paper Solarflare High-Performance Computing (HPC) Applications
Solarflare High-Performance Computing (HPC) Applications 10G Ethernet: Now Ready for Low-Latency HPC Applications Solarflare extends the benefits of its low-latency, high-bandwidth 10GbE server adapters
FLOW-3D Performance Benchmark and Profiling. September 2012
FLOW-3D Performance Benchmark and Profiling September 2012 Note The following research was performed under the HPC Advisory Council activities Participating vendors: FLOW-3D, Dell, Intel, Mellanox Compute
Ultra-Low Latency, High Density 48 port Switch and Adapter Testing
Ultra-Low Latency, High Density 48 port Switch and Adapter Testing Testing conducted by Solarflare and Force10 shows that ultra low latency application level performance can be achieved with commercially
Intel I340 Ethernet Dual Port and Quad Port Server Adapters for System x Product Guide
Intel I340 Ethernet Dual Port and Quad Port Server Adapters for System x Product Guide Based on the new Intel 82580 Gigabit Ethernet Controller, the Intel Ethernet Dual Port and Quad Port Server Adapters
Informatica Ultra Messaging SMX Shared-Memory Transport
White Paper Informatica Ultra Messaging SMX Shared-Memory Transport Breaking the 100-Nanosecond Latency Barrier with Benchmark-Proven Performance This document contains Confidential, Proprietary and Trade
Lustre Networking BY PETER J. BRAAM
Lustre Networking BY PETER J. BRAAM A WHITE PAPER FROM CLUSTER FILE SYSTEMS, INC. APRIL 2007 Audience Architects of HPC clusters Abstract This paper provides architects of HPC clusters with information
Solution Brief July 2014. All-Flash Server-Side Storage for Oracle Real Application Clusters (RAC) on Oracle Linux
Solution Brief July 2014 All-Flash Server-Side Storage for Oracle Real Application Clusters (RAC) on Oracle Linux Traditional SAN storage systems cannot keep up with growing application performance needs.
State of the Art Cloud Infrastructure
State of the Art Cloud Infrastructure Motti Beck, Director Enterprise Market Development WHD Global I April 2014 Next Generation Data Centers Require Fast, Smart Interconnect Software Defined Networks
LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu
3G Converged-NICs A Platform for Server I/O to Converged Networks
White Paper 3G Converged-NICs A Platform for Server I/O to Converged Networks This document helps those responsible for connecting servers to networks achieve network convergence by providing an overview
Emulex 8Gb Fibre Channel Expansion Card (CIOv) for IBM BladeCenter IBM BladeCenter at-a-glance guide
Emulex 8Gb Fibre Channel Expansion Card (CIOv) for IBM BladeCenter IBM BladeCenter at-a-glance guide The Emulex 8Gb Fibre Channel Expansion Card (CIOv) for IBM BladeCenter enables high-performance connection
Low-latency market data delivery to seize competitive advantage. WebSphere Front Office for Financial Markets: Fast, scalable access to market data.
Low-latency market data delivery to seize competitive advantage WebSphere Front Office for Financial Markets: Fast, scalable access to market data. Data from SIAC, NASDAQ and NYSE indicates a 158% increase
Communications Server for Linux
Communications Server for Linux SNA connectivity ^business on demand software Multiple types of connectivity exist within the Communications Server for Linux. CSLinux_snaconn.ppt Page 1 of 10 SNA connectivity
Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA
WHITE PAPER April 2014 Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA Executive Summary...1 Background...2 File Systems Architecture...2 Network Architecture...3 IBM BigInsights...5
VMWARE WHITE PAPER 1
1 VMWARE WHITE PAPER Introduction This paper outlines the considerations that affect network throughput. The paper examines the applications deployed on top of a virtual infrastructure and discusses the
New!! - Higher performance for Windows and UNIX environments
New!! - Higher performance for Windows and UNIX environments The IBM TotalStorage Network Attached Storage Gateway 300 (NAS Gateway 300) is designed to act as a gateway between a storage area network (SAN)
A Platform Built for Server Virtualization: Cisco Unified Computing System
A Platform Built for Server Virtualization: Cisco Unified Computing System What You Will Learn This document discusses how the core features of the Cisco Unified Computing System contribute to the ease
SMB Direct for SQL Server and Private Cloud
SMB Direct for SQL Server and Private Cloud Increased Performance, Higher Scalability and Extreme Resiliency June, 2014 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server
Performance Evaluation of Linux Bridge
Performance Evaluation of Linux Bridge James T. Yu School of Computer Science, Telecommunications, and Information System (CTI) DePaul University ABSTRACT This paper studies a unique network feature, Ethernet
QLogic 4Gb Fibre Channel Expansion Card (CIOv) for IBM BladeCenter IBM BladeCenter at-a-glance guide
QLogic 4Gb Fibre Channel Expansion Card (CIOv) for IBM BladeCenter IBM BladeCenter at-a-glance guide The QLogic 4Gb Fibre Channel Expansion Card (CIOv) for BladeCenter enables you to quickly and simply
Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks. An Oracle White Paper April 2003
Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks An Oracle White Paper April 2003 Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building
Low Latency Test Report Ultra-Low Latency 10GbE Switch and Adapter Testing Bruce Tolley, PhD, Solarflare
Ultra-Low Latency 10GbE Switch and Adapter Testing Bruce Tolley, PhD, Solarflare Testing conducted by Solarflare and Fujitsu shows that ultra low latency application-level performance can be achieved with
Cisco Performance Visibility Manager 1.0.1
Cisco Performance Visibility Manager 1.0.1 Cisco Performance Visibility Manager (PVM) is a proactive network- and applicationperformance monitoring, reporting, and troubleshooting system for maximizing
Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck
Sockets vs. RDMA Interface over 1-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Pavan Balaji Hemal V. Shah D. K. Panda Network Based Computing Lab Computer Science and Engineering
Enterprise Application Performance Management: An End-to-End Perspective
SETLabs Briefings VOL 4 NO 2 Oct - Dec 2006 Enterprise Application Performance Management: An End-to-End Perspective By Vishy Narayan With rapidly evolving technology, continued improvements in performance
SAS deployment on IBM Power servers with IBM PowerVM dedicated-donating LPARs
SAS deployment on IBM Power servers with IBM PowerVM dedicated-donating LPARs Narayana Pattipati IBM Systems and Technology Group ISV Enablement January 2013 Table of contents Abstract... 1 IBM PowerVM
I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology
I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology Reduce I/O cost and power by 40 50% Reduce I/O real estate needs in blade servers through consolidation Maintain
Choosing the Best Network Interface Card Mellanox ConnectX -3 Pro EN vs. Intel X520
COMPETITIVE BRIEF August 2014 Choosing the Best Network Interface Card Mellanox ConnectX -3 Pro EN vs. Intel X520 Introduction: How to Choose a Network Interface Card...1 Comparison: Mellanox ConnectX
Boosting Data Transfer with TCP Offload Engine Technology
Boosting Data Transfer with TCP Offload Engine Technology on Ninth-Generation Dell PowerEdge Servers TCP/IP Offload Engine () technology makes its debut in the ninth generation of Dell PowerEdge servers,
Cloud-ready network architecture
IBM Systems and Technology Thought Leadership White Paper May 2011 Cloud-ready network architecture 2 Cloud-ready network architecture Contents 3 High bandwidth with low latency 4 Converged communications
D1.2 Network Load Balancing
D1. Network Load Balancing Ronald van der Pol, Freek Dijkstra, Igor Idziejczak, and Mark Meijerink SARA Computing and Networking Services, Science Park 11, 9 XG Amsterdam, The Netherlands June [email protected],[email protected],
Comparing the Network Performance of Windows File Sharing Environments
Technical Report Comparing the Network Performance of Windows File Sharing Environments Dan Chilton, Srinivas Addanki, NetApp September 2010 TR-3869 EXECUTIVE SUMMARY This technical report presents the
Block based, file-based, combination. Component based, solution based
The Wide Spread Role of 10-Gigabit Ethernet in Storage This paper provides an overview of SAN and NAS storage solutions, highlights the ubiquitous role of 10 Gigabit Ethernet in these solutions, and illustrates
IBM RATIONAL PERFORMANCE TESTER
IBM RATIONAL PERFORMANCE TESTER Today, a major portion of newly developed enterprise applications is based on Internet connectivity of a geographically distributed work force that all need on-line access
Introduction to Cloud Design Four Design Principals For IaaS
WHITE PAPER Introduction to Cloud Design Four Design Principals For IaaS What is a Cloud...1 Why Mellanox for the Cloud...2 Design Considerations in Building an IaaS Cloud...2 Summary...4 What is a Cloud
Virtualizing SQL Server 2008 Using EMC VNX Series and Microsoft Windows Server 2008 R2 Hyper-V. Reference Architecture
Virtualizing SQL Server 2008 Using EMC VNX Series and Microsoft Windows Server 2008 R2 Hyper-V Copyright 2011 EMC Corporation. All rights reserved. Published February, 2011 EMC believes the information
Rapid Data Backup and Restore Using NFS on IBM ProtecTIER TS7620 Deduplication Appliance Express IBM Redbooks Solution Guide
Rapid Data Backup and Restore Using NFS on IBM ProtecTIER TS7620 Deduplication Appliance Express IBM Redbooks Solution Guide This IBM Redbooks Solution Guide provides an overview of how data backup and
Windows 8 SMB 2.2 File Sharing Performance
Windows 8 SMB 2.2 File Sharing Performance Abstract This paper provides a preliminary analysis of the performance capabilities of the Server Message Block (SMB) 2.2 file sharing protocol with 10 gigabit
Building Enterprise-Class Storage Using 40GbE
Building Enterprise-Class Storage Using 40GbE Unified Storage Hardware Solution using T5 Executive Summary This white paper focuses on providing benchmarking results that highlight the Chelsio T5 performance
Connecting the Clouds
Connecting the Clouds Mellanox Connected Clouds Mellanox s Ethernet and InfiniBand interconnects enable and enhance worldleading cloud infrastructures around the globe. Utilizing Mellanox s fast server
ECLIPSE Performance Benchmarks and Profiling. January 2009
ECLIPSE Performance Benchmarks and Profiling January 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox, Schlumberger HPC Advisory Council Cluster
High Performance Data-Transfers in Grid Environment using GridFTP over InfiniBand
High Performance Data-Transfers in Grid Environment using GridFTP over InfiniBand Hari Subramoni *, Ping Lai *, Raj Kettimuthu **, Dhabaleswar. K. (DK) Panda * * Computer Science and Engineering Department
Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014
Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet Anand Rangaswamy September 2014 Storage Developer Conference Mellanox Overview Ticker: MLNX Leading provider of high-throughput,
BRIDGING EMC ISILON NAS ON IP TO INFINIBAND NETWORKS WITH MELLANOX SWITCHX
White Paper BRIDGING EMC ISILON NAS ON IP TO INFINIBAND NETWORKS WITH Abstract This white paper explains how to configure a Mellanox SwitchX Series switch to bridge the external network of an EMC Isilon
Dell Reference Configuration for Hortonworks Data Platform
Dell Reference Configuration for Hortonworks Data Platform A Quick Reference Configuration Guide Armando Acosta Hadoop Product Manager Dell Revolutionary Cloud and Big Data Group Kris Applegate Solution
Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking
Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking Burjiz Soorty School of Computing and Mathematical Sciences Auckland University of Technology Auckland, New Zealand
Windows TCP Chimney: Network Protocol Offload for Optimal Application Scalability and Manageability
White Paper Windows TCP Chimney: Network Protocol Offload for Optimal Application Scalability and Manageability The new TCP Chimney Offload Architecture from Microsoft enables offload of the TCP protocol
1000-Channel IP System Architecture for DSS
Solution Blueprint Intel Core i5 Processor Intel Core i7 Processor Intel Xeon Processor Intel Digital Security Surveillance 1000-Channel IP System Architecture for DSS NUUO*, Qsan*, and Intel deliver a
Intel DPDK Boosts Server Appliance Performance White Paper
Intel DPDK Boosts Server Appliance Performance Intel DPDK Boosts Server Appliance Performance Introduction As network speeds increase to 40G and above, both in the enterprise and data center, the bottlenecks
Virtual Compute Appliance Frequently Asked Questions
General Overview What is Oracle s Virtual Compute Appliance? Oracle s Virtual Compute Appliance is an integrated, wire once, software-defined infrastructure system designed for rapid deployment of both
Demartek June 2012. Broadcom FCoE/iSCSI and IP Networking Adapter Evaluation. Introduction. Evaluation Environment
June 212 FCoE/iSCSI and IP Networking Adapter Evaluation Evaluation report prepared under contract with Corporation Introduction Enterprises are moving towards 1 Gigabit networking infrastructures and
Improving Grid Processing Efficiency through Compute-Data Confluence
Solution Brief GemFire* Symphony* Intel Xeon processor Improving Grid Processing Efficiency through Compute-Data Confluence A benchmark report featuring GemStone Systems, Intel Corporation and Platform
Title. Click to edit Master text styles Second level Third level
Title Click to edit Master text styles Second level Third level IBM s Vision For The New Enterprise Data Center Subram Natarajan Senior Consultant, STG Asia Pacific [email protected] Multiple
Cloud-Based Apps Drive the Need for Frequency-Flexible Clock Generators in Converged Data Center Networks
Cloud-Based Apps Drive the Need for Frequency-Flexible Generators in Converged Data Center Networks Introduction By Phil Callahan, Senior Marketing Manager, Timing Products, Silicon Labs Skyrocketing network
Enabling High performance Big Data platform with RDMA
Enabling High performance Big Data platform with RDMA Tong Liu HPC Advisory Council Oct 7 th, 2014 Shortcomings of Hadoop Administration tooling Performance Reliability SQL support Backup and recovery
An Oracle White Paper December 2010. Consolidating and Virtualizing Datacenter Networks with Oracle s Network Fabric
An Oracle White Paper December 2010 Consolidating and Virtualizing Datacenter Networks with Oracle s Network Fabric Introduction... 1 Today s Datacenter Challenges... 2 Oracle s Network Fabric... 3 Maximizing
Intel Ethernet Switch Converged Enhanced Ethernet (CEE) and Datacenter Bridging (DCB) Using Intel Ethernet Switch Family Switches
Intel Ethernet Switch Converged Enhanced Ethernet (CEE) and Datacenter Bridging (DCB) Using Intel Ethernet Switch Family Switches February, 2009 Legal INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION
Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers
Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers White Paper rev. 2015-11-27 2015 FlashGrid Inc. 1 www.flashgrid.io Abstract Oracle Real Application Clusters (RAC)
Extreme Networks: Building Cloud-Scale Networks Using Open Fabric Architectures A SOLUTION WHITE PAPER
Extreme Networks: Building Cloud-Scale Networks Using Open Fabric Architectures A SOLUTION WHITE PAPER WHITE PAPER Building Cloud- Scale Networks Abstract TABLE OF CONTENTS Introduction 2 Open Fabric-Based
Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1
Performance Study Performance Characteristics of and RDM VMware ESX Server 3.0.1 VMware ESX Server offers three choices for managing disk access in a virtual machine VMware Virtual Machine File System
LS DYNA Performance Benchmarks and Profiling. January 2009
LS DYNA Performance Benchmarks and Profiling January 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox HPC Advisory Council Cluster Center The
Achieving a High-Performance Virtual Network Infrastructure with PLUMgrid IO Visor & Mellanox ConnectX -3 Pro
Achieving a High-Performance Virtual Network Infrastructure with PLUMgrid IO Visor & Mellanox ConnectX -3 Pro Whitepaper What s wrong with today s clouds? Compute and storage virtualization has enabled
Upgrading Data Center Network Architecture to 10 Gigabit Ethernet
Intel IT IT Best Practices Data Centers January 2011 Upgrading Data Center Network Architecture to 10 Gigabit Ethernet Executive Overview Upgrading our network architecture will optimize our data center
3 Red Hat Enterprise Linux 6 Consolidation
Whitepaper Consolidation EXECUTIVE SUMMARY At this time of massive and disruptive technological changes where applications must be nimbly deployed on physical, virtual, and cloud infrastructure, Red Hat
