Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014



Similar documents
SMB Direct for SQL Server and Private Cloud

Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 SMB Direct

State of the Art Cloud Infrastructure

High Throughput File Servers with SMB Direct, Using the 3 Flavors of RDMA network adapters

SMB Advanced Networking for Fault Tolerance and Performance. Jose Barreto Principal Program Managers Microsoft Corporation

Hyper-V over SMB Remote File Storage support in Windows Server 8 Hyper-V. Jose Barreto Principal Program Manager Microsoft Corporation

Mellanox Accelerated Storage Solutions

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Storage, Cloud, Web 2.0, Big Data Driving Growth

Advancing Applications Performance With InfiniBand

Hyper-V over SMB: Remote File Storage Support in Windows Server 2012 Hyper-V. Jose Barreto Principal Program Manager Microsoft Corporation

Building a Scalable Storage with InfiniBand

Enabling High performance Big Data platform with RDMA

Microsoft Windows Server Hyper-V in a Flash

Connecting the Clouds

Microsoft Windows Server in a Flash

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

Replacing SAN with High Performance Windows Share over a Converged Network

High Performance OpenStack Cloud. Eli Karpilovski Cloud Advisory Council Chairman

StarWind Virtual SAN for Microsoft SOFS

Microsoft SQL Server 2014 in a Flash

Microsoft Hybrid Cloud IaaS Platforms

Mellanox Academy Online Training (E-learning)

Microsoft Windows Server Hyper-V in a Flash

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

Microsoft SMB Running Over RDMA in Windows Server 8

Bringing the Public Cloud to Your Data Center

Windows 8 SMB 2.2 File Sharing Performance

I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology

Solution Brief July All-Flash Server-Side Storage for Oracle Real Application Clusters (RAC) on Oracle Linux

Driving Revenue Growth Across Multiple Markets

IOmark- VDI. Nimbus Data Gemini Test Report: VDI a Test Report Date: 6, September

Windows TCP Chimney: Network Protocol Offload for Optimal Application Scalability and Manageability

RoCE vs. iwarp Competitive Analysis

Michael Kagan.

Building Enterprise-Class Storage Using 40GbE

Choices for implementing SMB 3 on non Windows Servers Dilip Naik HvNAS Pty Ltd Australians good at NAS protocols!

Performance Analysis: Scale-Out File Server Cluster with Windows Server 2012 R2 Date: December 2014 Author: Mike Leone, ESG Lab Analyst

InfiniBand vs Fibre Channel Throughput. Figure 1. InfiniBand vs 2Gb/s Fibre Channel Single Port Storage Throughput to Disk Media

Choosing the Best Network Interface Card Mellanox ConnectX -3 Pro EN vs. Intel X520

Overview: X5 Generation Database Machines

Can High-Performance Interconnects Benefit Memcached and Hadoop?

Storage Spaces. Storage Spaces

LSI SAS inside 60% of servers. 21 million LSI SAS & MegaRAID solutions shipped over last 3 years. 9 out of 10 top server vendors use MegaRAID

Connecting Flash in Cloud Storage

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA

Choosing the Best Network Interface Card for Cloud Mellanox ConnectX -3 Pro EN vs. Intel XL710

From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller

Introduction to Cloud Design Four Design Principals For IaaS

Optimizing Large Arrays with StoneFly Storage Concentrators

Block based, file-based, combination. Component based, solution based

Oracle Exadata: The World s Fastest Database Machine Exadata Database Machine Architecture

Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks. An Oracle White Paper April 2003

Building a cost-effective and high-performing public cloud. Sander Cruiming, founder Cloud Provider

Diablo and VMware TM powering SQL Server TM in Virtual SAN TM. A Diablo Technologies Whitepaper. May 2015

PCI Express Impact on Storage Architectures and Future Data Centers. Ron Emerick, Oracle Corporation

WHITE PAPER. How To Build a SAN. The Essential Guide for Turning Your Windows Server Into Shared Storage on Your IP Network

3G Converged-NICs A Platform for Server I/O to Converged Networks

Software Defined Microsoft. PRESENTATION TITLE GOES HERE Siddhartha Roy Cloud + Enterprise Division Microsoft Corporation

The Evolution of Microsoft SQL Server: The right time for Violin flash Memory Arrays

SOLUTION BRIEF AUGUST All-Flash Server-Side Storage for Oracle Real Application Clusters (RAC) on Oracle Linux

Microsoft SQL Server 2014 in a Flash

The last 18 months. AutoScale. IaaS. BizTalk Services Hyper-V Disaster Recovery Support. Multi-Factor Auth. Hyper-V Recovery.

InfiniBand in the Enterprise Data Center

Private cloud computing advances

InfiniBand Update Addressing new I/O challenges in HPC, Cloud, and Web 2.0 infrastructures. Brian Sparks IBTA Marketing Working Group Co-Chair

How To Connect Virtual Fibre Channel To A Virtual Box On A Hyperv Virtual Machine

Windows Server Infrastructure for SQL Server

Doubling the I/O Performance of VMware vsphere 4.1

New Storage System Solutions

Storage at a Distance; Using RoCE as a WAN Transport

InfiniBand Software and Protocols Enable Seamless Off-the-shelf Applications Deployment

Scale-Out File Server. Subtitle

2009 Oracle Corporation 1

Business in a flash! The All-Flash Datacenter of the Future Today! Storagetechnology Schloss Fleesensee

ECLIPSE Performance Benchmarks and Profiling. January 2009

EDUCATION. PCI Express, InfiniBand and Storage Ron Emerick, Sun Microsystems Paul Millard, Xyratex Corporation

broadberry.co.uk/storage-servers

Cloud Optimize Your IT

Virtual Compute Appliance Frequently Asked Questions

ZD-XL SQL Accelerator 1.5

Pre-Conference Seminar E: Flash Storage Networking

PCI Express Impact on Storage Architectures. Ron Emerick, Sun Microsystems

Virtualizing Microsoft SQL Server on Dell XC Series Web-scale Converged Appliances Powered by Nutanix Software. Dell XC Series Tech Note

Broadcom Ethernet Network Controller Enhanced Virtualization Functionality

iscsi Top Ten Top Ten reasons to use Emulex OneConnect iscsi adapters

HP 3PAR StoreServ 8000 Storage - what s new

Dell Converged Infrastructure

Architecting a High Performance Storage System

PCI Express and Storage. Ron Emerick, Sun Microsystems

Long-Haul System Family. Highest Levels of RDMA Scalability, Simplified Distance Networks Manageability, Maximum System Productivity

Windows Server Infrastructure and Storage for SQL Server & Hyper-V

Stovepipes to Clouds. Rick Reid Principal Engineer SGI Federal by SGI Federal. Published by The Aerospace Corporation with permission.

LS DYNA Performance Benchmarks and Profiling. January 2009

High Performance MySQL Cluster Cloud Reference Architecture using 16 Gbps Fibre Channel and Solid State Storage Technology

Integration of Microsoft Hyper-V and Coraid Ethernet SAN Storage. White Paper

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

Introduction. Need for ever-increasing storage scalability. Arista and Panasas provide a unique Cloud Storage solution

Ethernet: THE Converged Network Ethernet Alliance Demonstration as SC 09

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute

Transcription:

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet Anand Rangaswamy September 2014 Storage Developer Conference

Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server and storage interconnect FDR 56Gb/s InfiniBand 10/40/56GbE Ethernet Switches and Adapters Company headquarters: Yokneam, Israel; Sunnyvale, California ~1,432 employees* worldwide * As of March 2014 2

This image cannot currently be displayed. Leadership in Strategic Markets with an End to End Portfolio Big Data DB/Enterprise Cloud Storage Web 2.0 End-to-End - Cloud, Enterprise, and Storage Optimized - InfiniBand and Ethernet Portfolio ICs Adapter Cards Switches/Gateways Host/Fabric Software Metro / WAN Cables/Modules 3

Mellanox Unique Value Proposition - RDMA ZERO Copy Remote Data Transfer USER KERNEL Application Buffer Application Buffer HARDWARE Kernel Bypass Low Latency, High Performance Data Transfers Protocol Offload InfiniBand - 56Gb/s RoCE* 40Gb/s * RDMA over Converged Ethernet 4

RDMA How it Works HARDWARE KERNEL USER Application 1 OS Buffer 1 Buffer 1 Buffer 1 NIC Buffer 1 RDMA over InfiniBand or Ethernet HCA TCP/IP HCA Buffer 1 Buffer 1 Buffer 1 Buffer 1 Application 2 OS NIC RACK 1 RACK 2 5

Infiniband Acronyms 2002 10 Gb/s 2005 20 Gb/s 2008 40 Gb/s 2011 56 Gb/s 2014 100 Gb/s SDR DDR QDR FDR EDR VPI* FDR InfiniBand New Features and Capabilities Performance / Scalability Switch port bandwidth 56Gb/s InfiniBand Router and IB-Eth bridges First VPI silicon in the World Reliability / Efficiency Link bit encoding 64/66 Forward Error Correction Low power consumption 6

Microsoft Windows Server 2012 R2 SMB Direct New class of enterprise file storage with RDMA support File Client File Server Lower latency, higher throughout, lower CPU overhead Application User Kernel Fibre Channel replacement at a lower cost and higher performance SMB Client SMB Server Leverages Windows Server 2012 R2 Mellanox inbox InfiniBand & Ethernet RDMA drivers Accelerates Microsoft Hyper-V over SMB and Microsoft SQL Server over SMB solutions Network w/ RDMA support RDMA Adapter Network w/ RDMA support RDMA Adapter NTFS SCSI Disk 5X Performance Improvement Over 10GbE 7

Measuring SMB Direct Performance SMB Client IO Micro Benchmark SMB Client IO Micro Benchmark SMB Client IO Micro Benchmark 10GbE QDR IB FDR IB SMB Server 10GbE SMB Server QDR IB SMB Server FDR IB Single Server IO Micro Benchmark Fusion Fusion Fusion Fusion-IO IO IO IO Fusion Fusion Fusion Fusion-IO IO IO IO Fusion Fusion Fusion Fusion-IO IO IO IO Fusion Fusion Fusion Fusion-IO IO IO IO 8

Performance Comparison 9

Microsoft Delivers High-End Storage at a Fraction of the Cost FDR 56Gb/s InfiniBand delivers 5X higher throughput with 50% less CPU overhead vs. 10GbE Native Throughput Performance over FDR InfiniBand 10

Performance and Resiliency Similar to a Traditional SAN Scale-Out File Server Cluster using Windows Server 2012 R2 SMB Direct 1.1 Million IOPS for 8KB random reads 700K IOPS for 8KB random writes 100Gbps aggregate read bandwidth * 92Gbps aggregate write bandwidth Source: http://www.mellanox.com/related-docs/applications/achieving-1.1m-iops-over-smb-direct.pdf *Cluster Shared Volumes FDR InfiniBand enables Native Performance 11

Mellanox Testing 8KB I/O Echostreams flash array with 24 s IOPS (1000's IOs/sec) 700 600 500 400 IOPS (1000's IOs/sec) 300 200 100 0 Local s 10 GE TCP 10 GE RoCE 1x Link 10 GE RoCE 2x links 40 GE TCP 40 GE RoCE 1x link 40 GE RoCE 2x links IB 54Gbps FDR 1x link IB 54Gbps FDR 2x links IB 54Gbps FDR 3x links 12 12

Mellanox Testing 32KB I/O Echostreams flash array with 24 IOPS (1000's IOs/sec) 300 250 200 150 IOPS (1000's IOs/sec) 100 50 0 Local s 10 GE 10 GE RoCE 1x Link 10 GE RoCE 2x links 40 GE 40 GE RoCE 1x link 40 GE RoCE 2x links IB 54Gbps FDR 1x link IB 54Gbps FDR 2x links 13 13

Mellanox Testing 512KB I/O Echostreams flash array with 24 s IOPS (1000's IOs/sec) 20.0 18.0 16.0 14.0 12.0 10.0 IOPS (1000's IOs/sec) 8.0 6.0 4.0 2.0 0.0 Local s 10 GE 10 GE 1x RoCE 10 GE RoCE 2x links 40 GE 40 GE RoCE 1x link 40 GE RoCE 2x links IB 54Gbps FDR 1x link IB 54Gbps FDR 2x links 14 14

Hyper-V Live Migration on SMB Direct with Mellanox Interconnect SMB as a transport for Live Migration of VMs Delivers the power of SMB to provide: RDMA (SMB Direct) Streaming over multiple NICs (SMB Multichannel) Provides highest bandwidth and lowest latency Seconds 70 60 50 40 30 Live Migration Times* New in Windows Server 2012 R2 20 10 0 * Lower is better Live migration can stream over multiple networks for improved bandwidth Live Migration can take advantage of high speed networking RDMA enables offloading CPU resources to NIC during live migration Source: TechED 13 Opening Keynote Session with Satya Nadella 15

Success Story: Edgenet Reduced complexity vs. FC storage management Creates a more manageable storage subsystem - SMB easier to provision than Fibre Channel devices The storage server manages all the connections - Client Servers only need to know one path via SMB 10x Acceleration vs FC Allows consolidation of many FC arrays over SMB Live migration between servers at record speeds Mellanox VPI solutions Provides user access to the Hyper-V hosted application Source: http://www.mellanox.com/related-docs/case_studies/cs_edgenet_boost_performance_lower_cost.pdf 16

SMB Direct Storage Solutions 17

Introducing The Windows Flash Array Violin Memory, Inc. 2014 CONFIDENTIAL 18

Windows in a Flash! >1 year of Joint Violin - Microsoft Development SMB Direct for Microsoft applications RDMA as a Storage Superhighway Transformative Speed in Storage Consolidation and cost savings Enables file server functions Increases server utilization Encryption Dedupe Live Migration Mirroring Scale-Out Violin Memory, Inc. 2014 CONFIDENTIAL 19

Beta Site Performance: SQL* and Hyper-V 2x Increase SQL Writes 54% Increase SQL Reads 41% Increase Hyper-V Writes IOPS 1.2M 1.0M 800K 600K 400K WFA AFA WFA AFA WFA AFA 30% Decrease Application Server CPU Utilization 50% Decrease Latency at Load 200K SQL READS* SQL WRITES* HYPER-V WRITES SMB Direct reduces CPU load up to 30% Latency averages 500 microseconds at Load *All Microsoft SQL results based on SQLIO load test with Infiniband connectivity Violin Memory, Inc. 2014 CONFIDENTIAL 20

DataOn Cluster-in-a-Box over RDMA Extremely economical Simplified architecture Easy to install appliance CiB interconnects: ConnectX-3 40GbE ConnectX-3 56Gbs InfiniBand Deployed at: 21

EchoStreams: InfiniBand Enables Near Linear Scalability File Client (SMB 3.0) File Server (SMB 3.0) RDMA NIC RDMA NIC SQLIO RDMA NIC RDMA NIC RDMA NIC RDMA NIC 8KB random reads from a mirrored space (disk) ~600,000 IOPS SAS HBA SAS SAS HBA SAS Storage Spaces SAS HBA SAS SAS HBA SAS SAS HBA SAS SAS HBA SAS 8KB random reads from cache (RAM) ~1,000,000 IOPS JBOD JBOD JBOD JBOD JBOD JBOD 32KB random reads from a mirrored space (disk) ~500,000 IOPS ~16.5 GBytes/sec 22

Iron Networks IRON POD Preconfigured Microsoft Hyper-V cloud IaaS hardware platform Full integrated converged data center solution Iron Pod Description Fault tolerant Mellanox 40/56GbE server and storage network - Fully redundant network for high availability ConnectX-3 Pro adapters for RDMA and NVGRE acceleration - Dual 40/56GbE network adapters More info about Iron Networks: http://www.mellanox.com/relateddocs/solutions/iron_networks_microsoft_fast_track_architecture.pdf http://ironnetworks.com/microsoft-converged-infrastructure/ironpod-cloud- Building-Blocks http://ironnetworks.com/microsoft-converged-infrastructure/ironpod-system-500 23

Questions? 24

Thank You

Accelerating Microsoft SQL 2012 Parallel Data Warehouse V2 Analyze 1 Petabyte of Data in 1 Second Up to 100X faster performance than legacy data warehouse queries Up to 50X faster data query, up to 2X the data loading rate Unlimited storage scalability for future proofing Accelerated by Mellanox FDR 56Gb/s InfiniBand end-to-end solutions 26

Microsoft SQL Server Appliance 10X Faster & 50% Lower Capital Cost Parallel Data Warehouse Appliance V1 Parallel Data Warehouse Appliance V2 Control Node Pure hardware costs are ~50% lower Mgmt. Node Price per raw TB is close to 70% lower due to higher capacity LZ 70% more disk I/O bandwidth Backup Node Ethernet, InfiniBand & Fibre Channel InfiniBand Estimated total HW component list $ price: $1M 160 cores on 10 compute nodes 1.28 TB of RAM on compute Up to 30 TB of temp DB Up to 150 TB of user data 128 cores on 8 compute nodes 2TB of RAM on compute Up to 168 TB of temp DB Up to 1PB of user data Estimated total HW component list $ price: $500K 27