Interconnect Your Future Enabling the Best Datacenter Return on Investment. TOP500 Supercomputers, November 2015



Similar documents
Interconnect Your Future Enabling the Best Datacenter Return on Investment. TOP500 Supercomputers, June 2016

InfiniBand Strengthens Leadership as the High-Speed Interconnect Of Choice

Advancing Applications Performance With InfiniBand

Interconnecting Future DoE leadership systems

Storage, Cloud, Web 2.0, Big Data Driving Growth

InfiniBand Update Addressing new I/O challenges in HPC, Cloud, and Web 2.0 infrastructures. Brian Sparks IBTA Marketing Working Group Co-Chair

InfiniBand Takes Market Share from Ethernet and Proprietary Interconnects Top500 Supercomputers, Nov 2006

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK

State of the Art Cloud Infrastructure

SMB Direct for SQL Server and Private Cloud

Interoperability Testing and iwarp Performance. Whitepaper

LS DYNA Performance Benchmarks and Profiling. January 2009

White Paper Solarflare High-Performance Computing (HPC) Applications

Michael Kagan.

Mississippi State University High Performance Computing Collaboratory Brief Overview. Trey Breckenridge Director, HPC

HPC Update: Engagement Model

Enabling the Use of Data

Driving Revenue Growth Across Multiple Markets

ECLIPSE Performance Benchmarks and Profiling. January 2009

Hyperscale. The new frontier for HPC. Philippe Trautmann. HPC/POD Sales Manager EMEA March 13th, 2011

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Supercomputing Resources in BSC, RES and PRACE

Connecting the Clouds

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

Mellanox Technologies

Mellanox Academy Online Training (E-learning)

Choosing the Best Network Interface Card Mellanox ConnectX -3 Pro EN vs. Intel X520

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

The following InfiniBand products based on Mellanox technology are available for the HP BladeSystem c-class from HP:

Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks. An Oracle White Paper April 2003

第 十 三 回 PCクラスタシンポジウム. Cray クラスタ 製 品 のご 紹 介 クレイ ジャパン インク

HP APOLLO SYSTEMS FOR HPC AND BIG DATA ANALYTICS NG JIT KEONG TECHNICAL MARKETING MANAGER HP SERVER, SOUTH EAST ASIA

Sun Constellation System: The Open Petascale Computing Architecture

MELLANOX TECHNOLOGIES, LTD.

PRIMERGY server-based High Performance Computing solutions

Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca

Pedraforca: ARM + GPU prototype

HPC Market Update, HPC Trends In the Oil/Gas Sector and IDC's Top 10 Predictions for Earl Joseph, HPC Program Vice President

New Storage System Solutions

Hadoop on the Gordon Data Intensive Cluster

Comparing the performance of the Landmark Nexus reservoir simulator on HP servers

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

Mellanox Technologies, Ltd. Announces Record Quarterly Results

Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband

Cisco for SAP HANA Scale-Out Solution on Cisco UCS with NetApp Storage

3G Converged-NICs A Platform for Server I/O to Converged Networks

HPC-related R&D in 863 Program

Supercomputing Clusters with RapidIO Interconnect Fabric

Introduction to Infiniband. Hussein N. Harake, Performance U! Winter School

HPC & Big Data THE TIME HAS COME FOR A SCALABLE FRAMEWORK

High Performance Computing in CST STUDIO SUITE

Mellanox Accelerated Storage Solutions

FPGA Acceleration using OpenCL & PCIe Accelerators MEW 25

Jezelf Groen Rekenen met Supercomputers

Jean-Pierre Panziera Teratec 2011

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/ CAE Associates

Can High-Performance Interconnects Benefit Memcached and Hadoop?

Building a cost-effective and high-performing public cloud. Sander Cruiming, founder Cloud Provider

Using the Intel Xeon Phi (with the Stampede Supercomputer) ISC 13 Tutorial

Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 SMB Direct

Ultra Low Latency Data Center Switches and iwarp Network Interface Cards

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Scaling from Workstation to Cluster for Compute-Intensive Applications

Mellanox Achieves Record Quarterly and Annual Revenue

Stovepipes to Clouds. Rick Reid Principal Engineer SGI Federal by SGI Federal. Published by The Aerospace Corporation with permission.

Long-Haul System Family. Highest Levels of RDMA Scalability, Simplified Distance Networks Manageability, Maximum System Productivity

Accelerating From Cluster to Cloud: Overview of RDMA on Windows HPC. Wenhao Wu Program Manager Windows HPC team

The Future of Cloud Networking. Idris T. Vasi

ECLIPSE Best Practices Performance, Productivity, Efficiency. March 2009

Simplifying Big Data Deployments in Cloud Environments with Mellanox Interconnects and QualiSystems Orchestration Solutions

EDUCATION. PCI Express, InfiniBand and Storage Ron Emerick, Sun Microsystems Paul Millard, Xyratex Corporation

Parallel Programming Survey

Building Enterprise-Class Storage Using 40GbE

Deep Learning GPU-Based Hardware Platform

Accelerating Spark with RDMA for Big Data Processing: Early Experiences

Power Saving Features in Mellanox Products

Kriterien für ein PetaFlop System

InfiniBand Switch System Family. Highest Levels of Scalability, Simplified Network Manageability, Maximum System Productivity

Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Levels of Data Center IT Performance, Efficiency and Scalability

PCI Express and Storage. Ron Emerick, Sun Microsystems

Dr. Raju Namburu Computational Sciences Campaign U.S. Army Research Laboratory. The Nation s Premier Laboratory for Land Forces UNCLASSIFIED

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing

SGI High Performance Computing

I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology

Panasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory

Barry Bolding, Ph.D. VP, Cray Product Division

SX1012: High Performance Small Scale Top-of-Rack Switch

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

Summit and Sierra Supercomputers:

Supercomputing Status und Trends (Conference Report) Peter Wegner

Why Use 16Gb Fibre Channel with Windows Server 2012 Deployments

Scientific Computing Data Management Visions

BSC - Barcelona Supercomputer Center

Exploiting Remote Memory Operations to Design Efficient Reconfiguration for Shared Data-Centers over InfiniBand

Virtual Compute Appliance Frequently Asked Questions

Building a Top500-class Supercomputing Cluster at LNS-BUAP

Next Gen Data Center. KwaiSeng Consulting Systems Engineer

JuRoPA. Jülich Research on Petaflop Architecture. One Year on. Hugo R. Falter, COO Lee J Porter, Engineering

Introduction. Need for ever-increasing storage scalability. Arista and Panasas provide a unique Cloud Storage solution

Transcription:

Interconnect Your Future Enabling the Best Datacenter Return on Investment TOP500 Supercomputers, November 2015

InfiniBand FDR and EDR Continue Growth and Leadership The Most Used Interconnect On The TOP500 Supercomputing List EDR InfiniBand Now Used in 4 Systems on the List, with More EDR Systems Being Deployed by End of 2015 Will Result in More Than 2X the Number of EDR Systems Worldwide Compared to June 2015 Mellanox Delivers Up To 99.8% System Efficiency with InfiniBand, 4X higher versus Ethernet Delivering highest datacenter return on investment For the first time many Chinese-based Ethernet Cloud/Web2.0 systems submitted and ranked on the list Altering the trends perspective of high-performance computing systems on the TOP500 for Ethernet/InfiniBand connectivity 2015 Mellanox Technologies 2

TOP500 Performance Trends 76% CAGR Explosive high-performance computing market growth Clusters continue to dominate with 85% of the TOP500 list Mellanox InfiniBand Solutions Provide the Highest Systems Utilization on the TOP500 2015 Mellanox Technologies 3

TOP500 Major Highlights InfiniBand is the most used interconnect on the TOP500 list, with 237 systems, or 47.4% of the list FDR InfiniBand is the most used technology on the TOP500 list, connecting 163 systems 16% increase from Nov 14 to Nov 15 EDR InfiniBand now used in 4 systems on the List, with more EDR systems being deployed by end of 2015 More than 2X the number of EDR systems expected compared to July 2015 InfiniBand enables the most efficient system on the list with 99.8% efficiency record! Delivering highest datacenter Return on Investment, 4X higher than 10GbE, 33% higher than Cray interconnect InfiniBand enables the top 9 most efficient systems on the list, and 28 of the TOP30 InfiniBand is the most used interconnect for Petascale systems with 33 systems 45% of the Petascale systems, nearly 40% higher versus closest competition Mellanox InfiniBand is the Interconnect of Choice for HPC Infrastructures 2015 Mellanox Technologies 4

InfiniBand in the TOP500 InfiniBand is the most used interconnect technology for high-performance computing InfiniBand accelerates 237 systems, 47.4% of the list FDR InfiniBand is the most used technology on the TOP500 list, connecting 163 systems 16% increase from Nov 14 to Nov 15 EDR InfiniBand now used in 4 systems on the List, with more EDR systems being deployed by end of 2015 Expected to be deployed in more than 2X the number of EDR systems worldwide compared to June 2015 InfiniBand connects the most powerful clusters 33 of the Petascale-performance systems, 45% of the Petascale systems, nearly 40% higher versus closest competition InfiniBand is the most used interconnect technology in each category - TOP100/200/300/400/500 Connects 43% (43 systems) of the TOP100 while Ethernet only 7% (7 systems) Connects 44.5% (89 systems) of the TOP200 while Ethernet only 23.5% (47 systems) Connects 44.3% (133 systems) of the TOP300 while Ethernet only 32% (96 systems) Connects 45.3% (181 systems) of the TOP400 while Ethernet only 35.8% (143 systems) Connects 47.4% (237 systems) of the TOP500, Ethernet connects 36.2% (181 systems) 2015 Mellanox Technologies 5

Mellanox in the TOP500 Mellanox EDR InfiniBand is the fastest interconnect solution on the TOP500 100Gb/s throughput, 150 million messages per second, less than 0.7usec latency Mellanox InfiniBand is the most efficient interconnect technology on the list Enables the highest system utilization on the TOP500 99.8% system efficiency Enables the top 9 highest utilized systems on the TOP500 list Mellanox InfiniBand is the only Petascale-proven, standard interconnect solution Connects 33 out of the 76 Petaflop-performance systems on the list (45%) - 40% more systems versus closest competition Connects 1.4X the number of Cray based systems in the TOP100, 5X in TOP500 Mellanox s end-to-end scalable solutions accelerate GPU-based systems GPUDirect RDMA technology enables faster communications and higher performance Paving The Road to Exascale Performance 2015 Mellanox Technologies 6

TOP500 Interconnect Trends InfiniBand is The De-Facto Interconnect Solution for Performance Demanding Applications The Most Used Interconnect Solution for High-Performance Infrastructures 2015 Mellanox Technologies 7

TOP500 Petascale-Performance Systems Mellanox InfiniBand is the Interconnect of Choice for Petascale Computing Accelerates 45% of the Petaflop Systems 2015 Mellanox Technologies 8

TOP500: Market Segments InfiniBand Accelerates 62% of HPC Research Systems InfiniBand Accelerates 80% of HPC Academic Systems InfiniBand Accelerates 25% of Cloud, Web2.0 etc. Industry Systems Mellanox InfiniBand is the Interconnect of Choice for HPC Infrastructures 2015 Mellanox Technologies 9

TOP500 InfiniBand Accelerated Systems Number of Mellanox FDR InfiniBand systems grew 16% from Nov 14 to Nov 15; EDR InfiniBand now used in 4 systems on the TOP500 More EDR Systems Being Deployed by End of 2015 Will Result in More Than Double the Number of EDR Systems Worldwide Compared to June 2015 2015 Mellanox Technologies 10

InfiniBand versus Ethernet TOP100, 200, 300, 400, 500 InfiniBand is The Most Used Interconnect of the TOP100, 200, 300, 400 Supercomputers Due to Superior Performance, Scalability, Efficiency and Return-On-Investment 2015 Mellanox Technologies 11

InfiniBand s Unsurpassed System Efficiency The new Chinese based Cloud/Wed2.0 10GbE systems show extremely poor efficiency of 20-40%, not suitable for HPC applications Average Efficiency InfiniBand: 85% Cray: 74% 10GbE: 57% GigE: 43% InfiniBand is Key For the System Efficiency, Mellanox Delivers Up To 99.8% System Efficiency with InfiniBand More Than 4X Higher Efficiency Versus the New Additions of 10GbE Chinese-based Cloud/Wed2.0 Systems 2015 Mellanox Technologies 12

TOP500 Interconnect Placement InfiniBand is the Best Price/Performance Interconnect for Every Scale of HPC Systems 2015 Mellanox Technologies 13

InfiniBand Performance Trends InfiniBand-Connected CPUs Grew 37% From Nov 14 to Nov 15 InfiniBand-based System Performance Grew 42% From Nov 14 to Nov 15 Mellanox InfiniBand is the Most Efficient and Scalable Interconnect Solution Driving Factors: Performance, Efficiency, Scalability, Many-Many Cores 2015 Mellanox Technologies 14

Proud to Accelerate Future DOE Leadership Systems ( CORAL ) Summit System Sierra System Paving the Road to Exascale 2015 Mellanox Technologies 15

High-Performance Designed 100Gb/s Interconnect Solutions 100Gb/s Adapter, 0.7us latency 150 million messages per second (10 / 25 / 40 / 50 / 56 / 100Gb/s) 36 EDR (100Gb/s) Ports, <90ns Latency Throughput of 7.2Tb/s 7.02 Billion msg/sec (195M msg/sec/port) 32 100GbE Ports, 64 25/50GbE Ports (10 / 25 / 40 / 50 / 100GbE) Throughput of 6.4Tb/s Transceivers Active Optical and Copper Cables (10 / 25 / 40 / 50 / 56 / 100Gb/s) VCSELs, Silicon Photonics and Copper 2015 Mellanox Technologies 16

Interconnect Solutions Leadership Software and Hardware Comprehensive End-to-End Interconnect Software Products ICs Comprehensive End-to-End InfiniBand and Ethernet Portfolio Adapter Cards Switches/Gateways Software and Services Metro / WAN Cables/Modules At the Speeds of 10, 25, 40, 50, 56 and 100 Gigabit per Second 2015 Mellanox Technologies 17

Mellanox InfiniBand Solutions, Generation Lead over the Competition Smart Network For Smart Systems RDMA, Acceleration Engines, Programmability Higher Performance Unlimited Scalability Higher Resiliency Proven! 100 Gb/s Link Speed 2014 200 Gb/s Link Speed 2017 Gain Competitive Advantage Today Protect Your Future Power Consumption Per Switch Port Message Rate Switch Latency Scalability CPU efficiency 25% 44% 20% 2X Lower Higher Lower Higher 2015 Mellanox Technologies 18

End-to-End Interconnect Solutions for All Platforms Highest Performance and Scalability for X86, Power, GPU, ARM and FPGA-based Compute and Storage Platforms 10, 20, 25, 40, 50, 56 and 100Gb/s Speeds X86 Open POWER GPU ARM FPGA Smart Interconnect to Unleash The Power of All Compute Architectures 2015 Mellanox Technologies 19

Technology Roadmap One-Generation Lead over the Competition Mellanox 20Gbs 40Gbs 56Gbs 100Gbs 200Gbs Terascale Petascale Exascale 3 rd TOP500 2003 Virginia Tech (Apple) 1 st Roadrunner Mellanox Connected 2000 2005 2010 2015 2020 2015 Mellanox Technologies 20

Texas Advanced Computing Center (TACC) EDR InfiniBand System Mellanox End-to-End 100Gb/s EDR InfiniBand Accelerates Research And Applications for TACC Users 2015 Mellanox Technologies 21

Lenovo Innovation Center - EDR InfiniBand System LENOX, EDR InfiniBand connected system at the Lenovo HPC innovation center, Germany EDR InfiniBand provides ~25% higher system performance versus FDR InfiniBand on Graph500 At 128nodes EDR Mellanox Connected 2015 Mellanox Technologies 22

Shanghai Supercomputer Center EDR InfiniBand Magic Cube II supercomputer Sugon Cluster TC4600E system Mellanox end-to-end EDR InfiniBand EDR Mellanox Connected 2015 Mellanox Technologies 23

HPC Clouds Performance Demands Mellanox Solutions San Diego Supercomputing Center Comet System (2015) to Leverage Mellanox InfiniBand to Build HPC Cloud 2015 Mellanox Technologies 24

Mellanox Interconnect Advantages Proven, scalable and high performance end-to-end connectivity Flexible, support all compute architectures: x86, Power, ARM, GPU, FPGA etc. Standards-based (InfiniBand, Ethernet), supported by large eco-system Offloading architecture RDMA, application acceleration engines etc. Flexible topologies: Fat Tree, mesh, 3D Torus, Dragonfly+, etc. Converged I/O compute, storage, management on single fabric Backward and future compatible Speed-Up Your Present, Protect Your Future Paving The Road to Exascale Computing Together 2015 Mellanox Technologies 25

Texas Advanced Computing Center/Univ. of Texas - #10 Stampede system 6,000+ nodes (Dell), 462462 cores, Intel Phi co-processors 5.2 Petaflops Mellanox end-to-end FDR InfiniBand Petaflop Mellanox Connected 2015 Mellanox Technologies 26

NASA Ames Research Center - #13 Pleiades system SGI Altix ICE 20K InfiniBand nodes 3.4 sustained Petaflop performance Mellanox end-to-end FDR and QDR InfiniBand Supports variety of scientific and engineering projects Coupled atmosphere-ocean models Future space vehicle design Large-scale dark matter halos and galaxy evolution Petaflop Mellanox Connected 2015 Mellanox Technologies 27

Exploration & Production ENI S.p.A. - #19 HPC2 system IBM idataplex DX360M4 NVIDIA K20x GPUs 3.2 Petaflops sustained Petaflop performance Mellanox end-to-end FDR InfiniBand Petaflop Mellanox Connected 2015 Mellanox Technologies 28

Leibniz Rechenzentrum SuperMUC - #23 IBM idataplex and Intel Sandy Bridge 147456 cores Mellanox end-to-end FDR InfiniBand solutions 2.9 sustained Petaflop performance The fastest supercomputer in Europe 91% efficiency Petaflop Mellanox Connected 2015 Mellanox Technologies 29

Tokyo Institute of Technology - #25 TSUBAME 2.0, first Petaflop system in Japan 2.8 PF performance HP ProLiant SL390s G7 1400 servers Mellanox 40Gb/s InfiniBand Petaflop Mellanox Connected 2015 Mellanox Technologies 30

DOE/SC/Pacific Northwest National Laboratory - #27 Cascade system Atipa Visione IF442 Blade Server 2.5 sustained Petaflop performance Mellanox end-to-end InfiniBand FDR Intel Xeon Phi 5110P accelerator Petaflop Mellanox Connected 2015 Mellanox Technologies 31

Total Exploration Production - #33 Pangea system SGI Altix X system, 110400 cores Mellanox FDR InfiniBand 2 sustained Petaflop performance 91% efficiency Petaflop Mellanox Connected 2015 Mellanox Technologies 32

Grand Equipement National de Calcul Intensif - Centre Informatique National de l'enseignement Suprieur (GENCI-CINES) - #44 Occigen system 1.6 sustained Petaflop performance Bull bullx DLC Mellanox end-to-end FDR InfiniBand Petaflop Mellanox Connected 2015 Mellanox Technologies 33

Air Force Research Laboratory - #51 Spirit system SGI Altix X system, 74584 cores Mellanox FDR InfiniBand 1.4 sustained Petaflop performance 92.5% efficiency Petaflop Mellanox Connected 2015 Mellanox Technologies 34

CEA/TGCC-GENCI - #53 Bull Bullx B510, Intel Sandy Bridge 77184 cores Mellanox end-to-end FDR InfiniBand solutions 1.36 sustained Petaflop performance Petaflop Mellanox Connected 2015 Mellanox Technologies 35

Max-Planck-Gesellschaft MPI/IPP - # 56 IBM idataplex DX360M4 Mellanox end-to-end FDR InfiniBand solutions 1.3 sustained Petaflop performance Petaflop Mellanox Connected 2015 Mellanox Technologies 36

National Supercomputing Centre in Shenzhen (NSCS) - #57 Dawning TC3600 Blade Supercomputer 5200 nodes, 120640 cores, NVIDIA GPUs Mellanox end-to-end 40Gb/s InfiniBand solutions ConnectX-2 and IS5000 switches 1.27 sustained Petaflop performance The first Petaflop systems in China Petaflop Mellanox Connected 2015 Mellanox Technologies 37

NCAR (National Center for Atmospheric Research) - #58 Yellowstone system 1.3 sustained Petaflop performance 72,288 processor cores, 4,518 nodes (IBM) Mellanox end-to-end FDR InfiniBand, full fat tree, single plane Petaflop Mellanox Connected 2015 Mellanox Technologies 38

International Fusion Energy Research Centre (IFERC), EU(F4E) - Japan Broader Approach collaboration - #60 Bull Bullx B510, Intel Sandy Bridge 70560 cores 1.24 sustained Petaflop performance Mellanox end-to-end InfiniBand solutions Petaflop Mellanox Connected 2015 Mellanox Technologies 39

SURFsara - #69 The Cartesius system - the Dutch supercomputer Bull Bullx DLC B710/B720 system Mellanox end-to-end InfiniBand solutions 1.05 sustained Petaflop performance Petaflop Mellanox Connected 2015 Mellanox Technologies 40

Commissariat a l'energie Atomique (CEA) - #74 Tera 100, first Petaflop system in Europe - 1.05 PF performance 4,300 Bull S Series servers 140,000 Intel Xeon 7500 processing cores 300TB of central memory, 20PB of storage Mellanox 40Gb/s InfiniBand Petaflop Mellanox Connected 2015 Mellanox Technologies 41

Australian National University - #85 Fujitsu PRIMERGY CX250 S1 Mellanox FDR 56Gb/s InfiniBand end-to-end 980 Tflops performance 2015 Mellanox Technologies 42

Purdue University - #86 Conte system HP Cluster Platform SL250s Gen8 Intel Xeon E5-2670 8C 2.6GHz Intel Xeon Phi 5110P accelerator Total of 77,520 cores Mellanox FDR 56Gb/s InfiniBand end-to-end Mellanox Connect-IB InfiniBand adapters Mellanox MetroX long Haul InfiniBand solution 980 Tflops performance 2015 Mellanox Technologies 43

Barcelona Supercomputing Center - #92 MareNostrum 3 system 1.1 Petaflops peak performance ~50K cores, 91% efficiency Mellanox FDR InfiniBand 2015 Mellanox Technologies 44

Thank You