PCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters. from One Stop Systems (OSS)



Similar documents
PCI Express Impact on Storage Architectures and Future Data Centers. Ron Emerick, Oracle Corporation

PCI Express and Storage. Ron Emerick, Sun Microsystems

EDUCATION. PCI Express, InfiniBand and Storage Ron Emerick, Sun Microsystems Paul Millard, Xyratex Corporation

PCI Express Impact on Storage Architectures and Future Data Centers. Ron Emerick, Oracle Corporation

PCI Express Impact on Storage Architectures. Ron Emerick, Sun Microsystems

How PCI Express Works (by Tracy V. Wilson)

Storage Architectures. Ron Emerick, Oracle Corporation

PCI Express Impact on Storage Architectures and Future Data Centers

HPC Update: Engagement Model

Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband

GPU System Architecture. Alan Gray EPCC The University of Edinburgh

How to build a high speed PCI Express bus expansion system using the Max Express product family 1

Motherboard- based Servers versus ATCA- based Servers

Sun Microsystems Special Promotions for Education and Research January 9, 2007

PCI Express* Ethernet Networking

SUN HARDWARE FROM ORACLE: PRICING FOR EDUCATION

PCI Express IO Virtualization Overview

IBM System x family brochure

Architecting High-Speed Data Streaming Systems. Sujit Basu

UCS M-Series Modular Servers

Sun Constellation System: The Open Petascale Computing Architecture

PCI Express Supersedes SAS and SATA in Storage

HP PCIe IO Accelerator For Proliant Rackmount Servers And BladeSystems

SUN HARDWARE FROM ORACLE: PRICING FOR EDUCATION

QuickSpecs. HP Integrity cx2620 Server. Overview

The Bus (PCI and PCI-Express)

Communicating with devices

VTrak SATA RAID Storage System

A-CLASS The rack-level supercomputer platform with hot-water cooling

Cloud Data Center Acceleration 2015

HUAWEI Tecal E9000 Converged Infrastructure Blade Server

HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK

Maximizing Server Storage Performance with PCI Express and Serial Attached SCSI. Article for InfoStor November 2003 Paul Griffith Adaptec, Inc.

PRIMERGY server-based High Performance Computing solutions

Evaluation Report: HP Blade Server and HP MSA 16GFC Storage Evaluation

Arrow ECS sp. z o.o. Oracle Partner Academy training environment with Oracle Virtualization. Oracle Partner HUB

StorageBox High Performance NVMe JBOF

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.

UNIFIED HYBRID STORAGE. Performance, Availability and Scale for Any SAN and NAS Workload in Your Environment

Findings in High-Speed OrthoMosaic

Accelerating CFD using OpenFOAM with GPUs

A Smart Investment for Flexible, Modular and Scalable Blade Architecture Designed for High-Performance Computing.

CPU. Motherboard RAM. Power Supply. Storage. Optical Drives

power rid B ge C o m p u t e r

LS DYNA Performance Benchmarks and Profiling. January 2009

Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 SMB Direct

Optimized dual-use server and high-end workstation performance

Stovepipes to Clouds. Rick Reid Principal Engineer SGI Federal by SGI Federal. Published by The Aerospace Corporation with permission.

SMB Direct for SQL Server and Private Cloud

Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance

RAID. RAID 0 No redundancy ( AID?) Just stripe data over multiple disks But it does improve performance. Chapter 6 Storage and Other I/O Topics 29

Copyright 2013, Oracle and/or its affiliates. All rights reserved.

nappliance Network Virtualization Gateways

SATA Evolves SATA Specification v3.2 FMS 2013

4 Channel 6-Port SATA 6Gb/s PCIe RAID Host Card

PCI vs. PCI Express vs. AGP

Network Storage Appliance

OpenPOWER Outlook AXEL KOEHLER SR. SOLUTION ARCHITECT HPC

IOmark- VDI. Nimbus Data Gemini Test Report: VDI a Test Report Date: 6, September

New levels of efficiency and optimized design. The latest Intel CPUs. 2+1 expandability in UP 1U

Oracle Exadata: The World s Fastest Database Machine Exadata Database Machine Architecture

Terms of Reference Microsoft Exchange and Domain Controller/ AD implementation

APACHE HADOOP PLATFORM HARDWARE INFRASTRUCTURE SOLUTIONS

ECLIPSE Performance Benchmarks and Profiling. January 2009

1. Specifiers may alternately wish to include this specification in the following sections:

Optimizing GPU-based application performance for the HP for the HP ProLiant SL390s G7 server

Iron Networks Network Virtualization Gateways

præsentation oktober 2011

Integrating PCI Express into the PXI Backplane

SAPPHIRE TOXIC R9 270X 2GB GDDR5 WITH BOOST

Power Redundancy. I/O Connectivity Blade Compatibility KVM Support

NEC Micro Modular Server Introduction. NEC November 18,2014

The Advantages of Multi-Port Network Adapters in an SWsoft Virtual Environment

Post-production Video Editing Solution Guide with Microsoft SMB 3 File Serving AssuredSAN 4000

Headline in Arial Bold 30pt. The Need For Speed. Rick Reid Principal Engineer SGI

STORAGE HIGH SPEED INTERCONNECTS HIGH PERFORMANCE COMPUTING VISUALISATION GPU COMPUTING

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

Hardware RAID vs. Software RAID: Which Implementation is Best for my Application?

DD670, DD860, and DD890 Hardware Overview

How To Use An Amd Ramfire R7 With A 4Gb Memory Card With A 2Gb Memory Chip With A 3D Graphics Card With An 8Gb Card With 2Gb Graphics Card (With 2D) And A 2D Video Card With

SAN TECHNICAL - DETAILS/ SPECIFICATIONS

ioscale: The Holy Grail for Hyperscale

PCI Express SATA III RAID Controller Card with Mini-SAS Connector (SFF-8087) - HyperDuo SSD Tiering

Thematic Unit of Excellence on Computational Materials Science Solid State and Structural Chemistry Unit, Indian Institute of Science

Xserve Transition Guide. November 2010

Clusters: Mainstream Technology for CAE

Chapter 4 System Unit Components. Discovering Computers Your Interactive Guide to the Digital World

Cisco 7816-I5 Media Convergence Server

HUAWEI TECHNOLOGIES CO., LTD. HUAWEI FusionServer X6800 Data Center Server

PCI Express 2.0 SATA III RAID Controller Card with Internal Mini-SAS SFF-8087 Connector

PCI Express Basic Info *This info also applies to Laptops

Building Clusters for Gromacs and other HPC applications

Intel RAID SSD Cache Controller RCS25ZB040

Cisco for SAP HANA Scale-Out Solution on Cisco UCS with NetApp Storage

760 Veterans Circle, Warminster, PA Technical Proposal. Submitted by: ACT/Technico 760 Veterans Circle Warminster, PA

NGSS Data Consolidator (DC)

TS500-E5. Configuration Guide

HIGH AVAILABILITY IN THE DATACENTER

PCI Express Overview. And, by the way, they need to do it in less time.

HP 3PAR StoreServ 8000 Storage - what s new

Transcription:

PCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters from One Stop Systems (OSS)

PCIe Over Cable

PCIe provides greater performance 8 7 6 5 GBytes/s 4 3 2 1 0 EISA PCI 32/33 PCI 64/66 PCI-X 64/133 AGP 8X Gb E- net 10Gb E-net PCIe x1 PCIe x4 PCIe x8 PCIe x16 Bus

PCIe over Cable Comparison versus Ethernet Price 1 Gb Ethernet 10 Gb Ethernet PCIe over Cable 2.5Gb to 80Gb PCIe Performance 3 to 80 times faster than 1Gb Ethernet PCIe Cost Source OSS Adapters: $100 to $700 Cables: $30 to $300 Switches: $600 to $1,200 PCIe cables Heavy-duty well shielded cables All cables are cross-over style PCIe best suited for small, local networks Performance

PCIe vs Infiniband 40Gb/s InfiniBand Bundle 36-port 40Gb/s InfiniBand switch 8 Single-Port 40Gb/s PCI Express 2.0 InfiniBand HCA cards 8 2 meters long, copper cables, Total Price: $10K 80Gb/s PCIe Bundle 10-port 80Gb/s PCIe switch 80Gb host adapter and 2m cable Eight 20Gb/s PCI Express 2.0 host adapters with 2m cables Total Price: $6K

PCI Express Basics: Two Architectures Tree One CPU and multiple I/O boards Network Multi CPUs, Multi I/O Requires special H/W and S/W CPU CPU CPU CPU I/O I/O I/O Switch I/O I/O I/O

PCI Express Basics Lanes The Key to Performance Tx CPU I/O Rx x1 Lane Point-to-Point connections no arbitration Each lane consists of two differential pairs Separate Transmit and Receive pairs 2.5 or 5.0 Gb/s rate per pair Components auto-detect max clock rate Multiple lanes are used to increase performance x1 5 Gb/s x1 (pronounced by one ) x4 20 Gb/s x8 40 Gb/s x16 80 Gb/s

Tree Architecture I/O Expansion Host system RAID array PCIe I/O expansion system PCIe switch CPCI/CPCIe IO expansion

PCIe host cable adapters PCIe cost cable adapters PCIe x4 PCIe x8 PCIe x8 PCIe x16 PCIe cables

Upstream Adapters PC, laptop and industrial form factors

Downstream Adapters and Devices Creating downstream PCIe endpoints 9 PCIe board adapters 9 Backplane interface boards 9 Subsystems with PCIe cable inputs 9 Backplanes with PCIe cable inputs

Direct Attached Expansion Kits

Direct Attached Multi-port Switches Extends PCIe bus to multiple downstream sub-systems One upstream link to multiple downstream links Gen 1 and Gen 2 versions

HPC requires Substantial infrastructure with: Long-life, redundant servers GPU accelerators for math co-processing High speed storage or Solid State Disk (SSD) appliances High-speed connectivity

GPU Server AMD-based MB Eight GPU s/ssd s Could also used as NAS SSD appliance 1U PCIe switch One x16g2 uplink Nine x4g2 downlinks Two to eight servers Server to server communication 20Gb PCIe 10Gb Ethernet 20Gb Infiniband Multiple 1U or 2U GPU/SSD appliances Two to eight GPU's/SSD s per appliance 80Gb/s connectivity to server Network connectivity Server to Switch 80Gb/s Server to Server at 20Gb/s GPU/SSD appliance GPU/SSD appliance Architecture GPU Server 1U switch 1U servers Server to appliance connectivity at 80Gb/s

Latest server technology Longer life cycles from rugged servers reduce overall cost and downtime Reduced depth allows better fit in shallow racks Superior cooling and power Latest technology motherboards and processors provide wide range of processing options Dual 5500-series Nehalem quad-core or six-core processors Up to 96GB DRAM 2TB to 5TB disk drive capacity

1U PCIe switch One 80Gb upstream interface Nine 20Gb downstream interfaces Redundant servers with ExpressNet Server to server communication Windows or Linux OS Server redundancy Network connectivity at 10-20Gb/s 1U PCIe switch

Multiple GPUs support many users simultaneously in virtual networks Appliance provides necessary cooling and power not found in servers for optimal operation and significant reduction in downtime Hot swappable appliances provide redundancy GPU Appliance Server to Server at 20Gb/s GPU/SSD appliance GPU/SSD appliance 1U switch 1U servers

GPUs off-load high-end graphics and rendering GPUs from system processors GPUs provide rich media and 3D graphics to virtual desktops Typical GPUs AMD FireStream 9270 Processing power: single precision: 1.2 TFLOPS double precision: 240 GFLOPS AMD FireStream 9250 Processing power: single precision: 1.2 TFLOPS double precision: 240 GFLOPS ATI Radeon HD 5870 Processing power: single precision: 2.72 TFLOPS double precision: 544 GFLOPS

Solid State Disk (SSD) appliance 1U appliance One appliance per server Up to four 640GB SSD boards Server can access 2.5TB storage 2U appliance One appliance per two servers Up to eight 640GB SSD boards Each server can access 2.5TB storage 1U or 2U GPU/SSD appliance Up to 4 GPUs and 4 SSD boards per server 80Gb connectivity Each appliance can employ 4 to 8 Fusion-IO iodrive Duo boards (for example)

Combined GPU and SSD appliance Each 1U GPU/SSD appliance supports: Up to 2 GPU s per server Up to 1.25TB SSD storage per server Each 2U GPU/SSD appliance With 1U server supports: Up to 2 GPU s Up to 1.25TB SSD storage With 3U server supports: Up to 4 GPU s Up to 2.5TB SSD storage

2U Integrated Server 10TFLOP server integrates 9 Motherboard with dual AMD six-core processors Up to four double-wide AMD 9270 or HD5870 GPUs Istanbul-based motherboard Up to four double wide or eight single wide GPUs 9 2.72TFLOPS each or 10TFLOPS total processing power Dual 1500 watt power supplies Four SATA/SAS hot swappable disk drives Superior cooling 9 12 chassis fans 9 4 power supply fans Four hot swap disk drives Dual 1500-watt power supplies

Lower cost, lower latency, and less overhead than Infiniband Greater throughput and lower cost than 10Gb Ethernet Server to server communication over 20Gb PCI Express 80Gb connectivity between GPU appliance and server Network connectivity Server to Switch 80Gb/s PCIe connectivity Server to Server at 20Gb/s GPU/SSD appliance GPU/SSD appliance GPU Server/ 1U switch 1U servers Server to appliance connectivity at 80Gb/s

Latest technology server supporting up to eight GPU s and/or SSD s Long-life, latest technology, redundant 1U servers 20Gb/s PCI Express connectivity between servers 1U or 2U GPU/SSD appliances with 80Gb/s connection to severs The Future of HPC Clusters Network connectivity Server to Switch 80Gb/s Server to Server at 20Gb/s GPU/SSD appliance GPU/SSD appliance GPU Server 1U switch 1U servers Server to appliance connectivity at 80Gb/s

Direct Attached RAID Arrays 4-drive RAID PCIe x4 or 10Gb/s to RAID controller 12-drive RAID PCIe x8 or 20Gb/s to RAID controller 16-drive RAID PCIe x8 or 20Gb/s to RAID controller

Possible Storage Configurations Host cable adapter PCIe expansion kit with RAID board Downstream cable adapter and backplane

Possible Storage Configurations PCIe backplane with RAID board 3 PCIe x4 slots PCIex4 and x8 cable connectors

GPU Computing/RAID Sub-System 3U server 80Gb PCIe 80Gb PCIe over cable 1U PCIe switch 1U GPU Accelerator (includes 2 GPU s) 20Gb PCIe 20Gb PCIe 20Gb PCIe 20Gb PCIe 20Gb PCIe 20Gb PCIe 20Gb PCIe 20Gb PCIe RAID Array 1 RAID Array 3 RAID Array 5 RAID Array 7 RAID Array 2 RAID Array 4 RAID Array 6 RAID Array 8

Global GPU/RAID System 10Gb Ethernet switches to the outside world 1Gb Ethernet connections between nodes Node 1 Node 2 Node 3 Node 4 1Gb Ethernet connections to redundant switches

Summary PCI Express over cable operates from 10Gb/s to 80Gb/s Wide assortment of PCIe adapters, switches, and modules available PCI Express bus can be expanded from PC to IO or from PC to PC The expanding HPC market requires: redundant servers, multiple GPU s, high speed storage, and high-speed connectivity Servers with multiple high speed IO slots provide required bandwidth for GPU s and high speed storage GPU s and high speed storage can be attached to existing servers PCIe over cable provides the most economical high speed connectivity