Comet - High performance virtual clusters to support the long-tail of science.

Size: px
Start display at page:

Download "Comet - High performance virtual clusters to support the long-tail of science."

Transcription

1 Comet - High performance virtual clusters to support the long-tail of science. Philip M. Papadopoulos, Ph.D. San Diego Supercomputer Center California Institute for telecommunications and Information Technologies (Calit2) University of California, San Diego

2 HPC for the 99%

3 Comet is funded by U.S. National Science Foundation expand the use of high end resources to a much larger and more diverse community support the entire spectrum of NSF communities... promote a more comprehensive and balanced portfolio include research communities that are not users of traditional HPC systems. The long tail of science needs HPC

4 Cumulative Usage Fraction of All Jobs Charged in 2012 Millions of XD SUs Charged Jobs and SUs at various scales across NSF resources 99% of jobs run on NSF s HPC resources in 2012 used < 2048 cores 100% 90% 80% 70% 60% And consumed ~50% of the total core-hours across NSF resources 50% 40% 30% 20% 10% 0% One node Percentage of Jobs (Left Axis) SUs Charged (Right Axis) K 2K 4K 8K 16K Job Size (Cores) Job Size (Cores)

5 Comet Will Serve the 99%

6 Comet: System Characteristics Available 1Q 2015 Total flops ~1.9 PF (AVX2) Dell primary integrator Intel next-gen processors, former codename Haswell, with AVX2 Aeon storage vendor Mellanox FDR InfiniBand Standard compute nodes Dual 12-core Haswell processors 128 GB DDR4 DRAM (64 GB/socket!) 320 GB SSD (local scratch, VMs) GPU nodes Four NVIDIA GPUs/node Large-memory nodes (2Q 2015) 1.5 TB DRAM Four Haswell processors/node Hybrid fat-tree topology FDR (56 Gbps) InfiniBand Rack-level (72 nodes) full bisection bandwidth 4:1 oversubscription inter-rack Performance Storage 7 PB, 200 GB/s Scratch & Persistent Storage Durable Storage (reliability) 6 PB, 100 GB/s Gateway hosting nodes and VM image repository 100 Gbps external connectivity to Internet2 & ESNet

7 Comet Architecture Node-Local 72 HSWL Storage 320 GB 18 N racks 72 HSWL 320 GB 72 HSWL N GPU 4 Large- Memory 7x 36-port FDR in each rack wired as full fat-tree. 4:1 over subscription between racks. 18 FDR FDR 36p FDR 36p 18 Mid-tier Performance Storage 7 PB, 200 GB/s FDR 72 FDR 40GbE IB Core (2x) 72 Bridge (4x) 72 Arista 40GbE (2x) GbE 10GbE Durable Storage 6 PB, 100 GB/s R&E Network Access Data Movers Internet 2 Juniper 100 Gbps Arista 40GbE (2x) Data Mover (4x) Additional Support Components (not shown for clarity) NFS Servers, Virtual Image Repository, Gateway/Portal Hosting Nodes, Login Nodes, Ethernet Management Network, Rocks Management Nodes

8 Design Decision - Network Each Rack 72 nodes (144 CPUs, 1728 Cores) Fully-connected FDR IB (2-level Clos-Topology) 144 In-rack Cables 4:1 oversubscription between racks 18 inter-rack cables Supports the large majority of jobs with no performance degradation 3-level network for complete cluster worst-case latency similar to a much smaller cluster Reduced cost.

9 SSDs building on Gordon success Based on our experiences with Gordon, a number of applications will benefits from continued access to flash Applications that generate large numbers of temp files Computational finance analysis of multiple markets (NASDAQ, etc.) Text analytics word correlations in Google Ngram data Computational chemistry codes that write one- and twoelectron integral files to scratch Structural mechanics codes (e.g. Abaqus), which generate stiffness matrices that don t fit into memory

10 Large memory nodes While most user applications will run well on the standard compute nodes, a few domains will benefit from the large memory (1.5 TB nodes) De novo genome assembly: ALLPATHS-LG, SOAPdenovo, Velvet Finite-element calculations: Abaqus Visualization of large data sets

11 GPU nodes Comet s GPU nodes will serve a number of domains Molecular dynamics applications have been one of the biggest GPU success stories. Packages include Amber, CHARMM, Gromacs and NAMD Applications that depend heavily on linear algebra Image and signal processing

12 Key Comet Strategies Target modest-scale users and new users/communities: goal of 10,000 users/year! Support capacity computing, with a system optimized for small/modest-scale jobs and quicker resource response using allocation/scheduling policies Build upon and expand efforts with Science Gateways, encouraging gateway usage and hosting via software and operating policies Provide a virtualized environment to support development of customized software stacks, virtual environments, and project control of workspaces

13 Comet will serve a large number of users, including new communities/disciplines Allocations/scheduling policies to optimize for high throughput of many modest-scale jobs (leveraging Trestles experience) Optimized for rack-level jobs but cross-rack jobs feasible Optimized for throughput (ala Trestles) Per-project allocations caps to ensure large numbers of users Rapid access for start-ups with one-day account generation Limits on job sizes, with possibility of exceptions Gateway-friendly environment: Science gateways reach large communities w/ easy user access e.g. CIPRES gateway alone currently accounts for ~25% of all users of NSF resources, with 3,000 new users/year and ~5,000 users/year Virtualization provides low barriers to entry (see later charts)

14 Changing the face of XSEDE HPC users System design and policies Allocations, scheduling and security policies which favor gateways Support gateway middleware and gateway hosting machines Customized environments with high-performance virtualization Flexible allocations for bursty usage patterns Shared node runs for small jobs, user-settable reservations Third party apps Leverage and augment investments elsewhere FutureGrid experience, image packaging, training, on-ramp XSEDE (ECSS NIP & Gateways, TEOS, Campus Champions) Build off established successes supporting new communities Example-based documentation in Comet focus areas Unique HPC University contributions to enable community growth

15 Virtualization Environment Leveraging expertise of Indiana U/ FutureGrid team VM jobs scheduled just like batch jobs (not conventional cloud environment with immediate elastic access) VMs will be easy on-ramp for new users/communities, including low porting time Flexible software environments for new communities and apps VM repository/library Virtual HPC cluster (multi-node) with near-native IB latency and minimal overhead (SRIOV)

16 Single Root I/O Virtualization in HPC Problem: complex workflows demand increasing flexibility from HPC platforms Pro: Virtualization flexibility Con: Virtualization IO performance loss (e.g., excessive DMA interrupts) Solution: SR-IOV and Mellanox ConnectX-3 InfiniBand HCAs One physical function (PF) multiple virtual functions (VF), each with own DMA streams, memory space, interrupts Allows DMA to bypass hypervisor to VMs

17 More on Single Root IO Virtualization PCIe is the I/O bus of modern x86 servers The I/O controller is integrated into microprocessors The I/O complex is called single-root if there is only one controller for the bus I/O Virtualization Allow a single I/O device (e.g. Network, Disk Controller) to appear as multiple independent I/O devices These are called virtual functions Each virtual function can be independently controlled

18 Benchmark comparisons of SR-IOV Cluster v AWS (pre-haswell) Hardware/Software Configuration Native, SR-IOV Platform Rocks 6.1 (EL6) Virtualization via kvm CPUs 2x Xeon E (2.2GHz) 16 cores per node Amazon EC2 Amazon Linux (EL6) cc2.8xlarge Instances 2x Xeon E (2.6GHz) 16 cores per node RAM 64 GB DDR3 DRAM 60.5 DDR3 DRAM Interconnect QDR4X InfiniBand Mellanox ConnectX-3 (MT27500) Intel VT-d, SR-IOV enabled in firmware, kernel, drivers mlx4_core 1.1 Mellanox OFED 2.0 HCA firmware GbE common placement group

19 SRIOV Latency approaches Native HW SR-IOV < 30% overhead for Messages < 128 bytes < 10% overhead for eager send/recv Overhead 0% for bandwidth-limited regime Amazon EC2 > 5000% worse latency Time dependent (noisy) 19 OSU Microbenchmarks (3.9, osu_latency)

20 Bandwidth Unimpaired Native vs. SRIOV SR-IOV < 2% bandwidth loss over entire range > 95% peak bandwidth Amazon EC2 < 35% peak bandwidth 900% to 2500% worse bandwidth than virtualized InfiniBand 20 OSU Microbenchmarks (3.9, osu_bw)

21 Weather Modeling 15% Overhead 96-core (6-node) calculation Nearest-neighbor communication Scalable algorithms SR-IOV incurs modest (15%) performance hit...but still still 20% faster *** than Amazon WRF hr forecast *** 20% faster despite SR-IOV cluster having 20% slower CPUs

22 Quantum ESPRESSO: 28% overhead 48-core (3 node) calculation CG matrix inversion (irregular comm.) 3D FFT matrix transposes (All-to-all communication) 28% slower w/ SR-IOV SR-IOV still > 500% faster *** than EC2 Quantum Espresso DEISA AUSURF112 benchmark *** 20% faster despite SR-IOV cluster having 20% slower CPUs

23 If bandwidth unimpaired why Falloff for App performance Latency. Measured microbenchmark for latency shows 10-30% overhead. However, this is variable. Can be as much as 100%. Why? Hypervisor/Physical node scheduling. Expect advances in software to improve this.

24 SR-IOV is a huge step forward in highperformance virtualization Shows substantial improvement in latency over Amazon EC2, and it provides nearly zero bandwidth overhead Benchmark application performance confirms significant improvement over EC2 SR-IOV lowers performance barrier to virtualizing the interconnect and makes fully virtualized HPC clusters viable Comet will deliver virtualized HPC to new/non-traditional communities that need flexibility without major loss of performance

25 High-Performance Virtualization on Comet Mellanox FDR InfiniBand HCAs with SR-IOV Rocks and KVM to manage Virtual Machines and Clusters Flexibility to support complex science gateways and web-based workflow engines Custom compute appliances and virtual clusters developed with FutureGrid and their existing expertise

26 Virtual Clusters Overlay Physical Cluster with User-Owned High Performance Clusters Virtual Cluster 1 Virtual Cluster 2

27 Virtual Cluster Characteristics User-Owned and Defined Looks like bare metal cluster to user Low overhead (latency and bandwidth) for a virtualized Infiniband interface Single Root IO Virtualization Schedule compatibility with standard HPC batch jobs A node of virtual cluster is a virtual machine. That VM looks like one element of a parallel job to the scheduler Persistence of Disk State of virtual nodes across multiple boot sequences

28 Some Interesting Logistics for Virtual Clusters Scheduling Can we co-schedule virtual cluster nodes with regular HPC jobs? How do we efficiently handle disk images the make up the nodes of the virtual cluster?

29 Managing the Physical Infrastructure

30 Virtual frontend container Virtual Cluster Anatomy Private network segmentation: 10GB using VLAN Infiniband PKEY Generic compute nodes Public network Virtual Frontend Private network 10GB VLAN MMM PKEY LLLL VLAN NNN PKEY JJJJ VLAN MMM PKEY LLLL Virtual Compute Virtual Compute Virtual Frontend VLAN NNN PKEY JJJJ Infiniband VLAN NNN PKEY JJJJ Virtual Compute

31 VM Disk Management Each VM gets a 36 GB disk (Small SCSI) Disk images are persistent through reboots Two central NASes store all disk images VMs can be allocated on all compute nodes dependent on availability (scheduler) Two solutions: o o iscsi (Network mounted disk) Disk replication on nodes

32 VM Disk Management iscsi NAS This is what OpenStack Supports Big Issue: Bandwidth Bottleneck at NAS Compute nodes Targets Virtual computex iqn com.nas-0-0-vm-compute-x

33 A hybrid solution via replication Initial boot of any cluster node uses an iscsi disk(call this a node disk) on the NAS During normal operation, Comet moves a node disk to the physical host that is running the node VM. And then disconnects from the NAS o All Node disk operation is local to the physical host o Fundamentally enables scale out w/o a $1M NAS At Shutdown, any changes made to the node disk (now on the physical host) are migrated back to the NAS, ready for next boot

34 1.a Init Disk NAS iscsi mount on NAS enables virtual compute node to boot immediately. Read operations from NAS Write operations to local disk Compute nodes Targets Virtual compute-x iqn com.nas-0-0-vm-compute-x Replicate Disk

35 NAS 1.b Init Disk During boot, the disk image on the NAS is migrated to the physical host. Read-only and read/write are then merged into one local disk iscsi mount is disconnected Compute nodes Targets Virtual compute-x

36 NAS 2. Steady State During normal operation Node disk is snapshot Incremental snapshots sent to NAS (replicate back to NAS) Timing/load/experiment will tell us how often we can do this Compute nodes Targets Virtual compute-x

37 3. Release Disk NAS Compute nodes Targets Power off Virtual compute-x At shutdown, any unsynched changes are send back to NAS When the last snapshot is sent, the Virtual compute node can be rebooted on another system

38 Software to implement is under development Rocks Roll so that it can be a part of any physical Rocks-defined cluster Uses ZFS for disk image storage on NAS and hosting nodes RabbitMQ AMQP (Asynchronous Message Q Protocol) Pika - library for communication with RabbitMQ from Python

39 Full Virtualization isn t the only choice: Containers Comet supports fully-virtualized clusters OS of cluster can be almost anything: Windows, Linux, Solaris, (not Mac-OS) Containers are a different way to virtualize the file system and some other elements Network, inter-process communication In Solaris for more than a decade Newly popular in Linux with the docker project Containers must run the same kernel as the host operating system Networking and device support not as flexible (yet) Containers have much smaller software footprints

40 Full vs. Container Virtualization Full Virtualization (KVM) Container Virtualization (Docker) Physical Host Kernel HW Virtual Host Virtual Host Physical Network Fully-Virtualized have independent kernels/os. Hardware is universal. Memory/CPU defined by def of virtual HW Physical Host Kernel partial /proc Network Bridge Container 1 Container 2 Physical Network Containers Inherits the Host kernel and elements of its hardware. Need cgroups to limit container memory /cpu usage

41 Still need to Configure Virtual Systems Why Docker (container) virtualization popular Very space efficient if the changes to the base OS file system are small. Changes can be 100 s of megabytes instead of Gbytes Network (latency) performance and/or network topology is less important (topology needed for virtual clusters) Given how quickly (and relatively lightweight) Docker brings up virtual environments, this will be addressed A system is DEFINED by the contents of the file system System libraries, application code, configuration Docker and Full-virtualization need to be configured (No Free Lunch)

42 Running: Almost the same ~ Rocks-created server running in Docker Container Notice: uptime, processor and date created rocks-created server running fully-virtualized (KVM) Notice: # cpus, uptime

43 Summary Comet: HPC for the 99% Expand capability to enable virtual clusters Not generic cloud computing Advances in virtualization techniques continue Comet will support fully-virtualized clusters Container-based virtualized (Docker) become quite popular

44

SR-IOV: Performance Benefits for Virtualized Interconnects!

SR-IOV: Performance Benefits for Virtualized Interconnects! SR-IOV: Performance Benefits for Virtualized Interconnects! Glenn K. Lockwood! Mahidhar Tatineni! Rick Wagner!! July 15, XSEDE14, Atlanta! Background! High Performance Computing (HPC) reaching beyond traditional

More information

Hadoop on the Gordon Data Intensive Cluster

Hadoop on the Gordon Data Intensive Cluster Hadoop on the Gordon Data Intensive Cluster Amit Majumdar, Scientific Computing Applications Mahidhar Tatineni, HPC User Services San Diego Supercomputer Center University of California San Diego Dec 18,

More information

SMB Direct for SQL Server and Private Cloud

SMB Direct for SQL Server and Private Cloud SMB Direct for SQL Server and Private Cloud Increased Performance, Higher Scalability and Extreme Resiliency June, 2014 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server

More information

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency WHITE PAPER Solving I/O Bottlenecks to Enable Superior Cloud Efficiency Overview...1 Mellanox I/O Virtualization Features and Benefits...2 Summary...6 Overview We already have 8 or even 16 cores on one

More information

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014 Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet Anand Rangaswamy September 2014 Storage Developer Conference Mellanox Overview Ticker: MLNX Leading provider of high-throughput,

More information

High Performance OpenStack Cloud. Eli Karpilovski Cloud Advisory Council Chairman

High Performance OpenStack Cloud. Eli Karpilovski Cloud Advisory Council Chairman High Performance OpenStack Cloud Eli Karpilovski Cloud Advisory Council Chairman Cloud Advisory Council Our Mission Development of next generation cloud architecture Providing open specification for cloud

More information

State of the Art Cloud Infrastructure

State of the Art Cloud Infrastructure State of the Art Cloud Infrastructure Motti Beck, Director Enterprise Market Development WHD Global I April 2014 Next Generation Data Centers Require Fast, Smart Interconnect Software Defined Networks

More information

Perspec'ves on SDN. Roadmap to SDN Workshop, LBL

Perspec'ves on SDN. Roadmap to SDN Workshop, LBL Perspec'ves on SDN Roadmap to SDN Workshop, LBL Philip Papadopoulos San Diego Supercomputer Center California Ins8tute for Telecommunica8ons and Informa8on Technology University of California, San Diego

More information

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products MaxDeploy Ready Hyper- Converged Virtualization Solution With SanDisk Fusion iomemory products MaxDeploy Ready products are configured and tested for support with Maxta software- defined storage and with

More information

Cloud Computing. Alex Crawford Ben Johnstone

Cloud Computing. Alex Crawford Ben Johnstone Cloud Computing Alex Crawford Ben Johnstone Overview What is cloud computing? Amazon EC2 Performance Conclusions What is the Cloud? A large cluster of machines o Economies of scale [1] Customers use a

More information

Achieving a High-Performance Virtual Network Infrastructure with PLUMgrid IO Visor & Mellanox ConnectX -3 Pro

Achieving a High-Performance Virtual Network Infrastructure with PLUMgrid IO Visor & Mellanox ConnectX -3 Pro Achieving a High-Performance Virtual Network Infrastructure with PLUMgrid IO Visor & Mellanox ConnectX -3 Pro Whitepaper What s wrong with today s clouds? Compute and storage virtualization has enabled

More information

Simplifying Big Data Deployments in Cloud Environments with Mellanox Interconnects and QualiSystems Orchestration Solutions

Simplifying Big Data Deployments in Cloud Environments with Mellanox Interconnects and QualiSystems Orchestration Solutions Simplifying Big Data Deployments in Cloud Environments with Mellanox Interconnects and QualiSystems Orchestration Solutions 64% of organizations were investing or planning to invest on Big Data technology

More information

Enabling Technologies for Distributed and Cloud Computing

Enabling Technologies for Distributed and Cloud Computing Enabling Technologies for Distributed and Cloud Computing Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF Multi-core CPUs and Multithreading

More information

Enabling Technologies for Distributed Computing

Enabling Technologies for Distributed Computing Enabling Technologies for Distributed Computing Dr. Sanjay P. Ahuja, Ph.D. Fidelity National Financial Distinguished Professor of CIS School of Computing, UNF Multi-core CPUs and Multithreading Technologies

More information

Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 SMB Direct

Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 SMB Direct Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 Direct Increased Performance, Scaling and Resiliency July 2012 Motti Beck, Director, Enterprise Market Development [email protected]

More information

Scientific Computing Data Management Visions

Scientific Computing Data Management Visions Scientific Computing Data Management Visions ELI-Tango Workshop Szeged, 24-25 February 2015 Péter Szász Group Leader Scientific Computing Group ELI-ALPS Scientific Computing Group Responsibilities Data

More information

Realizing the next step in storage/converged architectures

Realizing the next step in storage/converged architectures Realizing the next step in storage/converged architectures Imagine having the same data access and processing power of an entire Facebook like datacenter in a single rack of servers Flash Memory Summit

More information

Building a Top500-class Supercomputing Cluster at LNS-BUAP

Building a Top500-class Supercomputing Cluster at LNS-BUAP Building a Top500-class Supercomputing Cluster at LNS-BUAP Dr. José Luis Ricardo Chávez Dr. Humberto Salazar Ibargüen Dr. Enrique Varela Carlos Laboratorio Nacional de Supercómputo Benemérita Universidad

More information

Zadara Storage Cloud A whitepaper. @ZadaraStorage

Zadara Storage Cloud A whitepaper. @ZadaraStorage Zadara Storage Cloud A whitepaper @ZadaraStorage Zadara delivers two solutions to its customers: On- premises storage arrays Storage as a service from 31 locations globally (and counting) Some Zadara customers

More information

Accelerating I/O- Intensive Applications in IT Infrastructure with Innodisk FlexiArray Flash Appliance. Alex Ho, Product Manager Innodisk Corporation

Accelerating I/O- Intensive Applications in IT Infrastructure with Innodisk FlexiArray Flash Appliance. Alex Ho, Product Manager Innodisk Corporation Accelerating I/O- Intensive Applications in IT Infrastructure with Innodisk FlexiArray Flash Appliance Alex Ho, Product Manager Innodisk Corporation Outline Innodisk Introduction Industry Trend & Challenge

More information

HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK

HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK Steve Oberlin CTO, Accelerated Computing US to Build Two Flagship Supercomputers SUMMIT SIERRA Partnership for Science 100-300 PFLOPS Peak Performance

More information

Application and Micro-benchmark Performance using MVAPICH2-X on SDSC Gordon Cluster

Application and Micro-benchmark Performance using MVAPICH2-X on SDSC Gordon Cluster Application and Micro-benchmark Performance using MVAPICH2-X on SDSC Gordon Cluster Mahidhar Tatineni ([email protected]) MVAPICH User Group Meeting August 27, 2014 NSF grants: OCI #0910847 Gordon: A Data

More information

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes Anthony Kenisky, VP of North America Sales About Appro Over 20 Years of Experience 1991 2000 OEM Server Manufacturer 2001-2007

More information

Sun Constellation System: The Open Petascale Computing Architecture

Sun Constellation System: The Open Petascale Computing Architecture CAS2K7 13 September, 2007 Sun Constellation System: The Open Petascale Computing Architecture John Fragalla Senior HPC Technical Specialist Global Systems Practice Sun Microsystems, Inc. 25 Years of Technical

More information

Overview: X5 Generation Database Machines

Overview: X5 Generation Database Machines Overview: X5 Generation Database Machines Spend Less by Doing More Spend Less by Paying Less Rob Kolb Exadata X5-2 Exadata X4-8 SuperCluster T5-8 SuperCluster M6-32 Big Memory Machine Oracle Exadata Database

More information

Toward a practical HPC Cloud : Performance tuning of a virtualized HPC cluster

Toward a practical HPC Cloud : Performance tuning of a virtualized HPC cluster Toward a practical HPC Cloud : Performance tuning of a virtualized HPC cluster Ryousei Takano Information Technology Research Institute, National Institute of Advanced Industrial Science and Technology

More information

SURFsara HPC Cloud Workshop

SURFsara HPC Cloud Workshop SURFsara HPC Cloud Workshop doc.hpccloud.surfsara.nl UvA workshop 2016-01-25 UvA HPC Course Jan 2016 Anatoli Danezi, Markus van Dijk [email protected] Agenda Introduction and Overview (current

More information

SR-IOV In High Performance Computing

SR-IOV In High Performance Computing SR-IOV In High Performance Computing Hoot Thompson & Dan Duffy NASA Goddard Space Flight Center Greenbelt, MD 20771 [email protected] [email protected] www.nccs.nasa.gov Focus on the research side

More information

Mellanox Academy Online Training (E-learning)

Mellanox Academy Online Training (E-learning) Mellanox Academy Online Training (E-learning) 2013-2014 30 P age Mellanox offers a variety of training methods and learning solutions for instructor-led training classes and remote online learning (e-learning),

More information

Enabling High performance Big Data platform with RDMA

Enabling High performance Big Data platform with RDMA Enabling High performance Big Data platform with RDMA Tong Liu HPC Advisory Council Oct 7 th, 2014 Shortcomings of Hadoop Administration tooling Performance Reliability SQL support Backup and recovery

More information

CON9577 Performance Optimizations for Cloud Infrastructure as a Service

CON9577 Performance Optimizations for Cloud Infrastructure as a Service CON9577 Performance Optimizations for Cloud Infrastructure as a Service John Falkenthal, Software Development Sr. Director - Oracle VM SPARC and VirtualBox Jeff Savit, Senior Principal Technical Product

More information

Cloud Computing through Virtualization and HPC technologies

Cloud Computing through Virtualization and HPC technologies Cloud Computing through Virtualization and HPC technologies William Lu, Ph.D. 1 Agenda Cloud Computing & HPC A Case of HPC Implementation Application Performance in VM Summary 2 Cloud Computing & HPC HPC

More information

Solid State Storage in Massive Data Environments Erik Eyberg

Solid State Storage in Massive Data Environments Erik Eyberg Solid State Storage in Massive Data Environments Erik Eyberg Senior Analyst Texas Memory Systems, Inc. Agenda Taxonomy Performance Considerations Reliability Considerations Q&A Solid State Storage Taxonomy

More information

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION A DIABLO WHITE PAPER AUGUST 2014 Ricky Trigalo Director of Business Development Virtualization, Diablo Technologies

More information

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks WHITE PAPER July 2014 Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks Contents Executive Summary...2 Background...3 InfiniteGraph...3 High Performance

More information

FLOW-3D Performance Benchmark and Profiling. September 2012

FLOW-3D Performance Benchmark and Profiling. September 2012 FLOW-3D Performance Benchmark and Profiling September 2012 Note The following research was performed under the HPC Advisory Council activities Participating vendors: FLOW-3D, Dell, Intel, Mellanox Compute

More information

Open Cirrus: Towards an Open Source Cloud Stack

Open Cirrus: Towards an Open Source Cloud Stack Open Cirrus: Towards an Open Source Cloud Stack Karlsruhe Institute of Technology (KIT) HPC2010, Cetraro, June 2010 Marcel Kunze KIT University of the State of Baden-Württemberg and National Laboratory

More information

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance 11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu

More information

Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers

Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers White Paper rev. 2015-11-27 2015 FlashGrid Inc. 1 www.flashgrid.io Abstract Oracle Real Application Clusters (RAC)

More information

Virtual Compute Appliance Frequently Asked Questions

Virtual Compute Appliance Frequently Asked Questions General Overview What is Oracle s Virtual Compute Appliance? Oracle s Virtual Compute Appliance is an integrated, wire once, software-defined infrastructure system designed for rapid deployment of both

More information

Can High-Performance Interconnects Benefit Memcached and Hadoop?

Can High-Performance Interconnects Benefit Memcached and Hadoop? Can High-Performance Interconnects Benefit Memcached and Hadoop? D. K. Panda and Sayantan Sur Network-Based Computing Laboratory Department of Computer Science and Engineering The Ohio State University,

More information

I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology

I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology Reduce I/O cost and power by 40 50% Reduce I/O real estate needs in blade servers through consolidation Maintain

More information

LS DYNA Performance Benchmarks and Profiling. January 2009

LS DYNA Performance Benchmarks and Profiling. January 2009 LS DYNA Performance Benchmarks and Profiling January 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox HPC Advisory Council Cluster Center The

More information

High Performance Computing in CST STUDIO SUITE

High Performance Computing in CST STUDIO SUITE High Performance Computing in CST STUDIO SUITE Felix Wolfheimer GPU Computing Performance Speedup 18 16 14 12 10 8 6 4 2 0 Promo offer for EUC participants: 25% discount for K40 cards Speedup of Solver

More information

Workshop on Parallel and Distributed Scientific and Engineering Computing, Shanghai, 25 May 2012

Workshop on Parallel and Distributed Scientific and Engineering Computing, Shanghai, 25 May 2012 Scientific Application Performance on HPC, Private and Public Cloud Resources: A Case Study Using Climate, Cardiac Model Codes and the NPB Benchmark Suite Peter Strazdins (Research School of Computer Science),

More information

Boas Betzler. Planet. Globally Distributed IaaS Platform Examples AWS and SoftLayer. November 9, 2015. 20014 IBM Corporation

Boas Betzler. Planet. Globally Distributed IaaS Platform Examples AWS and SoftLayer. November 9, 2015. 20014 IBM Corporation Boas Betzler Cloud IBM Distinguished Computing Engineer for a Smarter Planet Globally Distributed IaaS Platform Examples AWS and SoftLayer November 9, 2015 20014 IBM Corporation Building Data Centers The

More information

Configuration Maximums

Configuration Maximums Topic Configuration s VMware vsphere 5.1 When you select and configure your virtual and physical equipment, you must stay at or below the maximums supported by vsphere 5.1. The limits presented in the

More information

MaxDeploy Hyper- Converged Reference Architecture Solution Brief

MaxDeploy Hyper- Converged Reference Architecture Solution Brief MaxDeploy Hyper- Converged Reference Architecture Solution Brief MaxDeploy Reference Architecture solutions are configured and tested for support with Maxta software- defined storage and with industry

More information

IOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org

IOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org IOmark- VDI Nimbus Data Gemini Test Report: VDI- 130906- a Test Copyright 2010-2013 Evaluator Group, Inc. All rights reserved. IOmark- VDI, IOmark- VDI, VDI- IOmark, and IOmark are trademarks of Evaluator

More information

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays Red Hat Performance Engineering Version 1.0 August 2013 1801 Varsity Drive Raleigh NC

More information

StACC: St Andrews Cloud Computing Co laboratory. A Performance Comparison of Clouds. Amazon EC2 and Ubuntu Enterprise Cloud

StACC: St Andrews Cloud Computing Co laboratory. A Performance Comparison of Clouds. Amazon EC2 and Ubuntu Enterprise Cloud StACC: St Andrews Cloud Computing Co laboratory A Performance Comparison of Clouds Amazon EC2 and Ubuntu Enterprise Cloud Jonathan S Ward StACC (pronounced like 'stack') is a research collaboration launched

More information

Building a Scalable Storage with InfiniBand

Building a Scalable Storage with InfiniBand WHITE PAPER Building a Scalable Storage with InfiniBand The Problem...1 Traditional Solutions and their Inherent Problems...2 InfiniBand as a Key Advantage...3 VSA Enables Solutions from a Core Technology...5

More information

Broadcom Ethernet Network Controller Enhanced Virtualization Functionality

Broadcom Ethernet Network Controller Enhanced Virtualization Functionality White Paper Broadcom Ethernet Network Controller Enhanced Virtualization Functionality Advancements in VMware virtualization technology coupled with the increasing processing capability of hardware platforms

More information

Optimizing Web Infrastructure on Intel Architecture

Optimizing Web Infrastructure on Intel Architecture White Paper Intel Processors for Web Architectures Optimizing Web Infrastructure on Intel Architecture Executive Summary and Purpose of this Paper Today s data center infrastructures must adapt to mobile

More information

Datacenter Operating Systems

Datacenter Operating Systems Datacenter Operating Systems CSE451 Simon Peter With thanks to Timothy Roscoe (ETH Zurich) Autumn 2015 This Lecture What s a datacenter Why datacenters Types of datacenters Hyperscale datacenters Major

More information

Where IT perceptions are reality. Test Report. OCe14000 Performance. Featuring Emulex OCe14102 Network Adapters Emulex XE100 Offload Engine

Where IT perceptions are reality. Test Report. OCe14000 Performance. Featuring Emulex OCe14102 Network Adapters Emulex XE100 Offload Engine Where IT perceptions are reality Test Report OCe14000 Performance Featuring Emulex OCe14102 Network Adapters Emulex XE100 Offload Engine Document # TEST2014001 v9, October 2014 Copyright 2014 IT Brand

More information

Oracle Maximum Availability Architecture with Exadata Database Machine. Morana Kobal Butković Principal Sales Consultant Oracle Hrvatska

Oracle Maximum Availability Architecture with Exadata Database Machine. Morana Kobal Butković Principal Sales Consultant Oracle Hrvatska Oracle Maximum Availability Architecture with Exadata Database Machine Morana Kobal Butković Principal Sales Consultant Oracle Hrvatska MAA is Oracle s Availability Blueprint Oracle s MAA is a best practices

More information

Hadoop: Embracing future hardware

Hadoop: Embracing future hardware Hadoop: Embracing future hardware Suresh Srinivas @suresh_m_s Page 1 About Me Architect & Founder at Hortonworks Long time Apache Hadoop committer and PMC member Designed and developed many key Hadoop

More information

New Data Center architecture

New Data Center architecture New Data Center architecture DigitPA Conference 2010, Rome, Italy Silvano Gai Consulting Professor Stanford University Fellow Cisco Systems 1 Cloud Computing The current buzzword ;-) Your computing is

More information

Deep Dive on SimpliVity s OmniStack A Technical Whitepaper

Deep Dive on SimpliVity s OmniStack A Technical Whitepaper Deep Dive on SimpliVity s OmniStack A Technical Whitepaper By Hans De Leenheer and Stephen Foskett August 2013 1 Introduction This paper is an in-depth look at OmniStack, the technology that powers SimpliVity

More information

Increasing Flash Throughput for Big Data Applications (Data Management Track)

Increasing Flash Throughput for Big Data Applications (Data Management Track) Scale Simplify Optimize Evolve Increasing Flash Throughput for Big Data Applications (Data Management Track) Flash Memory 1 Industry Context Addressing the challenge A proposed solution Review of the Benefits

More information

Microsoft Windows Server Hyper-V in a Flash

Microsoft Windows Server Hyper-V in a Flash Microsoft Windows Server Hyper-V in a Flash Combine Violin s enterprise-class storage arrays with the ease and flexibility of Windows Storage Server in an integrated solution to achieve higher density,

More information

The Future of Computing Cisco Unified Computing System. Markus Kunstmann Channels Systems Engineer

The Future of Computing Cisco Unified Computing System. Markus Kunstmann Channels Systems Engineer The Future of Computing Cisco Unified Computing System Markus Kunstmann Channels Systems Engineer 2009 Cisco Systems, Inc. All rights reserved. Data Centers Are under Increasing Pressure Collaboration

More information

SQL Server Virtualization

SQL Server Virtualization The Essential Guide to SQL Server Virtualization S p o n s o r e d b y Virtualization in the Enterprise Today most organizations understand the importance of implementing virtualization. Virtualization

More information

Masters Project Proposal

Masters Project Proposal Masters Project Proposal Virtual Machine Storage Performance Using SR-IOV by Michael J. Kopps Committee Members and Signatures Approved By Date Advisor: Dr. Jia Rao Committee Member: Dr. Xiabo Zhou Committee

More information

Technical Paper. Moving SAS Applications from a Physical to a Virtual VMware Environment

Technical Paper. Moving SAS Applications from a Physical to a Virtual VMware Environment Technical Paper Moving SAS Applications from a Physical to a Virtual VMware Environment Release Information Content Version: April 2015. Trademarks and Patents SAS Institute Inc., SAS Campus Drive, Cary,

More information

SURFsara HPC Cloud Workshop

SURFsara HPC Cloud Workshop SURFsara HPC Cloud Workshop www.cloud.sara.nl Tutorial 2014-06-11 UvA HPC and Big Data Course June 2014 Anatoli Danezi, Markus van Dijk [email protected] Agenda Introduction and Overview (current

More information

Stovepipes to Clouds. Rick Reid Principal Engineer SGI Federal. 2013 by SGI Federal. Published by The Aerospace Corporation with permission.

Stovepipes to Clouds. Rick Reid Principal Engineer SGI Federal. 2013 by SGI Federal. Published by The Aerospace Corporation with permission. Stovepipes to Clouds Rick Reid Principal Engineer SGI Federal 2013 by SGI Federal. Published by The Aerospace Corporation with permission. Agenda Stovepipe Characteristics Why we Built Stovepipes Cluster

More information

Introduction to Infiniband. Hussein N. Harake, Performance U! Winter School

Introduction to Infiniband. Hussein N. Harake, Performance U! Winter School Introduction to Infiniband Hussein N. Harake, Performance U! Winter School Agenda Definition of Infiniband Features Hardware Facts Layers OFED Stack OpenSM Tools and Utilities Topologies Infiniband Roadmap

More information

ECLIPSE Best Practices Performance, Productivity, Efficiency. March 2009

ECLIPSE Best Practices Performance, Productivity, Efficiency. March 2009 ECLIPSE Best Practices Performance, Productivity, Efficiency March 29 ECLIPSE Performance, Productivity, Efficiency The following research was performed under the HPC Advisory Council activities HPC Advisory

More information

Performance Beyond PCI Express: Moving Storage to The Memory Bus A Technical Whitepaper

Performance Beyond PCI Express: Moving Storage to The Memory Bus A Technical Whitepaper : Moving Storage to The Memory Bus A Technical Whitepaper By Stephen Foskett April 2014 2 Introduction In the quest to eliminate bottlenecks and improve system performance, the state of the art has continually

More information

HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW

HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW 757 Maleta Lane, Suite 201 Castle Rock, CO 80108 Brett Weninger, Managing Director [email protected] Dave Smelker, Managing Principal [email protected]

More information

Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca

Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca Carlo Cavazzoni CINECA Supercomputing Application & Innovation www.cineca.it 21 Aprile 2015 FERMI Name: Fermi Architecture: BlueGene/Q

More information

RED HAT ENTERPRISE VIRTUALIZATION

RED HAT ENTERPRISE VIRTUALIZATION Giuseppe Paterno' Solution Architect Jan 2010 Red Hat Milestones October 1994 Red Hat Linux June 2004 Red Hat Global File System August 2005 Red Hat Certificate System & Dir. Server April 2006 JBoss April

More information

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...

More information

Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck

Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Sockets vs. RDMA Interface over 1-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Pavan Balaji Hemal V. Shah D. K. Panda Network Based Computing Lab Computer Science and Engineering

More information

Virtual Machine Monitors. Dr. Marc E. Fiuczynski Research Scholar Princeton University

Virtual Machine Monitors. Dr. Marc E. Fiuczynski Research Scholar Princeton University Virtual Machine Monitors Dr. Marc E. Fiuczynski Research Scholar Princeton University Introduction Have been around since 1960 s on mainframes used for multitasking Good example VM/370 Have resurfaced

More information

Building Clusters for Gromacs and other HPC applications

Building Clusters for Gromacs and other HPC applications Building Clusters for Gromacs and other HPC applications Erik Lindahl [email protected] CBR Outline: Clusters Clusters vs. small networks of machines Why do YOU need a cluster? Computer hardware Network

More information

Storage Architectures. Ron Emerick, Oracle Corporation

Storage Architectures. Ron Emerick, Oracle Corporation PCI Express PRESENTATION and Its TITLE Interfaces GOES HERE to Flash Storage Architectures Ron Emerick, Oracle Corporation SNIA Legal Notice The material contained in this tutorial is copyrighted by the

More information

Interconnect Analysis: 10GigE and InfiniBand in High Performance Computing

Interconnect Analysis: 10GigE and InfiniBand in High Performance Computing Interconnect Analysis: 10GigE and InfiniBand in High Performance Computing WHITE PAPER Highlights: There is a large number of HPC applications that need the lowest possible latency for best performance

More information

Data Center Op+miza+on

Data Center Op+miza+on Data Center Op+miza+on Sept 2014 Jitender Sunke VP Applications, ITC Holdings Ajay Arora Sr. Director, Centroid Systems Justin Youngs Principal Architect, Oracle 1 Agenda! Introductions! Oracle VCA An

More information

IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud

IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud February 25, 2014 1 Agenda v Mapping clients needs to cloud technologies v Addressing your pain

More information

Dell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture. Dell Compellent Product Specialist Team

Dell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture. Dell Compellent Product Specialist Team Dell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture Dell Compellent Product Specialist Team THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL

More information

Ceph Optimization on All Flash Storage

Ceph Optimization on All Flash Storage Ceph Optimization on All Flash Storage Somnath Roy Lead Developer, SanDisk Corporation Santa Clara, CA 1 Forward-Looking Statements During our meeting today we may make forward-looking statements. Any

More information

Mellanox Accelerated Storage Solutions

Mellanox Accelerated Storage Solutions Mellanox Accelerated Storage Solutions Moving Data Efficiently In an era of exponential data growth, storage infrastructures are being pushed to the limits of their capacity and data delivery capabilities.

More information

UCS M-Series Modular Servers

UCS M-Series Modular Servers UCS M-Series Modular Servers The Next Wave of UCS Innovation Marian Klas Cisco Systems June 2015 Cisco UCS - Powering Applications at Every Scale Edge-Scale Computing Cloud-Scale Computing Seamlessly Extend

More information

When Does Colocation Become Competitive With The Public Cloud? WHITE PAPER SEPTEMBER 2014

When Does Colocation Become Competitive With The Public Cloud? WHITE PAPER SEPTEMBER 2014 When Does Colocation Become Competitive With The Public Cloud? WHITE PAPER SEPTEMBER 2014 Table of Contents Executive Summary... 2 Case Study: Amazon Ec2 Vs In-House Private Cloud... 3 Aim... 3 Participants...

More information

How To Speed Up A Flash Flash Storage System With The Hyperq Memory Router

How To Speed Up A Flash Flash Storage System With The Hyperq Memory Router HyperQ Hybrid Flash Storage Made Easy White Paper Parsec Labs, LLC. 7101 Northland Circle North, Suite 105 Brooklyn Park, MN 55428 USA 1-763-219-8811 www.parseclabs.com [email protected] [email protected]

More information

Full and Para Virtualization

Full and Para Virtualization Full and Para Virtualization Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF x86 Hardware Virtualization The x86 architecture offers four levels

More information

StorPool Distributed Storage Software Technical Overview

StorPool Distributed Storage Software Technical Overview StorPool Distributed Storage Software Technical Overview StorPool 2015 Page 1 of 8 StorPool Overview StorPool is distributed storage software. It pools the attached storage (hard disks or SSDs) of standard

More information

Cloud Computing and the Internet. Conferenza GARR 2010

Cloud Computing and the Internet. Conferenza GARR 2010 Cloud Computing and the Internet Conferenza GARR 2010 Cloud Computing The current buzzword ;-) Your computing is in the cloud! Provide computing as a utility Similar to Electricity, Water, Phone service,

More information

Oracle Exadata: The World s Fastest Database Machine Exadata Database Machine Architecture

Oracle Exadata: The World s Fastest Database Machine Exadata Database Machine Architecture Oracle Exadata: The World s Fastest Database Machine Exadata Database Machine Architecture Ron Weiss, Exadata Product Management Exadata Database Machine Best Platform to Run the

More information

When Does Colocation Become Competitive With The Public Cloud?

When Does Colocation Become Competitive With The Public Cloud? When Does Colocation Become Competitive With The Public Cloud? PLEXXI WHITE PAPER Affinity Networking for Data Centers and Clouds Table of Contents EXECUTIVE SUMMARY... 2 CASE STUDY: AMAZON EC2 vs IN-HOUSE

More information

Interoperability Testing and iwarp Performance. Whitepaper

Interoperability Testing and iwarp Performance. Whitepaper Interoperability Testing and iwarp Performance Whitepaper Interoperability Testing and iwarp Performance Introduction In tests conducted at the Chelsio facility, results demonstrate successful interoperability

More information

FlashSoft Software from SanDisk : Accelerating Virtual Infrastructures

FlashSoft Software from SanDisk : Accelerating Virtual Infrastructures Technology Insight Paper FlashSoft Software from SanDisk : Accelerating Virtual Infrastructures By Leah Schoeb January 16, 2013 FlashSoft Software from SanDisk: Accelerating Virtual Infrastructures 1 FlashSoft

More information

Integrated Grid Solutions. and Greenplum

Integrated Grid Solutions. and Greenplum EMC Perspective Integrated Grid Solutions from SAS, EMC Isilon and Greenplum Introduction Intensifying competitive pressure and vast growth in the capabilities of analytic computing platforms are driving

More information