HPC Cloud. Focus on your research. Floris Sluiter Project leader SARA

Similar documents
SURFsara HPC Cloud Workshop

High Performance Computing in CST STUDIO SUITE

SURFsara HPC Cloud Workshop

Cloud Computing through Virtualization and HPC technologies

Clusters: Mainstream Technology for CAE

PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN

Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca

Cloud Computing. Alex Crawford Ben Johnstone

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/ CAE Associates

Sun Constellation System: The Open Petascale Computing Architecture

wu.cloud: Insights Gained from Operating a Private Cloud System

BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

Integration of Virtualized Workernodes in Batch Queueing Systems The ViBatch Concept

Introduction to High Performance Cluster Computing. Cluster Training for UCL Part 1

Scalable Cloud Computing Solutions for Next Generation Sequencing Data

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database

Michael Kagan.

Silviu Panica, Marian Neagul, Daniela Zaharie and Dana Petcu (Romania)

Efficient Load Balancing using VM Migration by QEMU-KVM

Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar

HPC performance applications on Virtual Clusters

Recommended hardware system configurations for ANSYS users

numascale White Paper The Numascale Solution: Extreme BIG DATA Computing Hardware Accellerated Data Intensive Computing By: Einar Rustad ABSTRACT

White Paper The Numascale Solution: Extreme BIG DATA Computing

Understanding the Benefits of IBM SPSS Statistics Server

Self service for software development tools

Software Scalability Issues in Large Clusters

Brocade and EMC Solution for Microsoft Hyper-V and SharePoint Clusters

Virtualization of a Cluster Batch System

Seeking Opportunities for Hardware Acceleration in Big Data Analytics

Big Data Challenges in Bioinformatics

CS 3530 Operating Systems. L02 OS Intro Part 1 Dr. Ken Hoganson

System Models for Distributed and Cloud Computing

How To Compare Amazon Ec2 To A Supercomputer For Scientific Applications

Preparation Guide. How to prepare your environment for an OnApp Cloud v3.0 (beta) deployment.

AMD SEAMICRO OPENSTACK BLUEPRINTS CLOUD- IN- A- BOX OCTOBER 2013

Outline. High Performance Computing (HPC) Big Data meets HPC. Case Studies: Some facts about Big Data Technologies HPC and Big Data converging

White Paper. Recording Server Virtualization

INDIAN INSTITUTE OF TECHNOLOGY KANPUR Department of Mechanical Engineering

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

Parallel Computing with MATLAB

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago


Scala Storage Scale-Out Clustered Storage White Paper

1 Bull, 2011 Bull Extreme Computing

Virtual Computing and VMWare. Module 4

Development of Monitoring and Analysis Tools for the Huawei Cloud Storage

COMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1)

A Novel Cloud Based Elastic Framework for Big Data Preprocessing

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

Pivot3 Reference Architecture for VMware View Version 1.03

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Intro to Virtualization

New Data Center architecture

Parallels Plesk Automation

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

HPC Update: Engagement Model

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

HPC Cluster Decisions and ANSYS Configuration Best Practices. Diana Collier Lead Systems Support Specialist Houston UGM May 2014

Sense Making in an IOT World: Sensor Data Analysis with Deep Learning

Virtual Machine Monitors. Dr. Marc E. Fiuczynski Research Scholar Princeton University

Lecture 2 Cloud Computing & Virtualization. Cloud Application Development (SE808, School of Software, Sun Yat-Sen University) Yabo (Arber) Xu

PRIMERGY server-based High Performance Computing solutions

Neptune. A Domain Specific Language for Deploying HPC Software on Cloud Platforms. Chris Bunch Navraj Chohan Chandra Krintz Khawaja Shams

Purchase of High Performance Computing (HPC) Central Compute Resources by Northwestern Researchers

Can High-Performance Interconnects Benefit Memcached and Hadoop?

Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage

Comparing the performance of the Landmark Nexus reservoir simulator on HP servers

Recent Advances in HPC for Structural Mechanics Simulations

Oracle Software. Hardware. Training. Consulting. Mythics Complete.

Dynamic Resource allocation in Cloud

Sheepdog: distributed storage system for QEMU

An Oracle White Paper July Oracle Primavera Contract Management, Business Intelligence Publisher Edition-Sizing Guide

Workshop on Parallel and Distributed Scientific and Engineering Computing, Shanghai, 25 May 2012

Accelerating Simulation & Analysis with Hybrid GPU Parallelization and Cloud Computing

On Demand Satellite Image Processing

Altix Usage and Application Programming. Welcome and Introduction

Microsoft Hyper-V chose a Primary Server Virtualization Platform

HP reference configuration for entry-level SAS Grid Manager solutions

Microsoft Private Cloud Fast Track

Enabling Technologies for Distributed Computing

Scaling from 1 PC to a super computer using Mascot

SMB Direct for SQL Server and Private Cloud

Dell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture. Dell Compellent Product Specialist Team

How To Speed Up A Flash Flash Storage System With The Hyperq Memory Router

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC Denver

Exploiting Remote Memory Operations to Design Efficient Reconfiguration for Shared Data-Centers over InfiniBand

Cluster Computing at HRI

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

Transcription:

HPC Cloud Focus on your research Floris Sluiter Project leader SARA

Why an HPC Cloud? Christophe Blanchet, IDB - Infrastructure Distributing Biology: Big task to port them all to your favorite architecture Is true for many scientific fields HPC Cloud: Provides a different solution to porting applications Provides the possibility to scale up your Desktop environment to HPC

What is High Performance Cloud Computing? Cloud Computing: Self Service Dynamically Scalable Computing Facilities Cloud computing is not about new technology, it is about new uses of technology HPC Cloud is a Cloud built with high end HPC hardware Through virtualization: Better seperation between users, allowing more administrative rights Very fast to set up custom environments, saving application porting time Users can fully define and control a private HPC cluster

The SARA / BiG Grid Cloud project The beginning Pilot Started in 2009 from a powerpoint presentation 3 months with 10 users on 48 core cluster Participation was a price in a competition Very succesfull Preparing for Operations Now 2010 BiG Grid grants funding for project 2010 2011 Product development 50 usergroups on 128 core cluster 2011 Big Grid grants funding for HPC hardware Presenting a production infrastructure

User participation 30 involved in Beta testing nr. Title Core Hours Storage Objective Group/instiute 14 samples * 2 vms * Cloud computing for sequence 2-4 cores * 2 days = Run a set of prepared vm's for different and specific Bacterial Genomics, CMBI 1 assembly 5000 10-100GB / VM sequence assembly tasks Nijmegen Cloud computing for a multimethod perspective study of construction of (cyber)space and 2 place 2000 (+) 75-100GB Analyse 20 million Flickr Geocoded data points Uva, GPIO institute asses cloud technology potential and efficiency on 3 Urban Flood Simulation 1500 1 GB Further ported develop Urban a user-friendly Flood simulation desktop modules environment UvA, Computational Science A user friendly cloud-based running in the cloud supporting modelling, testing and Computational Geo-ecology, 4 inverse modelling environment testing 1GB / VM large scale running of model. UvA Real life HPC cloud computing Microarray Department, experiences for MicroArray Test, development and acquire real life experiences Integrative BioInformatics Unit, 5 analyses 8000 150GB using vm's for microarray analysis UvA Customized pipelines for the up to 1TB of data -> Configure a customized virtual infrastructure for MRI Biomedical Imaging Group, 6 processing of MRI brain data? transferred out quickly. image processing pipelines Rotterdam, Erasmus MC Cloud computing for historical map collections: access and 7VM's of 500 GB = 3.5 Set up distributed, decentralized autonomous 7 georeferencing? TB georeferencing data delivery system. Department of Geography, UvA Parallellization of MT3DMS for 64 cores, schaling modeling contaminant transport at experiments / * 80 Goal, investigate massive parallell scaling for code 8 large scale hours = 5000 hours 1 TB speed-up Estimate an execution time of existing bioinformatics Deltares An imputation pipeline on Grid pipelines and, in particular, heavy imputation pipelines Groningen Bioinformatics 9 Gain 20TB on a new HPC cloud Center, university of groningen Regional Atmospheric Soaring Demonstrate how cloud computing eliminates porting Computational Geo-ecology, 10 Prediction 320 20GB problems. UvA Extraction of Social Signals from Pattern Recognition Laboratory, 11 Analysis of video next generation 160 630GB Video Feature extraction TU Delft sequencing data from mouse Run analysis pipeline to create mouse model for 12 tumors? 150-300GB genome analysis Chris Klijn, NKI

User participation 30 involved in Beta testing nr. Title Core Hours Storage Objective Group/instiute 14 samples * 2 vms * Cloud computing for sequence 2-4 cores * 2 days = Run a set of prepared vm's for different and specific Bacterial Genomics, CMBI 1 assembly 5000 10-100GB / VM sequence assembly tasks Nijmegen Cloud computing for a multimethod perspective study of construction of (cyber)space and 2 place 2000 (+) 75-100GB Analyse 20 million Flickr Geocoded data points Uva, GPIO institute Field # projects asses cloud technology potential and efficiency on 3 Urban Flood Simulation 1500 1 GB Further ported develop Urban a user-friendly Flood simulation desktop modules environment UvA, Computational Science A user friendly cloud-based running in the cloud supporting modelling, testing and Computational Geo-ecology, 4 inverse modelling environment testing 1GB / VM large scale running of model. UvA Real life HPC cloud computing Bioinformatics 14 Microarray Department, experiences for MicroArray Test, development and acquire real life experiences Integrative BioInformatics Unit, 5 analyses 8000 150GB using vm's for microarray analysis UvA Ecology 4 Customized pipelines for the up to 1TB of data -> Configure a customized virtual infrastructure for MRI Biomedical Imaging Group, 6 processing of MRI brain data? transferred out quickly. image processing pipelines Rotterdam, Erasmus MC Cloud computing for historical map collections: access and 7VM's of 500 GB = 3.5 Set up distributed, decentralized autonomous 7 georeferencing? TB georeferencing data delivery system. Department of Geography, UvA Geography 3 Parallellization of MT3DMS for 64 cores, schaling modeling contaminant transport at experiments / * 80 Goal, investigate massive parallell scaling for code 8 large scale hours Computer = 5000 hours 1 TB science speed-up Deltares Estimate an execution time of existing bioinformatics 4 An imputation pipeline on Grid pipelines and, in particular, heavy imputation pipelines Groningen Bioinformatics 9 Gain 20TB on a new HPC cloud Center, university of groningen Alpha/Gamma 5 Regional Atmospheric Soaring Demonstrate how cloud computing eliminates porting Computational Geo-ecology, 10 Prediction 320 20GB problems. UvA Extraction of Social Signals from Pattern Recognition Laboratory, 11 Analysis of video next generation 160 630GB Video Feature extraction TU Delft sequencing data from mouse Run analysis pipeline to create mouse model for 12 tumors? 150-300GB genome analysis Chris Klijn, NKI

HPC Cloud Project Publicity & dissemination In 2009 we were the first in the global Grid community to offer this to end users We started international interest group and presented at many events (ECEE, EGI, Sienna Cloud Scape, etc) and institutions We are in the top 5 search results for HPC Cloud and HPC Cloud Computing at Google News articles on paper and in online publications (HPCCloud HPCWire Computable)

HPC Cloud Architecture I make clouds 1 Node: CPU Intel 2.13 GHz 32 cores (Xeon-E7 "Westmere-EX") RAM 256 Gbyte "Local disk" 10 Tbyte Ethernet 4*10GE Total System (19 nodes) 608 cores RAM 4,75TB 96 ports 10GE, 1-hop, non-blocking interconnect Novel architecture 400TB shared storage (ISCSI,NFS,CIFS,WebDAV...) 11.5K specints / 5TFlops Low latecy 10GE switch chassis at the heart, with mountable storage Per node High mem, high core, 4* network (but smaller partitions possible) Fast and dynamic changes in configurations and partitions Security and safety through strict separation of users + monitoring network Invested in Expandability Network switch and storage controllers can be expanded to support more compute nodes Future architecture focus might change, depending on usage patterns Virtualization overhead <10% Very flexible for any application. Just great is often good enough Platform and tools: Redmine collaboration portal Custom GUI (Open Source) Open Nebula + custom add-ons

HPC Cloud Architecture vs other HPC Architectures System Node Characteristics Huygens National Super CPU Power6, 4.7Ghz, 32/64 cores RAM 128/256 GB "local disk" 8Tbyte Infiniband 8*20 Gbit/s Network, storage, and compute are fully balanced resulting in a highly optimized system for I/Ointensive, parallel applications LISA National Compute Cluster CPU Intel 2.26Ghz 8/12 cores RAM 24 Gbyte Local disk 65Gbyte/200Gbyte Infiniband 20Gbit/s Network, storage, and compute are optimized for efficient parallel applications Grid CPU Intel 2.2 Ghz 8 cores RAM 24 Gbyte Local disk 300Gbyte Ethernet 1 Gbit/s Network, storage, and compute are optimized for high throughput of data and single core data processing Cloud CPU Intel 2.13 GHz 32 cores (Xeon-E7 "Westmere-EX") RAM 256 Gbyte "Local disk" 10 Tbyte Ethernet 4*10GE Network, storage, and compute are dynamically scalable and configurable for both large and small workloads

HPC Cloud Application types Applications with different requirements can co-exist on the same physical host and in the same virtual cluster Single node (remote desktop on HPC node) Master with workers (standard cluster) Token server Pipelines/workflows example: MSWindows+Linux 24/7 Services that start workers HPC Type Examples Requirements Compute Intensive Data intensive Communication intensive Memory intensive Continuous services Monte Carlo simulations and parameter optimizations, etc Signal/Image processing in Astronomy, Remote Sensing, Medical Imaging, DNA matching, Pattern matching, etc Particle Physics, MPI, etc DNA assembly, etc Databases, webservers, webservices CPU Cycles I/O to data (SAN File Servers) Fast interconnect network Large (Shared) RAM Dynamically scalable

Virtual HPC Cluster A true HPC cluster with dedicated storage, network and compute nodes User has full control Custom environment examples A much bigger workstation and/or many identical workstations Closed source software Different environments in same cluster Example: combine Linux with windows nodes High mem, disk, I/O, CPU and in any combination Long running jobs

GUI + Collaboration Portal

For current and future users Most current users only asked a few support questions when they start to use it, it is easy to do support on this system ;-) In the Pilot and Beta phase, the hardware was not ideal for this type of system and applications. The new hardware is/will be much bigger and better! On the HPC Cloud: Do not change your HPC applications for the system, but rather adapt the environment they run in The result will be: a shorter time to a solution

Results from the participant survey Number of persons Background Intentions and expectations 50 Academic and research 20 Technical Developers Operations Use it! Copy it! 30 Industry Sell it! 25 VIPs Important innovation!

Results from the participant survey Number of persons Background Intentions and expectations 50 Academic and research 20 Technical Developers Operations Use it! Copy it! 30 Industry Sell it! 25 VIPs Important innovation! 40% sell it to the other 40%, overseen by 20%

A dwarf standing on the shoulders of Giants! Many people contributed to this project BiG Grid Collaboration, funding and support Too many colleagues to mention people from all the groups within Sara and many of the colleagues at Nikhef had active involvement too make this a succes. It is great to work with so many experts willing and able to contribute! The users, who helped a great deal shaping the experience the HPC Cloud is by trying it out, sending bug reports, feature requests and asking for more of it!

Thank you! Questions? www.cloud.sara.nl Photo's: http://cloudappreciationsociety.org/