Kriterien für ein PetaFlop System

Similar documents
The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC Denver

Sun Constellation System: The Open Petascale Computing Architecture

Cluster Implementation and Management; Scheduling

JuRoPA. Jülich Research on Petaflop Architecture. One Year on. Hugo R. Falter, COO Lee J Porter, Engineering

GPFS und HPSS am HLRS

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Current Status of FEFS for the K computer

New Storage System Solutions

FLOW-3D Performance Benchmark and Profiling. September 2012

HPC Update: Engagement Model

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Parallel Programming Survey

HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect

JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert

Cray DVS: Data Virtualization Service

Clusters: Mainstream Technology for CAE

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

Hadoop on the Gordon Data Intensive Cluster

Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband

GPU System Architecture. Alan Gray EPCC The University of Edinburgh

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products

Welcome to the. Jülich Supercomputing Centre. D. Rohe and N. Attig Jülich Supercomputing Centre (JSC), Forschungszentrum Jülich

Deploying and managing a Visualization Onera

HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK

Altix Usage and Application Programming. Welcome and Introduction

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

PRIMERGY server-based High Performance Computing solutions

Large File System Backup NERSC Global File System Experience

Building a Top500-class Supercomputing Cluster at LNS-BUAP

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007

Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca

Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer

HPC Cluster Decisions and ANSYS Configuration Best Practices. Diana Collier Lead Systems Support Specialist Houston UGM May 2014

How To Speed Up A Flash Flash Storage System With The Hyperq Memory Router

Jean-Pierre Panziera Teratec 2011

Data Analytics at NERSC. Joaquin Correa NERSC Data and Analytics Services

Lustre & Cluster. - monitoring the whole thing Erich Focht

ECLIPSE Best Practices Performance, Productivity, Efficiency. March 2009

HPC enabling of OpenFOAM R for CFD applications

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/ CAE Associates

Cosmological simulations on High Performance Computers

High Productivity Computing With Windows

Pedraforca: ARM + GPU prototype

Lessons learned from parallel file system operation

PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN

SGI High Performance Computing

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing

How To Compare Amazon Ec2 To A Supercomputer For Scientific Applications

LS DYNA Performance Benchmarks and Profiling. January 2009

SR-IOV In High Performance Computing

High Availability of the Polarion Server

Stovepipes to Clouds. Rick Reid Principal Engineer SGI Federal by SGI Federal. Published by The Aerospace Corporation with permission.

Accelerating From Cluster to Cloud: Overview of RDMA on Windows HPC. Wenhao Wu Program Manager Windows HPC team

STORAGE HIGH SPEED INTERCONNECTS HIGH PERFORMANCE COMPUTING VISUALISATION GPU COMPUTING

Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar

Building a Scalable Storage with InfiniBand

ECLIPSE Performance Benchmarks and Profiling. January 2009

Network Contention and Congestion Control: Lustre FineGrained Routing

KIT Site Report. Andreas Petzold. STEINBUCH CENTRE FOR COMPUTING - SCC

Resource Scheduling Best Practice in Hybrid Clusters

Barry Bolding, Ph.D. VP, Cray Product Division

David Rioja Redondo Telecommunication Engineer Englobe Technologies and Systems

OpenMP Programming on ScaleMP

PRACE hardware, software and services. David Henty, EPCC,

HPC-related R&D in 863 Program

IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud

Microsoft Compute Clusters in High Performance Technical Computing. Björn Tromsdorf, HPC Product Manager, Microsoft Corporation

Cray Gemini Interconnect. Technical University of Munich Parallel Programming Class of SS14 Denys Sobchyshak

Can High-Performance Interconnects Benefit Memcached and Hadoop?

Extreme Computing: The Bull Way

THE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC

Cloud Computing through Virtualization and HPC technologies

PCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters. from One Stop Systems (OSS)

Fine-grained File System Monitoring with Lustre Jobstat

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

NERSC File Systems and How to Use Them

Big Data and the Earth Observation and Climate Modelling Communities: JASMIN and CEMS

How Cineca supports IT

Software-defined Storage Architecture for Analytics Computing

Building Clusters for Gromacs and other HPC applications

Data Centric Systems (DCS)

Cloud Computing and Amazon Web Services

High Performance Computing in CST STUDIO SUITE

Recommended hardware system configurations for ANSYS users

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.

Optimizing GPU-based application performance for the HP for the HP ProLiant SL390s G7 server

Block based, file-based, combination. Component based, solution based

Transcription:

Kriterien für ein PetaFlop System Rainer Keller, HLRS :: :: ::

Context: Organizational HLRS is one of the three national supercomputing centers in Germany. The national supercomputing centers are working together in the Gauss Centre for Supercomputing GCS. GCS is the means to contribute to the Partnership for Advanced Computing in Europe (PRACE). All centers work within PRACE towards a European HPC Infrastructure and perform research with all PRACE partners towards Exascale computing. Additionally HLRS is responsible within PRACE and GCS for the support of the engineering community and the definition of the industrial offer. Kriterien für ein Petaflop System :: 2

Context: Systems in Stuttgart As a tradition, large variety of HPC systems offered: Cray XE6: Phase 1, Step0 Clusters, e.g.: Laki 62 TF BW-Grid 14 TF New architectures, e.g.: Cluster of Cell (2008) MD Grape (2007) SLES+CNL Vector MPPs Large ccnuma Kriterien für ein Petaflop System :: 3

Context: The petagcs Project (Phase1) The petagcs project is a BMBF funded project covering the national share for investment and operation of national supercomputing in Germany Covers currently Phase1 of all GCS centers Next phases will be covered in a similar way 50% co-funding is provided by the regional governments For HLRS: the Ministry of Science, Research and the Arts Baden-Württemberg Curie@CEA 105 TF Curie@CEA 1,6 PF Kriterien für ein Petaflop System :: 4

Context: Main User s Research Projects Aeroacoustics Aerodynamics Astrophysics Bioinformatics Combustion Fluid-Structure Interaction Helicopter Aerodynamics Meterology Medical Imaging Nanotechnology Solid State Physics Turbo Machinery Turbulence Phenomena Kriterien für ein Petaflop System :: 5

Constraints: Diversity User Community Performance Courses: MPI & OpenMP Fortran for Scientific Computing Coarray Fortran & UPC GPU Programming Single CPU Optimization CFD Early Developers Requires wide spread training & teaching expertise Scalability Workshops Tools Workshops Co-Development Hundreds of TeraFlop/s Demands for strong collaboration in particular with the vendor PetaFlop/s high-end capability Test users/beginners Federal project users Advanced users/ Heroes Kriterien für ein Petaflop System :: 6

Sustained Application Performance Memory Bound CPU Bound Disk IO intensive Open Source Communication intensive Grand Challenge Petascale Potential Demand large memory Important Application Domain Real Testcases Kriterien für ein Petaflop System :: 7

Tocal cost of ownership Kriterien für ein Petaflop System :: 8

Procurement Constraints: Summary Application performance in engineering domain is more important than peak performance. Programmability of the system (Software Stack/libraries, Support of diverse programming models, Available Tools, Adequate memory size, ratio and speed, ) is key! Performance of I/O to pre-/postprocess and visualisation is essential System architecture should be designed for highly scalable codes with low latency, low communication memory footprint and with high bandwidth interconnect Heterogeneous Customer Groups (Industry, Local, Regional, EU Wide) Future system must fit into the existing machine room within the planned update to 5MW Power & Cooling capacity Total Cost of Ownership is relevant and part of the selection criteria New System must support migration path for users of existing systems NEC SX-9: How to attract users of Vector Systems? NEC Nehalem: How to convince people to move to the new system? Vendors: Provider Collaboration capability needs special attention Kriterien für ein Petaflop System :: 9

Phase 1 Step 1: Hermit 1 3 rd PRACE TIER-0 System Kriterien für ein Petaflop System :: 10

Hermit: Infrastucture bring-up August 2010 November 2010 January 2011 February 2011 Finished: April 2011 Kriterien für ein Petaflop System :: 11

Phase 1 Step 0: Configuration Configuration of Test System: Cray XE6 (1 cabinet) 84 compute nodes AMD Magny Cours 8 core/socket, 2.0 GHz 32 GB memory (1344 total cores) 12 service nodes AMD Istanbul 6 core/socket, 2.2 GHz 16 GB memory (72 total cores) Kriterien für ein Petaflop System :: 12

Phase 1 Step 1: Overview Configuration: Peak Performace ~1PF Overall >3500 nodes Each node has 2 sockets AMD Interlagos @ 2.3GHz, 16 Cores each ~112k cores Nodes with 32GB and 64GB memory 2.7PB storage capacity @ ~150GB/s IO bandwidth External Access, Pre- & Postprocessing and Remote Visualization Nodes ~2MW maximal power consumption Essential: intensive support by on-site staff and a collaboration agreement between HLRS and CRAY Support for ISV Codes under CLE ( native ) or Cluster Compatibility Mode (CCM). Part of the acceptance tests are validation of sustained application performance spanning across the full system Kriterien für ein Petaflop System :: 13

Conceptual Architecture Cray XE6 Boot RAID SMW StorageSwitch Fabric (QDR) Login Servers Login Servers Login Servers MDS Lustre OSS HPSS Data Mover Management Server Login Servers Login Servers Compute Nodes MOM Nodes (SIO) Network Nodes (SIO) Router Nodes for Lustre & MPI (SIO) DVS Server Nodes for NGF (SIO) Boot, Syslog and System Database Nodes (SIO) Lustre global work Visualization Server Pre- & Postprocessing NAS - home Kriterien für ein Petaflop System :: 14

Phase 1 Step 1: Remote Visualization Servers Visualization Servers allow remote graphical access Full Hardware accelerated graphics through dedicated remote graphics hardware as well as VirtualGL Powerful Graphic Cards with GP-GPU post processing support beyond the capabilities of a typical enduser system Visualization of very large datasets without the need to move data outside the data center Direct link to parallel Visualization on the compute nodes Access to applications without the need for a local deployment 10 Gig-E :: 20.5.2011 24.03.2011 :: Kriterien für ein Petaflop System :: 15

Phase 1 Step1: Storage Solution ultra fast Lustre based solution for fast disk space 2.7 PB Lustre Workspace capacity 16 DDN SFA10k controllers Integrated into the overall HLRS environments MDS MDS 1 LNET 1. 2 IB Switch Fabric LNET 68 7x2+2RAID10 600GB FC Fast secure ultra fast HLRS wide Home Space (60TB) Local and fast storage capacity with 20TB for Pre- and Postprocessing servers OSS 1 OSS 64 safe Integrated with the existing HPSS system using a data mover concept 3840TB Total Raw Capacity Kriterien für ein Petaflop System :: 16

A glimpse on Phase 1 Step 2 Hermit 2 Step will run in parallel to Step 1 realising an integrated system Goal: maintain similar software stack for both installation steps Expected architectural changes: Aries interconnect Newest generation of CPUs Partially relying on accelerators Updated storage infrastructure Additional external servers Significantly increased sustained application performance Scheduled for Autumn 2013 Overall peak performance of complete Phase 1 will be >5PF Specification is driven by agreed sustained application performance and not peak performance tough Kriterien für ein Petaflop System :: 17

Thank You very much! Any Questions? :: :: ::