Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC



Similar documents
Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Clusters: Mainstream Technology for CAE

Cluster Implementation and Management; Scheduling

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN

PRIMERGY server-based High Performance Computing solutions

Cloud Computing through Virtualization and HPC technologies

Sun in HPC. Update for IDC HPC User Forum Tucson, AZ, Sept 2008

FLOW-3D Performance Benchmark and Profiling. September 2012

Building a Top500-class Supercomputing Cluster at LNS-BUAP

1 Bull, 2011 Bull Extreme Computing

Sun Constellation System: The Open Petascale Computing Architecture

Petascale Software Challenges. Piyush Chaudhary High Performance Computing

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC Denver

CRIBI. Calcolo Scientifico e Bioinformatica oggi Università di Padova 13 gennaio 2012

IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads

Improved LS-DYNA Performance on Sun Servers

A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures

Hadoop on the Gordon Data Intensive Cluster

HP reference configuration for entry-level SAS Grid Manager solutions

SRNWP Workshop. HP Solutions and Activities in Climate & Weather Research. Michael Riedmann European Performance Center

Current Status of FEFS for the K computer

ECLIPSE Performance Benchmarks and Profiling. January 2009

Quantum StorNext. Product Brief: Distributed LAN Client

Building Clusters for Gromacs and other HPC applications

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

Cluster Grid Interconects. Tony Kay Chief Architect Enterprise Grid and Networking

LS DYNA Performance Benchmarks and Profiling. January 2009

Oracle Maximum Availability Architecture with Exadata Database Machine. Morana Kobal Butković Principal Sales Consultant Oracle Hrvatska

Scalable NAS for Oracle: Gateway to the (NFS) future

Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre

Linux Cluster Computing An Administrator s Perspective

New Storage System Solutions

Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband

Data Center Op+miza+on

JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert

Altix Usage and Application Programming. Welcome and Introduction

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

HPC Update: Engagement Model

Michael Kagan.

Kriterien für ein PetaFlop System

Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar

Data Centric Systems (DCS)

Copyright 2013, Oracle and/or its affiliates. All rights reserved.

Using the Windows Cluster

SMB Direct for SQL Server and Private Cloud

Dell High-Performance Computing Clusters and Reservoir Simulation Research at UT Austin.

Deploying and managing a Visualization Onera

MOSIX: High performance Linux farm

Best Practices for Data Sharing in a Grid Distributed SAS Environment. Updated July 2010

Oracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/ CAE Associates

Silviu Panica, Marian Neagul, Daniela Zaharie and Dana Petcu (Romania)

SUN HPC SOFTWARE CLUSTERING MADE EASY

Fujitsu HPC Cluster Suite

Recommended hardware system configurations for ANSYS users

THE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC

Thematic Unit of Excellence on Computational Materials Science Solid State and Structural Chemistry Unit, Indian Institute of Science

INDIAN INSTITUTE OF TECHNOLOGY KANPUR Department of Mechanical Engineering

Purchase of High Performance Computing (HPC) Central Compute Resources by Northwestern Researchers

CMS Tier-3 cluster at NISER. Dr. Tania Moulik

Parallel Programming Survey

IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud

<Insert Picture Here> Refreshing Your Data Protection Environment with Next-Generation Architectures

Addendum No. 1 to Packet No Enterprise Data Storage Solution and Strategy for the Ingham County MIS Department

Monitoring Infrastructure for Superclusters: Experiences at MareNostrum

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

Simple Introduction to Clusters

PCI Express and Storage. Ron Emerick, Sun Microsystems

Distributed Operating Systems. Cluster Systems

LS-DYNA Scalability on Cray Supercomputers. Tin-Ting Zhu, Cray Inc. Jason Wang, Livermore Software Technology Corp.

High Availability Databases based on Oracle 10g RAC on Linux

MPI / ClusterTools Update and Plans

James Serra Sr BI Architect

10th TF-Storage Meeting

THE EXPAND PARALLEL FILE SYSTEM A FILE SYSTEM FOR CLUSTER AND GRID COMPUTING. José Daniel García Sánchez ARCOS Group University Carlos III of Madrid

PCI Express Impact on Storage Architectures and Future Data Centers. Ron Emerick, Oracle Corporation

Cray XT3 Supercomputer Scalable by Design CRAY XT3 DATASHEET

Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer

GPU System Architecture. Alan Gray EPCC The University of Edinburgh

SAN TECHNICAL - DETAILS/ SPECIFICATIONS

QuickSpecs. HP Integrity Virtual Machines (Integrity VM) Overview. Currently shipping versions:

High Performance Computing in CST STUDIO SUITE

The CNMS Computer Cluster

Cray DVS: Data Virtualization Service

Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca


CORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

Configuration Maximums

E4 UNIFIED STORAGE powered by Syneto

Transcription:

HPC Architecture End to End Alexandre Chauvin

Agenda HPC Software Stack Visualization National Scientific Center 2 Agenda

HPC Software Stack Alexandre Chauvin

Typical HPC Software Stack Externes LAN Typical HPC cluster configuration Compute nodes IO Nodes Interconnects Management Server Login Nodes Compute nodes High Speed Interconnect IO nodes Frontend / Login nodes Cluster Management Server Management LAN Application Application Application Resource Management MPI Development Tools Cluster Management Cluster File System Operating System Hardware Unix Required HPC Software Stack On the compute nodes OS (Windows, Linux, Unix) Parallel Libraries Development Tools For the full architecture High Speed IO Filesystem System Management Job Scheduler 4 Agenda

HPC Compute Nodes Typical operating systems for High Performance Computing Unix AIX, HP UX, Solaris Linux RHEL, SLES Windows To build applications, HPC users need compilers and mathematical libraries which usually depend on the hardware and operating system environment Compilers: Gnu, PGI, Intel, Pathscale, Mathematical Libraries: Goto, MKL, ACML, Mass, Unix To build parallel applications, HPC users usually need a version of MPI Message Passing Interface. The different implmentations depend on the network type, hardware vendor and operating system. OpenMPI, Mpich, LAM, HP MPI, PE, 5 Agenda

HPC System Management HPC System Administrators need System Management Software in order to remotely Easily deployed Operating systems with appropriate configuration Enables user accounts Allows predictive failures, security events and hardware and software errors Deploys new or updated software and operating system patches Cluster System Management Cluster management server Additionnaly tools such as Ganglia enables the HPC users and administrators to get a quick overview of the performance of the full system Screeenshot of Ganglia Monitoring 6 Agenda

Job Scheduler Job schedulers aim at imporving resource utilization and quelity of service The different metrics are: Resource Utilzation System Throughput Mean response time Different algorithms can be specified to the resource manager for scheduling jobs Based on # of processors required Estimated elapsed time required Job Priority Job Scheduler Source: www.dell.com 7 Agenda

HPC Software Stack High Performance Shared Filesystem HPC Environments requires shared filesystems accessible by all nodes or a pool of nodes with common purpose. HPC filesystem needs to be High Performance IO can represent a subsequent part of a HPC workload Scalable it must be possible to increase capacity and performance as a HPC systems and requirements can evolve Reliable no single point of failure Clients IO Network IO Servers (Data and/or Metadata) Enterprise Storage 8 Agenda

HPC Software Stack IO Parallelization Local Filesystem Not Shared Clients No parallelization HPC Filesystem Shared Parallelization Clients IO Network NFS Filesystem Shared No parallelization Clients IO Network IO Servers (Data and/or Metadata) IO Network NFS Server Storage Enterprise Storage 9 Agenda

GPFS Architecture Performance The performance of the GPFS subsystem is dependent of 3 different parts of its architecture Server Network SAN Network Disks Subsystem Clients IO Network Clients IO Network IO Servers (Data and/or Metadata) Enterprise Storage Client IO Requests Server Network Performance SAN Network Performance SAN Disk Performance IO Servers (Data and/or Metadata) SAN Network Each of them can be scaled to increase the IO capabilties. The goal is to have a balanced configuration across GPFS in term of performance. Enterprise Storage 10 Agenda

HPC Software Stack Multi Cluster Environment HPC File Systems and Job Scheduling can be shared among clusters at different sites This is used by major computing consortium such as Deisa DEISA is operating a heterogeneous HPC infrastructure currently formed by eleven European national supercomputing centres that are tightly interconnected by a dedicated high performance network. Among the Deisa projects, all the clusters are sharing the same data though Multi- Cluster GPFS Julich, Cineca, Idris and RZG have closely coupled job scheduling file systems The DEISA Global File System at European Scale 11 Agenda

Visualization Alexandre Chauvin

HPC Visualization Introduction Visualization is often the best way for HPC users to analyse the result of a large computation Typical Applications Computer-aided design Visualization of medical data Modeling and simulation Exploration of oil and gas Virtual training Remote collaboration Challenges Usually datasets are sent from HPC Data Center to analysts and are visualized locally on standalone worksation but HPC Computations writes large datasets Too large to be seen on a single computer, specially in high resolution mode Too large to be transferred through a WAN network Maybe too confidential to be transferred on the internet This method prevents collaboration between analysts WW 13 Agenda

HPC Software Stack Visualization Example #1: Dam Breaking EDF R&D SPARTACUS code 250 K / 2 M / 5 M Cells 1000 Time Steps 25 MB / File Loading Time: From 0.5 to 50 Secs / File 14 Agenda

Typical Visualization Architecture Data SW Network Network Data SW Data SW Data SW Networks Network Network Data SW Data Center Users 15 Agenda

HPC Visualization Architecture Visualization Cluster to render complex and large datasets this leads also to finer resolution of the final image WAN Collaboration on the WAN Networks Data SW Apps Image Transfer, not Data 16 Agenda

National Computing Center Alexandre Chauvin

Customer Case National Scientific Center Missions Helping national competitiveness for scientific researches needing very high compute capabilities. Dealing with More than 250 projects per year and 1000 users. Objectives for upgrade Having an always evolving platform to drive competitiveness Adding a highly performant architecture for scalar computation Integration to computing European consortium Current Computing Capabiltiies 10 Vector Supercomputers 8 Processor Systems and 64GB of Memory per system 10TB of IO disk space 1.28TFlops vector peak performance total SMP clusters 20 nodes of 32 processors @ 1.7GHz for OpenMP and MPI workloads 128GB of memory per node 1TB of IO disk space per node 96 nodes of 4 processors @ 1.5GHz and 8GB of memory for MPI workloads 10TB of shared IO Disk space 6.7TFlops peak performance total Also Pre-,post-processing capabilities Visualization 1000TB IO disk space for archiving 18 Agenda

New Computational Needs Requirements 19 Agenda

HPC Software Stack Requirements Analysis Compute performance to be improved by x20 Hybrid Architecture welcome Strong focus on IO capacity, usable performance, and availability Focus on RAS and usabillity Solution can be limited by Datacenter characterisitics, ie Power Consumption 20 Agenda

Solution Analysis SMP Cluster Hybrid solution possible based on SMP and MPP architectures Some numbers: Infiniband 1GB/s IO per link FC 4Gb 500MB/s IO per link SMP Cluster Characteristics SMP servers Based on 16 Dual core processor nodes @4.7GHz 54 of such nodes = 32TFlops Infiniband interconnect for MPI Network Infiniband Interconnect for global IO access 2 cards for redundancy SAN Fiber Channel 4Gb Connection for dedicated IO requirements 2 cards Switches 2 Infiniband switches for MPI communications 2 Infiniband switches for global IO connexion 2 SAN Fiber Channel switched for dedicated IO Dedicated Filesystem 400TBs of disks Vers Global IO 2 SAN Switch 140 ports 400TB of dedicated fs 21 Agenda

HPC Software Stack Solution Analysis MPP MPP Solution Characteristics Based on 10 Racks BlueGene/P 40960 procs @ 850MHz = 139TFlops 10Gb interconnect towards Global IO Filesystem 16 links for 2GB/s throughput performance To Global IO Filesystem 22 Agenda

Solution Analysis Global IO Filesystem High Speed shared interconnect based on GPFS software 2 different networks accessing the same data 16x 10Gb links for MPP 32x Infiniband links for SMP cluster GPFS servers are connected through 4 Fiber Channel 4Gb links to storage disks IO Subsystem Performance 16x 10Gb links gives 20GB/s theoritical throughput performance 4x 4Gb FC links gives 2GB/s theoritical throughput performance 8 disks bays for 16GB/s theoritical preformance 23 Agenda

Management Subsystem Job Scheduler Full HPC Management Subsystem Cluster Management Multi-Cluster Job Scheduling GPFS Cluster System Management Cluster management server 24 Agenda

Computing Solution Overview 56 nœuds de calcul SMP To European HPC Consortium Service Node BG Service Node BG Frontal BG Frontal BG 10 racks BG/P Réseau Myrinet 10G IB1 IB2 54 nœuds de calcul SMP Vers DEISA IB3 IB4 16 noeuds GPFS Connexion Réseau IDRIS 2 ports Ethernet 10G 5 ports Ethernet 1G 8 baies de disques pour 800 TO utiles et 16 GO/s IB 4x Ethernet 1G Ethernet 10G Myrinet 10G SAN FC 4 Gbits Nœud de management 400 TO pour FS locaux 8 GO/s 25 Agenda Réseau Admin 2 SAN switch 140 ports 2 nœuds interactifs