HPC platforms @ UL. http://hpc.uni.lu. Overview (as of 2013) and Usage. S. Varrette, PhD. University of Luxembourg, Luxembourg



Similar documents
Transforming the UL into a Big Data University. Current status and planned evolutions

A Holistic Model of the Energy-Efficiency of Hypervisors

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Cluster Implementation and Management; Scheduling

Parallel Programming Survey

Enabling Technologies for Distributed and Cloud Computing

Building a Top500-class Supercomputing Cluster at LNS-BUAP

Building Clusters for Gromacs and other HPC applications

Hadoop on the Gordon Data Intensive Cluster

SMB Direct for SQL Server and Private Cloud

Enabling Technologies for Distributed Computing

HPC Update: Engagement Model

Sun Constellation System: The Open Petascale Computing Architecture

HPC Growing Pains. Lessons learned from building a Top500 supercomputer

Sun in HPC. Update for IDC HPC User Forum Tucson, AZ, Sept 2008

Open-E Data Storage Software and Intel Modular Server a certified virtualization solution

Enabling Large-Scale Testing of IaaS Cloud Platforms on the Grid 5000 Testbed

E4 UNIFIED STORAGE powered by Syneto

Scientific Computing Data Management Visions

Deploying and managing a Visualization Onera

New Storage System Solutions

Lustre SMB Gateway. Integrating Lustre with Windows

Building a Linux Cluster

JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert

UNIFIED HYBRID STORAGE. Performance, Availability and Scale for Any SAN and NAS Workload in Your Environment

Stovepipes to Clouds. Rick Reid Principal Engineer SGI Federal by SGI Federal. Published by The Aerospace Corporation with permission.

Performance Evaluation of the XDEM framework on the OpenStack Cloud Computing Middleware

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

Performance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage

IOmark- VDI. Nimbus Data Gemini Test Report: VDI a Test Report Date: 6, September

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

Manual for using Super Computing Resources

Low-cost

CSC Yearly Team Meeting

MPI / ClusterTools Update and Plans

CORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/ CAE Associates


Performance in a Gluster System. Versions 3.1.x

The Asterope compute cluster

Current Status of FEFS for the K computer

White Paper. Recording Server Virtualization

Energy efficient computing on Embedded and Mobile devices. Nikola Rajovic, Nikola Puzovic, Lluis Vilanova, Carlos Villavieja, Alex Ramirez

Clusters: Mainstream Technology for CAE

10GbE Ethernet for Transfer Speeds of Over 1150MB/s

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar

Analysis of VDI Storage Performance During Bootstorm

Object storage in Cloud Computing and Embedded Processing

An introduction to Fyrkat

PRIMERGY server-based High Performance Computing solutions

Parallels Cloud Storage

Introduction to Gluster. Versions 3.0.x

HPC Advisory Council

Emulex OneConnect 10GbE NICs The Right Solution for NAS Deployments

Kriterien für ein PetaFlop System

Michael Kagan.

WHITE PAPER. How To Build a SAN. The Essential Guide for Turning Your Windows Server Into Shared Storage on Your IP Network

MESOS CB220. Cluster-in-a-Box. Network Storage Appliance. A Simple and Smart Way to Converged Storage with QCT MESOS CB220

Restricted Document. Pulsant Technical Specification

The functionality and advantages of a high-availability file server system

Boas Betzler. Planet. Globally Distributed IaaS Platform Examples AWS and SoftLayer. November 9, IBM Corporation

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

Cluster Computing at HRI

Extreme Computing: The Bull Way

Introduction to Cloud Computing

Large Scale Storage. Orlando Richards, Information Services LCFG Users Day, University of Edinburgh 18 th January 2013

Pivot3 Reference Architecture for VMware View Version 1.03

StorPool Distributed Storage Software Technical Overview

Scaling from Workstation to Cluster for Compute-Intensive Applications

HP reference configuration for entry-level SAS Grid Manager solutions

Distributed RAID Architectures for Cluster I/O Computing. Kai Hwang

VDI: What Does it Mean, Deploying challenges & Will It Save You Money?

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

Red Hat Enterprise Linux 6. Stanislav Polášek ELOS Technologies

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute

GTC Presentation March 19, Copyright 2012 Penguin Computing, Inc. All rights reserved

SAS Business Analytics. Base SAS for SAS 9.2

Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre

Lessons learned from parallel file system operation

HPC Software Requirements to Support an HPC Cluster Supercomputer

Fujitsu PRIMERGY Servers Portfolio

HP recommended configuration for Microsoft Exchange Server 2010: HP LeftHand P4000 SAN

An Introduction to High Performance Computing in the Department

Building Storage Service in a Private Cloud

Architecting a High Performance Storage System

Thematic Unit of Excellence on Computational Materials Science Solid State and Structural Chemistry Unit, Indian Institute of Science

Workshop on Parallel and Distributed Scientific and Engineering Computing, Shanghai, 25 May 2012

nexsan NAS just got faster, easier and more affordable.

IEEE Mass Storage Conference Vendor Reception Lake Tahoe, NV

Virtualization of the MS Exchange Server Environment

Linux Cluster Computing An Administrator s Perspective

The CNMS Computer Cluster

(Scale Out NAS System)

Designed for Maximum Accelerator Performance

Best Practices for Data Sharing in a Grid Distributed SAS Environment. Updated July 2010

CRIBI. Calcolo Scientifico e Bioinformatica oggi Università di Padova 13 gennaio 2012

Transcription:

HPC platforms @ UL Overview (as of 2013) and Usage http://hpc.uni.lu S. Varrette, PhD. University of Luxembourg, Luxembourg 1 / 66

Summary 1 Introduction 2 Overview of the Main HPC Components 3 HPC and Cloud Computing (CC) 4 The UL HPC platform 5 UL HPC in Practice: Toward an [Efficient] Win-Win Usage 2 / 66

Introduction Summary 1 Introduction 2 Overview of the Main HPC Components 3 HPC and Cloud Computing (CC) 4 The UL HPC platform 5 UL HPC in Practice: Toward an [Efficient] Win-Win Usage 3 / 66

Introduction Evolution of Computing Systems arpanet internet 1st Generation 2nd 3rd 4th 5th ENIAC Transistors Integrated Circuit Micro- Processor Beowulf Cluster Multi-Core Processor Cloud 180,000 tubes 30 t, 170 m 2 Replace tubes 1959: IBM 7090 Thousands of transistors in one circuit 1971: Intel 4004 Millions of transistors in one circuit 1989: Intel 80486 Multi-core processor 2005: Pentium D 150 Flops 33 KFlops 0.06 Mips 1 MFlops 74 MFlops 2 GFlops HW diversity 1946 1956 1963 1974 1980 1994 1998 2005 2010 4 / 66

Introduction Why High Performance Computing? The country that out-computes will be the one that out-competes. Council on Competitiveness Accelerate research by accelerating computations 14.4 GFlops 27.363TFlops (Dual-core i7 1.8GHz) (291computing nodes, 2944cores) Increase storage capacity Communicate faster 2TB (1 disk) 1042TB raw(444disks) 1 GbE (1 Gb/s) vs Infiniband QDR (40 Gb/s) 5 / 66

Introduction HPC at the Heart of our Daily Life Today... Research, Industry, Local Collectivities... Tomorrow: applied research, digital health, nano/bio techno S. Varrette, PhD. (UL) N HPC platforms @ UL 6 / 66

Introduction HPC at the Heart of National Strategies USA R&D program: 1G$/y for HPC 2005 2011 2014 DOE R&D budget: 12.7G$/y Japan 800 Me (Next Generation Supercomputer Program) 2008 2011 K supercomputer, first to break the 10 Pflops mark China massive investments (exascale program) since 2006 Russia 1.5G$ for the exascale program (T-Platform) India 1G$ program for exascale Indian machine 2012 EU 1.58G$ program for exascale 2012 7 / 66

Introduction HPC at the Heart of National Strategies USA R&D program: 1G$/y for HPC 2005 2011 2014 DOE R&D budget: 12.7G$/y Japan 800 Me (Next Generation Supercomputer Program) 2008 2011 K supercomputer, first to break the 10 Pflops mark China massive investments (exascale program) since 2006 Russia 1.5G$ for the exascale program (T-Platform) India 1G$ program for exascale Indian machine 2012 EU 1.58G$ program for exascale 2012 2012: 11.1G$ profits in the HPC technical server industry Record Revenues (10.3G$ in 2011) +7.7% [Source: IDC] 7 / 66

Overview of the Main HPC Components Summary 1 Introduction 2 Overview of the Main HPC Components 3 HPC and Cloud Computing (CC) 4 The UL HPC platform 5 UL HPC in Practice: Toward an [Efficient] Win-Win Usage 8 / 66

Overview of the Main HPC Components HPC Components: [GP]CPU CPU Always multi-core Ex: Intel Core i7-970 (July 2010) R peak 100 GFlops (DP) 6 cores @ 3.2GHz (32nm, 130W, 1170 millions transistors) GPU / GPGPU Always multi-core, optimized for vector processing Ex: Nvidia Tesla C2050 (July 2010) 448 cores @ 1.15GHz R peak 515 GFlops (DP) 10 Gflops for 50 e 9 / 66

Overview of the Main HPC Components HPC Components: Local Memory Larger, slower and cheaper L1 L2 L3 CPU - C a c - C a c - C a c Memory Bus Memory I/O Bus h h h Registers e e e register reference L1-cache (SRAM) reference L2-cache (SRAM) reference L3-cache (DRAM) reference Memory (DRAM) reference Disk memory reference Level: 1 2 3 4 Size: Speed: 500 bytes 64 KB to 8 MB 1 GB 1 TB sub ns 1-2 cycles 10 cycles 20 cycles hundreds cycles ten of thousands cycles SSD R/W: 560 MB/s; 85000 IOps HDD (SATA @ 7,2 krpm) R/W: 100 MB/s; 190 IOps 1500 e/tb 150 e/tb 10 / 66

Overview of the Main HPC Components HPC Components: Interconnect latency: time to send a minimal (0 byte) message from A to B bandwidth: max amount of data communicated per unit of time Technology Effective Bandwidth Latency Gigabit Ethernet 1 Gb/s 125 MB/s 40µs to 300µs Myrinet (Myri-10G) 9.6 Gb/s 1.2 GB/s 2.3µs 10 Gigabit Ethernet 10 Gb/s 1.25 GB/s 4µs to 5µs Infiniband QDR 40 Gb/s 5 GB/s 1.29µs to 2.6µs SGI NUMAlink 60 Gb/s 7.5 GB/s 1µs 11 / 66

Overview of the Main HPC Components HPC Components: Interconnect latency: time to send a minimal (0 byte) message from A to B bandwidth: max amount of data communicated per unit of time Technology Effective Bandwidth Latency Gigabit Ethernet 1 Gb/s 125 MB/s 40µs to 300µs Myrinet (Myri-10G) 9.6 Gb/s 1.2 GB/s 2.3µs 10 Gigabit Ethernet 10 Gb/s 1.25 GB/s 4µs to 5µs Infiniband QDR 40 Gb/s 5 GB/s 1.29µs to 2.6µs SGI NUMAlink 60 Gb/s 7.5 GB/s 1µs 11 / 66

Overview of the Main HPC Components HPC Components: Operating System Mainly Linux-based OS (91.4%) (Top500, Nov 2011)... or Unix based (6%) Reasons: stability prone to devels 12 / 66

Overview of the Main HPC Components HPC Components: Software Stack Remote connection to the platform: SSH User SSO: NIS or OpenLDAP-based Resource management: job/batch scheduler OAR, PBS, Torque, MOAB Cluster Suite (Automatic) Node Deployment: FAI (Fully Automatic Installation), Kickstart, Puppet, Chef, Kadeploy etc. Platform Monitoring: Nagios, Ganglia, Cacti etc. (eventually) Accounting: oarnodeaccounting, Gold allocation manager etc. 13 / 66

Overview of the Main HPC Components HPC Components: Data Management Storage architectural classes & I/O layers Application [Distributed] File system SATA SAS FC... iscsi... Network NFS CIFS AFP... Network DAS Interface SAN Interface NAS Interface SATA DAS Fiber Channel Ethernet/ Network Fiber Channel Ethernet/ Network SAS Fiber Channel SATA SAN File System NAS SAS SATA Fiber Channel SAS Fiber Channel 14 / 66

Overview of the Main HPC Components HPC Components: Data Management RAID standard levels 15 / 66

Overview of the Main HPC Components HPC Components: Data Management RAID combined levels 15 / 66

Overview of the Main HPC Components HPC Components: Data Management RAID combined levels 15 / 66

Overview of the Main HPC Components HPC Components: Data Management RAID combined levels Software vs. Hardware RAID management RAID Controller card performances differs! Basic (low cost): 300 MB/s; Advanced (expansive): 1,5 GB/s 15 / 66

Overview of the Main HPC Components HPC Components: Data Management File Systems Logical manner to store, organize, manipulate and access data. Disk file systems: FAT32, NTFS, HFS, ext3, ext4, xfs... Network file systems: NFS, SMB Distributed parallel file systems: HPC target data are stripped over multiple servers for high performance. generally add robust failover and recovery mechanisms Ex: Lustre, GPFS, FhGFS, GlusterFS... HPC storage make use of high density disk enclosures includes [redundant] RAID controllers 16 / 66

Overview of the Main HPC Components HPC Components: Data Center Definition (Data Center) Facility to house computer systems and associated components Basic storage component: rack (height: 42 RU) 17 / 66

Overview of the Main HPC Components HPC Components: Data Center Definition (Data Center) Facility to house computer systems and associated components Basic storage component: rack (height: 42 RU) Challenges: Power (UPS, battery), Cooling, Fire protection, Security Power/Heat dissipation per rack: HPC (computing) racks: 30-40 kw Storage racks: 15 kw Interconnect racks: 5 kw Power Usage Effectiveness PUE = Total facility power IT equipment power 17 / 66

Overview of the Main HPC Components HPC Components: Data Center 18 / 66

Overview of the Main HPC Components HPC Components: Summary HPC platforms involves: A data center / server room carefully designed Computing elements: CPU/GPGPU Interconnect elements Storage elements: HDD/SDD, disk enclosure, disks are virtually aggregated by RAID/LUNs/FS A flexible software stack Above all: expert system administrators... 19 / 66

HPC and Cloud Computing (CC) Summary 1 Introduction 2 Overview of the Main HPC Components 3 HPC and Cloud Computing (CC) 4 The UL HPC platform 5 UL HPC in Practice: Toward an [Efficient] Win-Win Usage 20 / 66

HPC and Cloud Computing (CC) CC characteristics in HPC context Horizontal scalability: perfect for replication/ HA (High Availability) best suited for runs with minimal communication and I/O nearly useless for true parallel/distributed HPC runs Cloud Data storage Data locality enforced for performance Data outsourcing vs. legal obligation to keep data local Accessibility, security challenges Virtualization layer lead to decreased performances huge overhead induced on I/O + no support of IB [Q E F]DR Cost effectiveness? 21 / 66

HPC and Cloud Computing (CC) Virtualization layer overhead 1000 Baseline Intel Baseline AMD Xen Intel Xen AMD KVM Intel KVM AMD VMWare ESXi Intel VMWare ESXi AMD Raw Result 100 10 HPL (TFlops) [1] M. Guzek, S. Varrette, V. Plugaru, J. E. Sanchez, and P. Bouvry. A Holistic Model of the Performance and the Energy-Efficiency of Hypervisors in an HPC Environment. LNCS EE-LSDS 13, Apr 2013. 22 / 66

HPC and Cloud Computing (CC) EC2 [Poor] performances Illustration on UnixBench index (the higher the better) Gist dedicated server 10 faster than m1.small instance (both 65$/m) IO throughput 29 higher for just $13 more a month. 23 / 66

HPC and Cloud Computing (CC) EC2 [Poor] performances NAS Parallel Benchmark (NPB) compared to ICHEC [1] [1] K Iqbal and E. Brazil. HPC vs. Cloud Benchmarking. An empirical evaluation of the performance and cost metrics. efiscal Workshop, 2012 24 / 66

HPC and Cloud Computing (CC) EC2 [Poor] performances NAS Parallel Benchmark (NPB) compared to ICHEC [1] [1] K Iqbal and E. Brazil. HPC vs. Cloud Benchmarking. An empirical evaluation of the performance and cost metrics. efiscal Workshop, 2012 24 / 66

HPC and Cloud Computing (CC) EC2 [Poor] performances NAS Parallel Benchmark (NPB) compared to ICHEC [1] [1] K Iqbal and E. Brazil. HPC vs. Cloud Benchmarking. An empirical evaluation of the performance and cost metrics. efiscal Workshop, 2012 24 / 66

HPC and Cloud Computing (CC) CC Cost effectiveness... Amazon HPC instances (as cc2.8xlarge) inherent restrictions Cluster Compute Eight Extra Large Instance 10 GbE interconnect only max 240 IPs Hardware cost coverage chaos+gaia usage: 11,154,125 CPUhour (1273 years) since 2007 15,06M$ on EC2* vs. 4 Me cumul. HW investment *cc2.8xlarge Cluster Compute Eight Extra Large Instance EU Data transfer cost (up to 0.12$ per GB) and performances 614,4$ to retrieve a single 5TB microscopic image Internet (GEANT network) to reach Ireland Amazon data center 25 / 66

CSO 23179 HPC and Cloud Computing (CC) The Magellan Report on Cloud Computing for Science U.S. Department of Energy Office of Advanced Scientific Computing Research (ASCR) December, 2011 Magellan Report: CC for Science Download Virtualization power of value for scientists enable fully customized environments flexible resource management Requires significant prog./sysadmin support Significant gaps and challenges exist managing VMs, workflows, data, security... Public clouds can be expansive more expansive than in-house large systems Ideal for resource consolidation HPC centers should continuously benchmark their computing cost DOE centers are cost competitive, (3-7 less expensive) when compared to commercial cloud providers 26 / 66

The UL HPC platform Summary 1 Introduction 2 Overview of the Main HPC Components 3 HPC and Cloud Computing (CC) 4 The UL HPC platform 5 UL HPC in Practice: Toward an [Efficient] Win-Win Usage 27 / 66

The UL HPC platform UL HPC platforms at a glance (2013) 2 geographic sites Kirchberg campus (AS.28, CS.43) LCSB building (Belval) 4 clusters: chaos+gaia, granduc, nyx. 291 nodes, 2944 cores, 27.363 TFlops 1042TB shared storage (raw capa.) 3 system administrators 4,091,010e (Cumul. HW Investment) since 2007 Hardware acquisition only 2,122,860e (excluding server rooms) Open-Source software stack SSH, LDAP, OAR, Puppet, Modules... 28 / 66

The UL HPC platform HPC server rooms 2009 CS.43 (Kirchberg campus) 14 racks, 100 m 2, 800,000e 2011 LCSB 6 th floor (Belval) 14 racks, 112 m 2, 1,100,000e 29 / 66

The UL HPC platform UL HPC Computing Nodes chaos Date Vendor Proc. Description #N #C R peak 2010 HP Intel Xeon L5640@2.26GHz 2 6C,24GB 32 384 3.472 TFlops 2011 Dell Intel Xeon L5640@2.26GHz 2 6C,24GB 16 192 1.736 TFlops 2012 Dell Intel Xeon X7560@2,26GHz 4 6C, 1TB 1 32 0.289 TFlops 2012 Dell Intel Xeon E5-2660@2.2GHz 2 8C,32GB 16 256 4.506 TFlops 2012 HP Intel Xeon E5-2660@2.2GHz 2 8C,32GB 16 256 4.506 TFlops chaos TOTAL: 81 1124 14.508 TFlops gaia 2011 Bull Intel Xeon L5640@2.26GHz 2 6C,24GB 72 864 7.811 TFlops 2012 Dell Intel Xeon E5-4640@2.4GHz 4 8C, 1TB 1 32 0.307 TFlops 2012 Bull Intel Xeon E7-4850@2GHz 16 10C,1TB 1 160 1.280 TFLops 2013 Viridis ARM A9 Cortex@1.1GHz 1 4C,4GB 96 384 0.422 TFlops gaia TOTAL: 170 1440 9.82 TFlops 2008 Dell Intel Xeon L5335@2GHz 2 4C,16GB 22 176 1.408 TFlops 2012 Dell Intel Xeon E5-2630L@2GHz 2 6C,24GB 16 192 1.536 TFlops granduc/petitprince TOTAL: 38 368 2.944 TFlops Testing cluster: nyx 2012 Dell Intel Xeon E5-2420@1.90GHz 1 6C,32GB 2 12 0.091 TFlops g5k TOTAL: 291 nodes, 2944 cores, 27.363 TFlops 30 / 66

The UL HPC platform UL HPC: General cluster organization Other Clusters network Site <sitename> Local Institution Network 10 GbE 10 GbE 1 GbE 1 GbE Site router Site access server Cluster A Adminfront OAR Puppet supervision Kadeploy etc... Fast local interconnect (Infiniband, 10GbE) Site Computing Nodes Cluster B NFS and/or Lustre Disk Enclosure Site Shared Storage Area S. Varrette, PhD. (UL) HPC platforms @ UL N 31 / 66

The UL HPC platform Ex: The chaos cluster Uni.lu (Kirchberg) Chaos cluster LCSB Belval (gaia cluster) Uni.lu 10 GbE 10 GbE 1 GbE Chaos cluster access CS.43 (416 cores) Computing Nodes Bull R423 (2U) Cisco Nexus C5010 10GbE (2*4c Intel Xeon E5630 @ 2.53GHz), RAM: 24GB 10 GbE Adminfront 10 GbE Dell PE R610 (2U) (2*4c Intel Xeon L5640 @ 2,26 GHz), RAM: 64GB NFS server FC8 FC8 [32 cores] (4*6c Intel Xeon X7560@2,26GHz), RAM:1TB 1x HP c7000 enclosure (10U) 32 blades HP BL2x220c G6 [384 cores] IB Dell R710 (2U) (2*4c Intel Xeon E5506 @ 2.13GHz), RAM: 24GB 1x Dell R910 (4U) Infiniband QDR 40 Gb/s (Min Hop) IB (2*6c Intel Xeon L5640@2.26GHz), RAM: 24GB IB AS.28 (708 cores) Computing Nodes NetApp E5486 (180 TB) 60 disks (3 TB SAS 7.2krpm) = 180 TB (raw) Multipathing over 2 controllers (Cache mirroring) 6 RAID6 LUNs (8+2 disks) = 144TB (lvm + xfs) Chaos cluster characteristics 2x HP SL6500 (8U) 16 blades SL230s Gen8 [256 cores] - Computing: 81 nodes, 1124 cores; Rpeak 14.508 TFlops 1x Dell M1000e enclosure (10U) 16 blades Dell M610 [196 cores] - Storage: 180 TB (NFS) + 180TB (NFS, backup) 1x Dell M1000e enclosure (10U) 16 blades Dell M620 [256 cores] (2*8c Intel Xeon E5-2660@2.2GHz), RAM: 32GB (2*6c Intel Xeon L5640@2.26GHz), RAM: 24GB (2*8c Intel Xeon E5-2660@2.2GHz), RAM: 32GB S. Varrette, PhD. (UL) HPC platforms @ UL N 32 / 66

The UL HPC platform Ex: The gaia cluster Uni.lu (Belval) Kirchberg (chaos cluster) Gaia cluster Gaia cluster characteristics Uni.lu - Computing: 170 nodes, 1440 cores; Rpeak 9,82 TFlops 10 GbE - Storage: 240 TB (NFS) + 180TB (NFS backup) + 240 TB (Lustre) 10 GbE 1 GbE Gaia cluster access Bull R423 (2U) Cisco Nexus C5010 10GbE (2*4c Intel Xeon L5620 @ 2,26 GHz), RAM: 16GB 10 GbE Adminfront LCSB Belval Computing nodes 1x BullX BCS enclosure (6U) 4 BullX S6030 [160 cores] Bull R423 (2U) (2*4c Intel Xeon L5620 @ 2,26 GHz), RAM: 16GB Columbus server (16*10c Intel Xeon E7-4850@2GHz), RAM: 1TB Infiniband QDR 40 Gb/s (Fat tree) IB 2x Viridis enclosure (4U) 96 ultra low-power SoC [384 cores] IB Bull R423 (2U) (1*4c ARM Cortex A9@1.1GHz), RAM: 4GB (2*4c Intel Xeon L5620 @ 2,26 GHz), RAM: 16GB 1x Dell R820 (4U) IB (4*8c Intel Xeon E5-4640@2.4GHz), RAM: 1TB MDS1 NFS server FC8 FC8 FC8 Bull R423 (2U) 5x Bullx B enclosure (35U) 60 BullX B500 [720 cores] (2*6c Intel Xeon L5640@2.26GHz), RAM: 24GB Bull R423 (2U) (2*4c Intel Xeon L5630@2,13 GHz), RAM: 24GB 12032 GPU cores FC8 20 disks (600 GB SAS 15krpm) Multipathing over 2 controllers (Cache mirroring) 2 RAID1 LUNs (10 disks) 6 TB (lvm + lustre) FC8 12 BullX B506 [144 cores] (2*6c Intel Xeon L5640@2.26GHz), RAM: 24GB 20 GPGPU Accelerator [12032 GPU cores] 4 Nvidia Tesla M2070 [448c] 20 Nvidia Tesla M2090 [512c] (2*4c Intel Xeon L5630@2,13 GHz), RAM: 96GB Nexsan E60 + E60X (240 TB) MDS2 OSS1 120 disks (2 TB SATA 7.2krpm) = 240 TB (raw) Multipathing over 2+2 controllers (Cache mirroring) 12 RAID6 LUNs (8+2 disks) = 192 TB (lvm + xfs) OSS2 Bull R423 (2U) (2*4c Intel Xeon L5630@2,13 GHz), RAM: 48GB FC8 Bull R423 (2U) (2*4c Intel Xeon L5630@2,13 GHz), RAM: 48GB 12032 GPU cores FC8 FC8 10 GbE FC8 (2*4c Intel Xeon L5630@2,13 GHz), RAM: 96GB Lustre Storage Bull R423 (2U) Nexsan E60 (4U, 12 TB) [32 cores] 12032 GPU cores 2*Nexsan E60 (2*4U, 2*120 TB) 2*60 disks (2 TB SATA 7.2krpm) = 240 TB (raw) 2*Multipathing over 2 controllers (Cache mirroring) 2*6 RAID6 LUNs (8+2 disks) = 2*96 TB (lvm + lustre) S. Varrette, PhD. (UL) HPC platforms @ UL N 33 / 66

The UL HPC platform Ex: Some racks of the gaia cluster 34 / 66

The UL HPC platform UL HPC Software Stack Characteristics Operating System: Linux Debian (CentOS on storage servers) Remote connection to the platform: SSH User SSO: OpenLDAP-based Resource management: job/batch scheduler: OAR (Automatic) Computing Node Deployment: FAI (Fully Automatic Installation) Puppet Kadeploy (chaos,gaia,nyx only) (granduc,petitprince/grid5000 only) Platform Monitoring: OAR Monika, OAR Drawgantt, Ganglia, Nagios, Puppet Dashboard etc. Commercial Softwares: Intel Cluster Studio XE, TotalView, Allinea DDT, Stata etc. 35 / 66

The UL HPC platform HPC in the Grande region and Around TFlops TB FTEs Country Name/Institute #Cores R peak Storage Manpower Luxembourg France Germany Belgium UK UL 2944 27.363 1042 3 CRP GL 800 6.21 144 1.5 TGCC Curie, CEA 77184 1667.2 5000 n/a LORIA, Nancy 3724 29.79 82 5.05 ROMEO, UCR, Reims 564 4.128 15 2 Juqueen, Juelich 393216 5033.2 448 n/a MPI, RZG 2556 14.1 n/a 5 URZ, (bwgrid),heidelberg 1140 10.125 32 9 UGent, VCS 4320 54.541 82 n/a CECI, UMons/UCL 2576 25.108 156 > 4 Darwin, Cambridge Univ 9728 202.3 20 n/a Legion, UCLondon 5632 45.056 192 6 Spain MareNostrum, BCS 33664 700.2 1900 14 36 / 66

The UL HPC platform Platform Monitoring Monika http://hpc.uni.lu/{chaos,gaia,granduc}/monika 37 / 66

The UL HPC platform Platform Monitoring Drawgantt http://hpc.uni.lu/{chaos,gaia,granduc}/drawgantt 37 / 66

The UL HPC platform Platform Monitoring Ganglia http://hpc.uni.lu/{chaos,gaia,granduc}/ganglia 37 / 66

The UL HPC platform Chronological Statistics 30 25 Chaos cluster Granduc cluster Gaia cluster Computing Requirements Evolution of UL HPC computing capacity Computing capacity [TFlops] 20 15 10 7.24 14.26 21.31 5 0 2.04 2.04 0.11 0.63 2006 2007 2008 2009 2010 2011 2012 38 / 66

The UL HPC platform Chronological Statistics 1600 1400 Storage (NFS) Storage (Lustre) Storage Requirements Evolution of UL HPC storage capacity 1200 Raw Storage capacity [TB] 1000 800 600 511 871 400 200 0 0 4.2 7.2 7.2 2006 2007 2008 2009 2010 2011 2012 31 38 / 66

The UL HPC platform Chronological Statistics UL HPC Yearly Hardware Investment 4,000,000 3,500,000 Server Rooms Interconnect Storage Computing nodes Cumulative hardware investment 4091kE Investment [Euro] VAT exclusive 3,000,000 2,500,000 2,000,000 1,500,000 1,000,000 1142kE 1298kE 3197kE 500,000 277kE 22kE 121kE 2006 2007 2008 2009 2010 2011 2012 38 / 66

The UL HPC platform 142 Registered Users (chaos+gaia only) 160 140 120 Evolution of registered users within UL internal clusters LCSB (Bio Medicine) URPM (Physics and Material Sciences) LBA (ex LSF) RUES (Engineering Science) SnT (Security and Trust) CSC Others (students etc.) Number of users 100 80 60 40 20 0 Jan 2008 Jul 2008 Jan 2009 Jul 2009 Jan 2010 Jul 2010 Jan 2011 Jul 2011 Jan 2012 Jul 2012 Jan 2013 39 / 66

The UL HPC platform 142 Registered Users (chaos+gaia only) LCSB (Bio Medicine) URPM (Physics/Material Sciences) LBA (ex LSF) RUES (Engineering Science) SnT (Security and Trust) CSC Others (students etc.) 100 6 7 11 13 18 19 25 27 27 27 33 36 39 47 52 55 62 68 81 97 110 126 142 80 Percentage (%) 60 40 20 0 Jan 2008 Jul 2008 Jan 2009 Jul 2009 Jan 2010 Jul 2010 Jan 2011 Jul 2011 Jan 2012 Jul 2012 Jan 2013 39 / 66

The UL HPC platform What s new since last year? Current capacity (2013) 291 nodes, 2944 cores, 27.363 TFlops 1042 TB (incl. backup) New Computing nodes for chaos and gaia +10 GPU nodes gaia-{63-72} +1 SMP node (16 procs/160 cores/1tb RAM) gaia-73 +1 big RAM node (4 procs/32 cores/1tb RAM) gaia-74 +16 HP SL nodes chaos: s-cluster1-{1-16} +16 Dell M620 chaos: e-cluster1-{1-16} Other computing nodes + 96 Viridis ARM nodes viridis-{1-48}, viridis-{101-148} +16 Dell M620 nodes Grid5000 petitprince-{1-16} Interconnect consolidation: 10 GbE switches + IB QDR (chaos) 40 / 66

The UL HPC platform What s new since last year? Current capacity (2013) 291 nodes, 2944 cores, 27.363 TFlops 1042 TB (incl. backup) Storage: 3 encl. feat. 60 disks (3TB), 6 x RAID 6 (8+2), xfs+lvm NFS chaos + cross-backup cartman (Kirchberg) / stan (Belval) Commercial software Intel Studio, // debuggers, Mathlab... New OAR policy for a more efficient usage of the platform restrict default jobs to promote container approach GitHUB [before] 10 jobs of 1 core [now] 1 job of 10 cores + GNU parallel better incentives to best-effort jobs for more resilient workflow project/long run management, big{mem,smp} Directory structure ($HOME,$WORK,$SCRATCH), Modules 40 / 66

The UL HPC platform 2013: Incoming Milestones Full website reformatting with improved doc/tutorials Training/Advert: UL HPC school (may-june 2013) OAR RESTful API cluster actions by standard HTTP operations better job monitoring (cost, power consumption etc.) Scalable primary backup (> 1 PB) solution Complement [on demand] cluster capacity (POST, GET, PUT, DELETE) investigate virtualization (Cloud / [K]VMs on the nodes) desktop grid on university TP rooms Job submission web-portal (Extreme Factory)? 41 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Summary 1 Introduction 2 Overview of the Main HPC Components 3 HPC and Cloud Computing (CC) 4 The UL HPC platform 5 UL HPC in Practice: Toward an [Efficient] Win-Win Usage 42 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage General Consideration The platform is *restricted* to UL members and is *shared* Everyone should be civic-minded. Just avoid the following behavior: (or you ll be banned) My work is the most important: I use all the resources for 1 month regularly clean your homedir from useless files Plan large scale experiments during night-time or week-ends try not to use more than 40 computing cores during working day... or use the besteffort queue User Charter Everyone must read and accept the user charter! https://hpc.uni.lu/documentation/user_charter 43 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage User Account Get an account: https://hpc.uni.lu/get_an_account With your account, you ll get: Access to the UL HPC wiki http://hpc.uni.lu/ Access to the UL HPC bug tracker http://hpc-tracker.uni.lu/ Subscribed to the mailing lists hpc-{users,platform}@uni.lu raise questions and concerns. Help us to make it a community! notification of platform maintenance on hpc-platform@uni.lu A nice way to reach workstation in the internal UL network (ProxyCommand) 44 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Typical Workflow on UL HPC resources 1 Connect to the frontend of a site/cluster ssh 2 (eventually) synchronize you code scp/rsync/svn/git 3 (eventually) Reserve a few interactive resources oarsub -I (eventually) Configure the resources kadeploy (eventually) Prepare your experiments gcc/icc/mpicc/javac/... Test your experiment on small size problem mpirun/java/bash... Free the resources 4 Reserve some resources oarsub 5 Run your experiment via a launcher script bash/python/perl/ruby... 6 Grab the results scp/rsync 7 Free the resources 45 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage UL HPC access *Restricted* to SSH connection with public key authentication on a non-standard port (8022) limits kiddie script scans/dictionary s attacks local homedir ~/.ssh/ Client remote homedir ~/.ssh/ SSH server config /etc/ssh/ Server id_dsa.pub authorized_keys sshd_config ssh_config id_dsa ssh_host_dsa_key.pub ssh_host_rsa_key.pub known_hosts ssh_host_dsa_key ssh_host_rsa_key OR 46 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage UL HPC SSH access ~/.ssh/config Host chaos-cluster Hostname access-chaos.uni.lu Host gaia-cluster Hostname access-gaia.uni.lu Host *-cluster User login Port 8022 ForwardAgent no $> ssh {chaos,gaia}-cluster $> ssh myworkstation When @ Home: $> ssh myworkstation.ext_ul Host myworkstation User localadmin Hostname myworkstation.uni.lux Host *.ext_ul ProxyCommand ssh -q gaia-cluster "nc -q 0 %h %p" Transferring data... $> rsync -avzu /devel/myproject chaos-cluster: (gaia)$> gaia_sync_home * (chaos)$> chaos_sync_home devel/ 47 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage UL HPC resource manager: OAR The OAR Batch Scheduler http://oar.imag.fr Versatile resource and task manager schedule jobs for users on the cluster resource OAR resource = a node or part of it (CPU/core) OAR job = execution time (walltime) on a set of resources 48 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage UL HPC resource manager: OAR The OAR Batch Scheduler http://oar.imag.fr Versatile resource and task manager schedule jobs for users on the cluster resource OAR resource = a node or part of it (CPU/core) OAR job = execution time (walltime) on a set of resources OAR main features includes: interactive vs. passive (aka. batch) jobs best effort jobs: use more resource, accept their release any time deploy jobs (Grid5000 only): deploy a customized OS environment... and have full (root) access to the resources powerful resource filtering/matching 48 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Main OAR commands oarsub submit/reserve a job (by default: 1 core for 2 hours) oardel delete a submitted job oarnodes shows the resources states oarstat shows information about running or planned jobs Submission interactive oarsub [options] -I passive oarsub [options] scriptname Each created job receive an identifier JobID Default passive job log files: OAR.JobID.std{out,err} You can make a reservation with -r "YYYY-MM-DD HH:MM:SS" 49 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Main OAR commands oarsub submit/reserve a job (by default: 1 core for 2 hours) oardel delete a submitted job oarnodes shows the resources states oarstat shows information about running or planned jobs Submission interactive oarsub [options] -I passive oarsub [options] scriptname Each created job receive an identifier JobID Default passive job log files: OAR.JobID.std{out,err} You can make a reservation with -r "YYYY-MM-DD HH:MM:SS" Direct access to nodes by ssh is forbidden: use oarsh instead 49 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage OAR job environment variables Once a job is created, some environments variables are defined: Variable $OAR_NODEFILE $OAR_JOB_ID $OAR_RESOURCE_PROPERTIES_FILE $OAR_JOB_NAME $OAR_PROJECT_NAME Description Filename which lists all reserved nodes for this job OAR job identifier Filename which lists all resources and their properties Name of the job given by the -n option of oarsub Job project name Useful for MPI jobs for instance: $> mpirun -machinefile $OAR_NODEFILE /path/to/myprog... Or to collect how many cores are reserved per node: $> cat $OAR_NODEFILE uniq -c 50 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage OAR job types Job Type Max Walltime (hour) Max #active jobs Max #active jobs per user interactive 12:00:00 10000 5 default 120:00:00 30000 10 besteffort 9000:00:00 10000 1000 cf /etc/oar/admission_rules/*.conf interactive: useful to test / prepare an experiment you get a shell on the first reserved resource best-effort vs. default: nearly unlimited constraints YET a besteffort job can be killed as soon as a default job as no other place to go enforce checkpointing (and/or idempotent) strategy 51 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Characterizing OAR resources Specifying wanted resources in a hierarchical manner Use the -l option of oarsub. Main constraints: enclosure=n number of enclosure nodes=n number of nodes core=n number of cores walltime=hh:mm:ss job s max duration Specifying OAR resource properties Use the -p option of oarsub: Syntax: -p "property= value " gpu= {YES,NO} has (or not) a GPU card host= fqdn full hostname of the resource network_address= hostname Short hostname of the resource (Chaos only) nodeclass= {k,b,h,d,r} Class of node 52 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage OAR (interactive) job examples 2 cores on 3 nodes (same enclosure) for 3h15: Total: 6 cores (frontend)$> oarsub -I -l /enclosure=1/nodes=3/core=2,walltime=3:15 53 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage OAR (interactive) job examples 2 cores on 3 nodes (same enclosure) for 3h15: Total: 6 cores (frontend)$> oarsub -I -l /enclosure=1/nodes=3/core=2,walltime=3:15 4 cores on a GPU node for 8 hours Total: 4 cores (frontend)$> oarsub -I -l /core=4,walltime=8 -p "gpu= YES " 53 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage OAR (interactive) job examples 2 cores on 3 nodes (same enclosure) for 3h15: Total: 6 cores (frontend)$> oarsub -I -l /enclosure=1/nodes=3/core=2,walltime=3:15 4 cores on a GPU node for 8 hours Total: 4 cores (frontend)$> oarsub -I -l /core=4,walltime=8 -p "gpu= YES " 2 nodes among the h-cluster1-* nodes (Chaos only) Total: 24 cores (frontend)$> oarsub -I -l nodes=2 -p "nodeclass= h " 53 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage OAR (interactive) job examples 2 cores on 3 nodes (same enclosure) for 3h15: Total: 6 cores (frontend)$> oarsub -I -l /enclosure=1/nodes=3/core=2,walltime=3:15 4 cores on a GPU node for 8 hours Total: 4 cores (frontend)$> oarsub -I -l /core=4,walltime=8 -p "gpu= YES " 2 nodes among the h-cluster1-* nodes (Chaos only) Total: 24 cores (frontend)$> oarsub -I -l nodes=2 -p "nodeclass= h " 4 cores on 2 GPU nodes + 20 cores on other nodes Total: 28 cores $> oarsub -I -l "{gpu= YES }/nodes=2/core=4 } {{ } +{gpu= NO }/core=20 } {{ } " 53 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage OAR (interactive) job examples 2 cores on 3 nodes (same enclosure) for 3h15: Total: 6 cores (frontend)$> oarsub -I -l /enclosure=1/nodes=3/core=2,walltime=3:15 4 cores on a GPU node for 8 hours Total: 4 cores (frontend)$> oarsub -I -l /core=4,walltime=8 -p "gpu= YES " 2 nodes among the h-cluster1-* nodes (Chaos only) Total: 24 cores (frontend)$> oarsub -I -l nodes=2 -p "nodeclass= h " 4 cores on 2 GPU nodes + 20 cores on other nodes Total: 28 cores $> oarsub -I -l "{gpu= YES }/nodes=2/core=4 } {{ } +{gpu= NO }/core=20 } {{ } " A full big SMP node Total: 160 cores on gaia-74 $> oarsub -t bigsmp -I l node=1 53 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Some other useful features of OAR Connect to a running job (frontend)$> oarsub -C JobID Cancel a job (frontend)$> oardel JobID Status of a jobs (frontend)$> oarstat -state -j JobID Get info on the nodes (frontend)$> oarnodes (frontend)$> oarnodes -l (frontend)$> oarnodes -s View the job (frontend)$> oarstat (frontend)$> oarstat -f -j JobID Run a best-effort job (frontend)$> oarsub -t besteffort... 54 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Designing efficient OAR job launchers Resources/Example https://github.com/ulhpc/launcher-scripts UL HPC grant access to parallel computing resources ideally: OpenMP/MPI/CUDA/OpenCL jobs if serial jobs/tasks: run them efficiently Avoid to submit purely serial jobs to the OAR queue a waste the computational power (11 out of 12 cores on gaia). use whole nodes by running at least 12 serial runs at once Key: understand difference between Task and OAR job 55 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Designing efficient OAR job launchers Methodical Design of Parallel Programs [Foster96] I. Foster, Designing and Building Parallel Programs. Addison Wesley, 1996. Available at: http://www.mcs.anl.gov/dbpp 56 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Serial tasks: BAD and NAIVE approach #OAR -l nodes=1 #OAR -n BADSerial #OAR -O BADSerial-%jobid%.log #OAR -E BADSerial-%jobid%.log if [ -f /etc/profile ]; then. /etc/profile fi # Now you can use: module load toto or cd $WORK [...] 57 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Serial tasks: BAD and NAIVE approach #OAR -l nodes=1 #OAR -n BADSerial #OAR -O BADSerial-%jobid%.log #OAR -E BADSerial-%jobid%.log if [ -f /etc/profile ]; then. /etc/profile fi # Now you can use: module load toto or cd $WORK [...] # Example 1: run in sequence $TASK 1...$TASK $NB_TASKS for i in seq 1 $NB_TASKS ; do $TASK $i done # Example 2: For each line of $ARG_TASK_FILE, run in sequence # $TASK <line1>... $TASK <lastline> while read line; do $TASK $line done < $ARG_TASK_FILE 57 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Serial tasks: A better approach (fork & wait) # Example 1: run in sequence $TASK 1...$TASK $NB_TASKS for i in seq 1 $NB_TASKS ; do $TASK $i & done wait # Example 2: For each line of $ARG_TASK_FILE, run in sequence # $TASK <line1>... $TASK <lastline> while read line; do $TASK $line & done < $ARG_TASK_FILE fi wait 58 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Serial tasks: A better approach (fork & wait) Different runs may not take the same time: load imbalance. 59 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Serial tasks with GNU Parallel 60 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Serial tasks with GNU Parallel ### Example 1: run in sequence $TASK 1...$TASK $NB_TASKS # On a single node seq $NB_TASKS parallel -u -j 12 $TASK {} # on many nodes seq $NB_TASKS parallel -tag -u -j 4 \\ -sshloginfile ${GP_SSHLOGINFILE}.task $TASK {} ### Example 2: For each line of $ARG_TASK_FILE, run in parallel # $TASK <line1>... $TASK <lastline> # On a single node cat $ARG_TASK_FILE parallel -u -j 12 -colsep $TASK {} # on many nodes cat $ARG_TASK_FILE parallel -tag -u -j 4 \\ -sshloginfile ${GP_SSHLOGINFILE}.task -colsep $TASK {} 61 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage MPI tasks: 3 Suites via module 1 OpenMPI http://www.open-mpi.org/ (node)$> module load OpenMPI (node)$> make (node)$> mpirun -hostfile $OAR_NODEFILE /path/to/mpi_prog 2 MVAPICH2 http://mvapich.cse.ohio-state.edu/overview/mvapich2 (node)$> module purge (node)$> module load MVAPICH2 (node)$> make clean && make (node)$> mpirun -hostfile $OAR_NODEFILE /path/to/mpi_prog 3 Intel Cluster Toolkit Compiler Edition (ictce for short): (node)$> module purge (node)$> module load ictce (node)$> make clean && make (node)$> mpirun -hostfile $OAR_NODEFILE /path/to/mpi_prog 62 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Last Challenges for a better efficiency Memory bottleneck A regular computing node have at least 2GB/core RAM Do 12-24 runs fit in the memory? If your job runs out of memory, it simply crashes Use fewer simultaneous runs if really needed! OR request a big memory machine (1TB RAM) $> oarsub -t bigmem... Or explore parallization (MPI, OpenMP, pthreads) Use $SCRATCH directory whenever you can gaia: shared between nodes (Lustre FS) chaos: NOT shared (/tmp) and clean at job end. 63 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Last Challenges for a better efficiency My favorite software is not installed on the cluster! Check if it does not exists via module If not: compile it in your home/work directory using GNU stow Share it to others: consider EasyBuild / ModuleFile General workflow for programs based on Autotools http://www.gnu.org/software/stow/ Get the software sources (version x.y.z) Compile and install it in your home/work directory (node)$>./configure [options] -prefix=$basedir/stow/mysoft.x.y.z (node)$> make && make install (node)$> cd $BASEDIR/stow && stow mysoft.x.y.z 64 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Last Challenges for a better efficiency Fault Tolerance Cluster maintenance from time to time 65 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Last Challenges for a better efficiency Fault Tolerance Cluster maintenance from time to time Reliability vs. Crash Faults in Distributed systems 1 0.8 Failing Probability F(t) 0.6 0.4 0.2 0 0 1000 2000 3000 4000 5000 Number of processors execution time: 1 day execution time: 5 days execution time: 10 days execution time: 20 days execution time: 30 days 65 / 66

UL HPC in Practice: Toward an [Efficient] Win-Win Usage Last Challenges for a better efficiency Fault Tolerance Cluster maintenance from time to time Reliability vs. Crash Faults in Distributed systems Fault Tolerance general strategy: checkpoint/rollback assumes a way to save the state of your program hints: OAR -signal -checkpoint -idempotent..., BLCR 65 / 66

Thank you for your attention... http://hpc.uni.lu 1 Introduction 2 Overview of the Main HPC Components 3 HPC and Cloud Computing (CC) 4 The UL HPC platform 5 UL HPC in Practice: Toward an [Efficient] Win-Win Usage Contacts: hpc-sysadmins@uni.lu 66 / 66