Access to the Federal High-Performance Computing-Centers



Similar documents
Architecture of Hitachi SR-8000

MEGWARE HPC Cluster am LRZ eine mehr als 12-jährige Zusammenarbeit. Prof. Dieter Kranzlmüller (LRZ)

Supercomputing Status und Trends (Conference Report) Peter Wegner

and RISC Optimization Techniques for the Hitachi SR8000 Architecture

How To Write A Parallel Computer Program

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Performance Characteristics of a Cost-Effective Medium-Sized Beowulf Cluster Supercomputer

A Scalable Ethernet Clos-Switch

Performance of the JMA NWP models on the PC cluster TSUBAME.

Parallel Programming Survey

Welcome to the. Jülich Supercomputing Centre. D. Rohe and N. Attig Jülich Supercomputing Centre (JSC), Forschungszentrum Jülich

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Building a Top500-class Supercomputing Cluster at LNS-BUAP

Kriterien für ein PetaFlop System

GPU System Architecture. Alan Gray EPCC The University of Edinburgh

OpenMP Programming on ScaleMP

Cosmological simulations on High Performance Computers

JuRoPA. Jülich Research on Petaflop Architecture. One Year on. Hugo R. Falter, COO Lee J Porter, Engineering

64-Bit versus 32-Bit CPUs in Scientific Computing

Extreme Scale Compu0ng at LRZ

Current Trend of Supercomputer Architecture

Extreme Scaling on Energy Efficient SuperMUC

Lecture 2 Parallel Programming Platforms

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

Introduction to High Performance Cluster Computing. Cluster Training for UCL Part 1

Supercomputing Resources in BSC, RES and PRACE

Altix Usage and Application Programming. Welcome and Introduction

A Brief Survery of Linux Performance Engineering. Philip J. Mucci University of Tennessee, Knoxville

BSC - Barcelona Supercomputer Center

Mixed Precision Iterative Refinement Methods Energy Efficiency on Hybrid Hardware Platforms

Building Clusters for Gromacs and other HPC applications

GPFS und HPSS am HLRS

Parallel Computing. Introduction

Performance of Various Computers Using Standard Linear Equations Software

Clusters: Mainstream Technology for CAE

PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN

Getting the Performance Out Of High Performance Computing. Getting the Performance Out Of High Performance Computing

OpenPOWER Outlook AXEL KOEHLER SR. SOLUTION ARCHITECT HPC

Cluster Computing at HRI

High performance parallel computing for Computational Fluid

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

Linux Cluster Computing An Administrator s Perspective

1 Bull, 2011 Bull Extreme Computing

HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK

bwgrid Treff MA/HD Sabine Richling, Heinz Kredel Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim 24.

SERVER CLUSTERING TECHNOLOGY & CONCEPT

LS DYNA Performance Benchmarks and Profiling. January 2009

The K computer: Project overview

1 DCSC/AU: HUGE. DeIC Sekretariat /RB. Bilag 1. DeIC (DCSC) Scientific Computing Installations

How Cineca supports IT

Virtualised MikroTik

High Performance Computing in the Multi-core Area

HPC Update: Engagement Model

High Performance Computing. Course Notes HPC Fundamentals

COMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1)

Wind-Tunnel Simulation using TAU on a PC-Cluster: Resources and Performance Stefan Melber-Wilkending / DLR Braunschweig

Cloud Computing through Virtualization and HPC technologies

Cray XT3 Supercomputer Scalable by Design CRAY XT3 DATASHEET

Performance Characteristics of Large SMP Machines

A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures

bwgrid Treff HD/MA Sabine Richling, Heinz Kredel Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim 30.

High Performance Computing within the AHRP

Experiences with HPC on Windows

Improved LS-DYNA Performance on Sun Servers

Evaluation of CUDA Fortran for the CFD code Strukti

Lecture 1: the anatomy of a supercomputer

ANALYSIS OF SUPERCOMPUTER DESIGN

Cray Gemini Interconnect. Technical University of Munich Parallel Programming Class of SS14 Denys Sobchyshak

Linux tools for debugging and profiling MPI codes

PCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters. from One Stop Systems (OSS)

High Performance Computing, an Introduction to

Logically a Linux cluster looks something like the following: Compute Nodes. user Head node. network

Thematic Unit of Excellence on Computational Materials Science Solid State and Structural Chemistry Unit, Indian Institute of Science

Parallel Programming and High-Performance Computing

GASPI A PGAS API for Scalable and Fault Tolerant Computing

RLX Technologies Server Blades

Grids Computing and Collaboration

ST810 Advanced Computing

Accelerating CFD using OpenFOAM with GPUs

Transcription:

Access to the Federal High-Performance Computing-Centers rabenseifner@hlrs.de University of Stuttgart High-Performance Computing-Center Stuttgart (HLRS) www.hlrs.de Slide 1 TOP 500 Nov. List German Sites, 1 st part federal thematic regional Rank 21 27 Installation Site Max-Planck-Gesellschaft MPI/IPP Leibniz Rechenzentrum Location Garching München R max 2010 =51% 1653 =82% Peak 3910 2016 Manufacturer Computer / Procs 1.3GHz/ 752 Hitachi SR8000-F1/168/ 168 Year of Installat ion 2000 / 31 Deutscher Wetterdienst Offenbach 1293 =67% 1920 IBM SP Power3 375MHz 16way/ 1280 2001 44 HLRN at Universitaet Hannover / RRZN Hannover 1038 =52% 1.3GHz/ 384 45 HLRN at ZIB/Konrad Zuse- Zentrum fuer Informationstechnik Berlin 1038 =52% 1.3GHz/ 384 49 DKRZ - Deutsches Klimarechenzentrum Hamburg 982 =96% 1024 NEC SX-6/128M16/ 128 64 Universitaet Heidelberg - IWR Heidelberg 825 =58% 1430 Megware HELICS AMD1.4GHz Myrinet/ 512 84 Deutscher Wetterdienst Offenbach 671 =69% 974 Cray Inc. T3E1200/ 812 1999 Jülich 135 Forschungszentrum Juelich Slide 2 / 12 173 Universitaet Aachen/RWTH Aachen 447 =69% 357 =69% 648 518 Cray Inc. T3E1200/ 540 SunFire15K/ 288 1999

TOP 500 Nov. List German Sites, 2 nd part federal thematic regional Rank 176 181 Installation Site Max-Planck-Gesellschaft MPI/IPP HWW/Universitaet Stuttgart Location Garching Stuttgart R max 355 =73% 341 =70% Peak 487 486 Manufacturer Computer / Procs Cray Inc. T3E/ 812 Cray Inc. T3E900/ 540 Year of Installat ion 1997 183 ZIB/Konrad Zuse-Zentrum fuer Informationstechnik Berlin 341 =70% 486 Cray Inc. T3E900/ 540 1999 302 Theoretical Chemistry, Ruhr- University Bochum Bochum 235.8 =66% 358.4 Megware AMD Athlon MP1600+ - SCI 2D-Torus/ 128 313 Universitaet Darmstadt Darmstadt 234.0 =47% 499.2 1.3GHz GigEth/ 96 315 Forschungszentrum Juelich Jülich 234 =72% 324 Cray Inc. T3E600/ 540 333 GWDG Göttingen 226 =67% 336 IBM SP Power3 375 MHz/ 224 2001 339 Technische Universitaet Chemnitz Chemnitz 221.6 =52% 424.0 Self-made CLIC PIII 800 MHz/ 530 2000 Göttingen 373 GWDG Slide 3 / 12 454 Universitaet Aachen/RWTH Aachen 198 =45% 197.3 442.4 259.2 IBM pseries 690 1.1GHz GigEth/ 96 SunFire15K/ 144 455 Universitaet Aachen/RWTH Aachen 197.3 259.2 SunFire15K/ 144 High Performance Computing in Germany Federal HLRS/HWW Stuttgart www.hlrs.de NIC Jülich www.kfa-juelich.de/nic LRZ/HLRB Munich www.lrz-muenchen.de Thematic (federal) RZ Garching www.rzg.mpg.de DKRZ/HLRE Hamburg www.dkrz.de DWD Offenbach www.dwd.de/de/technik/datenverarbeitung/ Regional ZIB Berlin www.zib.de RRZN Hannover www.rrzn.uni-hannover.de See also www.hlrs.de/hpc/centers.html and www.top500.org Slide 4 / 12

Architectures Parallel Vector Processor (PVP) & Symmetric Multi Processor (SMP) NEC SX-6, SX-5 CPUs Cray SV1, X1 Memory-Interconnect Memory MPP, massively parallel processor achieve Cray T3E-1200, 900, 600 Linux cluster Memory CPUs Node Interconnect Hierarchical: Cluster of SMP nodes Hitachi SR8000 IBM SP NEC SX cluster Linux cluster Node Interconnect nodes CPUs Memory-Interconnect Memory Slide 5 / 12 Vector Processing some platforms perform vector processing vector processor mainly a question of memory bandwidth how many load/stores per floating point operation? memory-interconnect for which memory access pattern the compiler supports the vector access the hardware supports the vector access Slide 6 / 12

HLRS Stuttgart: hww hww configuration (2003) NEC SX-4/40 H2 NEC SX-5/32 M2e Cray SV1/20 Fileserver Research networks debis networks hp V2250 Vaihingen 2 GBit/s d 20 km Untertürkheim SGI Origin 2000 Hitachi SR8000 Cray T3E-900/512 Slide 7 / 12 HLRS at Stuttgart High-Performance Computing-Center Stuttgart (HLRS) at University of Stuttgart Platforms: NEC SX-5/32 (128 GFlop/s peak) NEC SX-6/48 (550 GFlop/s peak) vector system, available ~ April 2004 Next generation NEC SX? (~ 12 TFlop/s peak) vector system, available ~ early 2005 Cray T3E 900-512 will be substituted by: IA32 and IA64 clusters Online-Proposals: www.hlrs.de > HW&Access > Access > proposal > test account (Schnupperaccount) Slide 8 / 12 NEC SX-6 NEC SX-5/32 M2e 2*16 PEs, 128 80 GByte Cray T3E 900-512 512 PEs, 461 64 GByte

NIC at Jülich John von Neumann-Institut für Computing (NIC) at Forschungszentrum Jülich Platforms: IBM Regatta (30x32 CPUs Power4+ 1.7 GHz, 128 GB/node, Federation Switch with 4 adapters per 32-CPU-SMP node, since Feb 11, 2004) Extension to 41 nodes: March 2004, 8.9 Tflop/s Cray T3E 1200-512 (will be switched off in Summer 2004) Cray SV1ex Online-Proposals: www.fz-juelich.de/nic/nic-e.html > Computing Time Deadlines April 30, for accounts valid July June (next year) Oct. 31, for accounts valid Jan. Dec. (next year) Slide 9 / 12 LRZ / HLRB at Munich: Hitachi SR8000 System: 168 nodes, 2.02 TFLOP/s peak 1.65 TFLOP/s Linpack 1.3 TB memory Node: 8 CPUs, 12 GFLOP/s 8 GB, SMP pseudo-vector ext. b/w: 950 MB/s CPU: 1.5 GFLOP/s, 375 MHz 4 GB/s memory b/w Installed: 1.Q 2000 at LRZ (112 nodes) Slide 10 / 12

Hitachi SR 8000-F1/112 (Rank 5 in TOP 500 / June 2000) Höchstleistungsrechner in Bayern (HLRB) at Leibniz-Rechenzentrum (LRZ) Platform: Hitachi SR8000-F1 (168 nodes, each with 8 processors) Online-Proposals: www.lrz.de/services/compute/hlrb/ > Online Project Proposal Submission Slide 11 / 12 High Performance Computing in Germany Summary Federal HLRS/HWW Stuttgart NIC Jülich LRZ/HLRB Munich Thematic (federal) RZ Garching DKRZ Hamburg DWD Offenbach Regional ZIB Berlin RRZN Hannover See also www.hlrs.de/hpc/centers.html www.hlrs.de >HW&Access >Access www.fz-juelich.de/ nic/nic-e.html www.lrz.de/services/ compute/hlrb/ Slide 12 / 12