High Performance Computing, an Introduction to
|
|
|
- Ross Lindsey
- 10 years ago
- Views:
Transcription
1 , an Introduction to Nicolas Renon, Ph. D, Research Engineer in Scientific Computations CALMIP - DTSI Université Paul Sabatier University of Toulouse ([email protected]) Toulouse University Computing Center : Page 1 Heat Equation in 1 D (diffusion) Page 2 1
2 Heat Equation in 1 D (diffusion) Implémentation(MATLAB) Explicit Scheme Page 3 Heat Equation in 1 D (diffusion) Page 4 2
3 time to solution : example #1 % Heat equa@on [0,1] EXPLICIT SCHEME %- - - maximum Time tmax=0.02; % mesh J=200; u_init=exp(- (x- 0.5).^2/(2*1/6^2)); ymin=- 0.1 ; ymax=1.5; t=0; % test 1 : Explicit scheme while (t<tmax) t=t+dt ; %t1=cpu@me; u=u+dt*(mu/dx^2*(pun(u,ud)- 2*u+mun(u,ug))+source(mu,x(2:J+1),t)); %temps=temps + cpu@me- t1; End Computa1on 1me size(j=) tmax=0,02 0,02 0,2 2, seconds tmax =0,2 0,28 162, seconds tmax=20 3,44 hours Page 5 time to solution : example #1 tmax=0, y = 0,0025e 2,1907x tmax=0,02 Expon. (tmax=0,02) Time to solution : 3 hours to reach solution, is it reasonable? Exponential behavior, why? Hardware? Page 6 3
4 26/11/13 What about size of the problem N= ? Courtesy of H. Neau Institut de Mecanique des Fluides de Toulouse Master Pro M2 IMAT - Octobre 2013 Université Paul Saba@er Page 7 Summary : High Performance Compu1ng - HPC HPC, what is it? What for? Concepts in HPC (processors and parallelism) Moore s Law, Amhdal s Law HPC computa1onal System and processor architecture Taxonomy : Shared Memory Machines Distributed Memory Machines Scalar, Superscalar, mule- core processors Code opemisaeon Parallel Computa1on Make it happen : Message Passing Interface MPI (+BE) OpenMP Basics : speed up, parallel efficiency and traps Master Pro M2 IMAT - Octobre 2013 Université Paul Saba@er Page 8 4
5 Numerical Simulation : 3rd Pilar of Science A chain relationship of 3 pilars : sic. Jack Dongarra Univ. Knoxville, Tennessee Validation / new theory theory Fluids Dynamic (Navier Stokes) Electromagnetism (Maxwell) Quantum Mechanics (Schrödinger) Experimenta@ons Observa@ons Large Hadron Collider (CERN geneva) ITER Hubble Validation Numerical Simula@on numerical implementation Approximations Page 9 Scientific computation numerical simulation thanks to Scientific computation : Efficient (optimum) execution of computer program Time to solution!!! Computer program (Shrödinger) Quantic Chemistry: Gaussian Atomic scale Material : VASP (Genomic) Blast (Mécanique) : Crash : LS-Dyna, Structure : ABAQUS, GETFEM++ (Navier-Stokes) Fluid Dynamic : OpenFOAM, Fluent, GETFEM ++ High Performance Compu@ng Mathema@cs Scientific Computation : transverse (all disciplines). Discretisation : spatial (FEM, Finite Volume, time (Explicit, implict scheme, Numerical Methods (Solve Linear Systems, Algebra, ) Page 10 5
6 HPC : Why? Problem is too : Complex Different scales Expensive Dangerous (nuclear) Competitiveness : Best product before others Page 11 HPC : where? Science Climate Change Astrophysics Biology, genomics, pharmacoloy Chemistry nano-technology Engineering Crash Earthquake CFD Combustion Economy : Financial Market Classified Cryptography Nuclear Page 12 6
7 Tantalum resolidification HPC : Research computation/memory and accuracy Page 13 Machine CURIE Programme Europe «PRACE» (France) Cosmos SImulation, from big-bang to nowdays Dark Energy Simulation Series (DEUSS) : 550 billions of Particules (N-body simulation) Go RAM (distributed) cores Jean- Michel ALIMI1, Yann RASERA 1, Pier- Stefano CORASANITI 1, Jérome COURTIN 1, Fabrice ROY 1, André FUZFA 1,2, Vincent BOUCHER 3,4, Romain TEYSSIER 4 1 Laboratoire Univers et Théories (LUTH), CNRS- UMR8102, Observatoire de Paris, Université Paris Diderot 2 GAMASCO, Université de NAMUR, Belgique 3 CP3, Université catholique de LOUVAIN, Belgique 4 Service d Astrophysique, CEA Saclay Page 14 7
8 26/11/13 Simulate Proteines (molecules) related to Alzheimer cores 960 TFLOP 2900 base fonctions LCPQ-IRSAMC CNRS/University of Toulouse / Université Paul Sabatier M. Caffarel - A. Scemama Machine CURIE Programme Europe «PRACE» (France) Master Pro M2 IMAT - Octobre 2013 Université Paul Saba@er Page 15 HPC : Competitiveness (Research and Industry) Proctor & Gamble use HPC for the Pringles? Master Pro M2 IMAT - Octobre 2013 Université Paul Saba@er Page 16 8
9 HPC : Competitivness (Research and Industry) capacity to Model a whole engine? Reduce Design Costs Faster in the marcket code coupling : Fluid and Mecanics Fluid and chemistry Have to be confident in the numerical solution!!! Page 17 HPC : Competitivness (Research and Industry) capacity to Model a whole aircraft? Size of a mesh? A part of the aircraft? A whole aircraft Page 18 9
10 HPC : Virtual Prototyping Virtual prototyping : run virtual test cheaper and faster CFD (Navier Stokes) => example : Brake cooling Codes : Fluent (commercial) OpenFOAM(free somware) FEM => Impact Crash Codes (commercial): Abaqus, Radioss CFD (combus@on) => In cylinder simula@on Codes (commercial): Ansys- CFX, AVBP (Academic/ commercial) Feel confident enough on the result!!! Page 19 Virtual Prototyping : Slide from PRACE workshop Septembre 2009 Virtual protoyping : run virtual test cheaper and faster Feel confident enough on the result!!! Page 20 10
11 26/11/13 Master Pro M2 IMAT - Octobre 2013 Université Paul Saba@er Page 21 HPC for experimentation Key Word : Data Deluge, «BigDATA» Large Hadron Collider : Higgs Boson Huge DATA management 40 millions of events per seconds GAFA (Google, Apple, Facebook, Amazon), and others Man Kind created data Peta-bytes in Peta-bytes in 2010 Credit : The Dtat deluge, Economist; Understanding Big Data, Eaton et al. Master Pro M2 IMAT - Octobre 2013 Université Paul Saba@er Page 22 11
12 HPC in a Process chain Reasearch & Developement Highly Parallel code FEM, CFD, etc Analysis/ Visualisa@on Supercompu@ng/ Data ManagementStorage Page 23 HPC : Operating High technology need to compute more and faster Solve the real World!! need to think new, new efficient technics/algorithm Concern the deepest level of program performance at processor (core) level massively parallel approach Computational Systems associated are hugely powerful what means powerful? what means performance? Applica1on (scien1fic problem to solve) HPC Mathema1cs Page 24 12
13 TOP 500 List TOP 500 : 2 1mes per year (June and November) the list of the most powerfull compu1ng Top500 Juin 2012 # # Page 25 HPC : TOP 500 List What means powerfull? we re dealing with : flop/s FLOATING OPERATION PER SECOND Simple opera1on (+,*) on x R There is a Peak performance (theory : Rpeak) There is a «for real» performance!! (R max) depend on what (applica1on/program) you re running on Benchmark : Linpack (Linear Package) Solve thanks to the system you re benching : Ax = b ; A R n R n, n > How far/close are you from the Rpeak( theory) power? TOP500 look at Rmax and not Rpeak Page 26 13
14 Top500 Juin 2012 TOP 500 List Rmax # # PF 10 PF Machine CURIE Programme Europe «PRACE» (France) Page 27 HPC Try to figure out what represent all this flops : basic idea: 1 kflop/s = 10 3 opera1ons on floats 1 page, each side with 5 column of 100 numbers process in 1 second 10 6 flop/s (1 Mflop) 1000 pages 2 «ream of paper» (10 cm high ; 3 inches high) Page 28 14
15 Calcul Haute Performance Basic idea: 10 9 flop/s (1 Gflop order of the power of 1 core ) ο( 1p) 1 column of 100 meter high (Second floor of Eiffel Tower) flop/s (1 Tflop «linle» parallel machine ) 1 col. 100 km (mesosphere, Meteores, aurora) ο 10 5 p flop/s (1 Pflop current 1Oth posi1on in TOP 500 ) km (1/3 moon- earth distance) flop/s (1 exaflop - imagine?) ( ) ο( 100p) Page 29 Japon : K-computer (Y2011), cores 10 PF Fujitsu RIKEN 8 PFlop (Kobe Japan) efficiency 93% 548 k core Sparc 10 Mw, 825 Mflop/W Single socket, 8C socket, 8 flop per core 6D tore Interconnect Page 30 15
16 Top500 June 2012 Calcul Haute Performance : TOP 500 List Moore s Law? # # MW 12 MW Nuclear Plant between 40 MW and 1450 MW Page 31 Machine CURIE Programme Europe «PRACE» (France) Top500 Juin MW # # Page 32 16
17 Machine haute Performance Page 33 (Research) : France / Europe European European Stage (PRACE): O(1000) TFLOP Na1onal Na1onal Stage (GENCI): O(100) TFLOP Toulouse Mesocentre Regional Stage : O(10) TFLOP Na1onal Weather Forecast (Meteo France) CNES CALMIP : Toulouse University Compu1ng Center Total (Petroleum Industry) PANGEA : 2PF GENCI : Grands Equipements Nationaux en Calcul Intensif Page 34 17
18 26/11/13 (Research) : Toulouse Toulouse University Compu@ng System (CALMIP) «HYPERION» 3500 cores / 38,57 TF Rpeak 223rd TOP 500 (November 2009,that s the middle age) Master Pro M2 IMAT - Octobre 2013 Université Paul Saba@er Page 35 Example in CFD : Industrial fluidised bed Neptune_CFD : cpu hours used in Y2010 on HYPERION Number of cores range in produc@on : 68c - 512c Courtesy of : O Simonin, H. Neau, Laviéville - Ins@tut de Mécanique des fluides de Toulouse - Université de Toulouse/ CNRS Master Pro M2 IMAT - Octobre 2013 Université Paul Saba@er Page 36 18
19 Laboratoire Evolution et Diversité Biologique (UM5174) Example in Life Science Fish studies in fish in two Rivers near by Toulouse fish studies cpu hour used mostly data set exploring (code in C! Good!) Chevaine Gandoise Vairon Page 37 HPC in Europe (Partnership for Advanced Computing in Europe) Top500 June 2012 FLOP #8 JUGENE 1PFLOP HERMIT 1PFLOP #9 CURIE 2PFLOP SUPERMUC 3PFLOP #4 European (PRACE) National (GENCI) MareNostrum 1PF FERMI 2.1PFLOP #7 Mesocentre (University/ Région) Page 38 19
20 TOP 500 Geographical repartition Page 39 HPC : Operationnal Technology For a given application/program : Get the best of the processor (core) HPC concepts Tunning code / processor architecture. Dedicated Scientific Librairies make the code Parallel : How far can we get? (algorithm) according to parallel efficiency What kind of Computing System (big computers : Supercomputer) High performance processors (compute power, memory (data) bandwidth), bande passante mémoire) A lot of memory (distributed, Shared) Tens of thousands of processors : interconnection? Page 40 20
21 Summary of the next part Units in HPC Processors (cores?) Moore s Law The Energy Problem efficiency of programs (Rmax vs. Rpeak) Parallelism Concepts Amdahl s Law Page 41 Units in HPC Flop : floa1ng opera1on (mult, add) opéra1on sur les nombres à virgule flonantes (nombre réels) 3, , E+08 Order of 1 Kflop/s 1 Kiloflop/s 10 3 Flop/sec Kilo from thousand in Greek 1 Mflop/s 1 Megaflop/s 10 6 Flop/sec Mega from large in greek 1 Gflop/s 1 Gigaflop/s 10 9 Flop/sec 1 Tflop/s 1 Téraflop/s Flop/sec 1 Pflop/s 1 Pétaflop/s Flop/sec 1 Eflop/s 1 Exaflop/s Flop/sec Processor/ core Departement system Current highest level To be reached eventually (?) Giga from Giant in greek Tera from Monster in greek Peta from five in greek( [10 3 ] 5 ) Exa from six in greek( [10 3 ] 6 ) Page 42 21
22 Units in HPC Data : memory (bytes) cache RAM (DIMMs) Hard Disk Order of 1 Ko(KB) 1 Kilobytes 10 3 bytes register 1 Mo (MB) 1 Megabytes 10 6 bytes Cache processor 1 Go (GB) 1 Gigabytes 10 9 bytes RAM 1 To (TB) 1 Térabytes bytes Disk 1 Po (PB) 1 Pétabytes bytes 1 octet (o) = 1 Byte (B) = 8 bit Huge File systems Page 43 Units in Big DATA Systems aggrega1on Order of 1 Po (PB) 1 Pétabyte bytes All letters delivered by America s postal service in one year 1 Eo (EB) 1 ExaByte bytes 10 Billion of copies of «Le Monde» 1 Zo (ZB) 1 Zettabyte bytes Total amount of information in existence 1 Yo (YB) 1 Yottabyte bytes? Credit : The Dtat deluge, Economist; Understanding Big Data, Eaton et al. Page 44 22
23 Latency «1me- to- reach» (roughly) time amount From to 1 nano-second 1 Kilobytes Computing units (in the core) register 100 nano-second 1 Megabytes core RAM Mico-second 1 Gigabytes Process process 100 Micro-second 1 Térabytes Computer computer bandwidth Access RAM Interconnexion 1 Mo/s 1 Megabytes/s 10 6 bytes/sec Interconnect 1 Go/s 1 Gigabytes/s 10 9 bytes/sec Fast interconnect/ RAM 1 To/s 1 Térabytes/s bytes/s Exists? Page 45 Processors Classical Moore Law Energy Problem (not good news for the users) The end of the «free lunch» Applica1ons, are they efficient in terms of HPC? Page 46 23
24 processor architecture FPU, ALU memory Control unit Comput unit register Data memory < Memory (cache) cache processor cycle Clock frequency = number of pulse per second 200 Mhz 200 Millions cycles per second Page 47 L g I Page 48 24
25 Moore(Intel), Moore Law is not a Physical law «something» double each 18 month «something» = transistors «something» related to «performance» this is an exponen1el Law could it last? (True for 30 years) emperical anyway Years The Nighties (end) Year 2000 (begining) Now Days Processeurs Pentium III Itanium Xeon (Quad core) Power 1GFlop 6Gflop 40Gflop Page 49 Energy Problem The end of the Race for Frequency, why : Thermal Dissipa1on : frequency*(tension 2 )*Cste (frequency) 3 Using new(old) technics for cooling : water Lots of electric supply (see TOP500), because lots of processors Smartphones, Tablets : need to compute faster with fewer energy (save baneries energy) Processeurs nowdays voltage : pocket lamp (1 volt) intensity : an oven (250 A) power : a light (100 wans) surface : a big stamp(10cm 2 ) Frequency stable around 3 Ghz Page 50 25
26 Needs always increasing (perpetual and exponen1al) Moore s Law: «something» increase increase performance under constraint (energy) parameters we can play with? nbre of transistors processors «smarter» (new architecture) other ways : accelerators (GPGPU, Xeon Phi) L g I Page 51 Current vendors choice : Increase number of transistors : Engrave : 65 nm, 45 nm, 32 nm (nano : 10-9 ) Constant (lower!) frequency Same architecture Conclusion/consequence : mul1- core/many- core processors Cache Cache core core core engrave size reduc@on (45 nm, ) 1 processor mono-core 1 processeurs ( or socket) multi-core (multi = 2, 4, 8, 16, ) Page 52 26
27 A core, what is it? That s compute (decode instruc1on and perform it) opera1ons/instruc1on Data : Registre (small memory buffers) Page 53 Problem number 1 «Free» gain of performance with increase of frequency «free» gain of performance with increase of the number of cores Mul1- core : If your applica1on/program is not parallel => NO GAINS! 4 cores «No Free Lunch» opera1on per seconds 2 cores Parallel pogram 1 core «Free Lunch» Sequen@al program. frequency Page 54 27
28 Problem number 2 : efficiency of your applica1on execu1on applica1on efficiency What we observe (roughly) : 10% of peak performance some applica1ons are doing bener some worth (really worth)! Page 55 Top500 Juin 2012 TOP 500 List Rmax/Rpeak # # % 90% Page 56 28
29 Problem number 2 : efficiency of your applica1on execu1on applica1on efficiency What we observe (roughly) : 10% of peak performance some applica1ons are doing bener some worth (really worth)! Some leads: Memory Instruc1ons per processor cycle poten1ally 4 or 8 opera1on per cycle Applica1on : 1,2-1,4 re1red opera1on per cycle Page 57 Memory Access Problem Why fast processors are slow? wai1ng for the data Equilibrium between compute velocity and data transfer velocity temps (ns) 10 période d'horloge cpu (ns) Mem acces time 1 0, année Mem acces time période d'horloge cpu (ns) Page 58 29
30 Try to minimize the problem : principle of a hierarchical memory small capacity fast access big capacity slow access memory Disk Time acces 300 ns memory RAM Memory DIMM 10 ns Cache level 2 (ou 3) 2 ns 1 ns Cache level 1 Register ship silicium Page 59 Parallelism Why? Automa1c? Speedup Parallel Efficiency (Scalability) and amdahl law Page 60 30
31 Parallel Compu1ng Solve problem faster Solve bigger problem (more realis1c) How? Make «processor s» work together «communica1on» problem Independant use of ressources Concepts in parallelism efficiency - speed- up amdahl law Parallel algorithm : granularity (lot of small tasks vs. Few big tasks) Load (compute) equilibrium and Synchronisa1on Harder to make a parallel program than a sequan1al one: but, things are gezng bener (technics), this is the aim of the training hardware gezng bener too parallelism is now everywhere (mul1- core/mul1- socket/processors laptop, GPGPU, ) Page 61 Automa1c parallelism (processor level, invisble for us, human ) ILP : Instruc1on parallelism Find independant instruc1ons and comput them in parallel Automa1c parallelism limited efficiency only at processor level So an explicit parallelisa1on have to be made Algorithm level Use of standard technics (API, Librairies) Fizng with a target(generic?) machine Page 62 31
32 Parallel applica1on : can use several ( ) cores/processors simultaneously If I am using 10 processors/cores, am I running 10 1mes faster? (do I get the solu1on 10 1mes faster?) Speedup = T Serial T Parallel Good efficiency in // if : speedup Number of processor Page 63 Applica1on scalability : speed- up evolu1on according to the number of processor 1000 Perfect efficiency 100 speed-up parfait Appli 1 Appli 2 Appli Nombre de processeurs Difficult : keep good efficiency with lots of processors Why? Algorithmic Aspects Compu1ng system capability Page 64 32
33 Prac1ce yourself (D.I.Y.) : Algorithm aspect Hypothesis 1 : Only a part of the applica1on(code) is parallel f p + f s =1 f p = parallel fraction of the code f s = sequential fraction of the code Hypothesis 2 : the parallel part is op1mal (perfect behaviour for any number of processor)! t N = f p N + f $ # s &t 1 " % t 1 = time to solution serial / sequential t N = time to solution for N processors Page 65 Evaluate theori1cal Speed- up, for a given number of processors N, knowing the serial part : f s = 1 10 Evaluate the limit (max.?): S( N) = t 1 t N f p + f s =1 lim S =? N What Does it means? Consequence : amdhal s law Page 66 33
34 amdahl law: time execution 1 core (sequential) 1 core (processor) Sequential part Parallel part 2 cores 4 cores n cores Time execution with n cores Page 67 amdahl law : lim N S = 1 f s speed-up parfait fs=0,1 fs=0,05 fp=0, Nombre de processeurs Page 68 34
35 26/11/13 Algorithm Aspect : split the problem One Domain (Mesh, i.e element) Basic Idea : Split in subdomain (create a par11on of the inital mesh) Each core/processor work on a sub- domain example : 64 cores (sub- domain) :=> element per core Mesh(domain) par11oned in sub mesh (domain) Master Pro M2 IMAT - Octobre 2013 Université Paul Saba@er Page 69 Algorithm Aspect : split the problem Parallel : TParallel = TComputing + TOverhead Overhead related to communica@on What we would like : Increase numbre of core/processor :=> decrease Time Control Overhead should increase not to much Master Pro M2 IMAT - Octobre 2013 Université Paul Saba@er Page 70 35
36 Distributed memory example in CFD : Industrial fluidized bed Time- to- solu@on and speed- up (exemple for 5 millions of Cells) HYPERION HYPERION C1 a : Al@x ICE Harpertown, Intel_MPI C1 b : Al@x ICE Harpertown, MPT C2 a : HYPERION Al@x ICE NHM, Intel MPI C2 b : HYPERION Al@x ICE NHM, MPT C3 : Cluster IB AMD Shanghaï, OpenMPI Courtesy of : O Simonin, H. Neau, Laviéville - Ins@tut de Mécanique des fluides de Toulouse - Université de Toulouse/ CNRS Page 71 Machine A t N A = time to solution for N processors Machine B t N B = time to solution for N processors Hypothesis 1 : Machine A is 10 Times slower than machine B t N A = 10.t N B compute speed up for both machine: S A A ( N) = t 1 t = 10 t 1 A N 10 t = t 1 B N t = S B B N B B ( N ) Same speed- up, S1ll have a the solu1on sooner with machine B Page 72 36
37 Calcul Haute Performance Loi de amdahl : Hypothèses : Seule une par1e de l applica1on est parallélisée La par1e parallélisée est op1male " t N = f p N + f % $ s' t 1 # & 1 = f p + f s S = t 1 t N = lim N S = 1 f s 1 f p N + f s t N = temps calcul N procésseurs f p = fraction parallèle f s = fraction séquentielle S = speedup Page 73 Prac1ce yourself : Hypothesis : Only a part of the applica1on(code) is parallel the parallel part is op1mal (perfect behaviour for any number of processor)! t N = f p N + f $ # s &t 1 ; 1= f p + f s " % give the speedup of the code with : f s = 1 10 t N = compute time for N processors f p = parallel fraction of the code f s = sequential fraction of the code S = speedup S =?? lim S =? N Consequence? amdhal s law Page 74 37
38 HPC: Complexity Page 75 time to solution : example #2 Matrix Product Basis of most (all) numerical method C, A,B Μ n m C = A B n c ij = Σ aik b kj k=1 do i = 1, n do j = 1, n do k = 1, n c(i,j) = c(i,j) + a(i,k) * b(k,j) end do end do end do n= (Unknows) A11 A12 A13 A21 A22 A23 A31 A32 A33 B11 B12 B13 B21 B22 B23 = B31 B32 B33 C11 C12 C13 C21 C22 C23 C31 C32 C33 how many matrix product Time to solu1on 10 2,98 hours ,76 hours ,40 days ,40 days ,97 days Page 76 38
39 Transport Equation in 1 D (convection) Page 77 Transport Equation in 1 D (convection) Explicit Scheme Page 78 39
40 Transport Equation in 1 D (convection) Page 79 40
CALMIP a Computing Meso-center for middle HPC academis Users
CALMIP a Computing Meso-center for middle HPC academis Users CALMIP : Calcul en Midi-Pyrénées Méso-centre de Calcul Boris Dintrans, President Operational Board CALMIP, Laboratoire d Astrophysique de Toulouse
High Performance Computing, an Introduction to
High Performance ing, an Introduction to Nicolas Renon, Ph. D, Research Engineer in Scientific ations CALMIP - DTSI Université Paul Sabatier University of Toulouse ([email protected]) Michel
Trends in High-Performance Computing for Power Grid Applications
Trends in High-Performance Computing for Power Grid Applications Franz Franchetti ECE, Carnegie Mellon University www.spiral.net Co-Founder, SpiralGen www.spiralgen.com This talk presents my personal views
COMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1)
COMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1) Mary Thomas Department of Computer Science Computational Science Research Center (CSRC) San Diego State University
Building a Top500-class Supercomputing Cluster at LNS-BUAP
Building a Top500-class Supercomputing Cluster at LNS-BUAP Dr. José Luis Ricardo Chávez Dr. Humberto Salazar Ibargüen Dr. Enrique Varela Carlos Laboratorio Nacional de Supercómputo Benemérita Universidad
YALES2 porting on the Xeon- Phi Early results
YALES2 porting on the Xeon- Phi Early results Othman Bouizi Ghislain Lartigue Innovation and Pathfinding Architecture Group in Europe, Exascale Lab. Paris CRIHAN - Demi-journée calcul intensif, 16 juin
Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer
Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer Stan Posey, MSc and Bill Loewe, PhD Panasas Inc., Fremont, CA, USA Paul Calleja, PhD University of Cambridge,
10- High Performance Compu5ng
10- High Performance Compu5ng (Herramientas Computacionales Avanzadas para la Inves6gación Aplicada) Rafael Palacios, Fernando de Cuadra MRE Contents Implemen8ng computa8onal tools 1. High Performance
Overview on Modern Accelerators and Programming Paradigms Ivan Giro7o [email protected]
Overview on Modern Accelerators and Programming Paradigms Ivan Giro7o [email protected] Informa(on & Communica(on Technology Sec(on (ICTS) Interna(onal Centre for Theore(cal Physics (ICTP) Mul(ple Socket
22S:295 Seminar in Applied Statistics High Performance Computing in Statistics
22S:295 Seminar in Applied Statistics High Performance Computing in Statistics Luke Tierney Department of Statistics & Actuarial Science University of Iowa August 30, 2007 Luke Tierney (U. of Iowa) HPC
Parallel Programming Survey
Christian Terboven 02.09.2014 / Aachen, Germany Stand: 26.08.2014 Version 2.3 IT Center der RWTH Aachen University Agenda Overview: Processor Microarchitecture Shared-Memory
Kriterien für ein PetaFlop System
Kriterien für ein PetaFlop System Rainer Keller, HLRS :: :: :: Context: Organizational HLRS is one of the three national supercomputing centers in Germany. The national supercomputing centers are working
HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK
HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK Steve Oberlin CTO, Accelerated Computing US to Build Two Flagship Supercomputers SUMMIT SIERRA Partnership for Science 100-300 PFLOPS Peak Performance
Data Centric Computing Revisited
Piyush Chaudhary Technical Computing Solutions Data Centric Computing Revisited SPXXL/SCICOMP Summer 2013 Bottom line: It is a time of Powerful Information Data volume is on the rise Dimensions of data
Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca
Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca Carlo Cavazzoni CINECA Supercomputing Application & Innovation www.cineca.it 21 Aprile 2015 FERMI Name: Fermi Architecture: BlueGene/Q
Logical Operations. Control Unit. Contents. Arithmetic Operations. Objectives. The Central Processing Unit: Arithmetic / Logic Unit.
Objectives The Central Processing Unit: What Goes on Inside the Computer Chapter 4 Identify the components of the central processing unit and how they work together and interact with memory Describe how
Jean-Pierre Panziera Teratec 2011
Technologies for the future HPC systems Jean-Pierre Panziera Teratec 2011 3 petaflop systems : TERA 100, CURIE & IFERC Tera100 Curie IFERC 1.25 PetaFlops 256 TB ory 30 PB disk storage 140 000+ Xeon cores
Lecture 11: Multi-Core and GPU. Multithreading. Integration of multiple processor cores on a single chip.
Lecture 11: Multi-Core and GPU Multi-core computers Multithreading GPUs General Purpose GPUs Zebo Peng, IDA, LiTH 1 Multi-Core System Integration of multiple processor cores on a single chip. To provide
Large Scale Simulation on Clusters using COMSOL 4.2
Large Scale Simulation on Clusters using COMSOL 4.2 Darrell W. Pepper 1 Xiuling Wang 2 Steven Senator 3 Joseph Lombardo 4 David Carrington 5 with David Kan and Ed Fontes 6 1 DVP-USAFA-UNLV, 2 Purdue-Calumet,
Introduction to Cloud Computing
Introduction to Cloud Computing Parallel Processing I 15 319, spring 2010 7 th Lecture, Feb 2 nd Majd F. Sakr Lecture Motivation Concurrency and why? Different flavors of parallel computing Get the basic
High Performance Computing. Course Notes 2007-2008. HPC Fundamentals
High Performance Computing Course Notes 2007-2008 2008 HPC Fundamentals Introduction What is High Performance Computing (HPC)? Difficult to define - it s a moving target. Later 1980s, a supercomputer performs
The K computer: Project overview
The Next-Generation Supercomputer The K computer: Project overview SHOJI, Fumiyoshi Next-Generation Supercomputer R&D Center, RIKEN The K computer Outline Project Overview System Configuration of the K
HPC Deployment of OpenFOAM in an Industrial Setting
HPC Deployment of OpenFOAM in an Industrial Setting Hrvoje Jasak [email protected] Wikki Ltd, United Kingdom PRACE Seminar: Industrial Usage of HPC Stockholm, Sweden, 28-29 March 2011 HPC Deployment
CS140: Parallel Scientific Computing. Class Introduction Tao Yang, UCSB Tuesday/Thursday. 11:00-12:15 GIRV 1115
CS140: Parallel Scientific Computing Class Introduction Tao Yang, UCSB Tuesday/Thursday. 11:00-12:15 GIRV 1115 1 CS 140 Course Information Instructor: Tao Yang (tyang@cs). Office Hours: T/Th 10-11(or email
Cosmological simulations on High Performance Computers
Cosmological simulations on High Performance Computers Cosmic Web Morphology and Topology Cosmological workshop meeting Warsaw, 12-17 July 2011 Maciej Cytowski Interdisciplinary Centre for Mathematical
High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/2015. 2015 CAE Associates
High Performance Computing (HPC) CAEA elearning Series Jonathan G. Dudley, Ph.D. 06/09/2015 2015 CAE Associates Agenda Introduction HPC Background Why HPC SMP vs. DMP Licensing HPC Terminology Types of
Performance Characteristics of a Cost-Effective Medium-Sized Beowulf Cluster Supercomputer
Res. Lett. Inf. Math. Sci., 2003, Vol.5, pp 1-10 Available online at http://iims.massey.ac.nz/research/letters/ 1 Performance Characteristics of a Cost-Effective Medium-Sized Beowulf Cluster Supercomputer
Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging
Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging In some markets and scenarios where competitive advantage is all about speed, speed is measured in micro- and even nano-seconds.
Accelerating Simulation & Analysis with Hybrid GPU Parallelization and Cloud Computing
Accelerating Simulation & Analysis with Hybrid GPU Parallelization and Cloud Computing Innovation Intelligence Devin Jensen August 2012 Altair Knows HPC Altair is the only company that: makes HPC tools
Parallel Computing. Introduction
Parallel Computing Introduction Thorsten Grahs, 14. April 2014 Administration Lecturer Dr. Thorsten Grahs (that s me) [email protected] Institute of Scientific Computing Room RZ 120 Lecture Monday 11:30-13:00
Introduction to the Mathematics of Big Data. Philippe B. Laval
Introduction to the Mathematics of Big Data Philippe B. Laval Fall 2015 Introduction In recent years, Big Data has become more than just a buzz word. Every major field of science, engineering, business,
Introduction to High Performance Cluster Computing. Cluster Training for UCL Part 1
Introduction to High Performance Cluster Computing Cluster Training for UCL Part 1 What is HPC HPC = High Performance Computing Includes Supercomputing HPCC = High Performance Cluster Computing Note: these
Performance of the JMA NWP models on the PC cluster TSUBAME.
Performance of the JMA NWP models on the PC cluster TSUBAME. K.Takenouchi 1), S.Yokoi 1), T.Hara 1) *, T.Aoki 2), C.Muroi 1), K.Aranami 1), K.Iwamura 1), Y.Aikawa 1) 1) Japan Meteorological Agency (JMA)
BSC - Barcelona Supercomputer Center
Objectives Research in Supercomputing and Computer Architecture Collaborate in R&D e-science projects with prestigious scientific teams Manage BSC supercomputers to accelerate relevant contributions to
LS-DYNA Scalability on Cray Supercomputers. Tin-Ting Zhu, Cray Inc. Jason Wang, Livermore Software Technology Corp.
LS-DYNA Scalability on Cray Supercomputers Tin-Ting Zhu, Cray Inc. Jason Wang, Livermore Software Technology Corp. WP-LS-DYNA-12213 www.cray.com Table of Contents Abstract... 3 Introduction... 3 Scalability
GPU System Architecture. Alan Gray EPCC The University of Edinburgh
GPU System Architecture EPCC The University of Edinburgh Outline Why do we want/need accelerators such as GPUs? GPU-CPU comparison Architectural reasons for GPU performance advantages GPU accelerated systems
2: Computer Performance
2: Computer Performance http://people.sc.fsu.edu/ jburkardt/presentations/ fdi 2008 lecture2.pdf... John Information Technology Department Virginia Tech... FDI Summer Track V: Parallel Programming 10-12
Chapter 4 System Unit Components. Discovering Computers 2012. Your Interactive Guide to the Digital World
Chapter 4 System Unit Components Discovering Computers 2012 Your Interactive Guide to the Digital World Objectives Overview Differentiate among various styles of system units on desktop computers, notebook
Lecture 1: the anatomy of a supercomputer
Where a calculator on the ENIAC is equipped with 18,000 vacuum tubes and weighs 30 tons, computers of the future may have only 1,000 vacuum tubes and perhaps weigh 1½ tons. Popular Mechanics, March 1949
A Very Brief History of High-Performance Computing
A Very Brief History of High-Performance Computing CPS343 Parallel and High Performance Computing Spring 2016 CPS343 (Parallel and HPC) A Very Brief History of High-Performance Computing Spring 2016 1
Parallelism and Cloud Computing
Parallelism and Cloud Computing Kai Shen Parallel Computing Parallel computing: Process sub tasks simultaneously so that work can be completed faster. For instances: divide the work of matrix multiplication
Lecture 3: Evaluating Computer Architectures. Software & Hardware: The Virtuous Cycle?
Lecture 3: Evaluating Computer Architectures Announcements - Reminder: Homework 1 due Thursday 2/2 Last Time technology back ground Computer elements Circuits and timing Virtuous cycle of the past and
High Performance Computing in CST STUDIO SUITE
High Performance Computing in CST STUDIO SUITE Felix Wolfheimer GPU Computing Performance Speedup 18 16 14 12 10 8 6 4 2 0 Promo offer for EUC participants: 25% discount for K40 cards Speedup of Solver
SGI High Performance Computing
SGI High Performance Computing Accelerate time to discovery, innovation, and profitability 2014 SGI SGI Company Proprietary 1 Typical Use Cases for SGI HPC Products Large scale-out, distributed memory
1 Bull, 2011 Bull Extreme Computing
1 Bull, 2011 Bull Extreme Computing Table of Contents HPC Overview. Cluster Overview. FLOPS. 2 Bull, 2011 Bull Extreme Computing HPC Overview Ares, Gerardo, HPC Team HPC concepts HPC: High Performance
Hardware-Aware Analysis and. Presentation Date: Sep 15 th 2009 Chrissie C. Cui
Hardware-Aware Analysis and Optimization of Stable Fluids Presentation Date: Sep 15 th 2009 Chrissie C. Cui Outline Introduction Highlights Flop and Bandwidth Analysis Mehrstellen Schemes Advection Caching
Discovering Computers 2011. Living in a Digital World
Discovering Computers 2011 Living in a Digital World Objectives Overview Differentiate among various styles of system units on desktop computers, notebook computers, and mobile devices Identify chips,
CHAPTER 2: HARDWARE BASICS: INSIDE THE BOX
CHAPTER 2: HARDWARE BASICS: INSIDE THE BOX Multiple Choice: 1. Processing information involves: A. accepting information from the outside world. B. communication with another computer. C. performing arithmetic
Surfing the Data Tsunami: A New Paradigm for Big Data Processing and Analytics
Surfing the Data Tsunami: A New Paradigm for Big Data Processing and Analytics Dr. Liangxiu Han Future Networks and Distributed Systems Group (FUNDS) School of Computing, Mathematics and Digital Technology,
Fast Multipole Method for particle interactions: an open source parallel library component
Fast Multipole Method for particle interactions: an open source parallel library component F. A. Cruz 1,M.G.Knepley 2,andL.A.Barba 1 1 Department of Mathematics, University of Bristol, University Walk,
SPARC64 VIIIfx: CPU for the K computer
SPARC64 VIIIfx: CPU for the K computer Toshio Yoshida Mikio Hondo Ryuji Kan Go Sugizaki SPARC64 VIIIfx, which was developed as a processor for the K computer, uses Fujitsu Semiconductor Ltd. s 45-nm CMOS
White Paper The Numascale Solution: Extreme BIG DATA Computing
White Paper The Numascale Solution: Extreme BIG DATA Computing By: Einar Rustad ABOUT THE AUTHOR Einar Rustad is CTO of Numascale and has a background as CPU, Computer Systems and HPC Systems De-signer
The Central Processing Unit:
The Central Processing Unit: What Goes on Inside the Computer Chapter 4 Objectives Identify the components of the central processing unit and how they work together and interact with memory Describe how
numascale White Paper The Numascale Solution: Extreme BIG DATA Computing Hardware Accellerated Data Intensive Computing By: Einar Rustad ABSTRACT
numascale Hardware Accellerated Data Intensive Computing White Paper The Numascale Solution: Extreme BIG DATA Computing By: Einar Rustad www.numascale.com Supemicro delivers 108 node system with Numascale
Big Data Challenges in Bioinformatics
Big Data Challenges in Bioinformatics BARCELONA SUPERCOMPUTING CENTER COMPUTER SCIENCE DEPARTMENT Autonomic Systems and ebusiness Pla?orms Jordi Torres [email protected] Talk outline! We talk about Petabyte?
Outline. High Performance Computing (HPC) Big Data meets HPC. Case Studies: Some facts about Big Data Technologies HPC and Big Data converging
Outline High Performance Computing (HPC) Towards exascale computing: a brief history Challenges in the exascale era Big Data meets HPC Some facts about Big Data Technologies HPC and Big Data converging
Data Management in the Cloud: Limitations and Opportunities. Annies Ductan
Data Management in the Cloud: Limitations and Opportunities Annies Ductan Discussion Outline: Introduc)on Overview Vision of Cloud Compu8ng Managing Data in The Cloud Cloud Characteris8cs Data Management
Design and Optimization of OpenFOAM-based CFD Applications for Hybrid and Heterogeneous HPC Platforms
Design and Optimization of OpenFOAM-based CFD Applications for Hybrid and Heterogeneous HPC Platforms Amani AlOnazi, David E. Keyes, Alexey Lastovetsky, Vladimir Rychkov Extreme Computing Research Center,
Barry Bolding, Ph.D. VP, Cray Product Division
Barry Bolding, Ph.D. VP, Cray Product Division 1 Corporate Overview Trends in Supercomputing Types of Supercomputing and Cray s Approach The Cloud The Exascale Challenge Conclusion 2 Slide 3 Seymour Cray
Sun Constellation System: The Open Petascale Computing Architecture
CAS2K7 13 September, 2007 Sun Constellation System: The Open Petascale Computing Architecture John Fragalla Senior HPC Technical Specialist Global Systems Practice Sun Microsystems, Inc. 25 Years of Technical
64-Bit versus 32-Bit CPUs in Scientific Computing
64-Bit versus 32-Bit CPUs in Scientific Computing Axel Kohlmeyer Lehrstuhl für Theoretische Chemie Ruhr-Universität Bochum March 2004 1/25 Outline 64-Bit and 32-Bit CPU Examples
Chapter 6. Inside the System Unit. What You Will Learn... Computers Are Your Future. What You Will Learn... Describing Hardware Performance
What You Will Learn... Computers Are Your Future Chapter 6 Understand how computers represent data Understand the measurements used to describe data transfer rates and data storage capacity List the components
Cluster Computing at HRI
Cluster Computing at HRI J.S.Bagla Harish-Chandra Research Institute, Chhatnag Road, Jhunsi, Allahabad 211019. E-mail: [email protected] 1 Introduction and some local history High performance computing
A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures
11 th International LS-DYNA Users Conference Computing Technology A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures Yih-Yih Lin Hewlett-Packard Company Abstract In this paper, the
Quiz for Chapter 1 Computer Abstractions and Technology 3.10
Date: 3.10 Not all questions are of equal difficulty. Please review the entire quiz first and then budget your time carefully. Name: Course: Solutions in Red 1. [15 points] Consider two different implementations,
High Performance Computing
High Performance Computing Introduction Salvatore Orlando 1 General information about the course Web Site: Register in moodle.unive.it OFFERTA FORMATIVA > Corsi di Laurea Magistrale > Dipartimento di Scienze
PRACE the European HPC Research Infrastructure. Carlos Mérida-Campos, Advisor of Spanish Member at PRACE Council
PRACE the European HPC Research Infrastructure Carlos Mérida-Campos, Advisor of Spanish Member at PRACE Council Barcelona, 6-June-2013 PRACE an European e-infrastructure & ESFRI-list item in operation
BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB
BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB Planet Size Data!? Gartner s 10 key IT trends for 2012 unstructured data will grow some 80% over the course of the next
GPU Hardware and Programming Models. Jeremy Appleyard, September 2015
GPU Hardware and Programming Models Jeremy Appleyard, September 2015 A brief history of GPUs In this talk Hardware Overview Programming Models Ask questions at any point! 2 A Brief History of GPUs 3 Once
Parallel Algorithm Engineering
Parallel Algorithm Engineering Kenneth S. Bøgh PhD Fellow Based on slides by Darius Sidlauskas Outline Background Current multicore architectures UMA vs NUMA The openmp framework Examples Software crisis
How To Compare Amazon Ec2 To A Supercomputer For Scientific Applications
Amazon Cloud Performance Compared David Adams Amazon EC2 performance comparison How does EC2 compare to traditional supercomputer for scientific applications? "Performance Analysis of High Performance
An Open Dynamic Big Data Driven Applica3on System Toolkit
An Open Dynamic Big Data Driven Applica3on System Toolkit Craig C. Douglas University of Wyoming and KAUST This research is supported in part by the Na3onal Science Founda3on and King Abdullah University
Unit 4: Performance & Benchmarking. Performance Metrics. This Unit. CIS 501: Computer Architecture. Performance: Latency vs.
This Unit CIS 501: Computer Architecture Unit 4: Performance & Benchmarking Metrics Latency and throughput Speedup Averaging CPU Performance Performance Pitfalls Slides'developed'by'Milo'Mar0n'&'Amir'Roth'at'the'University'of'Pennsylvania'
HPC enabling of OpenFOAM R for CFD applications
HPC enabling of OpenFOAM R for CFD applications Towards the exascale: OpenFOAM perspective Ivan Spisso 25-27 March 2015, Casalecchio di Reno, BOLOGNA. SuperComputing Applications and Innovation Department,
Introduction History Design Blue Gene/Q Job Scheduler Filesystem Power usage Performance Summary Sequoia is a petascale Blue Gene/Q supercomputer Being constructed by IBM for the National Nuclear Security
Big Data Streams. Analytics Challenges, Analysis, and Applications. Adel M. Alimi
Big Data Streams 1 Analytics Challenges, Analysis, and Applications Adel M. Alimi REGIM-Lab., University of Sfax, Tunisia http://adel.alimi.regim.org [email protected] 2 Evolution of Technology 3 Nano,
High-Performance Computing and Big Data Challenge
High-Performance Computing and Big Data Challenge Dr Violeta Holmes Matthew Newall The University of Huddersfield Outline High-Performance Computing E-Infrastructure Top500 -Tianhe-II UoH experience: HPC
Hadoop. http://hadoop.apache.org/ Sunday, November 25, 12
Hadoop http://hadoop.apache.org/ What Is Apache Hadoop? The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using
Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband
Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband A P P R O I N T E R N A T I O N A L I N C Steve Lyness Vice President, HPC Solutions Engineering [email protected] Company Overview
Lattice QCD Performance. on Multi core Linux Servers
Lattice QCD Performance on Multi core Linux Servers Yang Suli * Department of Physics, Peking University, Beijing, 100871 Abstract At the moment, lattice quantum chromodynamics (lattice QCD) is the most
The High Performance Internet of Things: using GVirtuS for gluing cloud computing and ubiquitous connected devices
WS on Models, Algorithms and Methodologies for Hierarchical Parallelism in new HPC Systems The High Performance Internet of Things: using GVirtuS for gluing cloud computing and ubiquitous connected devices
A GPU COMPUTING PLATFORM (SAGA) AND A CFD CODE ON GPU FOR AEROSPACE APPLICATIONS
A GPU COMPUTING PLATFORM (SAGA) AND A CFD CODE ON GPU FOR AEROSPACE APPLICATIONS SUDHAKARAN.G APCF, AERO, VSSC, ISRO 914712564742 [email protected] THOMAS.C.BABU APCF, AERO, VSSC, ISRO 914712565833
Building an energy dashboard. Energy measurement and visualization in current HPC systems
Building an energy dashboard Energy measurement and visualization in current HPC systems Thomas Geenen 1/58 [email protected] SURFsara The Dutch national HPC center 2H 2014 > 1PFlop GPGPU accelerators
Slide Set 8. for ENCM 369 Winter 2015 Lecture Section 01. Steve Norman, PhD, PEng
Slide Set 8 for ENCM 369 Winter 2015 Lecture Section 01 Steve Norman, PhD, PEng Electrical & Computer Engineering Schulich School of Engineering University of Calgary Winter Term, 2015 ENCM 369 W15 Section
Chapter 2 Parallel Architecture, Software And Performance
Chapter 2 Parallel Architecture, Software And Performance UCSB CS140, T. Yang, 2014 Modified from texbook slides Roadmap Parallel hardware Parallel software Input and output Performance Parallel program
Lecture 3: Modern GPUs A Hardware Perspective Mohamed Zahran (aka Z) [email protected] http://www.mzahran.com
CSCI-GA.3033-012 Graphics Processing Units (GPUs): Architecture and Programming Lecture 3: Modern GPUs A Hardware Perspective Mohamed Zahran (aka Z) [email protected] http://www.mzahran.com Modern GPU
Comparing the performance of the Landmark Nexus reservoir simulator on HP servers
WHITE PAPER Comparing the performance of the Landmark Nexus reservoir simulator on HP servers Landmark Software & Services SOFTWARE AND ASSET SOLUTIONS Comparing the performance of the Landmark Nexus
Next Generation GPU Architecture Code-named Fermi
Next Generation GPU Architecture Code-named Fermi The Soul of a Supercomputer in the Body of a GPU Why is NVIDIA at Super Computing? Graphics is a throughput problem paint every pixel within frame time
Supercomputing 2004 - Status und Trends (Conference Report) Peter Wegner
(Conference Report) Peter Wegner SC2004 conference Top500 List BG/L Moors Law, problems of recent architectures Solutions Interconnects Software Lattice QCD machines DESY @SC2004 QCDOC Conclusions Technical
Turbomachinery CFD on many-core platforms experiences and strategies
Turbomachinery CFD on many-core platforms experiences and strategies Graham Pullan Whittle Laboratory, Department of Engineering, University of Cambridge MUSAF Colloquium, CERFACS, Toulouse September 27-29
Price/performance Modern Memory Hierarchy
Lecture 21: Storage Administration Take QUIZ 15 over P&H 6.1-4, 6.8-9 before 11:59pm today Project: Cache Simulator, Due April 29, 2010 NEW OFFICE HOUR TIME: Tuesday 1-2, McKinley Last Time Exam discussion
~ Greetings from WSU CAPPLab ~
~ Greetings from WSU CAPPLab ~ Multicore with SMT/GPGPU provides the ultimate performance; at WSU CAPPLab, we can help! Dr. Abu Asaduzzaman, Assistant Professor and Director Wichita State University (WSU)
Supercomputing Resources in BSC, RES and PRACE
www.bsc.es Supercomputing Resources in BSC, RES and PRACE Sergi Girona, BSC-CNS Barcelona, 23 Septiembre 2015 ICTS 2014, un paso adelante para la RES Past RES members and resources BSC-CNS (MareNostrum)
