Editorial. Editorial. Publishers. Editor. Design

Size: px
Start display at page:

Download "Editorial. Editorial. Publishers. Editor. Design"

Transcription

1

2 Editorial Editorial Publishers Editor Design 2 3

3 Contents Contents Contents News Applications Systems Projects Centres Activities Courses

4 News News PRACE: Results of the 7 th Regular Call PRACE Projects meet in Varna

5 News News Interview with Prof. Dr.-Ing. Dr. h.c. Dr. h.c. Michael M. Resch

6 News News Prof. Michael M. Resch is the Chairman of the Board of Directors of GCS since May He was a cofounder of GCS. Prof. Resch is currently the director of the High Performance Computing Center Stuttgart (HLRS), the director of the Information Service Center (IZUS) of the University of Stuttgart and the director of the Institute for High Performance Computing (IHR) of the University of Stuttgart.

7 Applications Applications Numerical Method 2 2 Simulating the Life Cycle of molecular Clouds

8 Applications Applications 2 The typical Milky Way Disk

9 Applications Applications Disk Galaxies at different Gas Surface Densities: from low to high Redshift 2 2 Conclusions and Outlook References [1] Bigiel, F., Leroy, A., Walter, F., Brinks, E., de Blok, W. J. G., Madore, B., Thornley, M. D. [2] Brown, P. N., Byrne, G. D., Hindmarsh, A. C. 2 2 [3] Chabrier, G. [4] Clark, P. C., Glover, S. C. O., Klessen, R. S. [5] Evans, II, N. J. and et al. [6] Fryxell, B., et al. [7] Glover, S. C. O., Mac Low, M.-M. 2 2 [8] Górski, K. M., Hivon, E. 2 3 [9] Mac Low, M.-M., Klessen, R. S. 2 2 [10] Marinacci, F., Fraternali, F., Nipoti, C., Binney, J., Ciotti, L., Londrillo, P. 2 [11] Moster, B. P., Somerville, R. S., Maulbetsch, C., van den Bosch, F. C., Macciò, A. V., Naab, T., Oser, L. [12] Newman, S. F., et al. 2 [13] Tacconi, L., et al. 3

10 Applications Applications Gadget3: Numerical Simulation of Structure Formation in the Universe 3

11 Applications Applications References [1] Springel, V. [2] Springel, V., Yoshida, N., White, S.D.M. 2 2

12 Applications Applications The Numerical Challenge The Kane-Mele-Hubbard Model Numerical Simulation of Correlated Electron Systems

13 Applications Applications Results Impact of U for = 0 2

14 Applications Applications Acknowledgments [11] Hohenadler, M., Meng, Z. Y., Lang, T. C., Wessel, S., Muramatsu, A., Assaad, F. F. [12] Berg, E., Metlitski, M.A., Sachdev, S. [13] Rachel, S., Le Hur, K. [14] Assaad, F.F., Bercx, M., Hohenadler, M. Impact of U for > 0 References [1] Kane, C.L., Mele, E.J. [2] König, M., Wiedmann, S., Brüne, C., Roth, A., Buhmann, H., Molenkamp, L.W., Qi, X.-L., Zhang, S.-C. [3] Meng, Z.Y., Lang, T.C., Wessel, S., Assaad, F.F., Muramatsu, A. [4] Sorella, S., Otsuka, Y., Yunoki, S. [5] Assaad, F.F., Herbut, I.F. [15] Herbut, I.F. [16] Herbut, I.F., Jurii V., Vafek, O. [17] Yang, H.-Y., Albuquerque, A.F., Capponi, S., Läuchli, A.M., Schmidt, K.P. [18] Chang, C.-C., Scalettar, R.T. [19] Clark, B.K. [20] Sorella, S., Tosatti, E. [21] Paiva, T., Scalettar, R.T., Zheng, W., Singh, R.R.P., Oitmaa, J. [22] Qi, X.-L., Zhang, S.-C. [23] Ran, Y., Vishwanath, A., Lee, D.-H. [6] Hohenadler, M., Assaad, F.F. Outlook [7] Uehlinger, T., Jotzu, G., Messer, M., Greif, D., Hofstetter, W., Bissbort, U., Esslinger, T. [8] Assaad, F.F., Evertz, H.G. [9] Assaad, F.F. [10] Hohenadler, M., Lang, T.C., Assaad, F.F.

15 Applications Applications Highly-resolved numerical Simulations of bed-load Transport in a turbulent open Channel Flow Numerical Method Computational Setup Results

16 Applications Applications R uu r x / H Fix Ref FewPart LowSh G m ( x,0) Ref FewPart LowSh / H x Conclusions References [1] Dietrich, W.E., Kirchner, J.W., Ikeda, H., Iseya, F. [2] Shvidchenko, A.B., Pender, G. [3] Kempe, T., Fröhlich, J., [4] Kempe, T., Fröhlich, J., [5] Vowinckel, B., Kempe, T., Fröhlich, J., [6] Shields, A. [7] Yalin, M.S., Ferreira da Silva, A.M., [8] Vowinckel, B., Kempe, T., Fröhlich, J., Nikora, V.I. [9] Vowinckel, B., Kempe, T., Fröhlich, J.

17 Applications Applications How to fit the Local Universe into a Supercomputer? Recovering and Simulating Structures of the Local Universe 32 33

18 Applications Applications Acknowledgements References [1] Heß, S., Kitaura, F.-S., Gottlöber, S. [2] Kitaura, F.-S. [3] Kitaura, F.-S., Heß, S.

19 Applications Applications A scalable hybrid DFT/PMM-MD Approach for accurately simulating Biomolecules on SuperMUC

20 Applications Applications References [1] Senn, H.M., Thiel, W. [2] Schwörer, M., Breitenfeld, B., Tröster, P., Bauer, S., Lorenzen, K., Tavan, P., Mathias, G. [3] Lorenzen, K., Schwörer, M., Tröster, P., Mates, S., Tavan, P. [4] [5] Mathias, G., Baer, M.D. 2

21 Applications Applications Aircraft Wake Vortex Evolution during Approach and Landing With and without Plate Lines

22 Applications Applications 0 3 3

23 Applications Applications References [1] Holzäpfel, F., Steen, M. [2] Holzäpfel, F., Gerz, T., Frech, M., Tafferner, A., Köpp, F., Smalikho, I., Rahm, S., Hahn, K.-U., Schwarz, C. [3] Misaka, T., Holzäpfel, F., Gerz, T. [4] Stephan, A., Holzäpfel, F., Misaka, T. Links

24 Projects Projects Factories of the Future Resources, Technology, Infra- structure and Services for Simulation and Modelling FORTISSIMO

25 Projects Projects References [1] Sawyer, P. [2] Schubert, J. Core Project Partners [3]

26 Projects Projects Revisiting Dynamic Scheduling Techniques for HPC Infrastructures: The Approach of the DreamCloud Project [3] [4] Project Partners [1] References [1] [2] [3] [4] [2]

27 Projects Projects SkaSim - Scalable HPC-Codes for molecular Simulation in the Chemical Industry

28 Projects Projects POLCA: Programming Large Scale Heterogeneous Infrastructures

29 Projects Projects Programming with POLCA The POLCA Approach

30 Projects Projects What POLCA will provide Who is POLCA?

31 Projects Projects A flexible Framework for Energy and Performance Analysis of highly parallel Applications in a Supercomputing Centre Background Software

32 Projects Projects Summary References [1] [2] [3] Focht, E., Jeutter, A. [4] Guillen, C., Hesse, W., Brehm, M. [5] Focus of the Project Implementation [6] Treibig, J., Hager, G., Wellein, G. [7] [8] [9] 2

33 Projects Projects SIMOPEK Simulation and Optimization of Data Center Energy Flows from Cooling Networks taking into Account HPC Operation Scenarios External Influences/Constraints Data Center: Reduce Total Cost of Ownership Pillar 1 Building Infrastructure Pillar 2 HPC System Hardware Pillar 3 HPC System Software Pillar 4 HPC Applications Advanced Heat Reuse Technologies System Scheduler Neighboring Buildings Ulity Providers Infrastructure Aware Resource Management & Scheduling Modeling, Simulaon & Opmizaon Data Center Data Acquision Monitor Building Management & Infrastructure Hardware Management System Management Soware Performance Analysis Tools Infrastructure Monitoring System Hardware Monitoring System Soware Monitoring Performance Monitoring

34 Projects Projects External Influences/Constraints Neighboring Buildings Ulity Providers Data Center: Reduce Total Cost of Ownership Pillar 1 Building Infrastructure SIMOPEK Advanced Absorpon Cooling Pillar 2 HPC System Hardware SIMOPEK Power Consumpon Modeling, Simulaon & Opmizaon using MYNTS Pillar 3 HPC System Software Infrastructure Aware Resource Management & Scheduling SIMOPEK Data Collecon using PowerDam V.2.0 System Scheduler Pillar 4 HPC Applications FEPA Performance, Energy Modeling & Opmizaon FEPA Data Collecon Facts and Figures References [1] Wilde, T., Auweter, A., Shoukourian, H. [2] Shoukourian, H., Wilde, T., Auweter, A. [3] [4] Building Management & Infrastructure Hardware Management System Management Soware Performance Analysis Tools 2 2 Infrastructure Monitoring System Hardware Monitoring System Soware Monitoring Performance Monitoring

35 Projects Projects The Catwalk Project A quick Development Path for Performance Models

36 Projects Projects References [1] Calotoiu, A., Hoefler, T., Poke, M., Wolf, F. [2] Geimer, M., Wolf, F., Wylie, B.J.N., Ábrahám, E., Becker, D., Mohr, B. [3] an Mey, D., Biersdorff, S., Bischof, C., Diethelm, K., Eschweiler, D., Gerndt, M., Knüpfer, A., Lorenz, D., Malony, A.D., Nagel, W.E., Oleynik, Y., Rössel, C., Saviankou, P., Schmidl, D., Shende, S.S., Wagner, M., Wesarg, B., Wolf, F.

37 Projects Projects GROMEX Unified Longrange Electrostatics and Flexible Ionization Background of the Project Usability & Scalability Towards realistic Simulations

38 Projects Projects Project Partners References (1) Kabadshow, I., Dachsel, H. (2) Dachsel, H. (3) Donnini, S., Tegeler, F., Groenhof, G., Grubmüller, H. (4) Hess, B., Kutzner, C., van der Spoel, D., Lindahl, E. (5) Ullmann, R.T., Ullmann, G.M. 2

39 Projects Projects HOPSA A big Jump forward in HPC System and Application Monitoring Integration among the HOPSA Performance Analysis Tools The HOPSA Performance Tool Workflow

40 Projects Projects Conclusion Integration of System Data and Performance Analysis Tools

41 Projects Projects EU Project Partners (HOPSA-EU) Russian Project Partners (HOPSA-RU) References [1] Labarta, J., Girona, S., Pillet, V., Cortes, T., Gregoris, L. [5] an Mey, D., Biersdorff, S., Bischof, C., Diethelm, K., Eschweiler, D., Gerndt, M., Knüpfer, A., Lorenz, D., Malony, A.D., Nagel, W.E., Oleynik, Y., Rössel, C., Saviankou, P., Schmidl, D., Shende, S.S., Wagner, M., Wesarg, B., Wolf, F. [6] Servat, H., Llort, G., Giménez, J., Labarta, J. [7] [8] Adinets, A.V., Bryzgalov, P.A., Vad, V., Voevodin, V., Zhumatiy, S.A., Nikitenko, D.A. [2] Geimer, M., Wolf, F., Wylie, B.J.N., Abraham, E., Becker, D., Mohr, B. [3] Berg, E., Hagersten, E. [9] Mohr, B., Voevodin, V., Giménez, J., Hagersten, E., Knüpfer, A., Nikitenko, D.A., Nilsson, M., Servat, H., Shah, A., Winkler, F., Wolf, F., Zhujov, I. [10] [4] Nagel, W., Weber, M., Hoppe, H.-C., Solchenbach, K.

42 Systems Systems End of the HPC-FF Era

43 Systems Systems 2 References [1] [2] [3] [4]

44 Systems Systems JUROPA-3 - A Prototype for the Next-Generation HPC Cluster System Specifications

45 Systems Systems First Experiences with the Intel MIC Architecture at LRZ Intel MIC Architecture at LRZ Architectural Overview Programming Models Number of cores Frequency of cores GDDR5 memory size Number of hardware threads SIMD vector registere Flops/cycle Theoretical peak performance L2 cache per core

46 Systems Systems Benchmarks Acknowledgements References (1) Weinberg, V., (Editor) et al. (2) (3)

47 Systems The Extension of SuperMUC: Phase 2 SuperMUC Phase 1 Phase 2 Innovative Water Cooling Users from 25 European Countries Systems Financing References (1) (2) (3)

48 Centres Centres Leibniz Supercomputing Centre of the Bavarian Academy of Sciences and Humanities (Leibniz-Rechenzentrum, LRZ) provides comprehensive services to scientific and academic communities by: Research in HPC is carried out in collaboration with the distributed, statewide Competence Network for Technical and Scientific High Performance Computing in Bavaria (KONWIHR). Compute servers currently operated by LRZ are given in the following table System Size Peak Performance (TFlop/s) Purpose User Community Contact: Leibniz Supercomputing Centre Prof. Dr. Arndt Bode Boltzmannstr Garching near Munich Germany Phone

49 Centres Centres First German National Center Based on a long tradition in supercomputing at University of Stuttgart, HLRS (Höchstleistungsrechenzentrum Stuttgart) was founded in 1995 as the first German federal Centre for High Performance Computing. HLRS serves researchers at universities and research laboratories in Europe and Germany and their external and industrial partners with high-end computing power for engineering and scientific applications. Service for Industry Service provisioning for industry is done to gether with T-Systems, T-Systems sfr, and Porsche in the public-private joint venture hww (Höchstleistungsrechner für Wissenschaft und Wirtschaft). Through this co-operation industry always has acces to the most recent HPC technology. Bundling Competencies In order to bundle service resources in the state of Baden-Württemberg HLRS has teamed up with the Steinbuch Center for Computing of the Karlsruhe Institute of Technology. This collaboration has been implemented in the non-profit organization SICOS BW GmbH. World Class Research As one of the largest research centers for HPC HLRS takes a leading role in research. Participation in the German national initiative of excellence makes HLRS an outstanding place in the field. Contact: Höchstleistungs rechenzentrum Stuttgart (HLRS) Universität Stuttgart Prof. Dr.-Ing. Dr. h.c. Dr. h.c. Michael M. Resch Nobelstraße Stuttgart Germany Phone / Compute servers currently operated by HLRS System Cray XE6 "Hermit" (Q4 2011) NEC Cluster (Laki, Laki2) heterogenous compunting platform of 2 independent clusters Size 3,552 dual socket nodes with 113,664 AMD Interlagos cores 23 TB memory 9988 cores 911 nodes Peak Performance (TFlop/s) Purpose 1,045 Capability 170 Laki: Computing 120,5 TFlops Laki2: 47,2 TFlops User Community European and German Research Organizations and Industry German Universities, Research Institutes and Industry Hermit

50 Centres Centres Supercomputer-oriented research and development in selected fields of physics and other natural sciences by research groups of competence in supercomputing applications. Compute servers currently operated by JSC System Size Peak Performance (TFlop/s) Purpose User Community The Jülich Supercomputing Centre (JSC) at Forschungszentrum Jülich enables scientists and engineers to solve grand Implementation of strategic support infrastructures including communityoriented simulation laboratories and cross-sectional groups on mathematical methods and algorithms and parallel challenge problems of high complexity in science and engineering in collaborative infrastructures by means of supercom- performance tools, enabling the effective usage of the supercomputer resources. puting and Grid technologies. Provision of supercomputer resources Higher education for master and doctoral students in cooperation e.g. of the highest performance class for projects in science, research and industry in the fields of modeling and computer simulation including their methods. The selection of the projects is performed by an international peer-review procedure with the German Research School for Simulation Sciences. Contact: Jülich Supercomputing Centre (JSC) Forschungszentrum Jülich implemented by the John von Neumann Institute for Computing (NIC), a joint foundation of Forschungszentrum Jülich, Prof. Dr. Dr. Thomas Lippert Jülich Deutsches Elektronen-Synchrotron Germany DESY, and GSI Helmholtzzentrum für Phone Schwerionenforschung.

51 Activities Activities CECAM Tutorials at JSC CHANGES Workshop 2

52 Activities Activities Laboratory Experiments on Crowd Dynamics

53 Activities Activities JSC Guest Student Programme on Scientific Computing 2013

54 Activities Activities High-Q Club The highest scaling Codes on JUQUEEN Terra-Neo Gysela walberla PEPC PMG+PFASST Reference [1] Brömmel, D. dynqcd

55 Activities Activities Jülich Supercomputing Centre contributes to visionary Human Brain Project References [1] [2] [3]

56 Activities Activities Traffic and Granular Flow Conference celebrates 10 th Edition by returning to Jülich UNICORE Summit 2013 Reference [1] References [1] [2]

57 Activities Activities 3D Show at the Pharma Forum: Simulation and Visualization of the Airflow in Cleanrooms Links

58 Activities Activities The 17 th HLRS-NEC Workshop on Sustained Simulation Performance

59 Activities Activities ls1 mardyn - a Massively Parallel Molecular Simulation Code Scalability Neighbour Search Dynamic Load Balancing Partners

60 Activities Activities GCS at ISC 13 Review GCS Booth Highlights 2 Two GCS HPC systems amongst Top Ten of TOP500 ISC 13 Gauss Award Winner

61 Activities Activities Extreme Scaling Workshop at LRZ July 9-11, 2013: Running Real World Applications on more than 130,000 Cores on Super- MUC

62 Activities Activities LRZ Extreme Scale Benchmark and Optimization Suite 3 IBM MPI icc 12.1 Intel MPI icc ,000 atoms aquaporin, PME, 2 fs 2 M atoms ribosome, PME, 4 fs 12 M atoms peptides, PME, 2 fs Performance Results

63 Activities Activities HLRS Scientific Tutorials and Workshop Report and Outlook OpenACC Pro- gramming for Parallel Accelerated Supercomputers an alternative to CUDA from Cray perspective Cray XE6/ XC30 Optimization Workshops PRACE Advanced Training Centre Parallel Programming Workshop Iterative Solvers and Parallelization 2014 Workshop Announcements Scientific Conferences and Workshops at HLRS 12th HLRS/hww Workshop on Scalable Global Parallel File Systems (March/April 2014) 8th ZIH+HLRS Parallel Tools Workshop (date and location not yet fixed) High Performance Computing in Science and Engineering - The 17th Results and Review Workshop of the HPC Center Stuttgart (October 2014) IDC International HPC User Forum (October 2014) Parallel Programming Workshops: Training in Parallel Programming and CFD ISC and SC Tutorials Georg Hager, Gabriele Jost, Rolf Rabenseifner: Hybrid Parallel Programming with MPI & OpenMP. Tutorial 9 at the International Supercomputing Conference, ISC 13, Leipzig, June Georg Hager, Jan Treibig, Gerhard Wellein: Node-Level Performance Engineering. Tutorial 2 at the International Supercomputing Conference, ISC 13, Leipzig, June Rolf Rabenseifner, Georg Hager, Gabriele Jost: Hybrid MPI and OpenMP Parallel Programming. Half-day Tutorial at Super Computing 2013, SC13, Denver, Colorado, USA, November 17-22, Introduction to Computational Fluid Dynamics MPI & OpenMP Fortran for Scientific Computing Parallel Programming and Parallel Tools (TU Dresden, ZIH, February 24-27) Introduction to Computational Fluid Dynamics (HLRS, March 31 - April 4) Iterative Linear Solvers and Parallelization (HLRS, March 24-28) Cray XE6/XC30 Optimization Workshops (HLRS, March 17-20) (PATC) GPU Programming using CUDA (HLRS, April 7-9) Open ACC Programming for Parallel Accelerated Supercomputers (HLRS, April 10-11) (PATC) Unified Parallel C (UPC) and Co-Array Fortran (CAF) (HLRS, April 14-15) (PATC) Scientific Visualisation (HLRS, April 16-17) Parallel Programming with MPI & OpenMP (TU Hamburg-Harburg, July 28-30) Iterative Linear Solvers and Parallelization (LRZ, Garching, September 15-19) Introduction to Computational Fluid Dynamics (ZIMT Siegen, September/October) Unified Parallel C (UPC) and Co- Array Fortran (CAF) Message Passing Interface (MPI) for Beginners (HLRS, October 6-7) (PATC) Shared Memory Parallelization with OpenMP (HLRS, October 8) (PATC) Advanced Topics in Parallel Programming (HLRS, October 9-10) (PATC) Parallel Programming with MPI & OpenMP (FZ Jülich, JSC, December 1-3) Training in Programming Languages at HLRS Fortran for Scientific Computing (Dec 2-6, 2013 and Mar 10-14, 2014) (PATC) URLs: (PATC): This is a PRACE PATC course

64 GCS High Performance Computing Courses and Tutorials Parallel Programming with model, compilers, tools, monitoring, computational science, and also sessions will allow users to immediately Contents Parallel Programming with MPI, OpenMP and PETSc MPI, OpenMP, performance optimi- show how an awareness of the per- test and understand the language This course is targeted at scientists MPI, OpenMP, and Tools zation, mathematical software, and formance features of an application constructs. with little or no knowledge of the Date and Location application software. may lead to notable reductions in Fortran programming language, but Date and Location November 25-27, 2013 power consumption: Web Page needing it for participation in projects February 24-27, 2014 JSC, Forschungszentrum Jülich Web Page using a Fortran code base, for devel- Dresden, ZIH opment of their own codes, and for Contents events/sc-nov Second JUQUEEN Porting getting acquainted with additional Contents The focus is on programming models hierarchy and Tuning Workshop tools like debugger and syntax checker The focus is on programming models MPI, OpenMP, and PETSc. Hands-on Node-Level Performance (PATC course) as well as handling of compilers and MPI, OpenMP, and PETSc. Hands-on sessions (in C and Fortran) will allow Engineering heads libraries. The language is for the most sessions (in C and Fortran) will allow users to immediately test and (PATC course) Date and Location part treated at the level of the Fortran users to immediately test and understand the basic constructs of - The 3D Jacobi solver February 03-05, standard; features from Fortran understand the basic constructs of the Message Passing Interface (MPI) Date and Location - The Lattice-Boltzmann Method JSC, Forschungszentrum Jülich 2003 are limited to improvements on the Message Passing Interface (MPI) and the shared memory directives of December 03-04, Sparse Matrix-Vector Multiplication the elementary level. Advanced and the shared memory directives of OpenMP. Course language is English. LRZ Building, - Backprojection algorithm for CT Contents Fortran features like object-oriented OpenMP. The last day is dedicated to This course is organized by JSC in University Campus Garching, reconstruction The Blue Gene/Q petaflop super- programming or coarrays will be cov- tools for debugging and performance collaboration with HLRS. Presented near Munich, Boltzmannstr. 1 computer JUQUEEN marks another ered in a follow-on course in autumn. analysis of parallel applications. This by Dr. Rolf Rabenseifner, HLRS. Between each module, there is time quantum leap in supercomputer To consolidate the lecture material, course is organized by ZIH in collabo- Contents for Questions and Answers! performance at JSC. In order to use each day's approximately 4 hours of ration with HLRS. Web Page This course teaches performance this tool efficiently, special efforts by lecture are complemented by 3 hours engineering approaches on the Web Page the users are necessary, though. The of hands-on sessions. Web Page events/mpi compute node level. Performance aim of this hands-on workshop is to engineering as we define it is more compute/courses support current users of JUQUEEN Prerequisites Introduction to the than employing tools to identify in porting their software, in analyzing Course participants should have ba- Parallel Programming of Programming and Usage hotspots and bottlenecks. It is about Fortran for Scientific its performance, and in improving sic UNIX/Linux knowledge (login with High Performance Systems of the Supercomputer developing a thorough understanding Computing its efficiency. This course is a PATC secure shell, shell commands, basic Resources at Jülich of the interactions between software (PATC course) course (PRACE Advanced Training programming, vi or emacs editors). Dates and Location and hardware. This process must Centres). March 10-14, 2014 Date and Location start at the core, socket, and node Dates and Location Web Page RRZE building, University campus November 28-29, 2013 level, where the code gets executed December 02-06, 2013 and Web Page Erlangen, Martensstr. 1: Via video JSC, Forschungszentrum Jülich that does the actual computational March 10-14, pute/courses conference at LRZ if there is sufficient work. Once the architectural require- Stuttgart, HLRS events/juqueenpt14 interest. Contents ments of a code are understood and This course gives an overview of correlated with performance mea- Contents Programming with Fortran Contents the supercomputers JUROPA and surements, the potential benefit of This course is dedicated for scientists This course, a collaboration of JUQUEEN. Especially new users will optimizations can often be predicted. and students to learn (sequential) Dates and Locations Erlangen Regional Computing Centre learn how to program and use these We introduce a holistic node-level programming scientific applications February 03-07, 2014 (RRZE) and LRZ, is targeted at stu- systems efficiently. Topics discussed performance engineering strategy, with Fortran. The course teaches LRZ Building, University campus dents and scientists with interest in are: system architecture, usage apply it to different algorithms from newest Fortran standards. Hands-on Garching near Munich. programming modern HPC hardware,

65 GCS High Performance Computing Courses and Tutorials specifically the large scale parallel on these opportunities. From Monday and the shared memory directives of Introduction to Computa- Contents GPU Programming using computing systems available in Munich, to Wednesday, specialists from Cray OpenMP. This course is organized by tional Fluids Dynamics In this add-on course to the parallel CUDA Jülich and Stuttgart. will support you in your effort porting University of Kassel, HLRS, and IAG. programming course special topics and optimizing your application on our Date and Location are treated in more depth, in Date and Location Each day is comprised of approxi- Cray XE6. On Thursday, Georg Hager Web Page March 31 - April 04, 2014 particular performance analysis, I/O April 07-09, 2014 mately 4 hours of lectures and 3 and Jan Treibig from RRZE will present Stuttgart, HLRS and PGAS concepts. It is provided Stuttgart, HLRS hours of hands-on sessions. detailed information on optimizing in collaboration of Erlangen Regional codes on the multicore AMD Interlagos Eclipse: C/C++/Fortran Contents Computing Centre (RRZE) and LRZ Contents Web Page processor. Course language is English programming Numerical methods to solve the within KONWIHR. The course provides an introduction (if required). equations of Fluid Dynamics are pre- Each day is comprised of approxima- to the programming language CUDA, compute/courses Date and Location sented. The main focus is on explicit tely 5 hours of lectures and 2 hours which is used to write fast numeric Web Page March 25, 2014 Finite Volume schemes for the com- of hands-on sessions. algorithms for NVIDIA graphics proces- Cray XE6/XC 30 LRZ Building, University campus pressible Euler equations. Hands-on Day 1 sors (GPUs). Focus is on the basic Optimization Workshop Garching near Munich. sessions will manifest the content of Intel tools: MPI tracing and Checking usage of the language, the exploitation (PATC course) Iterative Linear Solvers and the lectures. Participants will learn Intel tools: OpenMP performance and of the most important features of the Parallelization Contents to implement the algorithms, but correctness. device (massive parallel computation, Date and Location This course is targeted at scientists also to apply existing software and to Day 2 shared memory, texture memory) and March 17-20, 2014 Dates and Location who wish to be introduced to pro- interpret the solutions correctly. Parallel I/O with MPI IO efficient usage of the hardware to Stuttgart, HLRS March 24-28, 2014 gramming C/C++/Fortran with the Methods and problems of paralleliza- Performance analysis with Scalasca. maximize performance. An overview Stuttgart, HLRS Eclipse C/C++ Development Tools tion are discussed. This course is Day 3 of the available development tools and Contents (CDT), or the Photran Plugin. Topics based on a lecture and practical Tuning I/O on LRZ's HPC systems. the advanced features of the language HLRS installed Hermit, a Cray XE6 September 15-19, 2014 covered include: awarded with the "Landeslehrpreis Portability of I/O: Binary files NetCDF is given. system with AMD Interlagos proces- Garching, LRZ Baden-Württemberg 2003" and or- HDF5. sors and a performance of 1 PFlop/s. ganized by HLRS, IAG, and University Day 4 Web Page We strongly encourage you to port Contents of Kassel. PGAS programming with coarray your applications to the new architec- The focus is on iterative and parallel Fortran and Unified Parallel C. ture as early as possible. To support solvers, the parallel programming Photran. Web Page PGAS hands on session. GPU Programming such effort we invite current and models MPI and OpenMP, and the (PATC course) future users to participate in special parallel middleware PETSc. Thereby, Prerequisites Prerequisites Cray XE6/XC30 Optimization Work- different modern Krylov Subspace Course participants should have basic Advanced Topics in High Good MPI and OpenMP knowledge Date and Location shops. With this course, we will give Methods (CG, GMRES, BiCGSTAB...) knowledge of the C and/or C++/ Performance Computing as presented in the course "Parallel April 07-09, 2014 all necessary information to move as well as highly efficient precondi- Fortran programming languages. (PATC course) programming of High Performance JSC, Forschungszentrum Jülich applications from the current NEC tioning techniques are presented Systems" (see above). SX-9, the Nehalem cluster, or other in the context of real life applica- Web Page Date and Location Contents systems to Hermit. Hermit provides tions. Hands-on sessions (in C and March 31 - April 03, 2014 Web Page Many-core programming is a very our users with a new level of perfor- Fortran) will allow users to immedi- compute/courses LRZ Building, University campus dynamic research area. Many scien- mance. To harvest this potential will ately test and understand the basic Garching near Munich. compute/courses tific applications have been ported to require all our efforts. We are look- constructs of iterative solvers, the GPU architectures during the past ing forward to working with our users Message Passing Interface (MPI) four years. We will give an introduction

66 GCS High Performance Computing Courses and Tutorials to CUDA, OpenCL, and multi-gpu will be presented, and interoperability Scientific Visualization computing since around 5 years. Advanced GPU Programming are: system architecture, usage programming using examples of in- of OpenACC directives with these Particularly GPGPUs have recently model, compilers, tools, monitoring, creasing complexity. After introducing and with CUDA will be demonstrated. Date and Location become very popular, however pro- Date and Location MPI, OpenMP, performance optimi- the basics the focus will be on Through application case studies and April 16-17, 2014 gramming GPGPUs using program- May 05-06, 2014 zation, mathematical software, and optimization and tuning of scientific tutorials, users will gain direct expe- Stuttgart, HLRS ming languages like CUDA or OpenCL JSC, Forschungszentrum Jülich application software. applications. rience of using OpenACC directives in is cumbersome and error-prone. This course is a PATC course (PRACE realistic applications. Contents Beyond introducing the basics of Contents Web Page Advanced Training Centres). Users may also bring their own codes This two day course is targeted at GPGPU-porogramming, we mainly Today's computers are commonly to discuss with Cray specialists or researchers with basic knowledge present OpenACC as an easier way equipped with multicore processors events/sc-may Web Page begin porting. in numerical simulation, who would to program GPUs using OpenMP-like and graphics processing units. To like to learn how to visualize their pragmas. Recently Intel developed make efficient use of these massively Parallel I/O and Portable events/gpu Web Page simulation results on the desktop but their own Many Integrated Core parallel compute resources advanced Data Formats also in Augmented Reality and Virtual (MIC) architecture which can be knowledge of architecture and pro- (PATC course) Open ACC Programming Environments. It will start with a programmed using standard paral- gramming models is indispensable. Cray XK Unified Parallel C (UPC) and short overview of scientific visualiza- lel programming techniques like This course focuses on finding and Date and Location (PATC course) Co-Array Fortran (CAF) tion, following a hands-on introduc- OpenMP and MPI. In the beginning eliminating bottlenecks using profiling May 21-23, 2014 (PATC course) tion to 3D desktop visualization with of 2013, the first production-level and advanced programming tech- JSC, Forschungszentrum Jülich Date and Location COVISE. On the second day, we will cards named Intel Xeon Phi came on niques, optimal usage of CPUs and April 10-11, 2014 Date and Location discuss how to build interactive 3D the market. The course discusses GPUs on a single node, and multi-gpu Contents Stuttgart, HLRS April 14-15, 2014 Models for Virtual Environments and various programming techniques for programming across multiple nodes. This course will introduce MPI parallel Stuttgart, HLRS how to set up an Augmented Reality Intel Xeon Phi and includes hands-on I/O and portable, self-describing data Contents visualization. session for both MIC and GPU pro- Web Page formats, such as HDF5 and NetCDF. This workshop will cover the pro- Contents gramming. The course is developed Participants should have experience gramming environment of the Cray Partitioned Global Address Space Web Page in collaboration with the Erlangen events/advgpu in parallel programming in general, XK7 hybrid supercomputer, which (PGAS) is a new model for parallel Regional Computing Centre (RRZE) and either C/C++ or Fortran in par- combines multicore CPUs with GPU programming. Unified Parallel C (UPC) within KONWIHR. Introduction to the ticular. This course is a PATC course accelerators. Attendees will learn and Co-Array Fortran (CAF) are PGAS Intel MIC&GPU Each day is comprised of approxi- Programming and Usage (PRACE Advanced Training Centres). about the directive-based OpenACC extensions to C and Fortran. PGAS Programming Workshop mately 5 hours of lectures and 2 of the Supercomputer programming model whose multi- languages allow any processor to (PATC course) hours of hands-on sessions. Resources at Jülich Web Page vendor support allows users to por- directly address memory/data on any tably develop applications for parallel other processors. Date and Location Prerequisites Date and Location events/parallelio accelerated supercomputers. Parallelism can be expressed more April 28-30, 2014 Good working knowledge of at least May 19-20, 2014 The workshop will also demonstrate easily compared to library based LRZ Building, University campus one of the standard HPC languages: JSC, Forschungszentrum Jülich how to use the Cray Programming approches as MPI. Hands-on sessions Garching, near Munich. Fortran 95, C or C++. Basic OpenMP Environment tools to identify CPU (in UPC and/or CAF) will allow users and MPI knowledge useful. Contents application bottlenecks, facilitate the to immediately test and understand Contents This course gives an overview of the OpenACC porting, provide accelerated the basic constructs of PGAS languages. With the rapidly growing demand for Web Page supercomputers JUROPA and performance feedback and to tune computing power new accelerator JUQUEEN. Especially new users will the ported applications. The Cray Web Page based architectures have entered compute/courses learn how to program and use these scientific libraries for accelerators the world of high performance systems efficiently. Topics discussed

Welcome to the. Jülich Supercomputing Centre. D. Rohe and N. Attig Jülich Supercomputing Centre (JSC), Forschungszentrum Jülich

Welcome to the. Jülich Supercomputing Centre. D. Rohe and N. Attig Jülich Supercomputing Centre (JSC), Forschungszentrum Jülich Mitglied der Helmholtz-Gemeinschaft Welcome to the Jülich Supercomputing Centre D. Rohe and N. Attig Jülich Supercomputing Centre (JSC), Forschungszentrum Jülich Schedule: Monday, May 19 13:00-13:30 Welcome

More information

Kriterien für ein PetaFlop System

Kriterien für ein PetaFlop System Kriterien für ein PetaFlop System Rainer Keller, HLRS :: :: :: Context: Organizational HLRS is one of the three national supercomputing centers in Germany. The national supercomputing centers are working

More information

HIGH PERFORMANCE CONSULTING COURSE OFFERINGS

HIGH PERFORMANCE CONSULTING COURSE OFFERINGS Performance 1(6) HIGH PERFORMANCE CONSULTING COURSE OFFERINGS LEARN TO TAKE ADVANTAGE OF POWERFUL GPU BASED ACCELERATOR TECHNOLOGY TODAY 2006 2013 Nvidia GPUs Intel CPUs CONTENTS Acronyms and Terminology...

More information

HIGH PERFORMANCE COMPUTING COMPETENCE CENTER BADEN-WÜRTTEMBERG

HIGH PERFORMANCE COMPUTING COMPETENCE CENTER BADEN-WÜRTTEMBERG HIGH PERFORMANCE COMPUTING COMPETENCE CENTER BADEN-WÜRTTEMBERG Contents High Performance Computing Competence Center Baden-Württemberg (hkz-bw)... 4 Vector Parallel Supercomputer NEC SX-6X... 8 Massively

More information

Accelerating Simulation & Analysis with Hybrid GPU Parallelization and Cloud Computing

Accelerating Simulation & Analysis with Hybrid GPU Parallelization and Cloud Computing Accelerating Simulation & Analysis with Hybrid GPU Parallelization and Cloud Computing Innovation Intelligence Devin Jensen August 2012 Altair Knows HPC Altair is the only company that: makes HPC tools

More information

Access, Documentation and Service Desk. Anupam Karmakar / Application Support Group / Astro Lab

Access, Documentation and Service Desk. Anupam Karmakar / Application Support Group / Astro Lab Access, Documentation and Service Desk Anupam Karmakar / Application Support Group / Astro Lab Time to get answer to these questions Who is allowed to use LRZ hardware? My file system is full. How can

More information

Unified Performance Data Collection with Score-P

Unified Performance Data Collection with Score-P Unified Performance Data Collection with Score-P Bert Wesarg 1) With contributions from Andreas Knüpfer 1), Christian Rössel 2), and Felix Wolf 3) 1) ZIH TU Dresden, 2) FZ Jülich, 3) GRS-SIM Aachen Fragmentation

More information

GASPI A PGAS API for Scalable and Fault Tolerant Computing

GASPI A PGAS API for Scalable and Fault Tolerant Computing GASPI A PGAS API for Scalable and Fault Tolerant Computing Specification of a general purpose API for one-sided and asynchronous communication and provision of libraries, tools, examples and best practices

More information

GPU System Architecture. Alan Gray EPCC The University of Edinburgh

GPU System Architecture. Alan Gray EPCC The University of Edinburgh GPU System Architecture EPCC The University of Edinburgh Outline Why do we want/need accelerators such as GPUs? GPU-CPU comparison Architectural reasons for GPU performance advantages GPU accelerated systems

More information

HPC Wales Skills Academy Course Catalogue 2015

HPC Wales Skills Academy Course Catalogue 2015 HPC Wales Skills Academy Course Catalogue 2015 Overview The HPC Wales Skills Academy provides a variety of courses and workshops aimed at building skills in High Performance Computing (HPC). Our courses

More information

Evaluation of CUDA Fortran for the CFD code Strukti

Evaluation of CUDA Fortran for the CFD code Strukti Evaluation of CUDA Fortran for the CFD code Strukti Practical term report from Stephan Soller High performance computing center Stuttgart 1 Stuttgart Media University 2 High performance computing center

More information

Part I Courses Syllabus

Part I Courses Syllabus Part I Courses Syllabus This document provides detailed information about the basic courses of the MHPC first part activities. The list of courses is the following 1.1 Scientific Programming Environment

More information

Unleashing the Performance Potential of GPUs for Atmospheric Dynamic Solvers

Unleashing the Performance Potential of GPUs for Atmospheric Dynamic Solvers Unleashing the Performance Potential of GPUs for Atmospheric Dynamic Solvers Haohuan Fu haohuan@tsinghua.edu.cn High Performance Geo-Computing (HPGC) Group Center for Earth System Science Tsinghua University

More information

GPGPU accelerated Computational Fluid Dynamics

GPGPU accelerated Computational Fluid Dynamics t e c h n i s c h e u n i v e r s i t ä t b r a u n s c h w e i g Carl-Friedrich Gauß Faculty GPGPU accelerated Computational Fluid Dynamics 5th GACM Colloquium on Computational Mechanics Hamburg Institute

More information

Programming models for heterogeneous computing. Manuel Ujaldón Nvidia CUDA Fellow and A/Prof. Computer Architecture Department University of Malaga

Programming models for heterogeneous computing. Manuel Ujaldón Nvidia CUDA Fellow and A/Prof. Computer Architecture Department University of Malaga Programming models for heterogeneous computing Manuel Ujaldón Nvidia CUDA Fellow and A/Prof. Computer Architecture Department University of Malaga Talk outline [30 slides] 1. Introduction [5 slides] 2.

More information

160 Numerical Methods and Programming, 2012, Vol. 13 (http://num-meth.srcc.msu.ru) UDC 004.021

160 Numerical Methods and Programming, 2012, Vol. 13 (http://num-meth.srcc.msu.ru) UDC 004.021 160 Numerical Methods and Programming, 2012, Vol. 13 (http://num-meth.srcc.msu.ru) UDC 004.021 JOB DIGEST: AN APPROACH TO DYNAMIC ANALYSIS OF JOB CHARACTERISTICS ON SUPERCOMPUTERS A.V. Adinets 1, P. A.

More information

Design and Optimization of a Portable Lattice Boltzmann Code for Heterogeneous Architectures

Design and Optimization of a Portable Lattice Boltzmann Code for Heterogeneous Architectures Design and Optimization of a Portable Lattice Boltzmann Code for Heterogeneous Architectures E Calore, S F Schifano, R Tripiccione Enrico Calore INFN Ferrara, Italy Perspectives of GPU Computing in Physics

More information

Big Data Visualization on the MIC

Big Data Visualization on the MIC Big Data Visualization on the MIC Tim Dykes School of Creative Technologies University of Portsmouth timothy.dykes@port.ac.uk Many-Core Seminar Series 26/02/14 Splotch Team Tim Dykes, University of Portsmouth

More information

Parallel Programming Survey

Parallel Programming Survey Christian Terboven 02.09.2014 / Aachen, Germany Stand: 26.08.2014 Version 2.3 IT Center der RWTH Aachen University Agenda Overview: Processor Microarchitecture Shared-Memory

More information

and RISC Optimization Techniques for the Hitachi SR8000 Architecture

and RISC Optimization Techniques for the Hitachi SR8000 Architecture 1 KONWIHR Project: Centre of Excellence for High Performance Computing Pseudo-Vectorization and RISC Optimization Techniques for the Hitachi SR8000 Architecture F. Deserno, G. Hager, F. Brechtefeld, G.

More information

Access to the Federal High-Performance Computing-Centers

Access to the Federal High-Performance Computing-Centers Access to the Federal High-Performance Computing-Centers rabenseifner@hlrs.de University of Stuttgart High-Performance Computing-Center Stuttgart (HLRS) www.hlrs.de Slide 1 TOP 500 Nov. List German Sites,

More information

Overview on Modern Accelerators and Programming Paradigms Ivan Giro7o igiro7o@ictp.it

Overview on Modern Accelerators and Programming Paradigms Ivan Giro7o igiro7o@ictp.it Overview on Modern Accelerators and Programming Paradigms Ivan Giro7o igiro7o@ictp.it Informa(on & Communica(on Technology Sec(on (ICTS) Interna(onal Centre for Theore(cal Physics (ICTP) Mul(ple Socket

More information

Extreme Scaling on Energy Efficient SuperMUC

Extreme Scaling on Energy Efficient SuperMUC Extreme Scaling on Energy Efficient SuperMUC Dieter Kranzlmüller Munich Network Management Team Ludwig- Maximilians- Universität München (LMU) & Leibniz SupercompuFng Centre (LRZ) of the Bavarian Academy

More information

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0)

TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx

More information

Challenges on Extreme Scale Computers - Complexity, Energy, Reliability

Challenges on Extreme Scale Computers - Complexity, Energy, Reliability Challenges on Extreme Scale Computers - Complexity, Energy, Reliability Dieter Kranzlmüller Munich Network Management Team Ludwig-Maximilians-Universität München (LMU) & Leibniz SupercompuFng Centre (LRZ)

More information

Jens Doleschal (TUD) Jens Doleschal (TUD) Jens Doleschal (TUD) 1.0 26/09/2012 Final version of the deliverable Jens Doleschal (TUD)

Jens Doleschal (TUD) Jens Doleschal (TUD) Jens Doleschal (TUD) 1.0 26/09/2012 Final version of the deliverable Jens Doleschal (TUD) Version Date Comments, Changes, Status Authors, contributors, reviewers 0.1 24/08/2012 First full version of the deliverable Jens Doleschal (TUD) 0.1 03/09/2012 Review Ben Hall (UCL) 0.1 13/09/2012 Review

More information

JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert

JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert Mitglied der Helmholtz-Gemeinschaft JUROPA Linux Cluster An Overview 19 May 2014 Ulrich Detert JuRoPA JuRoPA Jülich Research on Petaflop Architectures Bull, Sun, ParTec, Intel, Mellanox, Novell, FZJ JUROPA

More information

Introducing PgOpenCL A New PostgreSQL Procedural Language Unlocking the Power of the GPU! By Tim Child

Introducing PgOpenCL A New PostgreSQL Procedural Language Unlocking the Power of the GPU! By Tim Child Introducing A New PostgreSQL Procedural Language Unlocking the Power of the GPU! By Tim Child Bio Tim Child 35 years experience of software development Formerly VP Oracle Corporation VP BEA Systems Inc.

More information

Mixed Precision Iterative Refinement Methods Energy Efficiency on Hybrid Hardware Platforms

Mixed Precision Iterative Refinement Methods Energy Efficiency on Hybrid Hardware Platforms Mixed Precision Iterative Refinement Methods Energy Efficiency on Hybrid Hardware Platforms Björn Rocker Hamburg, June 17th 2010 Engineering Mathematics and Computing Lab (EMCL) KIT University of the State

More information

A Data Structure Oriented Monitoring Environment for Fortran OpenMP Programs

A Data Structure Oriented Monitoring Environment for Fortran OpenMP Programs A Data Structure Oriented Monitoring Environment for Fortran OpenMP Programs Edmond Kereku, Tianchao Li, Michael Gerndt, and Josef Weidendorfer Institut für Informatik, Technische Universität München,

More information

The Top Six Advantages of CUDA-Ready Clusters. Ian Lumb Bright Evangelist

The Top Six Advantages of CUDA-Ready Clusters. Ian Lumb Bright Evangelist The Top Six Advantages of CUDA-Ready Clusters Ian Lumb Bright Evangelist GTC Express Webinar January 21, 2015 We scientists are time-constrained, said Dr. Yamanaka. Our priority is our research, not managing

More information

Overview of HPC systems and software available within

Overview of HPC systems and software available within Overview of HPC systems and software available within Overview Available HPC Systems Ba Cy-Tera Available Visualization Facilities Software Environments HPC System at Bibliotheca Alexandrina SUN cluster

More information

Case Study on Productivity and Performance of GPGPUs

Case Study on Productivity and Performance of GPGPUs Case Study on Productivity and Performance of GPGPUs Sandra Wienke wienke@rz.rwth-aachen.de ZKI Arbeitskreis Supercomputing April 2012 Rechen- und Kommunikationszentrum (RZ) RWTH GPU-Cluster 56 Nvidia

More information

5x in 5 hours Porting SEISMIC_CPML using the PGI Accelerator Model

5x in 5 hours Porting SEISMIC_CPML using the PGI Accelerator Model 5x in 5 hours Porting SEISMIC_CPML using the PGI Accelerator Model C99, C++, F2003 Compilers Optimizing Vectorizing Parallelizing Graphical parallel tools PGDBG debugger PGPROF profiler Intel, AMD, NVIDIA

More information

Trends in High-Performance Computing for Power Grid Applications

Trends in High-Performance Computing for Power Grid Applications Trends in High-Performance Computing for Power Grid Applications Franz Franchetti ECE, Carnegie Mellon University www.spiral.net Co-Founder, SpiralGen www.spiralgen.com This talk presents my personal views

More information

HPC enabling of OpenFOAM R for CFD applications

HPC enabling of OpenFOAM R for CFD applications HPC enabling of OpenFOAM R for CFD applications Towards the exascale: OpenFOAM perspective Ivan Spisso 25-27 March 2015, Casalecchio di Reno, BOLOGNA. SuperComputing Applications and Innovation Department,

More information

Cosmological simulations on High Performance Computers

Cosmological simulations on High Performance Computers Cosmological simulations on High Performance Computers Cosmic Web Morphology and Topology Cosmological workshop meeting Warsaw, 12-17 July 2011 Maciej Cytowski Interdisciplinary Centre for Mathematical

More information

Altix Usage and Application Programming. Welcome and Introduction

Altix Usage and Application Programming. Welcome and Introduction Zentrum für Informationsdienste und Hochleistungsrechnen Altix Usage and Application Programming Welcome and Introduction Zellescher Weg 12 Tel. +49 351-463 - 35450 Dresden, November 30th 2005 Wolfgang

More information

David Rioja Redondo Telecommunication Engineer Englobe Technologies and Systems

David Rioja Redondo Telecommunication Engineer Englobe Technologies and Systems David Rioja Redondo Telecommunication Engineer Englobe Technologies and Systems About me David Rioja Redondo Telecommunication Engineer - Universidad de Alcalá >2 years building and managing clusters UPM

More information

JuRoPA. Jülich Research on Petaflop Architecture. One Year on. Hugo R. Falter, COO Lee J Porter, Engineering

JuRoPA. Jülich Research on Petaflop Architecture. One Year on. Hugo R. Falter, COO Lee J Porter, Engineering JuRoPA Jülich Research on Petaflop Architecture One Year on Hugo R. Falter, COO Lee J Porter, Engineering HPC Advisoy Counsil, Workshop 2010, Lugano 1 Outline The work of ParTec on JuRoPA (HF) Overview

More information

Turbomachinery CFD on many-core platforms experiences and strategies

Turbomachinery CFD on many-core platforms experiences and strategies Turbomachinery CFD on many-core platforms experiences and strategies Graham Pullan Whittle Laboratory, Department of Engineering, University of Cambridge MUSAF Colloquium, CERFACS, Toulouse September 27-29

More information

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC2013 - Denver

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC2013 - Denver 1 The PHI solution Fujitsu Industry Ready Intel XEON-PHI based solution SC2013 - Denver Industrial Application Challenges Most of existing scientific and technical applications Are written for legacy execution

More information

Supercomputing Resources in BSC, RES and PRACE

Supercomputing Resources in BSC, RES and PRACE www.bsc.es Supercomputing Resources in BSC, RES and PRACE Sergi Girona, BSC-CNS Barcelona, 23 Septiembre 2015 ICTS 2014, un paso adelante para la RES Past RES members and resources BSC-CNS (MareNostrum)

More information

Debugging in Heterogeneous Environments with TotalView. ECMWF HPC Workshop 30 th October 2014

Debugging in Heterogeneous Environments with TotalView. ECMWF HPC Workshop 30 th October 2014 Debugging in Heterogeneous Environments with TotalView ECMWF HPC Workshop 30 th October 2014 Agenda Introduction Challenges TotalView overview Advanced features Current work and future plans 2014 Rogue

More information

Applications to Computational Financial and GPU Computing. May 16th. Dr. Daniel Egloff +41 44 520 01 17 +41 79 430 03 61

Applications to Computational Financial and GPU Computing. May 16th. Dr. Daniel Egloff +41 44 520 01 17 +41 79 430 03 61 F# Applications to Computational Financial and GPU Computing May 16th Dr. Daniel Egloff +41 44 520 01 17 +41 79 430 03 61 Today! Why care about F#? Just another fashion?! Three success stories! How Alea.cuBase

More information

GRID Computing and Networks

GRID Computing and Networks A Member of the ExperTeam Group GRID Computing and Networks Karl Solchenbach Global IPv6 Summit Madrid, May 14, 2003 Pallas GmbH Hermülheimer Straße 10 D-50321 Brühl, Germany info@pallas.de http://www.pallas.com

More information

Application Performance Analysis Tools and Techniques

Application Performance Analysis Tools and Techniques Mitglied der Helmholtz-Gemeinschaft Application Performance Analysis Tools and Techniques 2012-06-27 Christian Rössel Jülich Supercomputing Centre c.roessel@fz-juelich.de EU-US HPC Summer School Dublin

More information

BIG CPU, BIG DATA. Solving the World s Toughest Computational Problems with Parallel Computing. Alan Kaminsky

BIG CPU, BIG DATA. Solving the World s Toughest Computational Problems with Parallel Computing. Alan Kaminsky Solving the World s Toughest Computational Problems with Parallel Computing Alan Kaminsky Solving the World s Toughest Computational Problems with Parallel Computing Alan Kaminsky Department of Computer

More information

PyFR: Bringing Next Generation Computational Fluid Dynamics to GPU Platforms

PyFR: Bringing Next Generation Computational Fluid Dynamics to GPU Platforms PyFR: Bringing Next Generation Computational Fluid Dynamics to GPU Platforms P. E. Vincent! Department of Aeronautics Imperial College London! 25 th March 2014 Overview Motivation Flux Reconstruction Many-Core

More information

Sourcery Overview & Virtual Machine Installation

Sourcery Overview & Virtual Machine Installation Sourcery Overview & Virtual Machine Installation Damian Rouson, Ph.D., P.E. Sourcery, Inc. www.sourceryinstitute.org Sourcery, Inc. About Us Sourcery, Inc., is a software consultancy founded by and for

More information

GPU Hardware and Programming Models. Jeremy Appleyard, September 2015

GPU Hardware and Programming Models. Jeremy Appleyard, September 2015 GPU Hardware and Programming Models Jeremy Appleyard, September 2015 A brief history of GPUs In this talk Hardware Overview Programming Models Ask questions at any point! 2 A Brief History of GPUs 3 Once

More information

Performance Evaluation of NAS Parallel Benchmarks on Intel Xeon Phi

Performance Evaluation of NAS Parallel Benchmarks on Intel Xeon Phi Performance Evaluation of NAS Parallel Benchmarks on Intel Xeon Phi ICPP 6 th International Workshop on Parallel Programming Models and Systems Software for High-End Computing October 1, 2013 Lyon, France

More information

Overview of HPC Resources at Vanderbilt

Overview of HPC Resources at Vanderbilt Overview of HPC Resources at Vanderbilt Will French Senior Application Developer and Research Computing Liaison Advanced Computing Center for Research and Education June 10, 2015 2 Computing Resources

More information

High Performance Computing in CST STUDIO SUITE

High Performance Computing in CST STUDIO SUITE High Performance Computing in CST STUDIO SUITE Felix Wolfheimer GPU Computing Performance Speedup 18 16 14 12 10 8 6 4 2 0 Promo offer for EUC participants: 25% discount for K40 cards Speedup of Solver

More information

Sun Constellation System: The Open Petascale Computing Architecture

Sun Constellation System: The Open Petascale Computing Architecture CAS2K7 13 September, 2007 Sun Constellation System: The Open Petascale Computing Architecture John Fragalla Senior HPC Technical Specialist Global Systems Practice Sun Microsystems, Inc. 25 Years of Technical

More information

Designing and Building Applications for Extreme Scale Systems CS598 William Gropp www.cs.illinois.edu/~wgropp

Designing and Building Applications for Extreme Scale Systems CS598 William Gropp www.cs.illinois.edu/~wgropp Designing and Building Applications for Extreme Scale Systems CS598 William Gropp www.cs.illinois.edu/~wgropp Welcome! Who am I? William (Bill) Gropp Professor of Computer Science One of the Creators of

More information

Enterprise HPC & Cloud Computing for Engineering Simulation. Barbara Hutchings Director, Strategic Partnerships ANSYS, Inc.

Enterprise HPC & Cloud Computing for Engineering Simulation. Barbara Hutchings Director, Strategic Partnerships ANSYS, Inc. Enterprise HPC & Cloud Computing for Engineering Simulation Barbara Hutchings Director, Strategic Partnerships ANSYS, Inc. Historical Perspective Evolution of Computing for Simulation Pendulum swing: Centralized

More information

walberla: A software framework for CFD applications

walberla: A software framework for CFD applications walberla: A software framework for CFD applications U. Rüde, S. Donath, C. Feichtinger, K. Iglberger, F. Deserno, M. Stürmer, C. Mihoubi, T. Preclic, D. Haspel (all LSS Erlangen), N. Thürey (LSS Erlangen/

More information

Recent and Future Activities in HPC and Scientific Data Management Siegfried Benkner

Recent and Future Activities in HPC and Scientific Data Management Siegfried Benkner Recent and Future Activities in HPC and Scientific Data Management Siegfried Benkner Research Group Scientific Computing Faculty of Computer Science University of Vienna AUSTRIA http://www.par.univie.ac.at

More information

walberla: A software framework for CFD applications on 300.000 Compute Cores

walberla: A software framework for CFD applications on 300.000 Compute Cores walberla: A software framework for CFD applications on 300.000 Compute Cores J. Götz (LSS Erlangen, jan.goetz@cs.fau.de), K. Iglberger, S. Donath, C. Feichtinger, U. Rüde Lehrstuhl für Informatik 10 (Systemsimulation)

More information

Performance Tools for System Monitoring

Performance Tools for System Monitoring Center for Information Services and High Performance Computing (ZIH) 01069 Dresden Performance Tools for System Monitoring 1st CHANGES Workshop, Jülich Zellescher Weg 12 Tel. +49 351-463 35450 September

More information

The GPU Accelerated Data Center. Marc Hamilton, August 27, 2015

The GPU Accelerated Data Center. Marc Hamilton, August 27, 2015 The GPU Accelerated Data Center Marc Hamilton, August 27, 2015 THE GPU-ACCELERATED DATA CENTER HPC DEEP LEARNING PC VIRTUALIZATION CLOUD GAMING RENDERING 2 Product design FROM ADVANCED RENDERING TO VIRTUAL

More information

A Case Study - Scaling Legacy Code on Next Generation Platforms

A Case Study - Scaling Legacy Code on Next Generation Platforms Available online at www.sciencedirect.com ScienceDirect Procedia Engineering 00 (2015) 000 000 www.elsevier.com/locate/procedia 24th International Meshing Roundtable (IMR24) A Case Study - Scaling Legacy

More information

Parallel Computing. Introduction

Parallel Computing. Introduction Parallel Computing Introduction Thorsten Grahs, 14. April 2014 Administration Lecturer Dr. Thorsten Grahs (that s me) t.grahs@tu-bs.de Institute of Scientific Computing Room RZ 120 Lecture Monday 11:30-13:00

More information

Extreme Scale Compu0ng at LRZ

Extreme Scale Compu0ng at LRZ Extreme Scale Compu0ng at LRZ Dieter Kranzlmüller Munich Network Management Team Ludwig- Maximilians- Universität München (LMU) & Leibniz SupercompuFng Centre (LRZ) of the Bavarian Academy of Sciences

More information

Search Strategies for Automatic Performance Analysis Tools

Search Strategies for Automatic Performance Analysis Tools Search Strategies for Automatic Performance Analysis Tools Michael Gerndt and Edmond Kereku Technische Universität München, Fakultät für Informatik I10, Boltzmannstr.3, 85748 Garching, Germany gerndt@in.tum.de

More information

CoolEmAll - Tools for realising an energy efficient data centre

CoolEmAll - Tools for realising an energy efficient data centre CoolEmAll - Tools for realising an energy efficient data centre Wolfgang Christmann christmann informationstechnik + medien GmbH & Co. KG www.christmann.info 1 Outline CoolEmAll project RECS system towards

More information

OpenMP Programming on ScaleMP

OpenMP Programming on ScaleMP OpenMP Programming on ScaleMP Dirk Schmidl schmidl@rz.rwth-aachen.de Rechen- und Kommunikationszentrum (RZ) MPI vs. OpenMP MPI distributed address space explicit message passing typically code redesign

More information

PRACE hardware, software and services. David Henty, EPCC, d.henty@epcc.ed.ac.uk

PRACE hardware, software and services. David Henty, EPCC, d.henty@epcc.ed.ac.uk PRACE hardware, software and services David Henty, EPCC, d.henty@epcc.ed.ac.uk Why? Weather, Climatology, Earth Science degree of warming, scenarios for our future climate. understand and predict ocean

More information

Recent Advances in HPC for Structural Mechanics Simulations

Recent Advances in HPC for Structural Mechanics Simulations Recent Advances in HPC for Structural Mechanics Simulations 1 Trends in Engineering Driving Demand for HPC Increase product performance and integrity in less time Consider more design variants Find the

More information

Parallel file I/O bottlenecks and solutions

Parallel file I/O bottlenecks and solutions Mitglied der Helmholtz-Gemeinschaft Parallel file I/O bottlenecks and solutions Views to Parallel I/O: Hardware, Software, Application Challenges at Large Scale Introduction SIONlib Pitfalls, Darshan,

More information

supercomputing. simplified.

supercomputing. simplified. supercomputing. simplified. INTRODUCING WINDOWS HPC SERVER 2008 R2 SUITE Windows HPC Server 2008 R2, Microsoft s third-generation HPC solution, provides a comprehensive and costeffective solution for harnessing

More information

Integrated Communication Systems

Integrated Communication Systems Integrated Communication Systems Courses, Research, and Thesis Topics Prof. Paul Müller University of Kaiserslautern Department of Computer Science Integrated Communication Systems ICSY http://www.icsy.de

More information

MEng, BSc Applied Computer Science

MEng, BSc Applied Computer Science School of Computing FACULTY OF ENGINEERING MEng, BSc Applied Computer Science Year 1 COMP1212 Computer Processor Effective programming depends on understanding not only how to give a machine instructions

More information

FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG

FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG INSTITUT FÜR INFORMATIK (MATHEMATISCHE MASCHINEN UND DATENVERARBEITUNG) Lehrstuhl für Informatik 10 (Systemsimulation) Massively Parallel Multilevel Finite

More information

Using GPUs in the Cloud for Scalable HPC in Engineering and Manufacturing March 26, 2014

Using GPUs in the Cloud for Scalable HPC in Engineering and Manufacturing March 26, 2014 Using GPUs in the Cloud for Scalable HPC in Engineering and Manufacturing March 26, 2014 David Pellerin, Business Development Principal Amazon Web Services David Hinz, Director Cloud and HPC Solutions

More information

Experiments in Unstructured Mesh Finite Element CFD Using CUDA

Experiments in Unstructured Mesh Finite Element CFD Using CUDA Experiments in Unstructured Mesh Finite Element CFD Using CUDA Graham Markall Software Performance Imperial College London http://www.doc.ic.ac.uk/~grm08 grm08@doc.ic.ac.uk Joint work with David Ham and

More information

High Performance Computing within the AHRP http://www.ahrp.info http://www.ahrp.info

High Performance Computing within the AHRP http://www.ahrp.info http://www.ahrp.info High Performance Computing within the AHRP http://www.ahrp.info http://www.ahrp.info The Alliance for HPC Rhineland-Palatinate! History, Goals and Tasks! Organization! Access to Resources! Training and

More information

The Uintah Framework: A Unified Heterogeneous Task Scheduling and Runtime System

The Uintah Framework: A Unified Heterogeneous Task Scheduling and Runtime System The Uintah Framework: A Unified Heterogeneous Task Scheduling and Runtime System Qingyu Meng, Alan Humphrey, Martin Berzins Thanks to: John Schmidt and J. Davison de St. Germain, SCI Institute Justin Luitjens

More information

Using WestGrid. Patrick Mann, Manager, Technical Operations Jan.15, 2014

Using WestGrid. Patrick Mann, Manager, Technical Operations Jan.15, 2014 Using WestGrid Patrick Mann, Manager, Technical Operations Jan.15, 2014 Winter 2014 Seminar Series Date Speaker Topic 5 February Gino DiLabio Molecular Modelling Using HPC and Gaussian 26 February Jonathan

More information

Productivity and HPC. Erik Hagersten, CTO, Rogue Wave Software AB Developing parallel, data-intensive applications is hard. We make it easier.

Productivity and HPC. Erik Hagersten, CTO, Rogue Wave Software AB Developing parallel, data-intensive applications is hard. We make it easier. Productivity and HPC Erik Hagersten, CTO, Rogue Wave Software AB Developing parallel, data-intensive applications is hard. We make it easier. Chief architect high-end servers Sun Microsystems 1994 1999

More information

ALPS - The Swiss Grand Challenge Programme on the Cray XT3. CUG 2007, Seattle Dominik Ulmer, CSCS

ALPS - The Swiss Grand Challenge Programme on the Cray XT3. CUG 2007, Seattle Dominik Ulmer, CSCS ALPS - The Swiss Grand Challenge Programme on the Cray XT3 CUG 2007, Seattle Dominik Ulmer, CSCS CSCS today Swiss National Supercomputing Centre, founded in 1991 as part of ETHZ. Since 2004, an autonomous

More information

Next Generation GPU Architecture Code-named Fermi

Next Generation GPU Architecture Code-named Fermi Next Generation GPU Architecture Code-named Fermi The Soul of a Supercomputer in the Body of a GPU Why is NVIDIA at Super Computing? Graphics is a throughput problem paint every pixel within frame time

More information

~ Greetings from WSU CAPPLab ~

~ Greetings from WSU CAPPLab ~ ~ Greetings from WSU CAPPLab ~ Multicore with SMT/GPGPU provides the ultimate performance; at WSU CAPPLab, we can help! Dr. Abu Asaduzzaman, Assistant Professor and Director Wichita State University (WSU)

More information

Next Generation Application Enablement Tools: A Framework for Automated Performance Analysis and Tuning

Next Generation Application Enablement Tools: A Framework for Automated Performance Analysis and Tuning Next Generation Application Enablement Tools: A Framework for Automated Performance Analysis and Tuning David Klepacki Advanced Computing Technology T.J. Watson Research Center 2008 Corporation ACTC Toolkit

More information

Introduction to Linux and Cluster Basics for the CCR General Computing Cluster

Introduction to Linux and Cluster Basics for the CCR General Computing Cluster Introduction to Linux and Cluster Basics for the CCR General Computing Cluster Cynthia Cornelius Center for Computational Research University at Buffalo, SUNY 701 Ellicott St Buffalo, NY 14203 Phone: 716-881-8959

More information

Recent Advances in Periscope for Performance Analysis and Tuning

Recent Advances in Periscope for Performance Analysis and Tuning Recent Advances in Periscope for Performance Analysis and Tuning Isaias Compres, Michael Firbach, Michael Gerndt Robert Mijakovic, Yury Oleynik, Ventsislav Petkov Technische Universität München Yury Oleynik,

More information

COMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1)

COMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1) COMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1) Mary Thomas Department of Computer Science Computational Science Research Center (CSRC) San Diego State University

More information

Program Grid and HPC5+ workshop

Program Grid and HPC5+ workshop Program Grid and HPC5+ workshop 24-30, Bahman 1391 Tuesday Wednesday 9.00-9.45 9.45-10.30 Break 11.00-11.45 11.45-12.30 Lunch 14.00-17.00 Workshop Rouhani Karimi MosalmanTabar Karimi G+MMT+K Opening IPM_Grid

More information

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC HPC Architecture End to End Alexandre Chauvin Agenda HPC Software Stack Visualization National Scientific Center 2 Agenda HPC Software Stack Alexandre Chauvin Typical HPC Software Stack Externes LAN Typical

More information

CERN openlab III. Major Review Platform CC. Sverre Jarp Alfio Lazzaro Julien Leduc Andrzej Nowak

CERN openlab III. Major Review Platform CC. Sverre Jarp Alfio Lazzaro Julien Leduc Andrzej Nowak CERN openlab III Major Review Platform CC Sverre Jarp Alfio Lazzaro Julien Leduc Andrzej Nowak Teaching (1) 3 workshops already held this year: Computer Architecture and Performance Tuning: 17/18 February

More information

MEng, BSc Computer Science with Artificial Intelligence

MEng, BSc Computer Science with Artificial Intelligence School of Computing FACULTY OF ENGINEERING MEng, BSc Computer Science with Artificial Intelligence Year 1 COMP1212 Computer Processor Effective programming depends on understanding not only how to give

More information

Fast Parallel Algorithms for Computational Bio-Medicine

Fast Parallel Algorithms for Computational Bio-Medicine Fast Parallel Algorithms for Computational Bio-Medicine H. Köstler, J. Habich, J. Götz, M. Stürmer, S. Donath, T. Gradl, D. Ritter, D. Bartuschat, C. Feichtinger, C. Mihoubi, K. Iglberger (LSS Erlangen)

More information

Introduction to grid technologies, parallel and cloud computing. Alaa Osama Allam Saida Saad Mohamed Mohamed Ibrahim Gaber

Introduction to grid technologies, parallel and cloud computing. Alaa Osama Allam Saida Saad Mohamed Mohamed Ibrahim Gaber Introduction to grid technologies, parallel and cloud computing Alaa Osama Allam Saida Saad Mohamed Mohamed Ibrahim Gaber OUTLINES Grid Computing Parallel programming technologies (MPI- Open MP-Cuda )

More information

CUDA programming on NVIDIA GPUs

CUDA programming on NVIDIA GPUs p. 1/21 on NVIDIA GPUs Mike Giles mike.giles@maths.ox.ac.uk Oxford University Mathematical Institute Oxford-Man Institute for Quantitative Finance Oxford eresearch Centre p. 2/21 Overview hardware view

More information

Anwendungsintegration und Workflows mit UNICORE 6

Anwendungsintegration und Workflows mit UNICORE 6 Mitglied der Helmholtz-Gemeinschaft Anwendungsintegration und Workflows mit UNICORE 6 Bernd Schuller und UNICORE-Team Jülich Supercomputing Centre, Forschungszentrum Jülich GmbH 26. November 2009 D-Grid

More information

HOPSA Project. Technical Report. A Workflow for Holistic Performance System Analysis

HOPSA Project. Technical Report. A Workflow for Holistic Performance System Analysis HOPSA Project Technical Report A Workflow for Holistic Performance System Analysis September 2012 Felix Wolf 1,2,3, Markus Geimer 2, Judit Gimenez 4, Juan Gonzalez 4, Erik Hagersten 5, Thomas Ilsche 6,

More information

Mitglied der Helmholtz-Gemeinschaft. System monitoring with LLview and the Parallel Tools Platform

Mitglied der Helmholtz-Gemeinschaft. System monitoring with LLview and the Parallel Tools Platform Mitglied der Helmholtz-Gemeinschaft System monitoring with LLview and the Parallel Tools Platform November 25, 2014 Carsten Karbach Content 1 LLview 2 Parallel Tools Platform (PTP) 3 Latest features 4

More information

Petascale Software Challenges. William Gropp www.cs.illinois.edu/~wgropp

Petascale Software Challenges. William Gropp www.cs.illinois.edu/~wgropp Petascale Software Challenges William Gropp www.cs.illinois.edu/~wgropp Petascale Software Challenges Why should you care? What are they? Which are different from non-petascale? What has changed since

More information

An Introduction to Parallel Computing/ Programming

An Introduction to Parallel Computing/ Programming An Introduction to Parallel Computing/ Programming Vicky Papadopoulou Lesta Astrophysics and High Performance Computing Research Group (http://ahpc.euc.ac.cy) Dep. of Computer Science and Engineering European

More information