OASIS3-MCT & Open-PALM: 2 open source codes couplers
|
|
|
- Abigayle Williams
- 10 years ago
- Views:
Transcription
1 OASIS3-MCT & Open-PALM: 2 open source codes couplers Anthony Thevenin ([email protected])
2 Code coupling issues Ê Why coupling scientific computing codes? Ä To treat a global system e.g. ocean - atmosphere coupling for Ocean + Atmosphere e climate modeling NEMO ARPEGE flux exchanges Ä To set up new applications from existing codes e.g. fluid - structure coupling for the research of the optimal position of the intakes in a combustion chamber Ä To build modular applications by splitting complex problems and assembling their elementary components e.g. data assimilation Page 2
3 Ê What does it imply? Code coupling issues Ä Drive the information exchanges between codes Ä Manage codes execution (serial or parallel) DATA exchanges Code1 Code2 Ê Commitments: Ä Easy to implement: non intrusive interfaces Ä Efficiency: no loss in performances Ä Portability: standard technical solutions Ä Flexibility: test different configurations with a few changes Ä Genericity: reuse components in other couplings Page 3
4 Ê First solution : code merging Technical coupling solutions Code 1: ocean model Program prog1 Call sub_prog2(in, out) end Code 2: atmosphere model Program prog2 Subroutine sub_prog2(in,out) end Ä Very efficient (memory exchange) and portable Ä BUT : c integration problems (no independency between codes) c No flexibility (problems in maintenance, evolution, ) c Potential memory wastes c parallelism of the master code imposed to the application Ä This is merging and not coupling Page 4
5 Technical coupling solutions Ê Second solution: use a communication protocol (MPI, CORBA, files, ) Code 1: ocean model Program prog1 Call XXX_send(prog2, data) end Code 2: atmosphere model Program prog2 Call XXX_recv(prog1, data) end Ä Can be very efficient and preserve existing codes (parallelism) Ä But: c Coupling is not generic c Good experience in parallel computing required c Lack of flexibility (interpolation, transformation, ) c Not always portable depending on the chosen mechanism c Too complex with more than two codes and many exchanges Page 5
6 Technical coupling solutions Ê Third solution: choose a coupler (or a coupling library) Ä A coupler is a software allowing data exchanges between several codes and managing their executions Ä It starts the executables in a sequence or in parallel Ä Data exchanges are invoked by general PUT/GET primitives Ä It uses a portable and efficient communication protocol (MPI) Ä It provides tools: c time and spatial interpolations, c parallel data redistribution, c monitoring, performance analysis COUPLER Code 1 Code 2 Code n Page 6
7 Technical coupling solutions Ê Static coupling (OASIS) Dynamic coupling (Open-PALM) Code1 Code2 Time Code1 Code2 Loop Code3 condition Code4 c Codes are started at the beginning of the simulation and exit together at the end of the application c A component can be launched and can release resources upon termination at any moment during the simulation. Computing resources (required memory and the number of concurrent processors) are handled by the coupler c Programs can be executed in loops or under logical conditions (faculty of describing complex coupling algorithms) Page 7
8 Technical coupling solutions Ê Managing data exchanges between components: end-point communications DATA Code1 Code2 Ä When a code produces a data potentially interesting for other components, it informs the coupler: c appeal to a generic Put primitive (non-blocking) Ä When a calculation code needs data, it tells the coupler that it is waiting for some information: c appeal to a generic Get primitive (blocking) Ä Total independence between the codes: c NO explicit indication of the receiving units on the producer side, NOR of the producer on the receivers side c replacement of a code by another one without interfering with the rest the application. Page 8
9 Technical coupling solutions Ê The coupler: probably the best solution to couple independently developed code Ä At run time, the component models remain separate executables and the coupler acts as a communication library linked to the models Ä Two levels of parallelism: c distributed components (codes can be parallel programs) c task parallelism (codes can be executed in parallel) Ä BUT: c multi-executable: possible waste of resources if sequential execution of the components is enforced c multi-executable: more difficult to debug; harder to manage for the OS c more or less efficient depending on the coupling algorithm Page 9
10 OASIS3 MCT, an open source codes coupler for massively parallel climate modeling Page 10
11 OASIS3 - MCT PRISM IS-ENES OASIS1 OASIS2 OASIS3 OASIS3-MCT OASIS4 Ê PRISM : Partnership for Research Infrastructure in Earth System Modeling Ê IS-ENES: InfraStructure for ENES : European project Ê Current developers : 2 permanents (CERFACS, CNRS) & 1 consultant Ê All sources are written in F90 and C Ê Open source product, distributed under a LGPL license Ê All external libraries used are public domain: MPI, NetCDF ; or open source: SCRIP (Los Alamos National Laboratory), MCT (Argonne National Laboratory) Page 11
12 OASIS3 - MCT Ê About 35 groups world-wide (climate modelling or operational monthly/seasonal forecasting): France: CERFACS, METEO-FRANCE, IPSL (LOCEAN, LMD, LSCE), OMP, LGGE, IFREMER Europe: ECMWF + Ec-Earth community Germany: MPI-M, IFM-GEOMAR, HZG, U. Frankfurt UK: MetOffice, NCAS/U. Reading, ICL Denmark: DMI Norway: U. Bergen Sweden: SMHI, U. Lund Ireland: ICHEC, NUI Galway The Netherland: KNMI Switzerland: ETH Zurich Italy: INGV, ENEA, CASPUR Czech_Republic :CHMI Spain: U. Castilla Tunisia: Inst. Nat. Met Japan: JMA, JAMSTEC China: IAP-CAS, Met. Nat. Centre, SCSIO Korea: KMA Australia: CSIRO New Zealand: NIWA Canada: RPN-Environment Canada, UQAM USA: Oregon State U., Hawaii U., JPL, MIT Peru: IGP OASIS3 is used in 5 European ESMs that participate in IPCC AR5 Page 12
13 OASIS3 - MCT Ê To use OASIS3-MCT: Ä Register on the website and download the sources Ä Compile and run the tutorial on the target platform Ê To configure OASIS3-MCT: Ä Identify the component models and their grids Ä Identify the coupling fields to be exchanged between those models Ä Identify the coupling parameters: - total duration - sources and targets symbolic names - exchange period - field transformations and interpolations c write a full namcouple configuration file Ä Insert calls of the OASIS3-MCT communication library Ê And then: Ä Compile OASIS3-MCT and link the component models with it Ä Start the models and let OASIS3-MCT manage the coupling exchanges Page 13
14 OASIS3 - MCT Ê Component model instrumentation: Ä Initialization: Ä Grid definition: Ä Local partition definition: Ä Coupling field declaration: Ä End of definition phase: call oasis_init_comp( )! call oasis_write_grid( )! call oasis_def_partition( )! call oasis_def_var( )! call oasis_enddef( )!! Ä Coupling field exchange: in model time stepping loop : call oasis_put(, date, field, )! call oasis_get(, date, field, )! c user s defined source or target c sending or receiving at appropriate time only c automatic averaging / accumulation if requested c automatic writing of coupling restart file at the end of run Ä Termination : call oasis_terminate( )! Page 14
15 OASIS3 - MCT Ê Communication: Ä Fully parallel communications between parallel models, based on MPI! pe1 pe2 pe3 model1 model2 Ä Interpolation of the coupling fields performed directly on the source or target component processes Ä Parallel redistribution of the coupling fields done directly from the source to the target component processes c without gathering the whole coupling field on an additional process as in the previous OASIS3 versions (no bottleneck) Ê I/O functionality: Ä Switch between coupled and forced mode! pe1 pe2 pe3 model1 file pe1 pe2 pe3 pe4 Page 15
16 Ê Interpolations: OASIS3 - MCT Ê Interpolations & transformations: Ä On 2D vector fields Ä On different types of grids : lat-lon, rotated, gaussian reduced, unstructured Ê Transformations: Ä Statistics, addition / multiplication by a scalar, global conservation x x1 x2 x3 x4 Nearest-neighbor bilinear conservative x source grid point target grid point Page 16
17 OASIS3 - MCT Ê Performance: 0,3500 Time for one ping-pong exchange on Curie (toy models) for OASIS-MCT and OASIS3.3 0,3000 Time (seconds) 0,2500 0,2000 0,1500 0,1000 0,0500 oasis3-mct #1 oasis3.3 #1 0, Number of cores Page 17
18 OASIS3 - MCT Ê Conclusions : Ê Perspectives: Ä First version of OASIS3-MCT released in August 2012 Ä Very simple to use for traditional OASIS3 users Ä OASIS3-MCT solves the problem of the coupling bottleneck observed at high number of cores for previous OASIS versions, with which the coupling fields had to be reassembled onto one (or more) coupler process to perform a mono-process interpolation Ä Excellent example of the benefit of shared software Ä Bi-cubic interpolation Ä Work on scalability at >1000 cores Ê Regular training sessions at CERFACS Page 18
19 Open-PALM : an open source codes coupler for massively parallel multiphysics/multi-components applications and dynamic algorithms Page 19
20 Ê History : Open-PALM Ä 1996: Mercator project: operational oceanography c data assimilation with a high resolution model c need for a modular and efficient tool to compose assimilation algorithm (component coupling approach) Ä Expertise at CERFACS: c OASIS coupler for climate modeling c parallel computations, data assimilation Ä Human power for the project: c 2 Mercator engineers c 1 CNRS engineer c 1 to 2 CERFACS engineers Ä 2000: first proto PALM_PROTO for Mercator Ä 2002: PALM_Research, SPMD version that simulates MPMD Ä 2003: PALM_MP, true MPMD Ä 2011: Open-PALM, open-source (LGPL) and collaboration with ONERA Page 20
21 Open-PALM Ê 60 groups use OpenPALM, half for data assimilation, half for multi physic and multi component simulations Ê > 300 people trained to OpenPALM Ê Various domains of use: - D.A.: Oceanography, hydrology, atmospheric chemistry, - Multi physic: hydrology, CFD, agriculture... Ê At CERFACS : 3 teams regularly use OpenPALM: GLOBC, CFD, AE, More than 15 regular users among the about 120 people, Most of CERFACS partners use the code or plan to use it : Météo-France, EDF, CNES, SAFRAN, ONERA Ê Current developers : CERFACS : 3 permanents & 1 post-doc ONERA : 1 permanent Page 21
22 Open-PALM Ê The graphical user interface PrePALM (CERFACS): Ä Definition of the coupling scheme Ä Monitoring and post-processing Ä Runs on standard PC or on clusters, allows to define coupling algorithms and prepare compilations process Ê PALM library (CERFACS): Ä Dynamic launching by the PALM driver which drives the parallel/sequential execution of the components Ä Management of data exchanges Ä The PALM library is compiled on the different machines where the applications are executed Ê CWIPI library (ONERA): Ä Manages fully parallel data exchanged based on distributed mesh definition Ä Interpolation between non coincident meshes c Open-PALM = PALM (GUI and library) + CWIPI Page 22
23 Open-PALM Ê PrePALM: a friendly and simple GUI for: Ä Representing the components Ä Handling the parallelism (the two levels) Ä Managing the resources Ä Describing loops and conditions Ä Describing the communications (exchanges) c implement and supervise coupled applications Ê With all the benefits of a GUI as, for instance: Ä Coherency checks on the exchanged data Ä Pictorial flow chart representation of the algorithm Ä Performance analysis, debugging and run-time monitoring Page 23
24 Open-PALM Page 24
25 Open-PALM TIME Page 25
26 Open-PALM Branch code: Control structures Fortran regions Branch: sequence of units and instructions c algorithm Unit: computational component Communications: described by connecting the plugs - Code - Module of a code - Data loaders (NetCDF) - Pre-defined unit Page 26
27 Open-PALM Ê Two levels of parallelism to take the best advantage from the intrinsic parallelism of an application (MPI+OpenMP possible) 1 Task parallelism: by drawing branches we program task parallelism Distributed components (each component is a MPI executable): parallel units Page 27
28 Open-PALM Ê Process management: Ä The PrePALM scheme is converted into a graph for the PALM driver Ä This driver schedules this graph in order to launch the components depending on the available resources Ä MPI-2 spawn as well as MPI-2 client/server are used for the dynamic launching of the units. MPI-1 launching available (choice during installation) Ê Target machines: Ä HPC or clusters Ä Designed for homogeneous computing but functionalities exist for heterogeneous computing (TCP/IP) Ä Open-PALM is ported on many architectures (BULL, NEC, CRAY, IBM, SGI, Linux, Mac OS X, Windows via Cygwin) Page 28
29 Open-PALM Ê Transformation of a code into a PALM unit: Ä Creation of an ID card (Read by PrePALM) Ä Replace PROGRAM (FORTRAN) by subroutine, or MAIN (C) by a function name Ä For a parallel unit: c skip the calls to MPI_INIT and MPI_FINALIZE c replace MPI_COMM_WORLD by PL_COMM_EXEC c replace the calls to STOP, EXIT by PALM_Abort Ä Add the PALM Primitives into the codes: c PALM_Put(args): to send data c PALM_Get(args): to receive data Ê Pre-requisites: Ä Codes must run under Linux/Unix Ä Code sources should be available AND have to be written in a compiled language (C, C++, FORTRAN 77 ou 90) OR Ä User functions (C or FORTRAN) with pointers on data provided by the code Page 29
30 Ê Algebra toolbox: Open-PALM Pre-defined units interfacing common mathematical libraries (BLAS, LAPACK, ) applied on the fly to objects exchanged between units (Interesting for units conversion, ). Load the unit from the Algebra Toolbox Execution and communications are managed as if it were a user unit Page 30
31 Ê Monitoring Tools: Open-PALM Ä Run time monitoring: PrePALM offers the opportunity of monitoring the computation evolution and the resources usage during the execution of an application. Ä Performances analysis: c CPU and elapsed time statistics for the user defined units c post-mortem replay of the algorithm progress in the graphical interface Ä Debugging tools: c a user defined debugging function, called on the PALM_Put and PALM_Get sides of an active communication to check the coherency of the objects, to track the execution and compute the global performances of the parallel execution c a PALM output file per branch with different verbosity levels set by the user for different classes of messages Page 31
32 Open-PALM Page 32
33 Open-PALM Page 33
34 Open-PALM Ê The CWIPI: library for data exchange with interpolation in parallel Ä Open source (LGPL), developed at ONERA since 2009 Ä Management of coupling between n parallel codes Ä Direct data exchange of interpolated fields between a source and a target meshes (non coincident and differently partitioned on several processes) Ä CWIPI primitives Interfaced for C/C++, Fortran and Python codes Ê Type of meshes: Ä 1D: edges Ä 2D: triangles, quadrangles, polygons Ä 3D: tetrahedrons, pyramids, prisms, hexahedra, polyhedra Ê Source and target mesh type can be the same or not Ê Primitives are called in the codes Page 34
35 Open-PALM Ê PALM is a software dealing with couplings or with complex applications built upon sub-modules while respecting the performances of the codes Ê Friendly and simple G.U.I. to describe, implement and supervise applications Ê Many interests: Ä Facilitate evolution and maintenance of complex applications and of its components, Ä Easy integration of new codes replacement, in existing applications (multi-physic coupling, collaborative development ) Ä Maximize the use of intrinsic parallelism of applications and algorithms Ê Regular training sessions at CERFACS (or in situ) Page 35
36 OASIS3-MCT & Open-PALM: 2 open source code couplers Anthony Thevenin ([email protected])
Part I Courses Syllabus
Part I Courses Syllabus This document provides detailed information about the basic courses of the MHPC first part activities. The list of courses is the following 1.1 Scientific Programming Environment
The PRISM software framework and the OASIS coupler
The PRISM software framework and the OASIS coupler Sophie Valcke 1, Reinhard Budich 2, Mick Carter 3, Eric Guilyardi 4, Marie-Alice Foujols 5, Michael Lautenschlager 6, René Redler 7, Lois Steenman-Clark
EC-Earth: new global earth system model
EC-Earth: new global earth system model Wilco Hazeleger Vincent v Gogh Global Climate Division/EC-Earth program KNMI, The Netherlands Amsterdam, December 2008 1 Amsterdam, December 2008 2 Observed climate
BLM 413E - Parallel Programming Lecture 3
BLM 413E - Parallel Programming Lecture 3 FSMVU Bilgisayar Mühendisliği Öğr. Gör. Musa AYDIN 14.10.2015 2015-2016 M.A. 1 Parallel Programming Models Parallel Programming Models Overview There are several
Collaborative modelling and concurrent scientific data analysis:
Collaborative modelling and concurrent scientific data analysis: Application case in space plasma environment with the Keridwen/SPIS- GEO Integrated Modelling Environment B. Thiebault 1, J. Forest 2, B.
HPC Wales Skills Academy Course Catalogue 2015
HPC Wales Skills Academy Course Catalogue 2015 Overview The HPC Wales Skills Academy provides a variety of courses and workshops aimed at building skills in High Performance Computing (HPC). Our courses
Sourcery Overview & Virtual Machine Installation
Sourcery Overview & Virtual Machine Installation Damian Rouson, Ph.D., P.E. Sourcery, Inc. www.sourceryinstitute.org Sourcery, Inc. About Us Sourcery, Inc., is a software consultancy founded by and for
SCADA/HMI MOVICON TRAINING COURSE PROGRAM
SCADA/HMI MOVICON TRAINING COURSE PROGRAM The Movicon training program includes the following courses: Basic Training Course: 1 day course at Progea head offices or authorized center. On location at client
How To Visualize Performance Data In A Computer Program
Performance Visualization Tools 1 Performance Visualization Tools Lecture Outline : Following Topics will be discussed Characteristics of Performance Visualization technique Commercial and Public Domain
Debugging in Heterogeneous Environments with TotalView. ECMWF HPC Workshop 30 th October 2014
Debugging in Heterogeneous Environments with TotalView ECMWF HPC Workshop 30 th October 2014 Agenda Introduction Challenges TotalView overview Advanced features Current work and future plans 2014 Rogue
Equalizer. Parallel OpenGL Application Framework. Stefan Eilemann, Eyescale Software GmbH
Equalizer Parallel OpenGL Application Framework Stefan Eilemann, Eyescale Software GmbH Outline Overview High-Performance Visualization Equalizer Competitive Environment Equalizer Features Scalability
HPC Deployment of OpenFOAM in an Industrial Setting
HPC Deployment of OpenFOAM in an Industrial Setting Hrvoje Jasak [email protected] Wikki Ltd, United Kingdom PRACE Seminar: Industrial Usage of HPC Stockholm, Sweden, 28-29 March 2011 HPC Deployment
Schedule WRF model executions in parallel computing environments using Python
Schedule WRF model executions in parallel computing environments using Python A.M. Guerrero-Higueras, E. García-Ortega and J.L. Sánchez Atmospheric Physics Group, University of León, León, Spain J. Lorenzana
How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx
Software Development around a Millisecond
Introduction Software Development around a Millisecond Geoffrey Fox In this column we consider software development methodologies with some emphasis on those relevant for large scale scientific computing.
High Performance Computing. Course Notes 2007-2008. HPC Fundamentals
High Performance Computing Course Notes 2007-2008 2008 HPC Fundamentals Introduction What is High Performance Computing (HPC)? Difficult to define - it s a moving target. Later 1980s, a supercomputer performs
Distributed Dynamic Load Balancing for Iterative-Stencil Applications
Distributed Dynamic Load Balancing for Iterative-Stencil Applications G. Dethier 1, P. Marchot 2 and P.A. de Marneffe 1 1 EECS Department, University of Liege, Belgium 2 Chemical Engineering Department,
Parallel Programming at the Exascale Era: A Case Study on Parallelizing Matrix Assembly For Unstructured Meshes
Parallel Programming at the Exascale Era: A Case Study on Parallelizing Matrix Assembly For Unstructured Meshes Eric Petit, Loïc Thebault, Quang V. Dinh May 2014 EXA2CT Consortium 2 WPs Organization Proto-Applications
Data Centric Systems (DCS)
Data Centric Systems (DCS) Architecture and Solutions for High Performance Computing, Big Data and High Performance Analytics High Performance Computing with Data Centric Systems 1 Data Centric Systems
Introduction to grid technologies, parallel and cloud computing. Alaa Osama Allam Saida Saad Mohamed Mohamed Ibrahim Gaber
Introduction to grid technologies, parallel and cloud computing Alaa Osama Allam Saida Saad Mohamed Mohamed Ibrahim Gaber OUTLINES Grid Computing Parallel programming technologies (MPI- Open MP-Cuda )
Parallel Analysis and Visualization on Cray Compute Node Linux
Parallel Analysis and Visualization on Cray Compute Node Linux David Pugmire, Oak Ridge National Laboratory and Hank Childs, Lawrence Livermore National Laboratory and Sean Ahern, Oak Ridge National Laboratory
Performance Monitoring of Parallel Scientific Applications
Performance Monitoring of Parallel Scientific Applications Abstract. David Skinner National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory This paper introduces an infrastructure
Decomposition into Parts. Software Engineering, Lecture 4. Data and Function Cohesion. Allocation of Functions and Data. Component Interfaces
Software Engineering, Lecture 4 Decomposition into suitable parts Cross cutting concerns Design patterns I will also give an example scenario that you are supposed to analyse and make synthesis from The
The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC2013 - Denver
1 The PHI solution Fujitsu Industry Ready Intel XEON-PHI based solution SC2013 - Denver Industrial Application Challenges Most of existing scientific and technical applications Are written for legacy execution
Kodak Remote Support System - RSS VPN
Kodak Graphic Communications Canada Company 4225 Kincaid Street Burnaby, B.C., Canada V5G 4P5 Tel: +1.604.451.2700 Fax: +1.604.437.9891 Request support: partnerplace.kodak.com http://graphics.kodak.com
1 Bull, 2011 Bull Extreme Computing
1 Bull, 2011 Bull Extreme Computing Table of Contents HPC Overview. Cluster Overview. FLOPS. 2 Bull, 2011 Bull Extreme Computing HPC Overview Ares, Gerardo, HPC Team HPC concepts HPC: High Performance
Cluster, Grid, Cloud Concepts
Cluster, Grid, Cloud Concepts Kalaiselvan.K Contents Section 1: Cluster Section 2: Grid Section 3: Cloud Cluster An Overview Need for a Cluster Cluster categorizations A computer cluster is a group of
Four Keys to Successful Multicore Optimization for Machine Vision. White Paper
Four Keys to Successful Multicore Optimization for Machine Vision White Paper Optimizing a machine vision application for multicore PCs can be a complex process with unpredictable results. Developers need
End-user Tools for Application Performance Analysis Using Hardware Counters
1 End-user Tools for Application Performance Analysis Using Hardware Counters K. London, J. Dongarra, S. Moore, P. Mucci, K. Seymour, T. Spencer Abstract One purpose of the end-user tools described in
P013 INTRODUCING A NEW GENERATION OF RESERVOIR SIMULATION SOFTWARE
1 P013 INTRODUCING A NEW GENERATION OF RESERVOIR SIMULATION SOFTWARE JEAN-MARC GRATIEN, JEAN-FRANÇOIS MAGRAS, PHILIPPE QUANDALLE, OLIVIER RICOIS 1&4, av. Bois-Préau. 92852 Rueil Malmaison Cedex. France
Designing and Building Applications for Extreme Scale Systems CS598 William Gropp www.cs.illinois.edu/~wgropp
Designing and Building Applications for Extreme Scale Systems CS598 William Gropp www.cs.illinois.edu/~wgropp Welcome! Who am I? William (Bill) Gropp Professor of Computer Science One of the Creators of
GEDAE TM - A Graphical Programming and Autocode Generation Tool for Signal Processor Applications
GEDAE TM - A Graphical Programming and Autocode Generation Tool for Signal Processor Applications Harris Z. Zebrowitz Lockheed Martin Advanced Technology Laboratories 1 Federal Street Camden, NJ 08102
Advanced discretisation techniques (a collection of first and second order schemes); Innovative algorithms and robust solvers for fast convergence.
New generation CFD Software APUS-CFD APUS-CFD is a fully interactive Arbitrary Polyhedral Unstructured Solver. APUS-CFD is a new generation of CFD software for modelling fluid flow and heat transfer in
Mariana Vertenstein. NCAR Earth System Laboratory CESM Software Engineering Group (CSEG) NCAR is sponsored by the National Science Foundation
Leveraging the New CESM1 CPL7 Architecture - Current and Future Challenges Mariana Vertenstein NCAR Earth System Laboratory CESM Software Engineering Group (CSEG) NCAR is sponsored by the National Science
Working with HPC and HTC Apps. Abhinav Thota Research Technologies Indiana University
Working with HPC and HTC Apps Abhinav Thota Research Technologies Indiana University Outline What are HPC apps? Working with typical HPC apps Compilers - Optimizations and libraries Installation Modules
Chapter 3: Operating-System Structures. Common System Components
Chapter 3: Operating-System Structures System Components Operating System Services System Calls System Programs System Structure Virtual Machines System Design and Implementation System Generation 3.1
Climate-Weather Modeling Studies Using a Prototype Global Cloud-System Resolving Model
ANL/ALCF/ESP-13/1 Climate-Weather Modeling Studies Using a Prototype Global Cloud-System Resolving Model ALCF-2 Early Science Program Technical Report Argonne Leadership Computing Facility About Argonne
LOAD BALANCING DISTRIBUTED OPERATING SYSTEMS, SCALABILITY, SS 2015. Hermann Härtig
LOAD BALANCING DISTRIBUTED OPERATING SYSTEMS, SCALABILITY, SS 2015 Hermann Härtig ISSUES starting points independent Unix processes and block synchronous execution who does it load migration mechanism
Linux tools for debugging and profiling MPI codes
Competence in High Performance Computing Linux tools for debugging and profiling MPI codes Werner Krotz-Vogel, Pallas GmbH MRCCS September 02000 Pallas GmbH Hermülheimer Straße 10 D-50321
Building an Inexpensive Parallel Computer
Res. Lett. Inf. Math. Sci., (2000) 1, 113-118 Available online at http://www.massey.ac.nz/~wwiims/rlims/ Building an Inexpensive Parallel Computer Lutz Grosz and Andre Barczak I.I.M.S., Massey University
MayaVi: A free tool for CFD data visualization
MayaVi: A free tool for CFD data visualization Prabhu Ramachandran Graduate Student, Dept. Aerospace Engg. IIT Madras, Chennai, 600 036. e mail: [email protected] Keywords: Visualization, CFD data,
Mathematical Libraries on JUQUEEN. JSC Training Course
Mitglied der Helmholtz-Gemeinschaft Mathematical Libraries on JUQUEEN JSC Training Course May 10, 2012 Outline General Informations Sequential Libraries, planned Parallel Libraries and Application Systems:
Applications to Computational Financial and GPU Computing. May 16th. Dr. Daniel Egloff +41 44 520 01 17 +41 79 430 03 61
F# Applications to Computational Financial and GPU Computing May 16th Dr. Daniel Egloff +41 44 520 01 17 +41 79 430 03 61 Today! Why care about F#? Just another fashion?! Three success stories! How Alea.cuBase
Introduction to Embedded Systems. Software Update Problem
Introduction to Embedded Systems CS/ECE 6780/5780 Al Davis logistics minor Today s topics: more software development issues 1 CS 5780 Software Update Problem Lab machines work let us know if they don t
CICS Transactions Measurement with no Pain
CICS Transactions Measurement with no Pain Prepared by Luiz Eduardo Gazola 4bears - Optimize Software, Brazil December 6 10, 2010 Orlando, Florida USA This paper presents a new approach for measuring CICS
JOURNAL OF OBJECT TECHNOLOGY
JOURNAL OF OBJECT TECHNOLOGY Online at http://www.jot.fm. Published by ETH Zurich, Chair of Software Engineering JOT, 2006 Vol. 5, No. 6, July - August 2006 On Assuring Software Quality and Curbing Software
How To Understand The Concept Of A Distributed System
Distributed Operating Systems Introduction Ewa Niewiadomska-Szynkiewicz and Adam Kozakiewicz [email protected], [email protected] Institute of Control and Computation Engineering Warsaw University of
HPC enabling of OpenFOAM R for CFD applications
HPC enabling of OpenFOAM R for CFD applications Towards the exascale: OpenFOAM perspective Ivan Spisso 25-27 March 2015, Casalecchio di Reno, BOLOGNA. SuperComputing Applications and Innovation Department,
ParFUM: A Parallel Framework for Unstructured Meshes. Aaron Becker, Isaac Dooley, Terry Wilmarth, Sayantan Chakravorty Charm++ Workshop 2008
ParFUM: A Parallel Framework for Unstructured Meshes Aaron Becker, Isaac Dooley, Terry Wilmarth, Sayantan Chakravorty Charm++ Workshop 2008 What is ParFUM? A framework for writing parallel finite element
Structural Health Monitoring Tools (SHMTools)
Structural Health Monitoring Tools (SHMTools) Getting Started LANL/UCSD Engineering Institute LA-CC-14-046 c Copyright 2014, Los Alamos National Security, LLC All rights reserved. May 30, 2014 Contents
OPERATING SYSTEM SERVICES
OPERATING SYSTEM SERVICES USER INTERFACE Command line interface(cli):uses text commands and a method for entering them Batch interface(bi):commands and directives to control those commands are entered
Thermo Scientific Compound Discoverer Software. A New Generation. of integrated solutions for small molecule structure ID
Thermo Scientific Compound Discoverer Software A New Generation of integrated solutions for small molecule structure ID Transforming Small Molecule Research Solutions for Small Molecule Structural ID Thermo
The Service Revolution software engineering without programming languages
The Service Revolution software engineering without programming languages Gustavo Alonso Institute for Pervasive Computing Department of Computer Science Swiss Federal Institute of Technology (ETH Zurich)
System Structures. Services Interface Structure
System Structures Services Interface Structure Operating system services (1) Operating system services (2) Functions that are helpful to the user User interface Command line interpreter Batch interface
Elemental functions: Writing data-parallel code in C/C++ using Intel Cilk Plus
Elemental functions: Writing data-parallel code in C/C++ using Intel Cilk Plus A simple C/C++ language extension construct for data parallel operations Robert Geva [email protected] Introduction Intel
CHAPTER 1: OPERATING SYSTEM FUNDAMENTALS
CHAPTER 1: OPERATING SYSTEM FUNDAMENTALS What is an operating? A collection of software modules to assist programmers in enhancing efficiency, flexibility, and robustness An Extended Machine from the users
BSC vision on Big Data and extreme scale computing
BSC vision on Big Data and extreme scale computing Jesus Labarta, Eduard Ayguade,, Fabrizio Gagliardi, Rosa M. Badia, Toni Cortes, Jordi Torres, Adrian Cristal, Osman Unsal, David Carrera, Yolanda Becerra,
EVOLUTION AND INDUSTRIALIZATION OF A SBAS REAL-TIME PERFORMANCE MONITORING TOOL (EVORA)
ENC 2015 Page 1 EVOLUTION AND INDUSTRIALIZATION OF A SBAS REAL-TIME PERFORMANCE MONITORING TOOL (EVORA) J. Doubek 1, M. Houdek 1, C. A. Pandele 2,3, A. Polivka 1, D. Chung 1 1 Iguassu Software Systems,
Building Technologies
Building Technologies : simple and secure danger management. Integration increases security! The increasing danger potential for people and assets makes the need for better protection increasingly more
FD4: A Framework for Highly Scalable Dynamic Load Balancing and Model Coupling
Center for Information Services and High Performance Computing (ZIH) FD4: A Framework for Highly Scalable Dynamic Load Balancing and Model Coupling Symposium on HPC and Data-Intensive Applications in Earth
Application Note: AN00141 xcore-xa - Application Development
Application Note: AN00141 xcore-xa - Application Development This application note shows how to create a simple example which targets the XMOS xcore-xa device and demonstrates how to build and run this
Chapter 18: Database System Architectures. Centralized Systems
Chapter 18: Database System Architectures! Centralized Systems! Client--Server Systems! Parallel Systems! Distributed Systems! Network Types 18.1 Centralized Systems! Run on a single computer system and
NetFlow Feature Acceleration
WHITE PAPER NetFlow Feature Acceleration Feature Description Rapid growth in Internet and intranet deployment and usage has created a major shift in both corporate and consumer computing paradigms. This
Using WestGrid. Patrick Mann, Manager, Technical Operations Jan.15, 2014
Using WestGrid Patrick Mann, Manager, Technical Operations Jan.15, 2014 Winter 2014 Seminar Series Date Speaker Topic 5 February Gino DiLabio Molecular Modelling Using HPC and Gaussian 26 February Jonathan
SCADAPack E ISaGRAF 3 User Manual
SCADAPack E ISaGRAF 3 User Manual 2 SCADAPack E ISaGRAF 3 User Manual Table of Contents Part I ISaGRAF 3 User Manual 3 1 Technical... Support 3 2 Safety... Information 4 3 Preface... 6 4 Overview... 8
INTEL PARALLEL STUDIO XE EVALUATION GUIDE
Introduction This guide will illustrate how you use Intel Parallel Studio XE to find the hotspots (areas that are taking a lot of time) in your application and then recompiling those parts to improve overall
Highly Scalable Dynamic Load Balancing in the Atmospheric Modeling System COSMO-SPECS+FD4
Center for Information Services and High Performance Computing (ZIH) Highly Scalable Dynamic Load Balancing in the Atmospheric Modeling System COSMO-SPECS+FD4 PARA 2010, June 9, Reykjavík, Iceland Matthias
Virtual Machines. www.viplavkambli.com
1 Virtual Machines A virtual machine (VM) is a "completely isolated guest operating system installation within a normal host operating system". Modern virtual machines are implemented with either software
Types Of Operating Systems
Types Of Operating Systems Date 10/01/2004 1/24/2004 Operating Systems 1 Brief history of OS design In the beginning OSes were runtime libraries The OS was just code you linked with your program and loaded
Full and Para Virtualization
Full and Para Virtualization Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF x86 Hardware Virtualization The x86 architecture offers four levels
So#ware Tools and Techniques for HPC, Clouds, and Server- Class SoCs Ron Brightwell
So#ware Tools and Techniques for HPC, Clouds, and Server- Class SoCs Ron Brightwell R&D Manager, Scalable System So#ware Department Sandia National Laboratories is a multi-program laboratory managed and
Eastern Washington University Department of Computer Science. Questionnaire for Prospective Masters in Computer Science Students
Eastern Washington University Department of Computer Science Questionnaire for Prospective Masters in Computer Science Students I. Personal Information Name: Last First M.I. Mailing Address: Permanent
Amazon EC2 Product Details Page 1 of 5
Amazon EC2 Product Details Page 1 of 5 Amazon EC2 Functionality Amazon EC2 presents a true virtual computing environment, allowing you to use web service interfaces to launch instances with a variety of
SCALABILITY OF CONTEXTUAL GENERALIZATION PROCESSING USING PARTITIONING AND PARALLELIZATION. Marc-Olivier Briat, Jean-Luc Monnot, Edith M.
SCALABILITY OF CONTEXTUAL GENERALIZATION PROCESSING USING PARTITIONING AND PARALLELIZATION Abstract Marc-Olivier Briat, Jean-Luc Monnot, Edith M. Punt Esri, Redlands, California, USA [email protected], [email protected],
Turbomachinery CFD on many-core platforms experiences and strategies
Turbomachinery CFD on many-core platforms experiences and strategies Graham Pullan Whittle Laboratory, Department of Engineering, University of Cambridge MUSAF Colloquium, CERFACS, Toulouse September 27-29
Centralized Systems. A Centralized Computer System. Chapter 18: Database System Architectures
Chapter 18: Database System Architectures Centralized Systems! Centralized Systems! Client--Server Systems! Parallel Systems! Distributed Systems! Network Types! Run on a single computer system and do
Deploying Clusters at Electricité de France. Jean-Yves Berthou
Electricit é Deploying Clusters at Workshop Operating Systems, Tools and Methods for High Performance Computing on Linux Clusters Jean-Yves Berthou Head of the Applied Scientific Computing Group EDF R&D
CISCO CONTENT SWITCHING MODULE SOFTWARE VERSION 4.1(1) FOR THE CISCO CATALYST 6500 SERIES SWITCH AND CISCO 7600 SERIES ROUTER
PRODUCT BULLETIN NO. 2438 CISCO CONTENT SWITCHING MODULE SOFTWARE VERSION 4.1(1) FOR THE CISCO CATALYST 6500 SERIES SWITCH AND CISCO 7600 SERIES ROUTER NEW FEATURES New features of the Cisco Content Switching
Introduction to Cloud Computing
Introduction to Cloud Computing Parallel Processing I 15 319, spring 2010 7 th Lecture, Feb 2 nd Majd F. Sakr Lecture Motivation Concurrency and why? Different flavors of parallel computing Get the basic
Client/Server Computing Distributed Processing, Client/Server, and Clusters
Client/Server Computing Distributed Processing, Client/Server, and Clusters Chapter 13 Client machines are generally single-user PCs or workstations that provide a highly userfriendly interface to the
Building Platform as a Service for Scientific Applications
Building Platform as a Service for Scientific Applications Moustafa AbdelBaky [email protected] Rutgers Discovery Informa=cs Ins=tute (RDI 2 ) The NSF Cloud and Autonomic Compu=ng Center Department
Notes and terms of conditions. Vendor shall note the following terms and conditions/ information before they submit their quote.
Specifications for ARINC 653 compliant RTOS & Development Environment Notes and terms of conditions Vendor shall note the following terms and conditions/ information before they submit their quote. 1.
Tools Page 1 of 13 ON PROGRAM TRANSLATION. A priori, we have two translation mechanisms available:
Tools Page 1 of 13 ON PROGRAM TRANSLATION A priori, we have two translation mechanisms available: Interpretation Compilation On interpretation: Statements are translated one at a time and executed immediately.
Origins of Operating Systems OS/360. Martin Grund HPI
Origins of Operating Systems OS/360 HPI Table of Contents IBM System 360 Functional Structure of OS/360 Virtual Machine Time Sharing 2 Welcome to Big Blue 3 IBM System 360 In 1964 IBM announced the IBM-360
Simplest Scalable Architecture
Simplest Scalable Architecture NOW Network Of Workstations Many types of Clusters (form HP s Dr. Bruce J. Walker) High Performance Clusters Beowulf; 1000 nodes; parallel programs; MPI Load-leveling Clusters
Objectives. Chapter 2: Operating-System Structures. Operating System Services (Cont.) Operating System Services. Operating System Services (Cont.
Objectives To describe the services an operating system provides to users, processes, and other systems To discuss the various ways of structuring an operating system Chapter 2: Operating-System Structures
Voice Internet Phone Gateway
Voice Internet Phone Gateway Quick Installation Guide IPC 1000 Series ARTDio Company Inc. Edition 1.0 Note: For more detailed hardware installation instructions, please refer to the IPC 1000 series User
Operating Systems 4 th Class
Operating Systems 4 th Class Lecture 1 Operating Systems Operating systems are essential part of any computer system. Therefore, a course in operating systems is an essential part of any computer science
Recent Advances in Periscope for Performance Analysis and Tuning
Recent Advances in Periscope for Performance Analysis and Tuning Isaias Compres, Michael Firbach, Michael Gerndt Robert Mijakovic, Yury Oleynik, Ventsislav Petkov Technische Universität München Yury Oleynik,
Last Class: OS and Computer Architecture. Last Class: OS and Computer Architecture
Last Class: OS and Computer Architecture System bus Network card CPU, memory, I/O devices, network card, system bus Lecture 3, page 1 Last Class: OS and Computer Architecture OS Service Protection Interrupts
COMP5426 Parallel and Distributed Computing. Distributed Systems: Client/Server and Clusters
COMP5426 Parallel and Distributed Computing Distributed Systems: Client/Server and Clusters Client/Server Computing Client Client machines are generally single-user workstations providing a user-friendly
