High Performance Computing at CEA
|
|
|
- Norman Blair
- 10 years ago
- Views:
Transcription
1 High Performance Computing at CEA Thierry Massard CEA-DAM Mascot 2012 Meeting 23/03/
2 www-hpc.cea.fr/en Paris CEA/DIF (Bruyères-Le-Châtel) Saclay CEA Supercomputing Complex Mascot 2012 Meeting 23/03/2012 2
3 Brief Presentation of CEA o CEA, namely Atomic and Alternative Energies Commission, is a French governmental research organization o employees o doctoral students o 3.9 B /year budget o scientific publications/year o 300+ European projects with CEA participation o Main focuses of CEA are: o Defense and global security o Energy o Information & Health technologies o Related and necessary fundamental research o Other outcomes o Teaching and knowledge dissemination o Technology transfer (patents, start-up/spin-off companies ) Mascot 2012 Meeting 23/03/2012 3
4 HPC a huge need for CEA Basic science and applied research for the Simulation program Applied research, industrial applications in the field of Energy and Health and Communication technology Fundamental research (Matter and Biology) ocomité Stratégique du Calcul Intensif 4/15 o22 octobre 2007 Mascot 2012 Meeting 23/03/2012 4
5 Fundamental Sciences of Matter Nanosciences Particles Physics Environmental risks tsunamis Global warming Fusion Structure of the Universe Mascot 2012 Meeting 23/03/2012 5
6 CEA Scientific Computing Complex Defense Computing Center TERA R&D on HPC Technologies Très Grand Centre de calcul du CEA TGCC CCRT Address the needs of CEA, French & European Research Communities and Industry in terms of High Performance Simulations While developing and assessing news technologies for more powerful and efficient HPC systems Joint Laboratories Extreme Computing Lab (BULL, CEA) Exascale Computing Research (INTEL, GENCI, CEA, UVSQ) Mascot 2012 Meeting 23/03/
7 Compute Complex Capacity Roadmap TERA TERA Tflops 100 TERA-10 TGCC 10 TERA-1 CCRT-B 1 CCRT-A 0, Mascot 2012 Meeting 23/03/
8 TERA Computing Centre Classified computing resources for the CEA/DAM Simulation Program Mascot 2012 Meeting 23/03/
9 TERA st european supercomputer in 2010 above the sustained Petaflops 1,25 Petaflop/s peak nodes coresintel Xeon Nehalem EX QDR Infiniband interconnect System software Open source 300 TB memory 20 PB storage disk 500 GB/s bandwidth toward the storage system High reliability machine to meet the strategic needs of CEA/DAM Mascot 2012 Meeting 23/03/2012 9
10 Grand Challenges on TERA 100 o We have taken advantage of the pre-production phase to «invite» a selection of frontier extreme computing cases : Part of the V&V process for the global system Reach the limit of the machine To observe the behavior of other software than those in CEA-DAM s portfolio o Sampling from the nov 2010 feb.2011 campaign : Massive parallellcalculation for the 3D transport of a sodium fast reactor using APOLLO3 R. Baron, C. Calvin, J. Dubois, J-J. Lautard, S. Salmons CEA/DEN/DANS/DM2S/SERMA up to cores Simulation of graphene growth on Silicon Carbibe (BigDFT) Thierry Deutsch, Luigi Genovese, Pascal Pochet, Eduardo Machado-Charry CEA/DSM/INAC Grenoble Classical Molecular Dynamics Simulations (STAMP) L. Soulard et al., CEA/DAM/DPTA up to cœurs Design of artiificial enzymes (STAMPS) G. Debret, M. Masella, P. Cuniasse, G. Collet CEA/DSV/iBiTecS Mascot 2012 Meeting 23/03/
11 Mascot 2012 Meeting 23/03/
12 Mascot 2012 Meeting 23/03/
13 Simulations de dynamique moléculaire TERA-100 La dynamique moléculaire résout l équation fondamentale de la dynamique pour un ensemble de N particules en interaction mutuelle. Pourvu que N soit suffisamment grand, elle peut permettre de faire le lien entre le calcul quantique et la mécanique des milieux continus. Performances du code Stamp sur TERA-100 entre 250 millions et un milliard d atomes, et jusqu à processeurs. Ejection de matière lors de la réflexion d un choc sur une surface libre. 125 millions d atomes, 4000 cœurs, potentiel interatomique assez simple mais temps physique de simulation très long (pour la dynamique moléculaire) : de l ordre de la nanoseconde. Plastification du diamant au passage d une onde de choc. Relativement peu d atomes (1,3 millions, 6400 cœurs) mais potentiel interatomique extrêmement complexe. Mascot 2012 Meeting 23/03/
14 TERA-100: Co-designed by BULL and CEA/DAM TERA-100 has been the first European computer to reach the Petaflops. General Purpose production system with a tremendous effectiveness and reliability (1.05 Pflops, 83,7% of the peak performance in a 22 hours run) Tera100 key figures : 220 racks in 650 m² 4370 servers interconnected with IB-QDR Intel X7560 processors compute cores 300 TB of memory 500 GB/s I/O bandwidth System software mostly based on Open source developments (Linux, LUSTRE, SLURM..) Mascot 2012 Meeting 23/03/
15 Supercomputers... A roadmap to Exaflops Joint Bull/CEA R&D to insure availability of supercomputers for the needs of the Simulation program EXA 1 N? Monde N? EU 1 Ef 2019 Ambitious goal : to maintain a european capabillity to design and build Supercomputers in Europe x30 TERA 1000 High ROI in technology and economy for the civilian applications x30 N? Monde N? EU 30Pf 2001 Closed DAM machines TERA N 5 monde N 1 EU 4,8 TF x10 TERA N 5 monde N 1 EU 63 TF x TERA 100 N 6 monde N 1 EU 1Pf TGCC Open CEA computing centers Mascot 2012 Meeting 23/03/2012 Direction 15 des
16 CCRT Computing Centre for Research and Technology A partnership approach to gather academic research and industrials high end simulation in the same computing centre Mascot 2012 Meeting 23/03/
17 Computing Center for Research and Technology o A successful business model since 2003 o CCRT major goals o Meet CEA & partners simulation needs for research o Synergy between research organizations, universities and industry Promoting exchanges and scientific collaborations between partners Usages of HPC, code scaling and optimization Centralized application support, workshops o Industrial partners o Multi-year agreements for a share based on total cost of ownership (sharing risk and opportunities) Mascot 2012 Meeting 23/03/
18 CCRT Partnership Overview (2011) ONERA 1,0% Techspace Aero 0,8% Snecma 8,9% INERIS* 1.5% Areva* Turbomeca 3% 8% Astrium 1,0% Cerfacs 0,7% DAM 12,2% DEN 17,2% EDF 22% DSV 5,5% DSM 18,2% Mascot 2012 Meeting 23/03/
19 Current CCRT Architecture (~200 Tflops) NEC Vector system Hybride (CPU/GPU) Bull Cluster BULL Itanium Cluster Extension CCRT Tflops 47.7 Tflops Mémoire 23.2 To Disques 420 To 6 * 10 Gb/s HP Post-processing Cluster 8 * 1Gb/s SX8 SX9 8 nodes 3 nodes 2 Tflops 4.8 Tflops 0.9 To 3 To 40 TB of Disk 8 * 10 Gb/s 8 * 10 Gb/s nœuds Bull 8736 Intel/Nehalem cores 3 GB/core 103 TFlops SGI Altix 450 DMF Level 1: 1 Po of DDN Sata disk 500 TB of disk 48serveurs Tesla S NVIDIA cores 4 GB/GPU 192 Tflops-SP 38 graphic nodes 3.2 TB of memory, 100 TB of disk Backbone Ethernet 10Gb/s Users 2 Gb/s NFS Server Level 2: SUN-SL8500 (9940 & T10000 tapes) Storage System Mascot 2012 Meeting 23/03/
20 TGCC Très Grand Centre de calcul du CEA A completely new computing centre built by CEA to host the French contribution to the PRACE Research Infrastructure and other high end HPC resources Mascot 2012 Meeting 23/03/
21 TGCC Overview Designed to host : o2nd system of the PRACE Research Infrastructure (European Tier0) o3rd generation of the CCRT (French - Tier1) ohigh speed network node (10+ GIGE links) Main characteristics of the infrastructure onew power line 225 kv / 60 MW oinitially configured for ~12 MW (7.5 MW of IT equipments) o2600 m² machine rooms floor space o Conference room for 200 people ooperated by CEA/DAM-Ile-de-France site teams Mascot 2012 Meeting 23/03/
22 Très Grand Centre de Calcul du CEA «in one glance» Mascot 2012 Meeting 23/03/
23 The European strategy for HPC Gouvernance européenne 21 pays membres dont la France représentée par GENCI Tier-0 European Centres Tier-1 National Centres > 1 Pflop/s Gouvernance nationale Tier-2 Regional Centres (Universities) < 50 Tflop/s Gouvernances régionales Equip@meso Mascot 2012 Meeting 23/03/
24 PRACE supercomputers JUGENE : IBM BG/P 1 PF GENCI@TGCC CURIE : BULL Bullx, 2 PF GSC@Juelich & CINECA IBM BG/Q systems (Fermi) GSC@HLRS Hermit : Cray XE6, 1 PF Mascot 2012 Meeting 23/03/2012 GSC@LRZ SuperMUC : IBM idataplex, 3 PF BSC IBM 24
25 Mai JUGENE : 360Mh PRACE : calls in progress Novembre 1st regular call JUGENE : 360Mh CURIE : 40Mh (nœuds larges) Mai Novembre Mai Novembre 2nd regular call JUGENE : 360 Mh CURIE : 30 Mh (NL), 125 Mh (NF) HERMIT : 160 Mh 3rd regular call JUGENE : 33 % PRACE CURIE : 80 % PRACE HERMIT : 33 % PRACE SUPERMUC : 50% PRACE FERMI :? MARENOSTRUM :? 1 er nov à mi janvier JUGENE : 300 Mh, HERMIT : 160 Mh, SuperMuc (LRZ) : 200Mh CURIE : 28Mh (NL), 188 Mh (NF), Marenostrum (BSC) : 135 Mh, FERMI (CINECA) : 300 Mh 4 th regular call Preparatory Access Call Appel à projets Mascot 2012 Meeting 23/03/2012 Allocation JUGENE : 10 Mh/an - CURIE : 5 Mh/an (NL), 20 Mh/an (NF) et 1.4 Mh/an (230kh en 2011) (NH) - HERMIT (tbd) Appel continu relevé tous les 3 mois 25
26 Curie: The BULL Tier0 System Mascot 2012 Meeting 23/03/
27 Curie configuration Fat nodes 360 BullX S6010 Intel NH EX 2,26 GHz cœurs 32 =>128 core/node 128 GB => 512 GB/node 4 GB/core 105 TFlops Thin nodes 5040 BullX B510 Intel Sandy Bridge 2,7 GHz cœurs 16 core/node 4 GB/core TFlops Hybrid nodes 144 Bullx B Nvidia M TFlops Interconnect Infiniband QDR 5 Po GB/s Lustre file system Mascot 2012 Meeting 23/03/
28 CEA Supercomputing Complex Summary A single operator mutualizing expertise developed at CEA/DAM A rich ecosystem connected from regional to worldwide scale Collaborations with academics and industry ranging from design of HPC system and their environment (I/O, storage..) to applications and usages through system and management software, algorithms and middleware TERATEC TGCC TERA CCRT Mascot 2012 Meeting 23/03/
29 ENJOY YOUR STAY AT TGCC! Mascot 2012 Meeting 23/03/
How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx
Supercomputing Resources in BSC, RES and PRACE
www.bsc.es Supercomputing Resources in BSC, RES and PRACE Sergi Girona, BSC-CNS Barcelona, 23 Septiembre 2015 ICTS 2014, un paso adelante para la RES Past RES members and resources BSC-CNS (MareNostrum)
Cosmological simulations on High Performance Computers
Cosmological simulations on High Performance Computers Cosmic Web Morphology and Topology Cosmological workshop meeting Warsaw, 12-17 July 2011 Maciej Cytowski Interdisciplinary Centre for Mathematical
Kriterien für ein PetaFlop System
Kriterien für ein PetaFlop System Rainer Keller, HLRS :: :: :: Context: Organizational HLRS is one of the three national supercomputing centers in Germany. The national supercomputing centers are working
PRACE hardware, software and services. David Henty, EPCC, [email protected]
PRACE hardware, software and services David Henty, EPCC, [email protected] Why? Weather, Climatology, Earth Science degree of warming, scenarios for our future climate. understand and predict ocean
Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales
Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes Anthony Kenisky, VP of North America Sales About Appro Over 20 Years of Experience 1991 2000 OEM Server Manufacturer 2001-2007
Interconnect Your Future Enabling the Best Datacenter Return on Investment. TOP500 Supercomputers, November 2015
Interconnect Your Future Enabling the Best Datacenter Return on Investment TOP500 Supercomputers, November 2015 InfiniBand FDR and EDR Continue Growth and Leadership The Most Used Interconnect On The TOP500
HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect [email protected]
HPC and Big Data EPCC The University of Edinburgh Adrian Jackson Technical Architect [email protected] EPCC Facilities Technology Transfer European Projects HPC Research Visitor Programmes Training
PRACE the European HPC Research Infrastructure. Carlos Mérida-Campos, Advisor of Spanish Member at PRACE Council
PRACE the European HPC Research Infrastructure Carlos Mérida-Campos, Advisor of Spanish Member at PRACE Council Barcelona, 6-June-2013 PRACE an European e-infrastructure & ESFRI-list item in operation
Extreme Scaling on Energy Efficient SuperMUC
Extreme Scaling on Energy Efficient SuperMUC Dieter Kranzlmüller Munich Network Management Team Ludwig- Maximilians- Universität München (LMU) & Leibniz SupercompuFng Centre (LRZ) of the Bavarian Academy
BSC - Barcelona Supercomputer Center
Objectives Research in Supercomputing and Computer Architecture Collaborate in R&D e-science projects with prestigious scientific teams Manage BSC supercomputers to accelerate relevant contributions to
Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca
Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca Carlo Cavazzoni CINECA Supercomputing Application & Innovation www.cineca.it 21 Aprile 2015 FERMI Name: Fermi Architecture: BlueGene/Q
Welcome to the. Jülich Supercomputing Centre. D. Rohe and N. Attig Jülich Supercomputing Centre (JSC), Forschungszentrum Jülich
Mitglied der Helmholtz-Gemeinschaft Welcome to the Jülich Supercomputing Centre D. Rohe and N. Attig Jülich Supercomputing Centre (JSC), Forschungszentrum Jülich Schedule: Monday, May 19 13:00-13:30 Welcome
HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK
HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK Steve Oberlin CTO, Accelerated Computing US to Build Two Flagship Supercomputers SUMMIT SIERRA Partnership for Science 100-300 PFLOPS Peak Performance
Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC
HPC Architecture End to End Alexandre Chauvin Agenda HPC Software Stack Visualization National Scientific Center 2 Agenda HPC Software Stack Alexandre Chauvin Typical HPC Software Stack Externes LAN Typical
Extreme Computing: The Bull Way
Architect of an Open World Extreme Computing: The Bull Way Dr.-Ing. Joachim Redmer, Director HPC ( [email protected] ) Bull today Bull is an Information Technology company, focusing on open and secure systems
HPC Update: Engagement Model
HPC Update: Engagement Model MIKE VILDIBILL Director, Strategic Engagements Sun Microsystems [email protected] Our Strategy Building a Comprehensive HPC Portfolio that Delivers Differentiated Customer Value
Barry Bolding, Ph.D. VP, Cray Product Division
Barry Bolding, Ph.D. VP, Cray Product Division 1 Corporate Overview Trends in Supercomputing Types of Supercomputing and Cray s Approach The Cloud The Exascale Challenge Conclusion 2 Slide 3 Seymour Cray
Mississippi State University High Performance Computing Collaboratory Brief Overview. Trey Breckenridge Director, HPC
Mississippi State University High Performance Computing Collaboratory Brief Overview Trey Breckenridge Director, HPC Mississippi State University Public university (Land Grant) founded in 1878 Traditional
International High Performance Computing. Troels Haugbølle Centre for Star and Planet Formation Niels Bohr Institute PRACE User Forum
International High Performance Computing Troels Haugbølle Centre for Star and Planet Formation Niels Bohr Institute PRACE User Forum Why International HPC? Large-scale science projects can require resources
Journée Mésochallenges 2015 SysFera and ROMEO Make Large-Scale CFD Simulations Only 3 Clicks Away
SysFera and ROMEO Make Large-Scale CFD Simulations Only 3 Clicks Away Benjamin Depardon SysFera Sydney Tekam Tech-Am ING Arnaud Renard ROMEO Manufacturing with HPC 98% of products will be developed digitally
Extreme Scale Compu0ng at LRZ
Extreme Scale Compu0ng at LRZ Dieter Kranzlmüller Munich Network Management Team Ludwig- Maximilians- Universität München (LMU) & Leibniz SupercompuFng Centre (LRZ) of the Bavarian Academy of Sciences
Transforming the UL into a Big Data University. Current status and planned evolutions
Transforming the UL into a Big Data University Current status and planned evolutions December 6th, 2013 UniGR Workshop - Big Data Sébastien Varrette, PhD Prof. Pascal Bouvry Prof. Volker Müller http://hpc.uni.lu
Jean-Pierre Panziera Teratec 2011
Technologies for the future HPC systems Jean-Pierre Panziera Teratec 2011 3 petaflop systems : TERA 100, CURIE & IFERC Tera100 Curie IFERC 1.25 PetaFlops 256 TB ory 30 PB disk storage 140 000+ Xeon cores
OpenMP Programming on ScaleMP
OpenMP Programming on ScaleMP Dirk Schmidl [email protected] Rechen- und Kommunikationszentrum (RZ) MPI vs. OpenMP MPI distributed address space explicit message passing typically code redesign
Sun Constellation System: The Open Petascale Computing Architecture
CAS2K7 13 September, 2007 Sun Constellation System: The Open Petascale Computing Architecture John Fragalla Senior HPC Technical Specialist Global Systems Practice Sun Microsystems, Inc. 25 Years of Technical
Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar
Computational infrastructure for NGS data analysis José Carbonell Caballero Pablo Escobar Computational infrastructure for NGS Cluster definition: A computer cluster is a group of linked computers, working
Hadoop on the Gordon Data Intensive Cluster
Hadoop on the Gordon Data Intensive Cluster Amit Majumdar, Scientific Computing Applications Mahidhar Tatineni, HPC User Services San Diego Supercomputer Center University of California San Diego Dec 18,
YALES2 porting on the Xeon- Phi Early results
YALES2 porting on the Xeon- Phi Early results Othman Bouizi Ghislain Lartigue Innovation and Pathfinding Architecture Group in Europe, Exascale Lab. Paris CRIHAN - Demi-journée calcul intensif, 16 juin
How Cineca supports IT
How Cineca supports IT Topics CINECA: an overview Systems and Services for Higher Education HPC for Research Activities and Industries Cineca: the Consortium Not For Profit Founded in 1969 HPC FERMI: TOP500
1 DCSC/AU: HUGE. DeIC Sekretariat 2013-03-12/RB. Bilag 1. DeIC (DCSC) Scientific Computing Installations
Bilag 1 2013-03-12/RB DeIC (DCSC) Scientific Computing Installations DeIC, previously DCSC, currently has a number of scientific computing installations, distributed at five regional operating centres.
HPC-related R&D in 863 Program
HPC-related R&D in 863 Program Depei Qian Sino-German Joint Software Institute (JSI) Beihang University Aug. 27, 2010 Outline The 863 key project on HPC and Grid Status and Next 5 years 863 efforts on
Jezelf Groen Rekenen met Supercomputers
Jezelf Groen Rekenen met Supercomputers Symposium Groene ICT en duurzaamheid: Nieuwe energie in het hoger onderwijs Walter Lioen Groepsleider Supercomputing About SURFsara SURFsara
CALMIP a Computing Meso-center for middle HPC academis Users
CALMIP a Computing Meso-center for middle HPC academis Users CALMIP : Calcul en Midi-Pyrénées Méso-centre de Calcul Boris Dintrans, President Operational Board CALMIP, Laboratoire d Astrophysique de Toulouse
Stovepipes to Clouds. Rick Reid Principal Engineer SGI Federal. 2013 by SGI Federal. Published by The Aerospace Corporation with permission.
Stovepipes to Clouds Rick Reid Principal Engineer SGI Federal 2013 by SGI Federal. Published by The Aerospace Corporation with permission. Agenda Stovepipe Characteristics Why we Built Stovepipes Cluster
JuRoPA. Jülich Research on Petaflop Architecture. One Year on. Hugo R. Falter, COO Lee J Porter, Engineering
JuRoPA Jülich Research on Petaflop Architecture One Year on Hugo R. Falter, COO Lee J Porter, Engineering HPC Advisoy Counsil, Workshop 2010, Lugano 1 Outline The work of ParTec on JuRoPA (HF) Overview
Accelerating Simulation & Analysis with Hybrid GPU Parallelization and Cloud Computing
Accelerating Simulation & Analysis with Hybrid GPU Parallelization and Cloud Computing Innovation Intelligence Devin Jensen August 2012 Altair Knows HPC Altair is the only company that: makes HPC tools
GPU System Architecture. Alan Gray EPCC The University of Edinburgh
GPU System Architecture EPCC The University of Edinburgh Outline Why do we want/need accelerators such as GPUs? GPU-CPU comparison Architectural reasons for GPU performance advantages GPU accelerated systems
Deploying and managing a Visualization Farm @ Onera
Deploying and managing a Visualization Farm @ Onera Onera Scientific Day - October, 3 2012 Network and computing department (DRI), Onera P.F. Berte [email protected] Plan Onera global HPC
Linux Cluster Computing An Administrator s Perspective
Linux Cluster Computing An Administrator s Perspective Robert Whitinger Traques LLC and High Performance Computing Center East Tennessee State University : http://lxer.com/pub/self2015_clusters.pdf 2015-Jun-14
Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007
Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms Cray User Group Meeting June 2007 Cray s Storage Strategy Background Broad range of HPC requirements
A National Computing Grid: FGI
A National Computing Grid: FGI Vera Hansper, Ulf Tigerstedt, Kimmo Mattila, Luis Alves 3/10/2012 FGI Grids in Finland : a short history 3/10/2012 FGI In the beginning, we had M-Grid Interest in Grid technology
Relations with ISV and Open Source. Stephane Requena GENCI [email protected]
Relations with ISV and Open Source Stephane Requena GENCI [email protected] Agenda of this session 09:15 09:30 Prof. Hrvoje Jasak: Director, Wikki Ltd. «HPC Deployment of OpenFOAM in an Industrial
MEGWARE HPC Cluster am LRZ eine mehr als 12-jährige Zusammenarbeit. Prof. Dieter Kranzlmüller (LRZ)
MEGWARE HPC Cluster am LRZ eine mehr als 12-jährige Zusammenarbeit Prof. Dieter Kranzlmüller (LRZ) LRZ HPC-Systems at the End of the UNIX-Era (Years 2000-2002) German national supercomputer Hitachi SR800
Introduction History Design Blue Gene/Q Job Scheduler Filesystem Power usage Performance Summary Sequoia is a petascale Blue Gene/Q supercomputer Being constructed by IBM for the National Nuclear Security
SGI High Performance Computing
SGI High Performance Computing Accelerate time to discovery, innovation, and profitability 2014 SGI SGI Company Proprietary 1 Typical Use Cases for SGI HPC Products Large scale-out, distributed memory
How To Compare Amazon Ec2 To A Supercomputer For Scientific Applications
Amazon Cloud Performance Compared David Adams Amazon EC2 performance comparison How does EC2 compare to traditional supercomputer for scientific applications? "Performance Analysis of High Performance
Estonian Scientific Computing Infrastructure (ETAIS)
Estonian Scientific Computing Infrastructure (ETAIS) Week #7 Hardi Teder [email protected] University of Tartu March 27th 2013 Overview Estonian Scientific Computing Infrastructure Estonian Research infrastructures
Cluster Implementation and Management; Scheduling
Cluster Implementation and Management; Scheduling CPS343 Parallel and High Performance Computing Spring 2013 CPS343 (Parallel and HPC) Cluster Implementation and Management; Scheduling Spring 2013 1 /
Parallel Programming Survey
Christian Terboven 02.09.2014 / Aachen, Germany Stand: 26.08.2014 Version 2.3 IT Center der RWTH Aachen University Agenda Overview: Processor Microarchitecture Shared-Memory
New Storage System Solutions
New Storage System Solutions Craig Prescott Research Computing May 2, 2013 Outline } Existing storage systems } Requirements and Solutions } Lustre } /scratch/lfs } Questions? Existing Storage Systems
1 Bull, 2011 Bull Extreme Computing
1 Bull, 2011 Bull Extreme Computing Table of Contents HPC Overview. Cluster Overview. FLOPS. 2 Bull, 2011 Bull Extreme Computing HPC Overview Ares, Gerardo, HPC Team HPC concepts HPC: High Performance
PRIMERGY server-based High Performance Computing solutions
PRIMERGY server-based High Performance Computing solutions PreSales - May 2010 - HPC Revenue OS & Processor Type Increasing standardization with shift in HPC to x86 with 70% in 2008.. HPC revenue by operating
Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance
Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance Hybrid Storage Performance Gains for IOPS and Bandwidth Utilizing Colfax Servers and Enmotus FuzeDrive Software NVMe Hybrid
LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu
Current Status of FEFS for the K computer
Current Status of FEFS for the K computer Shinji Sumimoto Fujitsu Limited Apr.24 2012 LUG2012@Austin Outline RIKEN and Fujitsu are jointly developing the K computer * Development continues with system
Altix Usage and Application Programming. Welcome and Introduction
Zentrum für Informationsdienste und Hochleistungsrechnen Altix Usage and Application Programming Welcome and Introduction Zellescher Weg 12 Tel. +49 351-463 - 35450 Dresden, November 30th 2005 Wolfgang
JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert
Mitglied der Helmholtz-Gemeinschaft JUROPA Linux Cluster An Overview 19 May 2014 Ulrich Detert JuRoPA JuRoPA Jülich Research on Petaflop Architectures Bull, Sun, ParTec, Intel, Mellanox, Novell, FZJ JUROPA
www.thinkparq.com www.beegfs.com
www.thinkparq.com www.beegfs.com KEY ASPECTS Maximum Flexibility Maximum Scalability BeeGFS supports a wide range of Linux distributions such as RHEL/Fedora, SLES/OpenSuse or Debian/Ubuntu as well as a
Cray s Storage History and Outlook Lustre+ Jason Goodman, Cray LUG 2015 - Denver
Cray s Storage History and Outlook Lustre+ Jason Goodman, Cray - Denver Agenda Cray History from Supercomputers to Lustre Where we are Today Cray Business OpenSFS Flashback to the Future SSDs, DVS, and
Data management challenges in todays Healthcare and Life Sciences ecosystems
Data management challenges in todays Healthcare and Life Sciences ecosystems Jose L. Alvarez Principal Engineer, WW Director Life Sciences [email protected] Evolution of Data Sets in Healthcare
Building a Top500-class Supercomputing Cluster at LNS-BUAP
Building a Top500-class Supercomputing Cluster at LNS-BUAP Dr. José Luis Ricardo Chávez Dr. Humberto Salazar Ibargüen Dr. Enrique Varela Carlos Laboratorio Nacional de Supercómputo Benemérita Universidad
Purchase of High Performance Computing (HPC) Central Compute Resources by Northwestern Researchers
Information Technology Purchase of High Performance Computing (HPC) Central Compute Resources by Northwestern Researchers Effective for FY2016 Purpose This document summarizes High Performance Computing
Scientific Computing Data Management Visions
Scientific Computing Data Management Visions ELI-Tango Workshop Szeged, 24-25 February 2015 Péter Szász Group Leader Scientific Computing Group ELI-ALPS Scientific Computing Group Responsibilities Data
FLOW-3D Performance Benchmark and Profiling. September 2012
FLOW-3D Performance Benchmark and Profiling September 2012 Note The following research was performed under the HPC Advisory Council activities Participating vendors: FLOW-3D, Dell, Intel, Mellanox Compute
Summit and Sierra Supercomputers:
Whitepaper Summit and Sierra Supercomputers: An Inside Look at the U.S. Department of Energy s New Pre-Exascale Systems November 2014 1 Contents New Flagship Supercomputers in U.S. to Pave Path to Exascale
News about HPC and Clouds @ Inria
News about HPC and Clouds @ Inria Claude Kirchner Advisor to the president 24/11/2014 Nov 24, 2014-2 Nov 24, 2014-3 Inria Research Centres Inria LILLE Nord Europe Inria PARIS - Rocquencourt Inria NANCY
The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC2013 - Denver
1 The PHI solution Fujitsu Industry Ready Intel XEON-PHI based solution SC2013 - Denver Industrial Application Challenges Most of existing scientific and technical applications Are written for legacy execution
GPFS und HPSS am HLRS
GPFS und HPSS am HLRS Peter W. Haas Archivierung im Bereich Höchstleistungsrechner Swisstopo, Bern 3. Juli 2009 1 High Performance Computing Center Stuttgart Table of Contents 1. What are GPFS and HPSS
Trends in High-Performance Computing for Power Grid Applications
Trends in High-Performance Computing for Power Grid Applications Franz Franchetti ECE, Carnegie Mellon University www.spiral.net Co-Founder, SpiralGen www.spiralgen.com This talk presents my personal views
The Hartree Centre helps businesses unlock the potential of HPC
The Hartree Centre helps businesses unlock the potential of HPC Fostering collaboration and innovation across UK industry with help from IBM Overview The need The Hartree Centre needs leading-edge computing
A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures
11 th International LS-DYNA Users Conference Computing Technology A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures Yih-Yih Lin Hewlett-Packard Company Abstract In this paper, the
Information about Pan-European HPC infrastructure PRACE. Vít Vondrák IT4Innovations
Information about Pan-European HPC infrastructure PRACE Vít Vondrák IT4Innovations Realizing the ESFRI Vision for an HPC RI European HPC- facilities at the top of an HPC provisioning pyramid Tier- 0: European
Supercomputer Center Management Challenges. Branislav Jansík
Supercomputer Center Management Challenges Branislav Jansík 2000-2004 PhD at KTH Stockholm Molecular properties for heavy element compounds Density Functional Theory for Molecular Properties 2004-2006
Energy efficient computing on Embedded and Mobile devices. Nikola Rajovic, Nikola Puzovic, Lluis Vilanova, Carlos Villavieja, Alex Ramirez
Energy efficient computing on Embedded and Mobile devices Nikola Rajovic, Nikola Puzovic, Lluis Vilanova, Carlos Villavieja, Alex Ramirez A brief look at the (outdated) Top500 list Most systems are built
Performance Evaluation of Amazon EC2 for NASA HPC Applications!
National Aeronautics and Space Administration Performance Evaluation of Amazon EC2 for NASA HPC Applications! Piyush Mehrotra!! J. Djomehri, S. Heistand, R. Hood, H. Jin, A. Lazanoff,! S. Saini, R. Biswas!
Hank Childs, University of Oregon
Exascale Analysis & Visualization: Get Ready For a Whole New World Sept. 16, 2015 Hank Childs, University of Oregon Before I forget VisIt: visualization and analysis for very big data DOE Workshop for
Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband
Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband A P P R O I N T E R N A T I O N A L I N C Steve Lyness Vice President, HPC Solutions Engineering [email protected] Company Overview
David Vicente Head of User Support BSC
www.bsc.es Programming MareNostrum III David Vicente Head of User Support BSC Agenda WEDNESDAY - 17-04-13 9:00 Introduction to BSC, PRACE PATC and this training 9:30 New MareNostrum III the views from
IBM System Cluster 1350 ANSYS Microsoft Windows Compute Cluster Server
IBM System Cluster 1350 ANSYS Microsoft Windows Compute Cluster Server IBM FLUENT Benchmark Results IBM & FLUENT Recommended Configurations IBM 16-Core BladeCenter S Cluster for FLUENT Systems: Up to Six
Access, Documentation and Service Desk. Anupam Karmakar / Application Support Group / Astro Lab
Access, Documentation and Service Desk Anupam Karmakar / Application Support Group / Astro Lab Time to get answer to these questions Who is allowed to use LRZ hardware? My file system is full. How can
Supercomputing 2004 - Status und Trends (Conference Report) Peter Wegner
(Conference Report) Peter Wegner SC2004 conference Top500 List BG/L Moors Law, problems of recent architectures Solutions Interconnects Software Lattice QCD machines DESY @SC2004 QCDOC Conclusions Technical
HCERES report on research unit:
Research units HCERES report on research unit: Centre Lasers Intenses et Applications CELIA Under the supervision of the following institutions and research bodies: Université de Bordeaux Commissariat
The K computer: Project overview
The Next-Generation Supercomputer The K computer: Project overview SHOJI, Fumiyoshi Next-Generation Supercomputer R&D Center, RIKEN The K computer Outline Project Overview System Configuration of the K
ALPS Supercomputing System A Scalable Supercomputer with Flexible Services
ALPS Supercomputing System A Scalable Supercomputer with Flexible Services 1 Abstract Supercomputing is moving from the realm of abstract to mainstream with more and more applications and research being
