Parallel I/O on JUQUEEN
|
|
|
- Vincent Gaines
- 10 years ago
- Views:
Transcription
1 Parallel I/O on JUQUEEN 3. February rd JUQUEEN Porting and Tuning Workshop Sebastian Lührs, Kay Thust Jülich Supercomputing Centre
2 Overview Blue Gene/Q I/O Hardware Overview, Cabling, I/O Services GPFS architecture and data path Pitfalls The Parallel I/O Software Stack Darshan Overview, Usage, Interpretation SIONlib Parallel HDF5 I/O Hints 3. February 2015 JUQUEEN Porting and Tuning Workshop
3 Blue Gene/Q Packaging Hierarchy I/O-Nodes JUQUEEN I/O Nodes: 248 (27x8 + 1x32) IBM February 2015 JUQUEEN Porting and Tuning Workshop
4 Blue Gene/Q: I/O-node cabling (8 ION/Rack) IBM February 2015 JUQUEEN Porting and Tuning Workshop
5 Blue Gene/Q: I/O Services Function shipping system calls to I/O-node Support NFS, GPFS, Lustre and PVFS2 filesystems Supports ratio of 8192:1 compute task to I/O-node Only 1 I/O-Proxy per compute node Standard communications protocol OFED verbs Using Torus DMA hardware for performance IBM February 2015 JUQUEEN Porting and Tuning Workshop
6 IBM General Parallel File System : Architecture and I/O Data Path on BG/Q IBM February 2015 JUQUEEN Porting and Tuning Workshop
7 Pitfall 1: Frequent flushing on small blocks Modern file systems in HPC have large file system blocks A flush on a file handle forces the file system to perform all pending write operations If application writes in small data blocks the same file system block it has to be read and written multiple times Performance degradation due to the inability to combine several write calls flush() 3. February 2015 JUQUEEN Porting and Tuning Workshop
8 Pitfall 2: Parallel Creation of Individual Files Jugene (BG/P) + GPFS: file create+open, one file per task versus one file per I/O-node Contention at node doing directory updates (directory meta-node) Pre-created files or own directory per task may help performance, but does not simplify file handling Complicates file management (e.g. archive) shared files are mandatory 3. February 2015 JUQUEEN Porting and Tuning Workshop
9 Pitfall 3: False sharing of file system blocks Parallel I/O to shared files (POSIX) data block file system block free file system block t 1 t 2 lock FS Block FS Block FS Block data task 1 lock data task 2 Data blocks of individual processes do not fill up a complete file system block Several processes share a file system block Exclusive access (e.g. write) must be serialized The more processes have to synchronize the more waiting time will propagate 3. February 2015 JUQUEEN Porting and Tuning Workshop
10 Pitfall 4: Number of Tasks per Shared File Meta-data wall on file level File meta-data management Locking I/Oclient FS blocks file i-node indirect blocks Example Blue Gene/P Jugene (72 racks) I/O forwarding nodes (ION) GPFS client on ION Solution: tasks : files ratio ~ const SIONlib: one file per ION implicit task-to-file mapp T/F: 512/1 or 1/1 T/F: 4096/1 T/F: 16384/1 3. February 2015 JUQUEEN Porting and Tuning Workshop
11 Pitfall 5: Portability Endianess (byte order) of binary data Example (32 bit): = Address Little Endian Big Endian Conversion of files might be necessary and expensive Solution: Choosing a portable data format (HDF5, NetCDF) 3. February 2015 JUQUEEN Porting and Tuning Workshop
12 The Parallel I/O Software Stack Parallel application P-HDF5 NetCDF-4 PNetCDF Shared file Tasklocal files MPI-I/O SIONlib POSIX I/O Parallel file system 3. February 2015 JUQUEEN Porting and Tuning Workshop
13 How to choose an I/O strategy? Performance considerations Portability Amount of data Frequency of reading/writing Scalability Different HPC architectures Data exchange with others Long-term storage E.g. use two formats and converters: Internal: Write/read data as-is Restart/checkpoint files External: Write/read data in non-decomposed format (portable, system-independent, self-describing) Workflows, Pre-, Postprocessing, Data exchange, 3. February 2015 JUQUEEN Porting and Tuning Workshop
14 Darshan: Overview Developed at ANL: HPC I/O Characterization Tool instruments I/O Uses modified versions of I/O libraries during linking No code changes needed only recompilation with wrapped compiler Helps analysing I/O patterns and identifying bottlenecks Analyses parallel I/O with MPI-IO and standard POSIX I/O Ready to use version available on JUQUEEN 3. February 2015 JUQUEEN Porting and Tuning Workshop
15 Darshan: Usage example on JUQUEEN Load module Recompile module load darshan make MPICC=mpicc.darshan Tell runjob where to save the output (in submit script) runjob... --envs \ DARSHAN_LOG_PATH=$HOME/darshanlogs Analyse output darshan-job-summary.pl mylog.darshan.gz evince mylog.pdf 3. February 2015 JUQUEEN Porting and Tuning Workshop
16 Darshan: Interpret the summary Average and statistical information on I/O patterns Relative time for I/O Most common access sizes Additional metrics File count I/O size histogram Timeline for read / write per task 3. February 2015 JUQUEEN Porting and Tuning Workshop
17 Serial SIONlib: Overview Shared file I/O with automatic file system block alignment One single or few large files (e.g. one per ION) Only open and close calls are collective Support for file coalescing For data sizes << file system block size Application t 1 t 2 t 3 Tasks t n-2 t n-1 t n program /* fopen() */ sid=sion_paropen_mpi( filename, bw, Physical multi-file Logical task-local files Parallel file system SIONlib /* fwrite(bindata,1,nbytes, fileptr) */ sion_fwrite(bindata,1,nbytes, sid); /* fclose() */ sion_parclose_mpi(sid) &numfiles, &chunksize, gcom, &lcom, &fileptr,...); 3. February 2015 JUQUEEN Porting and Tuning Workshop
18 Parallel HDF5: Overview Self-describing file format Hierarchical, filesystem-like data format Support of additional metadata for datasets Portable Builds on top of standard MPI-I/O or POSIX-I/O Move between system architectures Automatic conversion of data representation Move between parallel/serial applications e.g MPI processes MPI processes e.g. simulation post-processing Complex data selections possible for read or write e.g. read only sub-array of dataset with stride slide provided by Jens Henrik Göbbert 3. February 2015 JUQUEEN Porting and Tuning Workshop
19 Parallel HDF5: I/O Hints Align datasets to file system block size H5Pset_alignment(...); Use hyperslabs H5Sselect_hyperslab(...); Chunk datasets H5Pset_cache(...); H5Pset_chunk_cache(...); Increase transfer buffer H5Pset_buffer(...); Improve metadata handling H5Pset_istore_k(...); H5Pset_meta_block_size( ); P0 P1 P0 P1 slide provided by Jens Henrik Göbbert 3. February 2015 JUQUEEN Porting and Tuning Workshop
20 I/O Hints Reduce locking for MPI-IO export BGLOCKLESSMPIO_F_TYPE="0x ROMIO hints echo " IBM_largeblock_io true" > romio.hints echo " cb_buffer_size " >> romio.hints export ROMIO_HINTS=romio.hints Note: don t forget to add exports to runjob, via --exp-env X MPIX_Calls: splitting communicators per I/O-bridge FORTRAN: MPIX_PSET_SAME_COMM_CREATE(INTEGER pset_comm_same, INTEGER ierr) C: #include <mpix.h> int MPIX_Pset_same_comm_create( MPI_Comm *pset_comm_same ) sid=sion_paropen_mpi( filename, bw, &numfiles, &chunksize, gcom, &lcom,...); 3. February 2015 JUQUEEN Porting and Tuning Workshop
Parallel file I/O bottlenecks and solutions
Mitglied der Helmholtz-Gemeinschaft Parallel file I/O bottlenecks and solutions Views to Parallel I/O: Hardware, Software, Application Challenges at Large Scale Introduction SIONlib Pitfalls, Darshan,
THE EXPAND PARALLEL FILE SYSTEM A FILE SYSTEM FOR CLUSTER AND GRID COMPUTING. José Daniel García Sánchez ARCOS Group University Carlos III of Madrid
THE EXPAND PARALLEL FILE SYSTEM A FILE SYSTEM FOR CLUSTER AND GRID COMPUTING José Daniel García Sánchez ARCOS Group University Carlos III of Madrid Contents 2 The ARCOS Group. Expand motivation. Expand
System Software for High Performance Computing. Joe Izraelevitz
System Software for High Performance Computing Joe Izraelevitz Agenda Overview of Supercomputers Blue Gene/Q System LoadLeveler Job Scheduler General Parallel File System HPC at UR What is a Supercomputer?
Portable, Scalable, and High-Performance I/O Forwarding on Massively Parallel Systems. Jason Cope [email protected]
Portable, Scalable, and High-Performance I/O Forwarding on Massively Parallel Systems Jason Cope [email protected] Computation and I/O Performance Imbalance Leadership class computa:onal scale: >100,000
LLNL s Parallel I/O Testing Tools and Techniques for ASC Parallel File Systems
UCRL-CONF-203489 LAWRENCE LIVERMORE NATIONAL LABORATORY LLNL s Parallel I/O Testing Tools and Techniques for ASC Parallel File Systems W. E. Loewe, R. M. Hedges, T. T. McLarty, and C. J. Morrone April
Part I Courses Syllabus
Part I Courses Syllabus This document provides detailed information about the basic courses of the MHPC first part activities. The list of courses is the following 1.1 Scientific Programming Environment
COSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters
COSC 6374 Parallel I/O (I) I/O basics Fall 2012 Concept of a clusters Processor 1 local disks Compute node message passing network administrative network Memory Processor 2 Network card 1 Network card
Uncovering degraded application performance with LWM 2. Aamer Shah, Chih-Song Kuo, Lucas Theisen, Felix Wolf November 17, 2014
Uncovering degraded application performance with LWM 2 Aamer Shah, Chih-Song Kuo, Lucas Theisen, Felix Wolf November 17, 214 Motivation: Performance degradation Internal factors: Inefficient use of hardware
How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx
Parallel IO performance and scalability study on the PRACE CURIE supercomputer.
Available on-line at www.prace-ri.eu Partnership for Advanced Computing in Europe Parallel IO performance and scalability study on the PRACE CURIE supercomputer. Philippe Wautelet a,, Pierre Kestener a
Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC
HPC Architecture End to End Alexandre Chauvin Agenda HPC Software Stack Visualization National Scientific Center 2 Agenda HPC Software Stack Alexandre Chauvin Typical HPC Software Stack Externes LAN Typical
Introduction History Design Blue Gene/Q Job Scheduler Filesystem Power usage Performance Summary Sequoia is a petascale Blue Gene/Q supercomputer Being constructed by IBM for the National Nuclear Security
Hadoop Distributed File System. T-111.5550 Seminar On Multimedia 2009-11-11 Eero Kurkela
Hadoop Distributed File System T-111.5550 Seminar On Multimedia 2009-11-11 Eero Kurkela Agenda Introduction Flesh and bones of HDFS Architecture Accessing data Data replication strategy Fault tolerance
A highly configurable and efficient simulator for job schedulers on supercomputers
Mitglied der Helmholtz-Gemeinschaft A highly configurable and efficient simulator for job schedulers on supercomputers April 12, 2013 Carsten Karbach, Jülich Supercomputing Centre (JSC) Motivation Objective
A Survey of Shared File Systems
Technical Paper A Survey of Shared File Systems Determining the Best Choice for your Distributed Applications A Survey of Shared File Systems A Survey of Shared File Systems Table of Contents Introduction...
Operating System for the K computer
Operating System for the K computer Jun Moroo Masahiko Yamada Takeharu Kato For the K computer to achieve the world s highest performance, Fujitsu has worked on the following three performance improvements
ProTrack: A Simple Provenance-tracking Filesystem
ProTrack: A Simple Provenance-tracking Filesystem Somak Das Department of Electrical Engineering and Computer Science Massachusetts Institute of Technology [email protected] Abstract Provenance describes a file
Hadoop on the Gordon Data Intensive Cluster
Hadoop on the Gordon Data Intensive Cluster Amit Majumdar, Scientific Computing Applications Mahidhar Tatineni, HPC User Services San Diego Supercomputer Center University of California San Diego Dec 18,
IBM Platform Computing : infrastructure management for HPC solutions on OpenPOWER Jing Li, Software Development Manager IBM
IBM Platform Computing : infrastructure management for HPC solutions on OpenPOWER Jing Li, Software Development Manager IBM #OpenPOWERSummit Join the conversation at #OpenPOWERSummit 1 Scale-out and Cloud
Network File System (NFS) Pradipta De [email protected]
Network File System (NFS) Pradipta De [email protected] Today s Topic Network File System Type of Distributed file system NFS protocol NFS cache consistency issue CSE506: Ext Filesystem 2 NFS
Anwendungsintegration und Workflows mit UNICORE 6
Mitglied der Helmholtz-Gemeinschaft Anwendungsintegration und Workflows mit UNICORE 6 Bernd Schuller und UNICORE-Team Jülich Supercomputing Centre, Forschungszentrum Jülich GmbH 26. November 2009 D-Grid
LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu
GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"
GPFS Storage Server Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " Agenda" GPFS Overview" Classical versus GSS I/O Solution" GPFS Storage Server (GSS)" GPFS Native RAID
COSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters
COSC 6374 Parallel Computation Parallel I/O (I) I/O basics Spring 2008 Concept of a clusters Processor 1 local disks Compute node message passing network administrative network Memory Processor 2 Network
Mitglied der Helmholtz-Gemeinschaft. System monitoring with LLview and the Parallel Tools Platform
Mitglied der Helmholtz-Gemeinschaft System monitoring with LLview and the Parallel Tools Platform November 25, 2014 Carsten Karbach Content 1 LLview 2 Parallel Tools Platform (PTP) 3 Latest features 4
Current Status of FEFS for the K computer
Current Status of FEFS for the K computer Shinji Sumimoto Fujitsu Limited Apr.24 2012 LUG2012@Austin Outline RIKEN and Fujitsu are jointly developing the K computer * Development continues with system
Hopes and Facts in Evaluating the Performance of HPC-I/O on a Cloud Environment
Hopes and Facts in Evaluating the Performance of HPC-I/O on a Cloud Environment Pilar Gomez-Sanchez Computer Architecture and Operating Systems Department (CAOS) Universitat Autònoma de Barcelona, Bellaterra
Bottleneck Detection in Parallel File Systems with Trace-Based Performance Monitoring
Julian M. Kunkel - Euro-Par 2008 1/33 Bottleneck Detection in Parallel File Systems with Trace-Based Performance Monitoring Julian M. Kunkel Thomas Ludwig Institute for Computer Science Parallel and Distributed
Workshop on Parallel and Distributed Scientific and Engineering Computing, Shanghai, 25 May 2012
Scientific Application Performance on HPC, Private and Public Cloud Resources: A Case Study Using Climate, Cardiac Model Codes and the NPB Benchmark Suite Peter Strazdins (Research School of Computer Science),
Recent Advances in Periscope for Performance Analysis and Tuning
Recent Advances in Periscope for Performance Analysis and Tuning Isaias Compres, Michael Firbach, Michael Gerndt Robert Mijakovic, Yury Oleynik, Ventsislav Petkov Technische Universität München Yury Oleynik,
HadioFS: Improve the Performance of HDFS by Off-loading I/O to ADIOS. Xiaobing Li
HadioFS: Improve the Performance of HDFS by Off-loading I/O to ADIOS by Xiaobing Li A thesis submitted to the Graduate Faculty of Auburn University in partial fulfillment of the requirements for the Degree
Implementing the Hadoop Distributed File System Protocol on OneFS Jeff Hughes EMC Isilon
Implementing the Hadoop Distributed File System Protocol on OneFS Jeff Hughes EMC Isilon Outline Hadoop Overview OneFS Overview MapReduce + OneFS Details of isi_hdfs_d Wrap up & Questions 2 Hadoop Overview
Cluster Implementation and Management; Scheduling
Cluster Implementation and Management; Scheduling CPS343 Parallel and High Performance Computing Spring 2013 CPS343 (Parallel and HPC) Cluster Implementation and Management; Scheduling Spring 2013 1 /
Welcome to the. Jülich Supercomputing Centre. D. Rohe and N. Attig Jülich Supercomputing Centre (JSC), Forschungszentrum Jülich
Mitglied der Helmholtz-Gemeinschaft Welcome to the Jülich Supercomputing Centre D. Rohe and N. Attig Jülich Supercomputing Centre (JSC), Forschungszentrum Jülich Schedule: Monday, May 19 13:00-13:30 Welcome
Managing Complexity in Distributed Data Life Cycles Enhancing Scientific Discovery
Center for Information Services and High Performance Computing (ZIH) Managing Complexity in Distributed Data Life Cycles Enhancing Scientific Discovery Richard Grunzke*, Jens Krüger, Sandra Gesing, Sonja
I/O intensive applications: what are the main differences in the design of the HPC filesystems vs the MapReduce ones?
I/O intensive applications: what are the main differences in the design of the HPC filesystems vs the MapReduce ones? Matthieu Dorier, Radu Marius Tudoran Master 2 Research ENS Cachan - Brittany extension
EXPLOITING SHARED MEMORY TO IMPROVE PARALLEL I/O PERFORMANCE
EXPLOITING SHARED MEMORY TO IMPROVE PARALLEL I/O PERFORMANCE Andrew B. Hastings Sun Microsystems, Inc. Alok Choudhary Northwestern University September 19, 2006 This material is based on work supported
General Parallel File System (GPFS) Native RAID For 100,000-Disk Petascale Systems
General Parallel File System (GPFS) Native RAID For 100,000-Disk Petascale Systems Veera Deenadhayalan IBM Almaden Research Center 2011 IBM Corporation Hard Disk Rates Are Lagging There have been recent
David Rioja Redondo Telecommunication Engineer Englobe Technologies and Systems
David Rioja Redondo Telecommunication Engineer Englobe Technologies and Systems About me David Rioja Redondo Telecommunication Engineer - Universidad de Alcalá >2 years building and managing clusters UPM
Silviu Panica, Marian Neagul, Daniela Zaharie and Dana Petcu (Romania)
Silviu Panica, Marian Neagul, Daniela Zaharie and Dana Petcu (Romania) Outline Introduction EO challenges; EO and classical/cloud computing; EO Services The computing platform Cluster -> Grid -> Cloud
1 Bull, 2011 Bull Extreme Computing
1 Bull, 2011 Bull Extreme Computing Table of Contents HPC Overview. Cluster Overview. FLOPS. 2 Bull, 2011 Bull Extreme Computing HPC Overview Ares, Gerardo, HPC Team HPC concepts HPC: High Performance
Parallel Visualization of Petascale Simulation Results from GROMACS, NAMD and CP2K on IBM Blue Gene/P using VisIt Visualization Toolkit
Available online at www.prace-ri.eu Partnership for Advanced Computing in Europe Parallel Visualization of Petascale Simulation Results from GROMACS, NAMD and CP2K on IBM Blue Gene/P using VisIt Visualization
INF-110. GPFS Installation
INF-110 GPFS Installation Overview Plan the installation Before installing any software, it is important to plan the GPFS installation by choosing the hardware, deciding which kind of disk connectivity
Parallel I/O on Mira Venkat Vishwanath and Kevin Harms
Parallel I/O on Mira Venkat Vishwanath and Kevin Harms Argonne Na*onal Laboratory [email protected] ALCF-2 I/O Infrastructure Mira BG/Q Compute Resource Tukey Analysis Cluster 48K Nodes 768K Cores 10 PFlops
Running on Blue Gene/Q at Argonne Leadership Computing Facility (ALCF)
Running on Blue Gene/Q at Argonne Leadership Computing Facility (ALCF) ALCF Resources: Machines & Storage Mira (Production) IBM Blue Gene/Q 49,152 nodes / 786,432 cores 768 TB of memory Peak flop rate:
Performance Analysis and Optimization Tool
Performance Analysis and Optimization Tool Andres S. CHARIF-RUBIAL [email protected] Performance Analysis Team, University of Versailles http://www.maqao.org Introduction Performance Analysis Develop
CS2510 Computer Operating Systems
CS2510 Computer Operating Systems HADOOP Distributed File System Dr. Taieb Znati Computer Science Department University of Pittsburgh Outline HDF Design Issues HDFS Application Profile Block Abstraction
CS2510 Computer Operating Systems
CS2510 Computer Operating Systems HADOOP Distributed File System Dr. Taieb Znati Computer Science Department University of Pittsburgh Outline HDF Design Issues HDFS Application Profile Block Abstraction
High Performance Computing Facility Specifications, Policies and Usage. Supercomputer Project. Bibliotheca Alexandrina
High Performance Computing Facility Specifications, Policies and Usage Supercomputer Project Bibliotheca Alexandrina Bibliotheca Alexandrina 1/16 Topics Specifications Overview Site Policies Intel Compilers
Distributed File Systems
Distributed File Systems Paul Krzyzanowski Rutgers University October 28, 2012 1 Introduction The classic network file systems we examined, NFS, CIFS, AFS, Coda, were designed as client-server applications.
Chapter 6, The Operating System Machine Level
Chapter 6, The Operating System Machine Level 6.1 Virtual Memory 6.2 Virtual I/O Instructions 6.3 Virtual Instructions For Parallel Processing 6.4 Example Operating Systems 6.5 Summary Virtual Memory General
Application Performance Analysis Tools and Techniques
Mitglied der Helmholtz-Gemeinschaft Application Performance Analysis Tools and Techniques 2012-06-27 Christian Rössel Jülich Supercomputing Centre [email protected] EU-US HPC Summer School Dublin
Performance Analysis of Mixed Distributed Filesystem Workloads
Performance Analysis of Mixed Distributed Filesystem Workloads Esteban Molina-Estolano, Maya Gokhale, Carlos Maltzahn, John May, John Bent, Scott Brandt Motivation Hadoop-tailored filesystems (e.g. CloudStore)
Petascale Software Challenges. Piyush Chaudhary [email protected] High Performance Computing
Petascale Software Challenges Piyush Chaudhary [email protected] High Performance Computing Fundamental Observations Applications are struggling to realize growth in sustained performance at scale Reasons
HPC Software Requirements to Support an HPC Cluster Supercomputer
HPC Software Requirements to Support an HPC Cluster Supercomputer Susan Kraus, Cray Cluster Solutions Software Product Manager Maria McLaughlin, Cray Cluster Solutions Product Marketing Cray Inc. WP-CCS-Software01-0417
Google File System. Web and scalability
Google File System Web and scalability The web: - How big is the Web right now? No one knows. - Number of pages that are crawled: o 100,000 pages in 1994 o 8 million pages in 2005 - Crawlable pages might
InterferenceRemoval: Removing Interference of Disk Access for MPI Programs through Data Replication
InterferenceRemoval: Removing Interference of Disk Access for MPI Programs through Data Replication Xuechen Zhang and Song Jiang The ECE Department Wayne State University Detroit, MI, 4822, USA {xczhang,
CSAR: Cluster Storage with Adaptive Redundancy
CSAR: Cluster Storage with Adaptive Redundancy Manoj Pillai, Mario Lauria Department of Computer and Information Science The Ohio State University Columbus, OH, 4321 Email: pillai,[email protected]
JUBE. A Flexible, Application- and Platform-Independent Environment for Benchmarking
JUBE A Flexible, Application- and Platform-Independent Environment for Benchmarking Sebastian Lührs [email protected] Jülich Supercomputing Centre Cy-Tera/LinkSCEEM HPC Administrator Workshop, 19-21
Prepared By : Manoj Kumar Joshi & Vikas Sawhney
Prepared By : Manoj Kumar Joshi & Vikas Sawhney General Agenda Introduction to Hadoop Architecture Acknowledgement Thanks to all the authors who left their selfexplanatory images on the internet. Thanks
Mixing Hadoop and HPC Workloads on Parallel Filesystems
Mixing Hadoop and HPC Workloads on Parallel Filesystems Esteban Molina-Estolano *, Maya Gokhale, Carlos Maltzahn *, John May, John Bent, Scott Brandt * * UC Santa Cruz, ISSDM, PDSI Lawrence Livermore National
PART IV Performance oriented design, Performance testing, Performance tuning & Performance solutions. Outline. Performance oriented design
PART IV Performance oriented design, Performance testing, Performance tuning & Performance solutions Slide 1 Outline Principles for performance oriented design Performance testing Performance tuning General
NetApp High-Performance Computing Solution for Lustre: Solution Guide
Technical Report NetApp High-Performance Computing Solution for Lustre: Solution Guide Robert Lai, NetApp August 2012 TR-3997 TABLE OF CONTENTS 1 Introduction... 5 1.1 NetApp HPC Solution for Lustre Introduction...5
Data Centric Systems (DCS)
Data Centric Systems (DCS) Architecture and Solutions for High Performance Computing, Big Data and High Performance Analytics High Performance Computing with Data Centric Systems 1 Data Centric Systems
EZManage V4.0 Release Notes. Document revision 1.08 (15.12.2013)
EZManage V4.0 Release Notes Document revision 1.08 (15.12.2013) Release Features Feature #1- New UI New User Interface for every form including the ribbon controls that are similar to the Microsoft office
The scalability impact of a component-based software engineering framework on a growing SAMR toolkit: a case study
The scalability impact of a component-based software engineering framework on a growing SAMR toolkit: a case study Benjamin A. Allan and Jaideep Ray We examine a growing toolkit for parallel SAMR-based
Blue Gene Active Storage for High Performance BG/Q I/O and Scalable Data-centric Analytics
for High Performance BG/Q I/O and Scalable Data-centric Analytics Blake G. Fitch [email protected] The Extended, World Wide Active Storage Fabric Team Blake G. Fitch Robert S. Germain Michele Franceschini
JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert
Mitglied der Helmholtz-Gemeinschaft JUROPA Linux Cluster An Overview 19 May 2014 Ulrich Detert JuRoPA JuRoPA Jülich Research on Petaflop Architectures Bull, Sun, ParTec, Intel, Mellanox, Novell, FZJ JUROPA
Introduction to Highly Available NFS Server on scale out storage systems based on GlusterFS
Introduction to Highly Available NFS Server on scale out storage systems based on GlusterFS Soumya Koduri Red Hat Meghana Madhusudhan Red Hat AGENDA What is GlusterFS? Integration with NFS Ganesha Clustered
Easier - Faster - Better
Highest reliability, availability and serviceability ClusterStor gets you productive fast with robust professional service offerings available as part of solution delivery, including quality controlled
Parallel Processing using the LOTUS cluster
Parallel Processing using the LOTUS cluster Alison Pamment / Cristina del Cano Novales JASMIN/CEMS Workshop February 2015 Overview Parallelising data analysis LOTUS HPC Cluster Job submission on LOTUS
Advanced MPI. Hybrid programming, profiling and debugging of MPI applications. Hristo Iliev RZ. Rechen- und Kommunikationszentrum (RZ)
Advanced MPI Hybrid programming, profiling and debugging of MPI applications Hristo Iliev RZ Rechen- und Kommunikationszentrum (RZ) Agenda Halos (ghost cells) Hybrid programming Profiling of MPI applications
HPC @ CRIBI. Calcolo Scientifico e Bioinformatica oggi Università di Padova 13 gennaio 2012
HPC @ CRIBI Calcolo Scientifico e Bioinformatica oggi Università di Padova 13 gennaio 2012 what is exact? experience on advanced computational technologies a company lead by IT experts with a strong background
Cosmological simulations on High Performance Computers
Cosmological simulations on High Performance Computers Cosmic Web Morphology and Topology Cosmological workshop meeting Warsaw, 12-17 July 2011 Maciej Cytowski Interdisciplinary Centre for Mathematical
Storage Virtualization in Cloud
Storage Virtualization in Cloud Cloud Strategy Partners, LLC Sponsored by: IEEE Educational Activities and IEEE Cloud Computing Course Presenter s Biography This IEEE Cloud Computing tutorial has been
