Moab and TORQUE Highlights CUG 2015
|
|
|
- Helen Day
- 10 years ago
- Views:
Transcription
1 Moab and TORQUE Highlights CUG 2015 David Beer TORQUE Architect 28 Apr 2015 Gary D. Brown HPC Product Manager 1
2 Agenda NUMA-aware Heterogeneous Jobs Ascent Project Power Management and Energy Accounting DataWarp Integration Nitro HTC Scheduler Collaboration Invitation 2
3 NUMA-aware Heterogeneous Job Task Placement 3
4 NUMA-aware Job Task Placement Moab job submission options (msub) Task-oriented (tasks=nnn), not job- or nodeoriented Logical processors/task (lprocs=nn) Logical processor definition Cores (usecores on request) Threads (usethreads on request) Cores or threads (allowthreads default) Memory per task (memory=xxx) Other options per task (features, generic resources, accelerators, etc) TORQUE qsub supports same syntax 4
5 NUMA-aware Job Task Placement (continued) Places task exclusively on resources of specified locality node socket numa Compute node NUMA node (Intel Xeon Haswell and AMD Opteron) core thread Uses control groups (cgroups) to pin task to resources Processors Memory PCIe devices (GPUs, MICs) 5
6 NUMA-aware Job Task Placement (continued) Example job submission commands # Simple 1 task per core, regardless if hyper-threaded qsub L tasks=100:lprocs=1:usecores:place=core # Coupled-simulation (1 solid task, 2000 Cray fluid tasks) msub L tasks=1:usecores:memory=800gb:place=node:bigmem -L tasks=2000:usecores:place=core:memory=2gb:cray # 1 task using 1 thread per 2 threads (e.g., Intel MIC KNL) qsub L tasks=1200:lprocs=1:usethreads:place=thread=2 # 1 task per 2 sockets with 6 cores (3 / socket) and 1 GPU msub L tasks=100:lprocs=6:usecores:memory=200gb:gpu=1: place=socket=2 6
7 NUMA Example 1 Heterogeneous Nodes qsub L tasks=288:lprocs=1:usecores:place=core=2 mympijob 10 tasks per node 5 tasks per node 7
8 NUMA Example 2 GPU Cluster qsub -L tasks=16:lprocs=1:usecores:gpu=1:place=core mygpusim GPU/Memory Data Transfer Speed GPUs NUMA 0 NUMA 1 NUMA 2 NUMA 3 0/1 100% 85% 54% 48% 2/3 58% 33% 100% 86% 4/5 58% 33% 98% 85% 6/7 58% 33% 100% 86% Excellent performance Good performance Fair performance Poor performance 8
9 NUMA Example 3 Avionics Simulation msub -L tasks=1:lprocs=3:memory=5gib:usecores:place=core=4 # red System board -L tasks=4:lprocs=1:memory=10gib:usecores:place=core=2 # orange A board -L tasks=2:lprocs=2:memory=20gib:usecores:place=core=3 # yellow B board -L tasks=1:lprocs=4:memory=50gib:usecores:place=core=5 # green C board -L tasks=3:lprocs=8:memory=30gib:usecores:place=socket # blue D board -l naccesspolicy=singlejob AvionicSim.pbs 9
10 NUMA-aware Benefits Accurately place jobs on cores according to memory access NUMA-optimized, memory-bound workloads can run up to 2.5x faster More consistent job runtimes with predictable memory access times Much less job-to-job or task-to-task interference on each compute node User can define resources as needed by apps Request resources on a per-task type basis, allowing jobs to reserve only what they need Increases hardware ROI with higher job throughput due to shorter job runtimes 10
11 Ascent Project 11
12 Focus and Goals Main goals Improve scalability (>10x) Improve job throughput (>10x) Improve stability (rock-solid) Improve architecture Timeframe 24 months with four deliveries Moab 8.0 / TORQUE 5.0 Moab 8.1 / TORQUE 5.1 Moab 9.0 / TORQUE 6.0 Next version People Focused team of top engineers 12
13 Moab 8.0 Example 13
14 TORQUE 5.0 Results Reduced jitter by eliminating polling of compute nodes for job information Job submission rate of 250+ jobs per second for 100K jobs with no errors 14
15 Moab 8.1 Results Moab scheduling cycle time reduced by 55% Moab client command response delay reduced by 84% Moab job submission time reduced by 60% 15
16 Moab 9.0 Results (target of 29 July) Moab query client commands response delay reduced to sub-second 16
17 Power Management 17
18 CPU Clock Frequency Control Power management for running nodes Job submission cpuclock= option Absolute Clock Frequency Number cpuclock=2200 cpuclock=1800mhz Linux Power Governor Policy cpuclock=conservative Relative P-state Number cpuclock=0 cpuclock=p2 Can set in job scripts and Moab templates 18
19 Node Power State Management Additional Power States (C-states) Power management for idle nodes Suspend (C3/C6) Hibernate (S4) Shutdown (S3) 19
20 Power Management Control Architecture Moab On Off MWS pbsnodes pbs_server Suspend Hibernate Shutdown Power RM Plug-in NodePower Script On Off Wake-on-LAN Linux OS /sys/power TORQUE pbs_mom IPMI 20
21 Moab/Cray Power Management Control Architecture Moab On Off NodePower Script capmc On Shutdown SMW 21
22 Per-Job Energy Accounting Part of TORQUE/Cray integration End of job TORQUE checks for RUR-based energy consumption information Sums energy use for each job step (aprun) Sends energy_used job metric to Moab Moab passes job metric to Moab Accounting Manager for possible charging 22
23 DataWarp Integration 23
24 DataWarp Integration Project Supports all DataWarp storage types Per-Job DW storage Persistent DW storage (across multiple jobs) All DW storage scheduled by Moab Out-of-the-box functionality Some site-specific information configuration required 24
25 DataWarp Integration 2 per-job DW storage models supported Selectable by user at job submission 3-job Model Favors compute nodes as scarce resource Compute nodes allocated only during user job Prologue/Epilogue Model Favors DW storage as scarce resource Compute nodes allocated during entire DW storage allocation 25
26 3-Job Model Follows Moab Data-staging Model Uses Moab system jobs Automatically creates 3-job workflow DW creation/input data staging User job Output data staging/dw destruction 26
27 3-Job Model Architecture 27
28 Prologue/Epilogue Model Uses TORQUE prologue and epilogue Prologue Create job s DW storage allocation Stage input data to DW storage User job Epilogue Stage output data from DW storage Destroy job s DW storage allocation 28
29 Prologue/Epilogue Model Architecture 29
30 Persistent DataWarp Storage Moab tracks DW storage capacity Moab persistent DW storage reservation Persistent DataWarp storage management Created by reservation start trigger Destroyed by reservation end trigger Jobs get mount point 30
31 Nitro HTC Job Scheduler 31
32 Nitro Product High-throughput Computing (HTC) Usually short, single-core jobs Nitro HTC Scheduler Product Eliminates traditional scheduler overhead for HTC jobs Executes HTC jobs very fast Shorter the HTC jobs, the greater the speedup Submitted as normal batch job to HPC job scheduler Scheduler-agnostic 32
33 Nitro Architecture Nitro Coordinator Executes on first host allocated to Nitro job Schedules all HTC tasks defined in Tasks file Assigns tasks to Nitro workers for execution Nitro Workers Execute on all hosts allocated to Nitro job Worker executes assigned tasks as fast as possible Can run more tasks than cores on worker host Task File Log Files 33
34 Nitro Product Architecture 34
35 Nitro Job Progress Log In-progress and Final versions Nitro Job Information Start / Finish/ Elapsed Times Job ID Task / Job Progress Log / Completed Task Log File Paths Success/Failure/Timeout/Invalid Task Statistics Average Task Time Tasks per Second Worker Statistics 35
36 Completed Task Log All tasks Task Status Success Failure Timeout Invalid task definition Tab-delimited format Spreadsheet, database, etc Optional column headings Task Information Job ID Task ID Line Number Task Status Exit Code Host Name Task Start Time Task Duration User CPU Time System CPU Time Core Count Max Physical Memory Max Virtual Memory Task Name (optional) Labels (optional) Failure Message (optional) 36
37 Nitro Linger Mode --linger Command Line Option Coordinator continues execution after Task file end-of-file encountered Periodically polls Task file User appends new task definitions to Task file Nitro executes additional tasks Statistics Last 60 seconds 37
38 Collaboration Invitation: Malleable/Evolving Job API 38
39 Job Types and Scheduling Classification Description Rigid Moldable Malleable A job that requires a fixed set of resources, all of which the batch system must allocate to the job before it starts (fixed static resource allocation) A job that allows a variable set of resources for scheduling but requires a fixed set of resources to execute, which the batch system must allocate before it starts the job and which the job must discover in order to execute properly (variable static resource allocation) A job that allows a variable set of resources, which the batch system dynamically allocates and deallocates and of which allocation/deallocation the scheduler must inform the running job so it can adapt to the new resource allocation (unidirectional, scheduler-initiated and application-executed, variable dynamic resource allocation) Evolving Dynamic A job that dynamically requests and relinquishes resources during its runtime and of which the scheduler must be informed while the job is running so it can allocate or deallocate the resources (unidirectional, application-initiated and scheduler-executed, variable dynamic resource allocation) Adaptive Computing Definition A job where the scheduler or the job can dynamically initiate resource allocation changes during the job's runtime and of which the other must be informed while the job is running so they both keep the job's resources allocation synchronized (bidirectional, application- or scheduler-initiated and scheduler- or application-executed, variable dynamic resource allocation) 39
40 System Utilization with No Malleable or Evolving Jobs A B E G C D F
41 System Utilization with Malleable Job M and Evolving Job D M 1 M A 2 M3 B C M 4 D 2 D 1 M 5 M E F M 7 G 41
42 Standard Malleable/Evolving Job API Currently no standard application API Much research done Univ of Illinois at Urbana-Champaign (UIUC) Malleable jobs (Charm++ codes) German Research School for Simulation Sciences (RWTH Aachen) (group now at Technische Universität Darmstadt) Evolving jobs (AMR code with Maui/TORQUE) Invitation to start API collaboration Application Developers Scheduler Vendors 42
43 Questions? 43
Broadening Moab/TORQUE for Expanding User Needs
Broadening Moab/TORQUE for Expanding User Needs Gary D. Brown HPC Product Manager CUG 2016 1 2016 Adaptive Computing Enterprises, Inc. Agenda DataWarp Intel MIC KNL Viewpoint Web Portal User Portal Administrator
Resource Scheduling Best Practice in Hybrid Clusters
Available online at www.prace-ri.eu Partnership for Advanced Computing in Europe Resource Scheduling Best Practice in Hybrid Clusters C. Cavazzoni a, A. Federico b, D. Galetti a, G. Morelli b, A. Pieretti
A High Performance Computing Scheduling and Resource Management Primer
LLNL-TR-652476 A High Performance Computing Scheduling and Resource Management Primer D. H. Ahn, J. E. Garlick, M. A. Grondona, D. A. Lipari, R. R. Springmeyer March 31, 2014 Disclaimer This document was
How System Settings Impact PCIe SSD Performance
How System Settings Impact PCIe SSD Performance Suzanne Ferreira R&D Engineer Micron Technology, Inc. July, 2012 As solid state drives (SSDs) continue to gain ground in the enterprise server and storage
Running applications on the Cray XC30 4/12/2015
Running applications on the Cray XC30 4/12/2015 1 Running on compute nodes By default, users do not log in and run applications on the compute nodes directly. Instead they launch jobs on compute nodes
Batch Systems. provide a mechanism for submitting, launching, and tracking jobs on a shared resource
PBS INTERNALS PBS & TORQUE PBS (Portable Batch System)-software system for managing system resources on workstations, SMP systems, MPPs and vector computers. It was based on Network Queuing System (NQS)
LSKA 2010 Survey Report Job Scheduler
LSKA 2010 Survey Report Job Scheduler Graduate Institute of Communication Engineering {r98942067, r98942112}@ntu.edu.tw March 31, 2010 1. Motivation Recently, the computing becomes much more complex. However,
Overview of HPC Resources at Vanderbilt
Overview of HPC Resources at Vanderbilt Will French Senior Application Developer and Research Computing Liaison Advanced Computing Center for Research and Education June 10, 2015 2 Computing Resources
SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education www.accre.vanderbilt.
SLURM: Resource Management and Job Scheduling Software Advanced Computing Center for Research and Education www.accre.vanderbilt.edu Simple Linux Utility for Resource Management But it s also a job scheduler!
Miami University RedHawk Cluster Working with batch jobs on the Cluster
Miami University RedHawk Cluster Working with batch jobs on the Cluster The RedHawk cluster is a general purpose research computing resource available to support the research community at Miami University.
LS-DYNA Scalability on Cray Supercomputers. Tin-Ting Zhu, Cray Inc. Jason Wang, Livermore Software Technology Corp.
LS-DYNA Scalability on Cray Supercomputers Tin-Ting Zhu, Cray Inc. Jason Wang, Livermore Software Technology Corp. WP-LS-DYNA-12213 www.cray.com Table of Contents Abstract... 3 Introduction... 3 Scalability
Agenda. Context. System Power Management Issues. Power Capping Overview. Power capping participants. Recommendations
Power Capping Linux Agenda Context System Power Management Issues Power Capping Overview Power capping participants Recommendations Introduction of Linux Power Capping Framework 2 Power Hungry World Worldwide,
Optimizing GPU-based application performance for the HP for the HP ProLiant SL390s G7 server
Optimizing GPU-based application performance for the HP for the HP ProLiant SL390s G7 server Technology brief Introduction... 2 GPU-based computing... 2 ProLiant SL390s GPU-enabled architecture... 2 Optimizing
Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering
Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays Red Hat Performance Engineering Version 1.0 August 2013 1801 Varsity Drive Raleigh NC
Managing GPUs by Slurm. Massimo Benini HPC Advisory Council Switzerland Conference March 31 - April 3, 2014 Lugano
Managing GPUs by Slurm Massimo Benini HPC Advisory Council Switzerland Conference March 31 - April 3, 2014 Lugano Agenda General Slurm introduction Slurm@CSCS Generic Resource Scheduling for GPUs Resource
SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education www.accre.vanderbilt.
SLURM: Resource Management and Job Scheduling Software Advanced Computing Center for Research and Education www.accre.vanderbilt.edu Simple Linux Utility for Resource Management But it s also a job scheduler!
How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx
Using the Windows Cluster
Using the Windows Cluster Christian Terboven [email protected] aachen.de Center for Computing and Communication RWTH Aachen University Windows HPC 2008 (II) September 17, RWTH Aachen Agenda o Windows Cluster
Cluster Monitoring and Management Tools RAJAT PHULL, NVIDIA SOFTWARE ENGINEER ROB TODD, NVIDIA SOFTWARE ENGINEER
Cluster Monitoring and Management Tools RAJAT PHULL, NVIDIA SOFTWARE ENGINEER ROB TODD, NVIDIA SOFTWARE ENGINEER MANAGE GPUS IN THE CLUSTER Administrators, End users Middleware Engineers Monitoring/Management
IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud
IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud February 25, 2014 1 Agenda v Mapping clients needs to cloud technologies v Addressing your pain
Parallel Programming Survey
Christian Terboven 02.09.2014 / Aachen, Germany Stand: 26.08.2014 Version 2.3 IT Center der RWTH Aachen University Agenda Overview: Processor Microarchitecture Shared-Memory
RED HAT ENTERPRISE LINUX 7
RED HAT ENTERPRISE LINUX 7 TECHNICAL OVERVIEW Scott McCarty Senior Solutions Architect, Red Hat 01/12/2015 1 Performance Tuning Overview Little's Law: L = A x W (Queue Length = Average Arrival Rate x Wait
Job Scheduling with Moab Cluster Suite
Job Scheduling with Moab Cluster Suite IBM High Performance Computing February 2010 Y. Joanna Wong, Ph.D. [email protected] 2/22/2010 Workload Manager Torque Source: Adaptive Computing 2 Some terminology..
HP ProLiant DL580 Gen8 and HP LE PCIe Workload WHITE PAPER Accelerator 90TB Microsoft SQL Server Data Warehouse Fast Track Reference Architecture
WHITE PAPER HP ProLiant DL580 Gen8 and HP LE PCIe Workload WHITE PAPER Accelerator 90TB Microsoft SQL Server Data Warehouse Fast Track Reference Architecture Based on Microsoft SQL Server 2014 Data Warehouse
Microsoft HPC. V 1.0 José M. Cámara ([email protected])
Microsoft HPC V 1.0 José M. Cámara ([email protected]) Introduction Microsoft High Performance Computing Package addresses computing power from a rather different approach. It is mainly focused on commodity
159.735. Final Report. Cluster Scheduling. Submitted by: Priti Lohani 04244354
159.735 Final Report Cluster Scheduling Submitted by: Priti Lohani 04244354 1 Table of contents: 159.735... 1 Final Report... 1 Cluster Scheduling... 1 Table of contents:... 2 1. Introduction:... 3 1.1
Multi-core and Linux* Kernel
Multi-core and Linux* Kernel Suresh Siddha Intel Open Source Technology Center Abstract Semiconductor technological advances in the recent years have led to the inclusion of multiple CPU execution cores
Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging
Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging In some markets and scenarios where competitive advantage is all about speed, speed is measured in micro- and even nano-seconds.
Parallel Algorithm Engineering
Parallel Algorithm Engineering Kenneth S. Bøgh PhD Fellow Based on slides by Darius Sidlauskas Outline Background Current multicore architectures UMA vs NUMA The openmp framework Examples Software crisis
The Uintah Framework: A Unified Heterogeneous Task Scheduling and Runtime System
The Uintah Framework: A Unified Heterogeneous Task Scheduling and Runtime System Qingyu Meng, Alan Humphrey, Martin Berzins Thanks to: John Schmidt and J. Davison de St. Germain, SCI Institute Justin Luitjens
HPC Cluster Decisions and ANSYS Configuration Best Practices. Diana Collier Lead Systems Support Specialist Houston UGM May 2014
HPC Cluster Decisions and ANSYS Configuration Best Practices Diana Collier Lead Systems Support Specialist Houston UGM May 2014 1 Agenda Introduction Lead Systems Support Specialist Cluster Decisions Job
Power Management in the Linux Kernel
Power Management in the Linux Kernel Tate Hornbeck, Peter Hokanson 7 April 2011 Intel Open Source Technology Center Venkatesh Pallipadi Senior Staff Software Engineer 2001 - Joined Intel Processor and
Maximum performance, minimal risk for data warehousing
SYSTEM X SERVERS SOLUTION BRIEF Maximum performance, minimal risk for data warehousing Microsoft Data Warehouse Fast Track for SQL Server 2014 on System x3850 X6 (95TB) The rapid growth of technology has
Intel Media Server Studio - Metrics Monitor (v1.1.0) Reference Manual
Intel Media Server Studio - Metrics Monitor (v1.1.0) Reference Manual Overview Metrics Monitor is part of Intel Media Server Studio 2015 for Linux Server. Metrics Monitor is a user space shared library
DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION
DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION A DIABLO WHITE PAPER AUGUST 2014 Ricky Trigalo Director of Business Development Virtualization, Diablo Technologies
Contents. 2. cttctx Performance Test Utility... 8. 3. Server Side Plug-In... 9. 4. Index... 11. www.faircom.com All Rights Reserved.
c-treeace Load Test c-treeace Load Test Contents 1. Performance Test Description... 1 1.1 Login Info... 2 1.2 Create Tables... 3 1.3 Run Test... 4 1.4 Last Run Threads... 5 1.5 Total Results History...
GPU System Architecture. Alan Gray EPCC The University of Edinburgh
GPU System Architecture EPCC The University of Edinburgh Outline Why do we want/need accelerators such as GPUs? GPU-CPU comparison Architectural reasons for GPU performance advantages GPU accelerated systems
DELL. Virtual Desktop Infrastructure Study END-TO-END COMPUTING. Dell Enterprise Solutions Engineering
DELL Virtual Desktop Infrastructure Study END-TO-END COMPUTING Dell Enterprise Solutions Engineering 1 THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS AND TECHNICAL
Monitoring Infrastructure for Superclusters: Experiences at MareNostrum
ScicomP13 2007 SP-XXL Monitoring Infrastructure for Superclusters: Experiences at MareNostrum Garching, Munich Ernest Artiaga Performance Group BSC-CNS, Operations Outline BSC-CNS and MareNostrum Overview
Optimizing Shared Resource Contention in HPC Clusters
Optimizing Shared Resource Contention in HPC Clusters Sergey Blagodurov Simon Fraser University Alexandra Fedorova Simon Fraser University Abstract Contention for shared resources in HPC clusters occurs
Seeking Opportunities for Hardware Acceleration in Big Data Analytics
Seeking Opportunities for Hardware Acceleration in Big Data Analytics Paul Chow High-Performance Reconfigurable Computing Group Department of Electrical and Computer Engineering University of Toronto Who
Hybrid Cluster Management: Reducing Stress, increasing productivity and preparing for the future
Hybrid Cluster Management: Reducing Stress, increasing productivity and preparing for the future Clement Lau, Ph. D. Sales Director, APJ Bright Computing Agenda 1.Reduce 2.IncRease 3.PrepaRe Reduce System
Overview on Modern Accelerators and Programming Paradigms Ivan Giro7o [email protected]
Overview on Modern Accelerators and Programming Paradigms Ivan Giro7o [email protected] Informa(on & Communica(on Technology Sec(on (ICTS) Interna(onal Centre for Theore(cal Physics (ICTP) Mul(ple Socket
- An Essential Building Block for Stable and Reliable Compute Clusters
Ferdinand Geier ParTec Cluster Competence Center GmbH, V. 1.4, March 2005 Cluster Middleware - An Essential Building Block for Stable and Reliable Compute Clusters Contents: Compute Clusters a Real Alternative
Managing Adaptability in Heterogeneous Architectures through Performance Monitoring and Prediction
Managing Adaptability in Heterogeneous Architectures through Performance Monitoring and Prediction Cristina Silvano [email protected] Politecnico di Milano HiPEAC CSW Athens 2014 Motivations System
JBoss Data Grid Performance Study Comparing Java HotSpot to Azul Zing
JBoss Data Grid Performance Study Comparing Java HotSpot to Azul Zing January 2014 Legal Notices JBoss, Red Hat and their respective logos are trademarks or registered trademarks of Red Hat, Inc. Azul
Chapter 2: Getting Started
Chapter 2: Getting Started Once Partek Flow is installed, Chapter 2 will take the user to the next stage and describes the user interface and, of note, defines a number of terms required to understand
HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK
HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK Steve Oberlin CTO, Accelerated Computing US to Build Two Flagship Supercomputers SUMMIT SIERRA Partnership for Science 100-300 PFLOPS Peak Performance
Networking Virtualization Using FPGAs
Networking Virtualization Using FPGAs Russell Tessier, Deepak Unnikrishnan, Dong Yin, and Lixin Gao Reconfigurable Computing Group Department of Electrical and Computer Engineering University of Massachusetts,
Siebel System Monitoring and Diagnostics Guide. Version 8.0, Rev. B December 2011
Siebel System Monitoring and Diagnostics Guide Version 8.0, Rev. B December 2011 Copyright 2005, 2011 Oracle and/or its affiliates. All rights reserved. This software and related documentation are provided
Energy Efficient High Performance Computing Power Measurement Methodology. (version 1.2RC2)
Energy Efficient High Performance Computing Power Measurement Methodology (version 1.2RC2) Contents 1 Introduction 6 2 Checklist for Reporting Power Values 7 Quality Level..................................
OpenMP Programming on ScaleMP
OpenMP Programming on ScaleMP Dirk Schmidl [email protected] Rechen- und Kommunikationszentrum (RZ) MPI vs. OpenMP MPI distributed address space explicit message passing typically code redesign
Mitglied der Helmholtz-Gemeinschaft. System monitoring with LLview and the Parallel Tools Platform
Mitglied der Helmholtz-Gemeinschaft System monitoring with LLview and the Parallel Tools Platform November 25, 2014 Carsten Karbach Content 1 LLview 2 Parallel Tools Platform (PTP) 3 Latest features 4
Know your Cluster Bottlenecks and Maximize Performance
Know your Cluster Bottlenecks and Maximize Performance Hands-on training March 2013 Agenda Overview Performance Factors General System Configuration - PCI Express (PCIe) Capabilities - Memory Configuration
LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu
HP reference configuration for entry-level SAS Grid Manager solutions
HP reference configuration for entry-level SAS Grid Manager solutions Up to 864 simultaneous SAS jobs and more than 3 GB/s I/O throughput Technical white paper Table of contents Executive summary... 2
Informationsaustausch für Nutzer des Aachener HPC Clusters
Informationsaustausch für Nutzer des Aachener HPC Clusters Paul Kapinos, Marcus Wagner - 21.05.2015 Informationsaustausch für Nutzer des Aachener HPC Clusters Agenda (The RWTH Compute cluster) Project-based
Long-term monitoring of apparent latency in PREEMPT RT Linux real-time systems
Long-term monitoring of apparent latency in PREEMPT RT Linux real-time systems Carsten Emde Open Source Automation Development Lab (OSADL) eg Aichhalder Str. 39, 78713 Schramberg, Germany [email protected]
Siebel System Monitoring and Diagnostics Guide. Siebel Innovation Pack 2013 Version 8.1/8.2 September 2013
Siebel System Monitoring and Diagnostics Guide Siebel Innovation Pack 2013 Version 8.1/8.2 September 2013 Copyright 2005, 2013 Oracle and/or its affiliates. All rights reserved. This software and related
Performance Tuning Guidelines for PowerExchange for Microsoft Dynamics CRM
Performance Tuning Guidelines for PowerExchange for Microsoft Dynamics CRM 1993-2016 Informatica LLC. No part of this document may be reproduced or transmitted in any form, by any means (electronic, photocopying,
Performance Evaluation of NAS Parallel Benchmarks on Intel Xeon Phi
Performance Evaluation of NAS Parallel Benchmarks on Intel Xeon Phi ICPP 6 th International Workshop on Parallel Programming Models and Systems Software for High-End Computing October 1, 2013 Lyon, France
HP ProLiant BL660c Gen9 and Microsoft SQL Server 2014 technical brief
Technical white paper HP ProLiant BL660c Gen9 and Microsoft SQL Server 2014 technical brief Scale-up your Microsoft SQL Server environment to new heights Table of contents Executive summary... 2 Introduction...
FPGA-based Multithreading for In-Memory Hash Joins
FPGA-based Multithreading for In-Memory Hash Joins Robert J. Halstead, Ildar Absalyamov, Walid A. Najjar, Vassilis J. Tsotras University of California, Riverside Outline Background What are FPGAs Multithreaded
BSC vision on Big Data and extreme scale computing
BSC vision on Big Data and extreme scale computing Jesus Labarta, Eduard Ayguade,, Fabrizio Gagliardi, Rosa M. Badia, Toni Cortes, Jordi Torres, Adrian Cristal, Osman Unsal, David Carrera, Yolanda Becerra,
Cloud Computing with Red Hat Solutions. Sivaram Shunmugam Red Hat Asia Pacific Pte Ltd. [email protected]
Cloud Computing with Red Hat Solutions Sivaram Shunmugam Red Hat Asia Pacific Pte Ltd [email protected] Linux Automation Details Red Hat's Linux Automation strategy for next-generation IT infrastructure
Next Generation Operating Systems
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015 The end of CPU scaling Future computing challenges Power efficiency Performance == parallelism Cisco Confidential 2 Paradox of the
A Taxonomy and Survey of Energy-Efficient Data Centers and Cloud Computing Systems
A Taxonomy and Survey of Energy-Efficient Data Centers and Cloud Computing Systems Anton Beloglazov, Rajkumar Buyya, Young Choon Lee, and Albert Zomaya Present by Leping Wang 1/25/2012 Outline Background
An Oracle White Paper March 2013. Load Testing Best Practices for Oracle E- Business Suite using Oracle Application Testing Suite
An Oracle White Paper March 2013 Load Testing Best Practices for Oracle E- Business Suite using Oracle Application Testing Suite Executive Overview... 1 Introduction... 1 Oracle Load Testing Setup... 2
Energy-aware job scheduler for highperformance
Energy-aware job scheduler for highperformance computing 7.9.2011 Olli Mämmelä (VTT), Mikko Majanen (VTT), Robert Basmadjian (University of Passau), Hermann De Meer (University of Passau), André Giesler
PTC System Monitor Solution Training
PTC System Monitor Solution Training Patrick Kulenkamp June 2012 Agenda What is PTC System Monitor (PSM)? How does it work? Terminology PSM Configuration The PTC Integrity Implementation Drilling Down
HP PCIe IO Accelerator For Proliant Rackmount Servers And BladeSystems
WHITE PAPER HP PCIe IO Accelerator For Proliant Rackmount Servers And BladeSystems 951 SanDisk Drive, Milpitas, CA 95035 www.sandisk.com Table of Contents Overview & Features... 3 QuickSpecs...3 HP Supported
The QEMU/KVM Hypervisor
The /KVM Hypervisor Understanding what's powering your virtual machine Dr. David Alan Gilbert [email protected] 2015-10-14 Topics Hypervisors and where /KVM sits Components of a virtual machine KVM Devices:
Pros and Cons of HPC Cloud Computing
CloudStat 211 Pros and Cons of HPC Cloud Computing Nils gentschen Felde Motivation - Idea HPC Cluster HPC Cloud Cluster Management benefits of virtual HPC Dynamical sizing / partitioning Loadbalancing
Integration of Virtualized Workernodes in Batch Queueing Systems The ViBatch Concept
Integration of Virtualized Workernodes in Batch Queueing Systems, Dr. Armin Scheurer, Oliver Oberst, Prof. Günter Quast INSTITUT FÜR EXPERIMENTELLE KERNPHYSIK FAKULTÄT FÜR PHYSIK KIT University of the
RWTH GPU Cluster. Sandra Wienke [email protected] November 2012. Rechen- und Kommunikationszentrum (RZ) Fotos: Christian Iwainsky
RWTH GPU Cluster Fotos: Christian Iwainsky Sandra Wienke [email protected] November 2012 Rechen- und Kommunikationszentrum (RZ) The RWTH GPU Cluster GPU Cluster: 57 Nvidia Quadro 6000 (Fermi) innovative
Introduction to Cloud Computing
Introduction to Cloud Computing Parallel Processing I 15 319, spring 2010 7 th Lecture, Feb 2 nd Majd F. Sakr Lecture Motivation Concurrency and why? Different flavors of parallel computing Get the basic
Work Environment. David Tur HPC Expert. HPC Users Training September, 18th 2015
Work Environment David Tur HPC Expert HPC Users Training September, 18th 2015 1. Atlas Cluster: Accessing and using resources 2. Software Overview 3. Job Scheduler 1. Accessing Resources DIPC technicians
High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/2015. 2015 CAE Associates
High Performance Computing (HPC) CAEA elearning Series Jonathan G. Dudley, Ph.D. 06/09/2015 2015 CAE Associates Agenda Introduction HPC Background Why HPC SMP vs. DMP Licensing HPC Terminology Types of
Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage
White Paper Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage A Benchmark Report August 211 Background Objectivity/DB uses a powerful distributed processing architecture to manage
The Moab Scheduler. Dan Mazur, McGill HPC [email protected] Aug 23, 2013
The Moab Scheduler Dan Mazur, McGill HPC [email protected] Aug 23, 2013 1 Outline Fair Resource Sharing Fairness Priority Maximizing resource usage MAXPS fairness policy Minimizing queue times Should
Estimate Performance and Capacity Requirements for Workflow in SharePoint Server 2010
Estimate Performance and Capacity Requirements for Workflow in SharePoint Server 2010 This document is provided as-is. Information and views expressed in this document, including URL and other Internet
An HPC Application Deployment Model on Azure Cloud for SMEs
An HPC Application Deployment Model on Azure Cloud for SMEs Fan Ding CLOSER 2013, Aachen, Germany, May 9th,2013 Rechen- und Kommunikationszentrum (RZ) Agenda Motivation Windows Azure Relevant Technology
BIRT Document Transform
BIRT Document Transform BIRT Document Transform is the industry leader in enterprise-class, high-volume document transformation. It transforms and repurposes high-volume documents and print streams such
Simulation Platform Overview
Simulation Platform Overview Build, compute, and analyze simulations on demand www.rescale.com CASE STUDIES Companies in the aerospace and automotive industries use Rescale to run faster simulations Aerospace
Cisco EnergyWise Return on Investment with Cisco Unified IP Phones
Solution Overview Cisco EnergyWise Return on Investment with Cisco Unified IP Phones As organizations strive to reduce their energy consumption, Cisco has responded with an energy-management architecture
Running a Workflow on a PowerCenter Grid
Running a Workflow on a PowerCenter Grid 2010-2014 Informatica Corporation. No part of this document may be reproduced or transmitted in any form, by any means (electronic, photocopying, recording or otherwise)
Technical Paper. Moving SAS Applications from a Physical to a Virtual VMware Environment
Technical Paper Moving SAS Applications from a Physical to a Virtual VMware Environment Release Information Content Version: April 2015. Trademarks and Patents SAS Institute Inc., SAS Campus Drive, Cary,
NoSQL Performance Test In-Memory Performance Comparison of SequoiaDB, Cassandra, and MongoDB
bankmark UG (haftungsbeschränkt) Bahnhofstraße 1 9432 Passau Germany www.bankmark.de [email protected] T +49 851 25 49 49 F +49 851 25 49 499 NoSQL Performance Test In-Memory Performance Comparison of SequoiaDB,
