Ra - Batch Scripts. Timothy H. Kaiser, Ph.D. tkaiser@mines.edu
|
|
- Joella Flowers
- 7 years ago
- Views:
Transcription
1 Ra - Batch Scripts Timothy H. Kaiser, Ph.D. tkaiser@mines.edu
2 Jobs on Ra are Run via a Batch System Ra is a shared resource Purpose: Give fair access to all users Have control over where jobs are run Set limits on times and processor counts Allow exclusive access to nodes while you own them
3 Batch Systems Create a script that describes: What you want to do Resources Time required What resources (# nodes) you want How long you want the resources Submit you script to the queuing system Wait for some time for your job to start Your job runs and you get your output
4 Moab -Top level Batch System Product of Cluster Resources products/moab-cluster-suite/workloadmanager.php Marketing pitch from the web page: Moab Workload Manager is a policy-based job scheduler and event engine that enables utility-based computing for clusters. Moab Workload Manager combines intelligent scheduling of resources with advanced reservations to process jobs on the right resources at the right time. It also provides flexible policy and event engines that process workloads faster and in line with set business requirements and priorities.
5 Moab Set up queues policy Hands actual job submission to Torque Torque - a.k.a. pbs Torque is a descendent of the open source version of PBS Moab has the commands of pbs but adds some. The most important is msub
6 A Simple Script for a MPI job!/bin/bash #PBS -l nodes=1:ppn=8 #PBS -l walltime=02:00:00 #PBS -N testio #PBS -o stdout #PBS -e stderr #PBS -V #PBS -m abe #PBS -M tkaiser@mines.edu # cd $PBS_O_WORKDIR mpiexec -n 8 c_ex00 Scripts contain normal shell commands and comments designated with a # that are interpreted by PBS
7 Running and monitoring jobs ~/mydir]$ msub mpmd Use msub to submit the job qstat to see its status (Q,R,C) qdel to kill the job if needed 8452 ~/mydir]$ qstat 8452 Job id Name User Time Use S Queue ra testio tkaiser 0 R SHORT [tkaiser@ra ~/mydir]$ [tkaiser@ra ~/mydir]$ [tkaiser@ra ~/mydir]$ qstat -u tkaiser Job id Name User Time Use S Queue ra testio tkaiser 0 R SHORT [tkaiser@ra ~/mydir]$ [tkaiser@ra ~/mydir]$ [tkaiser@ra ~/mydir]$ qdel 8452
8 A Simple Script for a MPI job Command!/bin/bash #PBS -l nodes=1:ppn=8 #PBS -l walltime=02:00:00 #PBS -N testio #PBS -o stdout #PBS -e stderr #PBS -V #PBS -m abe #PBS -M tkaiser@mines.edu Comment We will run this job using the "bash" shell We want 1 node with 8 processors for a total of 8 processors We will run for up to 2 hours The name of our job is testio Standard output from our program will go to a file stdout Error output from our program will go to a file stderr Very important! Exports all environment variables from the submitting shell into the batch shell. Send mail on abort, begin, end use n to suppress Where to send to (cell phone?) # Not important. Just a separator line. cd $PBS_O_WORKDIR mpiexec -n 8 c_ex00 Very important! Go to directory $PBS_O_WORKDIR which is the directory which is where our script resides Run the MPI program c_ex00 on 8 computing cores.
9 #PBS -l More Script notes Used to specify resource requests Can have multiple per -l lines per script #PBS -l nodes=2:ppn=8:fat Ra has 16 Gbyte and 32 Gbyte nodes Specify the 32 Gbyte nodes #PBS -l naccesspolicy=singlejob Grant exclusive access to a node Not normally needed
10 PBS Variables Variable PBS_JOBID PBS_O_WORKDIR PBS_NNODES PBS_NODEFILE PBS_O_HOME PBS_JOBNAME PBS_QUEUE PBS_MOMPORT PBS_O_HOST PBS_O_LANG PBS_O_LOGNAME PBS_O_PATH PBS_O_SHELL PBS_JOBCOOKIE PBS_TASKNUM PBS_NODENUM Meaning unique PBS job ID jobs submission directory number of nodes requested file with list of allocated nodes home dir of submitting user user specified job name job queue active port for mom daemon host of currently running job language variable for job name of submitting user path to executables used in job script script shell job cookie number of tasks requested (see pbsdsh) node offset number (see pbsdsh)
11 Where s my Output? Standard Error and Standard Out don t show up until after your job is completed Output is temporarily stored in /tmp on your nodes You can ssh to your nodes and see your output Looking into changing this behavior, or at least make it optional You can pipe output to a file: mpiexec myjob > myjob.$pbs_jobid
12 Useful Moab Commands Command canceljob checkjob mdiag mjobctl mrsvctl mshow msub releasehold releaseres sethold showq showres showstart showstate cancel job Description provide detailed status report for specified job provide diagnostic reports for resources, workload, and scheduling control and modify job create, control and modify reservations displays various diagnostic messages about the system and job queues submit a job (Donʼt use qsub) release job defers and holds release reservations set job holds show queued jobs show existing reservations show estimates of when job can/will start show current state of resources
13 Useful Torque commands Command tracejob pbsnodes qalter qdel qhold qrls qrun qsig qstat pbsdsh qsub Description trace job actions and states recorded in TORQUE logs view/modify batch status of compute nodes modify queued batch jobs delete/cancel batch jobs hold batch jobs release batch job holds start a batch job send a signal to a batch job view queues and jobs launch tasks within a parallel job submit a job (Use msub instead)
14 Running Interactive
15 Interactive Runs It is possible to grab a collection of nodes and run parallel interactively Nodes are yours until you quit Might be useful during development Running a number of small parallel jobs Required (almost) for running parallel debuggers
16 Multistep Process Ask for the nodes using msub msub -q INTERACTIVE -I -V -l nodes=1 Go to your working directory Run your job(s) Logout to release your nodes You are automatically Logged on to the nodes you are given Ignore: cp: cannot stat `/opt/torque/mom_priv/jobs/73156.ra.local.sc': No such file or directory Don t abuse it, lest you be shot.
17 Example ~/guide]$msub -q INTERACTIVE -I -V -l nodes=1 -l naccesspolicy=singlejob msub: waiting for job 1280.ra.mines.edu to start msub: job 1280.ra.mines.edu ready ~]$cd guide ~/guide]$mpiexec -n 4 c_ex00 Hello from 0 of 4 on compute-9-8.local Hello from 1 of 4 on compute-9-8.local Hello from 2 of 4 on compute-9-8.local Hello from 3 of 4 on compute-9-8.local [tkaiser@compute-9-8 ~/guide]$ [tkaiser@compute-9-8 ~/guide]$mpiexec -n 16 c_ex00 Hello from 1 of 16 on compute-9-8.local Hello from 0 of 16 on compute-9-8.local Hello from 3 of 16 on compute-9-8.local Hello from 4 of 16 on compute-9-8.local Hello from 15 of 16 on compute-9-8.local Hello from 2 of 16 on compute-9-8.local Hello from 6 of 16 on compute-9-8.local [tkaiser@compute-9-8 ~/guide]$ [tkaiser@compute-9-8 ~/guide]$exit logout Ensures exclusive access msub: job 1280.ra.mines.edu completed [tkaiser@ra ~/guide]$
Batch Scripts for RA & Mio
Batch Scripts for RA & Mio Timothy H. Kaiser, Ph.D. tkaiser@mines.edu 1 Jobs are Run via a Batch System Ra and Mio are shared resources Purpose: Give fair access to all users Have control over where jobs
More informationJob Scheduling with Moab Cluster Suite
Job Scheduling with Moab Cluster Suite IBM High Performance Computing February 2010 Y. Joanna Wong, Ph.D. yjw@us.ibm.com 2/22/2010 Workload Manager Torque Source: Adaptive Computing 2 Some terminology..
More informationQuick Tutorial for Portable Batch System (PBS)
Quick Tutorial for Portable Batch System (PBS) The Portable Batch System (PBS) system is designed to manage the distribution of batch jobs and interactive sessions across the available nodes in the cluster.
More informationResource Management and Job Scheduling
Resource Management and Job Scheduling Jenett Tillotson Senior Cluster System Administrator Indiana University May 18 18-22 May 2015 1 Resource Managers Keep track of resources Nodes: CPUs, disk, memory,
More informationJob scheduler details
Job scheduler details Advanced Computing Center for Research & Education (ACCRE) Job scheduler details 1 / 25 Outline 1 Batch queue system overview 2 Torque and Moab 3 Submitting jobs (ACCRE) Job scheduler
More informationMiami University RedHawk Cluster Working with batch jobs on the Cluster
Miami University RedHawk Cluster Working with batch jobs on the Cluster The RedHawk cluster is a general purpose research computing resource available to support the research community at Miami University.
More informationHodor and Bran - Job Scheduling and PBS Scripts
Hodor and Bran - Job Scheduling and PBS Scripts UND Computational Research Center Now that you have your program compiled and your input file ready for processing, it s time to run your job on the cluster.
More informationSLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education www.accre.vanderbilt.
SLURM: Resource Management and Job Scheduling Software Advanced Computing Center for Research and Education www.accre.vanderbilt.edu Simple Linux Utility for Resource Management But it s also a job scheduler!
More informationJuropa. Batch Usage Introduction. May 2014 Chrysovalantis Paschoulas c.paschoulas@fz-juelich.de
Juropa Batch Usage Introduction May 2014 Chrysovalantis Paschoulas c.paschoulas@fz-juelich.de Batch System Usage Model A Batch System: monitors and controls the resources on the system manages and schedules
More informationSLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education www.accre.vanderbilt.
SLURM: Resource Management and Job Scheduling Software Advanced Computing Center for Research and Education www.accre.vanderbilt.edu Simple Linux Utility for Resource Management But it s also a job scheduler!
More informationPBS Tutorial. Fangrui Ma Universit of Nebraska-Lincoln. October 26th, 2007
PBS Tutorial Fangrui Ma Universit of Nebraska-Lincoln October 26th, 2007 Abstract In this tutorial we gave a brief introduction to using PBS Pro. We gave examples on how to write control script, and submit
More informationIntroduction to Sun Grid Engine (SGE)
Introduction to Sun Grid Engine (SGE) What is SGE? Sun Grid Engine (SGE) is an open source community effort to facilitate the adoption of distributed computing solutions. Sponsored by Sun Microsystems
More informationTORQUE Administrator s Guide. version 2.3
TORQUE Administrator s Guide version 2.3 Copyright 2009 Cluster Resources, Inc. All rights reserved. Trademarks Cluster Resources, Moab, Moab Workload Manager, Moab Cluster Manager, Moab Cluster Suite,
More informationNYUAD HPC Center Running Jobs
NYUAD HPC Center Running Jobs 1 Overview... Error! Bookmark not defined. 1.1 General List... Error! Bookmark not defined. 1.2 Compilers... Error! Bookmark not defined. 2 Loading Software... Error! Bookmark
More informationRA MPI Compilers Debuggers Profiling. March 25, 2009
RA MPI Compilers Debuggers Profiling March 25, 2009 Examples and Slides To download examples on RA 1. mkdir class 2. cd class 3. wget http://geco.mines.edu/workshop/class2/examples/examples.tgz 4. tar
More informationLinux für bwgrid. Sabine Richling, Heinz Kredel. Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim. 27.
Linux für bwgrid Sabine Richling, Heinz Kredel Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim 27. June 2011 Richling/Kredel (URZ/RUM) Linux für bwgrid FS 2011 1 / 33 Introduction
More informationWork Environment. David Tur HPC Expert. HPC Users Training September, 18th 2015
Work Environment David Tur HPC Expert HPC Users Training September, 18th 2015 1. Atlas Cluster: Accessing and using resources 2. Software Overview 3. Job Scheduler 1. Accessing Resources DIPC technicians
More informationIntroduction to Running Hadoop on the High Performance Clusters at the Center for Computational Research
Introduction to Running Hadoop on the High Performance Clusters at the Center for Computational Research Cynthia Cornelius Center for Computational Research University at Buffalo, SUNY 701 Ellicott St
More informationRunning applications on the Cray XC30 4/12/2015
Running applications on the Cray XC30 4/12/2015 1 Running on compute nodes By default, users do not log in and run applications on the compute nodes directly. Instead they launch jobs on compute nodes
More informationTutorial: Using WestGrid. Drew Leske Compute Canada/WestGrid Site Lead University of Victoria
Tutorial: Using WestGrid Drew Leske Compute Canada/WestGrid Site Lead University of Victoria Fall 2013 Seminar Series Date Speaker Topic 23 September Lindsay Sill Introduction to WestGrid 9 October Drew
More informationGuillimin HPC Users Meeting. Bryan Caron
November 13, 2014 Bryan Caron bryan.caron@mcgill.ca bryan.caron@calculquebec.ca McGill University / Calcul Québec / Compute Canada Montréal, QC Canada Outline Compute Canada News October Service Interruption
More informationGrid 101. Grid 101. Josh Hegie. grid@unr.edu http://hpc.unr.edu
Grid 101 Josh Hegie grid@unr.edu http://hpc.unr.edu Accessing the Grid Outline 1 Accessing the Grid 2 Working on the Grid 3 Submitting Jobs with SGE 4 Compiling 5 MPI 6 Questions? Accessing the Grid Logging
More informationGetting Started with HPC
Getting Started with HPC An Introduction to the Minerva High Performance Computing Resource 17 Sep 2013 Outline of Topics Introduction HPC Accounts Logging onto the HPC Clusters Common Linux Commands Storage
More informationSEE-GRID-SCI. Cluster installation and configuration. www.see-grid.eu. SEE-GRID-SCI Training Event, Yerevan, Armenia, 24-25 July 2008
SEE-GRID-SCI Cluster installation and configuration www.see-grid.eu SEE-GRID-SCI Training Event, Yerevan, Armenia, 24-25 July 2008 Mikayel Gyurjyan Institute for Informatics and Automation Problems National
More informationHow To Run A Cluster On A Linux Server On A Pcode 2.5.2.2 (Amd64) On A Microsoft Powerbook 2.6.2 2.4.2 On A Macbook 2 (Amd32)
UNIVERSIDAD REY JUAN CARLOS Máster Universitario en Software Libre Curso Académico 2012/2013 Campus Fuenlabrada, Madrid, España MSWL-THESIS - Proyecto Fin de Master Distributed Batch Processing Autor:
More informationUsing WestGrid. Patrick Mann, Manager, Technical Operations Jan.15, 2014
Using WestGrid Patrick Mann, Manager, Technical Operations Jan.15, 2014 Winter 2014 Seminar Series Date Speaker Topic 5 February Gino DiLabio Molecular Modelling Using HPC and Gaussian 26 February Jonathan
More informationCaltech Center for Advanced Computing Research System Guide: MRI2 Cluster (zwicky) January 2014
1. How to Get An Account CACR Accounts 2. How to Access the Machine Connect to the front end, zwicky.cacr.caltech.edu: ssh -l username zwicky.cacr.caltech.edu or ssh username@zwicky.cacr.caltech.edu Edits,
More informationThe Moab Scheduler. Dan Mazur, McGill HPC daniel.mazur@mcgill.ca Aug 23, 2013
The Moab Scheduler Dan Mazur, McGill HPC daniel.mazur@mcgill.ca Aug 23, 2013 1 Outline Fair Resource Sharing Fairness Priority Maximizing resource usage MAXPS fairness policy Minimizing queue times Should
More informationUsing Parallel Computing to Run Multiple Jobs
Beowulf Training Using Parallel Computing to Run Multiple Jobs Jeff Linderoth August 5, 2003 August 5, 2003 Beowulf Training Running Multiple Jobs Slide 1 Outline Introduction to Scheduling Software The
More informationIntroduction to the SGE/OGS batch-queuing system
Grid Computing Competence Center Introduction to the SGE/OGS batch-queuing system Riccardo Murri Grid Computing Competence Center, Organisch-Chemisches Institut, University of Zurich Oct. 6, 2011 The basic
More informationNEC HPC-Linux-Cluster
NEC HPC-Linux-Cluster Hardware configuration: 4 Front-end servers: each with SandyBridge-EP processors: 16 cores per node 128 GB memory 134 compute nodes: 112 nodes with SandyBridge-EP processors (16 cores
More informationRunning on Blue Gene/Q at Argonne Leadership Computing Facility (ALCF)
Running on Blue Gene/Q at Argonne Leadership Computing Facility (ALCF) ALCF Resources: Machines & Storage Mira (Production) IBM Blue Gene/Q 49,152 nodes / 786,432 cores 768 TB of memory Peak flop rate:
More informationUsing the Yale HPC Clusters
Using the Yale HPC Clusters Stephen Weston Robert Bjornson Yale Center for Research Computing Yale University Oct 2015 To get help Send an email to: hpc@yale.edu Read documentation at: http://research.computing.yale.edu/hpc-support
More informationHPC at IU Overview. Abhinav Thota Research Technologies Indiana University
HPC at IU Overview Abhinav Thota Research Technologies Indiana University What is HPC/cyberinfrastructure? Why should you care? Data sizes are growing Need to get to the solution faster Compute power is
More informationGrid Engine 6. Troubleshooting. BioTeam Inc. info@bioteam.net
Grid Engine 6 Troubleshooting BioTeam Inc. info@bioteam.net Grid Engine Troubleshooting There are two core problem types Job Level Cluster seems OK, example scripts work fine Some user jobs/apps fail Cluster
More informationJob Scheduling Explained More than you ever want to know about how jobs get scheduled on WestGrid systems...
Job Scheduling Explained More than you ever want to know about how jobs get scheduled on WestGrid systems... Martin Siegert, SFU Cluster Myths There are so many jobs in the queue - it will take ages until
More informationSubmitting and Running Jobs on the Cray XT5
Submitting and Running Jobs on the Cray XT5 Richard Gerber NERSC User Services RAGerber@lbl.gov Joint Cray XT5 Workshop UC-Berkeley Outline Hopper in blue; Jaguar in Orange; Kraken in Green XT5 Overview
More informationThe RWTH Compute Cluster Environment
The RWTH Compute Cluster Environment Tim Cramer 11.03.2013 Source: D. Both, Bull GmbH Rechen- und Kommunikationszentrum (RZ) How to login Frontends cluster.rz.rwth-aachen.de cluster-x.rz.rwth-aachen.de
More informationGrid Engine Basics. Table of Contents. Grid Engine Basics Version 1. (Formerly: Sun Grid Engine)
Grid Engine Basics (Formerly: Sun Grid Engine) Table of Contents Table of Contents Document Text Style Associations Prerequisites Terminology What is the Grid Engine (SGE)? Loading the SGE Module on Turing
More informationHigh-Performance Reservoir Risk Assessment (Jacta Cluster)
High-Performance Reservoir Risk Assessment (Jacta Cluster) SKUA-GOCAD 2013.1 Paradigm 2011.3 With Epos 4.1 Data Management Configuration Guide 2008 2013 Paradigm Ltd. or its affiliates and subsidiaries.
More informationBatch Systems. provide a mechanism for submitting, launching, and tracking jobs on a shared resource
PBS INTERNALS PBS & TORQUE PBS (Portable Batch System)-software system for managing system resources on workstations, SMP systems, MPPs and vector computers. It was based on Network Queuing System (NQS)
More informationIntroduction to SDSC systems and data analytics software packages "
Introduction to SDSC systems and data analytics software packages " Mahidhar Tatineni (mahidhar@sdsc.edu) SDSC Summer Institute August 05, 2013 Getting Started" System Access Logging in Linux/Mac Use available
More informationHPC system startup manual (version 1.30)
HPC system startup manual (version 1.30) Document change log Issue Date Change 1 12/1/2012 New document 2 10/22/2013 Added the information of supported OS 3 10/22/2013 Changed the example 1 for data download
More informationInstalling and running COMSOL on a Linux cluster
Installing and running COMSOL on a Linux cluster Introduction This quick guide explains how to install and operate COMSOL Multiphysics 5.0 on a Linux cluster. It is a complement to the COMSOL Installation
More informationHigh Performance Computing Facility Specifications, Policies and Usage. Supercomputer Project. Bibliotheca Alexandrina
High Performance Computing Facility Specifications, Policies and Usage Supercomputer Project Bibliotheca Alexandrina Bibliotheca Alexandrina 1/16 Topics Specifications Overview Site Policies Intel Compilers
More informationPBS Professional 12.1
PBS Professional 12.1 PBS Works is a division of PBS Professional 12.1 User s Guide, updated 5/16/13. Copyright 2003-2013 Altair Engineering, Inc. All rights reserved. PBS, PBS Works, PBS GridWorks, PBS
More informationParallel Debugging with DDT
Parallel Debugging with DDT Nate Woody 3/10/2009 www.cac.cornell.edu 1 Debugging Debugging is a methodical process of finding and reducing the number of bugs, or defects, in a computer program or a piece
More informationMartinos Center Compute Clusters
Intro What are the compute clusters How to gain access Housekeeping Usage Log In Submitting Jobs Queues Request CPUs/vmem Email Status I/O Interactive Dependencies Daisy Chain Wrapper Script In Progress
More informationFile System-Aware Job Scheduling with Moab
LLNL-TM-409992 File System-Aware Job Scheduling with Moab D.A. Lipari, P.D. Eckert July 8, 2009 This document was prepared as an account of work sponsored by an agency of the United States government.
More informationHow To Run A Steady Case On A Creeper
Crash Course Introduction to OpenFOAM Artur Lidtke University of Southampton akl1g09@soton.ac.uk November 4, 2014 Artur Lidtke Crash Course Introduction to OpenFOAM 1 / 32 What is OpenFOAM? Using OpenFOAM
More informationThe SUN ONE Grid Engine BATCH SYSTEM
The SUN ONE Grid Engine BATCH SYSTEM Juan Luis Chaves Sanabria Centro Nacional de Cálculo Científico (CeCalCULA) Latin American School in HPC on Linux Cluster October 27 November 07 2003 What is SGE? Is
More informationTORQUE Resource Manager
TORQUE Resource Manager Administrator Guide 4.2.5 September 2013 2013 Adaptive Computing Enterprises Inc. All rights reserved. Distribution of this document for commercial purposes in either hard or soft
More informationCluster@WU User s Manual
Cluster@WU User s Manual Stefan Theußl Martin Pacala September 29, 2014 1 Introduction and scope At the WU Wirtschaftsuniversität Wien the Research Institute for Computational Methods (Forschungsinstitut
More informationBeyond Windows: Using the Linux Servers and the Grid
Beyond Windows: Using the Linux Servers and the Grid Topics Linux Overview How to Login & Remote Access Passwords Staying Up-To-Date Network Drives Server List The Grid Useful Commands Linux Overview Linux
More informationPBS Professional 11.2. User s Guide. PBS Works is a division of
PBS Professional 11.2 User s Guide PBS Works is a division of PBS Professional User s Guide, Altair PBS Professional 11.2, Updated: 12/16/11. Edited by: Anne Urban Copyright 2003-2011 Altair Engineering,
More information1.0. User Manual For HPC Cluster at GIKI. Volume. Ghulam Ishaq Khan Institute of Engineering Sciences & Technology
Volume 1.0 FACULTY OF CUMPUTER SCIENCE & ENGINEERING Ghulam Ishaq Khan Institute of Engineering Sciences & Technology User Manual For HPC Cluster at GIKI Designed and prepared by Faculty of Computer Science
More informationAn Introduction to High Performance Computing in the Department
An Introduction to High Performance Computing in the Department Ashley Ford & Chris Jewell Department of Statistics University of Warwick October 30, 2012 1 Some Background 2 How is Buster used? 3 Software
More informationGrid Engine Users Guide. 2011.11p1 Edition
Grid Engine Users Guide 2011.11p1 Edition Grid Engine Users Guide : 2011.11p1 Edition Published Nov 01 2012 Copyright 2012 University of California and Scalable Systems This document is subject to the
More informationA High Performance Computing Scheduling and Resource Management Primer
LLNL-TR-652476 A High Performance Computing Scheduling and Resource Management Primer D. H. Ahn, J. E. Garlick, M. A. Grondona, D. A. Lipari, R. R. Springmeyer March 31, 2014 Disclaimer This document was
More informationGrid Engine Training Introduction
Grid Engine Training Jordi Blasco (jordi.blasco@xrqtc.org) 26-03-2012 Agenda 1 How it works? 2 History Current status future About the Grid Engine version of this training Documentation 3 Grid Engine internals
More informationHOD Scheduler. Table of contents
Table of contents 1 Introduction... 2 2 HOD Users... 2 2.1 Getting Started... 2 2.2 HOD Features...5 2.3 Troubleshooting... 14 3 HOD Administrators... 21 3.1 Getting Started... 22 3.2 Prerequisites...
More informationAltair. PBS Pro. User Guide 5.4. for UNIX, Linux, and Windows
Altair PBS Pro TM User Guide 5.4 for UNIX, Linux, and Windows Portable Batch System TM User Guide PBS-3BA01: Altair PBS Pro TM 5.4.2, Updated: December 15, 2004 Edited by: James Patton Jones Copyright
More informationHack the Gibson. John Fitzpatrick Luke Jennings. Exploiting Supercomputers. 44Con Edition September 2013. Public EXTERNAL
Hack the Gibson Exploiting Supercomputers 44Con Edition September 2013 John Fitzpatrick Luke Jennings Labs.mwrinfosecurity.com MWR Labs Labs.mwrinfosecurity.com MWR Labs 1 Outline Introduction Important
More informationUsing the Millipede cluster - I
Using the Millipede cluster - I Fokke Dijkstra, Bob Dröge High Performance Computing and Visualisation group Donald Smits Centre for Information Technology General introduction Course aimed at beginners
More informationThe CNMS Computer Cluster
The CNMS Computer Cluster This page describes the CNMS Computational Cluster, how to access it, and how to use it. Introduction (2014) The latest block of the CNMS Cluster (2010) Previous blocks of the
More informationStreamline Computing Linux Cluster User Training. ( Nottingham University)
1 Streamline Computing Linux Cluster User Training ( Nottingham University) 3 User Training Agenda System Overview System Access Description of Cluster Environment Code Development Job Schedulers Running
More informationThe Asterope compute cluster
The Asterope compute cluster ÅA has a small cluster named asterope.abo.fi with 8 compute nodes Each node has 2 Intel Xeon X5650 processors (6-core) with a total of 24 GB RAM 2 NVIDIA Tesla M2050 GPGPU
More informationMFCF Grad Session 2015
MFCF Grad Session 2015 Agenda Introduction Help Centre and requests Dept. Grad reps Linux clusters using R with MPI Remote applications Future computing direction Technical question and answer period MFCF
More informationCluster Computing With R
Cluster Computing With R Stowers Institute for Medical Research R/Bioconductor Discussion Group Earl F. Glynn Scientific Programmer 18 December 2007 1 Cluster Computing With R Accessing Linux Boxes from
More informationEfficient cluster computing
Efficient cluster computing Introduction to the Sun Grid Engine (SGE) queuing system Markus Rampp (RZG, MIGenAS) MPI for Evolutionary Anthropology Leipzig, Feb. 16, 2007 Outline Introduction Basic concepts:
More informationSGE Roll: Users Guide. Version @VERSION@ Edition
SGE Roll: Users Guide Version @VERSION@ Edition SGE Roll: Users Guide : Version @VERSION@ Edition Published Aug 2006 Copyright 2006 UC Regents, Scalable Systems Table of Contents Preface...i 1. Requirements...1
More informationMaui Administrator's Guide
Overview Maui Administrator's Guide Maui 3.2 Last Updated May 16 The Maui Scheduler can be thought of as a policy engine which allows sites control over when, where, and how resources such as processors,
More informationEnigma, Sun Grid Engine (SGE), and the Joint High Performance Computing Exchange (JHPCE) Cluster
Enigma, Sun Grid Engine (SGE), and the Joint High Performance Computing Exchange (JHPCE) Cluster http://www.biostat.jhsph.edu/bit/sge_lecture.ppt.pdf Marvin Newhouse Fernando J. Pineda The JHPCE staff:
More informationBatch Job Management with Torque/OpenPBS
Batch Job Management with Torque/OpenPBS The batch system on titan uses OpenPBS, a free customizable batch system. Jobs are submitted by users with qsub from titan.physics.umass.edu, and are scheduled
More informationIntroduction to Sun Grid Engine 5.3
CHAPTER 1 Introduction to Sun Grid Engine 5.3 This chapter provides background information about the Sun Grid Engine 5.3 system that is useful to users and administrators alike. In addition to a description
More informationHigh Performance Computing with Sun Grid Engine on the HPSCC cluster. Fernando J. Pineda
High Performance Computing with Sun Grid Engine on the HPSCC cluster Fernando J. Pineda HPSCC High Performance Scientific Computing Center (HPSCC) " The Johns Hopkins Service Center in the Dept. of Biostatistics
More informationLANL Computing Environment for PSAAP Partners
LANL Computing Environment for PSAAP Partners Robert Cunningham rtc@lanl.gov HPC Systems Group (HPC-3) July 2011 LANL Resources Available To Alliance Users Mapache is new, has a Lobo-like allocation Linux
More informationSubmitting batch jobs Slurm on ecgate. Xavi Abellan xavier.abellan@ecmwf.int User Support Section
Submitting batch jobs Slurm on ecgate Xavi Abellan xavier.abellan@ecmwf.int User Support Section Slide 1 Outline Interactive mode versus Batch mode Overview of the Slurm batch system on ecgate Batch basic
More informationManual for using Super Computing Resources
Manual for using Super Computing Resources Super Computing Research and Education Centre at Research Centre for Modeling and Simulation National University of Science and Technology H-12 Campus, Islamabad
More informationHeterogeneous Clustering- Operational and User Impacts
Heterogeneous Clustering- Operational and User Impacts Sarita Salm Sterling Software MS 258-6 Moffett Field, CA 94035.1000 sarita@nas.nasa.gov http :llscience.nas.nasa.govl~sarita ABSTRACT Heterogeneous
More informationTechnical Support. Copyright notice does not imply publication. For more information, contact Altair at: Web: www.pbsgridworks.com pbssales@altair.
A division of PBS Professional User s Guide, Altair PBS Professional 10.4, Updated: 4/ 22/10. Edited by: Anne Urban Copyright 2003-2010 Altair Engineering, Inc. All rights reserved. PBS, PBS Works, PBS
More informationPBS Professional 11.1
PBS Professional 11.1 PBS Works is a division of PBS Professional User s Guide, Altair PBS Professional 11.1, Updated: 7/ 1/11. Edited by: Anne Urban Copyright 2003-2011 Altair Engineering, Inc. All rights
More informationParallel Computing using MATLAB Distributed Compute Server ZORRO HPC
Parallel Computing using MATLAB Distributed Compute Server ZORRO HPC Goals of the session Overview of parallel MATLAB Why parallel MATLAB? Multiprocessing in MATLAB Parallel MATLAB using the Parallel Computing
More informationTable of Contents New User Orientation...1
Table of Contents New User Orientation...1 Introduction...1 Helpful Resources...3 HPC Environment Overview...4 Basic Tasks...10 Understanding and Managing Your Allocations...16 New User Orientation Introduction
More informationWestGrid. Handbook for Researchers at the University of Manitoba. January 2010
WestGrid Handbook for Researchers at the University of Manitoba January 2010 2 Table of Contents Table of Contents...3 1 Overview...5 1.1 This Guide... 5 1.2 WestGrid... 5 2 Information for Grant Applicants...6
More informationRocoto. HWRF Python Scripts Training Miami, FL November 19, 2015
Rocoto HWRF Python Scripts Training Miami, FL November 19, 2015 Outline Introduction to Rocoto How it works Overview and description of XML Effectively using Rocoto (run, boot, stat, check, rewind, logs)
More informationThe Application Level Placement Scheduler
The Application Level Placement Scheduler Michael Karo 1, Richard Lagerstrom 1, Marlys Kohnke 1, Carl Albing 1 Cray User Group May 8, 2006 Abstract Cray platforms present unique resource and workload management
More informationIntroduction to Grid Engine
Introduction to Grid Engine Workbook Edition 8 January 2011 Document reference: 3609-2011 Introduction to Grid Engine for ECDF Users Workbook Introduction to Grid Engine for ECDF Users Author: Brian Fletcher,
More informationUsing the Yale HPC Clusters
Using the Yale HPC Clusters Stephen Weston Robert Bjornson Yale Center for Research Computing Yale University Dec 2015 To get help Send an email to: hpc@yale.edu Read documentation at: http://research.computing.yale.edu/hpc-support
More informationUsing NeSI HPC Resources. NeSI Computational Science Team (support@nesi.org.nz)
NeSI Computational Science Team (support@nesi.org.nz) Outline 1 About Us About NeSI Our Facilities 2 Using the Cluster Suitable Work What to expect Parallel speedup Data Getting to the Login Node 3 Submitting
More informationITG Software Engineering
IBM WebSphere Administration 8.5 Course ID: Page 1 Last Updated 12/15/2014 WebSphere Administration 8.5 Course Overview: This 5 Day course will cover the administration and configuration of WebSphere 8.5.
More informationMatlab on a Supercomputer
Matlab on a Supercomputer Shelley L. Knuth Research Computing April 9, 2015 Outline Description of Matlab and supercomputing Interactive Matlab jobs Non-interactive Matlab jobs Parallel Computing Slides
More informationHigh Performance Computing
High Performance Computing at Stellenbosch University Gerhard Venter Outline 1 Background 2 Clusters 3 SU History 4 SU Cluster 5 Using the Cluster 6 Examples What is High Performance Computing? Wikipedia
More informationHigh Performance Computing at the Oak Ridge Leadership Computing Facility
Page 1 High Performance Computing at the Oak Ridge Leadership Computing Facility Page 2 Outline Our Mission Computer Systems: Present, Past, Future Challenges Along the Way Resources for Users Page 3 Our
More informationDebugging and Profiling Lab. Carlos Rosales, Kent Milfeld and Yaakoub Y. El Kharma carlos@tacc.utexas.edu
Debugging and Profiling Lab Carlos Rosales, Kent Milfeld and Yaakoub Y. El Kharma carlos@tacc.utexas.edu Setup Login to Ranger: - ssh -X username@ranger.tacc.utexas.edu Make sure you can export graphics
More informationSun Grid Engine, a new scheduler for EGEE
Sun Grid Engine, a new scheduler for EGEE G. Borges, M. David, J. Gomes, J. Lopez, P. Rey, A. Simon, C. Fernandez, D. Kant, K. M. Sephton IBERGRID Conference Santiago de Compostela, Spain 14, 15, 16 May
More informationBig Data Evaluator 2.1: User Guide
University of A Coruña Computer Architecture Group Big Data Evaluator 2.1: User Guide Authors: Jorge Veiga, Roberto R. Expósito, Guillermo L. Taboada and Juan Touriño May 5, 2016 Contents 1 Overview 3
More informationHow To Run A Tompouce Cluster On An Ipra (Inria) 2.5.5 (Sun) 2 (Sun Geserade) 2-5.4 (Sun-Ge) 2/5.2 (
Running Hadoop and Stratosphere jobs on TomPouce cluster 16 October 2013 TomPouce cluster TomPouce is a cluster of 20 calcula@on nodes = 240 cores Located in the Inria Turing building (École Polytechnique)
More informationIntroduction to HPC Workshop. Center for e-research (eresearch@nesi.org.nz)
Center for e-research (eresearch@nesi.org.nz) Outline 1 About Us About CER and NeSI The CS Team Our Facilities 2 Key Concepts What is a Cluster Parallel Programming Shared Memory Distributed Memory 3 Using
More informationGRID Computing: CAS Style
CS4CC3 Advanced Operating Systems Architectures Laboratory 7 GRID Computing: CAS Style campus trunk C.I.S. router "birkhoff" server The CAS Grid Computer 100BT ethernet node 1 "gigabyte" Ethernet switch
More information