Submitting batch jobs Slurm on ecgate. Xavi Abellan User Support Section
|
|
|
- Godfrey Morrison
- 9 years ago
- Views:
Transcription
1 Submitting batch jobs Slurm on ecgate Xavi Abellan User Support Section Slide 1
2 Outline Interactive mode versus Batch mode Overview of the Slurm batch system on ecgate Batch basic concepts Creating a batch job Basic job management Checking the batch system status Accessing the Slurm Accounting database Trouble-shooting Bonus: migration from LoadLeveler Slide 2
3 Interactive vs Batch When you login, the default shell on ecgate is either the Kornshell (ksh), Bash or the C-shell (csh). To run a script or a program interactively, enter the executable name and any necessary arguments at the system prompt. You can also run your job in background so that other commands can be executed at the same time $>./your-program arg1 arg2 $>./your-program arg1 arg2 & Slide 3
4 Interactive vs Batch But Background is not batch The program is still running interactively on the login node You share the node with the rest of the users The limits for interactive sessions still apply: CPU time limit of 30 min per process $> ulimit -a Interactive sessions should be limited to development tasks, editing files, compilation or very small tests Slide 4
5 Interactive vs Batch Login node Computing (batch) nodes Slide 5
6 Interactive vs Batch Login node Computing (batch) nodes Slide 6
7 Batch on ecgate We used LoadLeveler in the previous ecgate The batch system used on the current is Slurm: Cluster workload manager: Framework to execute and monitor batch work Resource allocation (where?) Scheduling (when?) Batch job: shell script that will run unattended, with some special directives describing the job itself Slide 7
8 How does it work? Cluster utilisation Queues and priorities Limits Check status Job submission Node info Job info Slurm Login node Computing (batch) nodes Slide 8
9 Quality of service (queues) In Slurm, QoS (Quality of Service) = queue The queues have an associated priority and have certain limits Standard queues available to all users QoS Description Priority Wall Time Limit Total Jobs User Jobs express Suitable for short jobs hours normal Suitable for most of the work. This is the default day long Suitable for long jobs days 32 4 Special queues with the access restricted to meet certain conditions QoS Description Priority Wall Time Limit Total Jobs User Jobs timecrit1 timecrit2 Automatically set by EcAccess for Time Critical Option 1 jobs Only for jobs belonging to Time Critical Option 2 suites hours hours Slide 9
10 Batch job script #!/bin/bash # The job name #SBATCH --job-name=helloworld # Set the error and output files #SBATCH --output=hello-%j.out #SBATCH --error=hello-%j.out # Set the initial working directory #SBATCH --workdir=/scratch/us/usxa # Choose the queue #SBATCH - qos=express # Wall clock time limit #SBATCH --time=00:05:00 # Send an on failure #SBATCH mail-type=fail # This is the job echo Hello World! sleep 30 A job is a shell script bash/ksh/csh Directives are shell comments: starting with #SBATCH Lowercase only No spaces in between No variable expansion All directives are optional System defaults in place Slide 10
11 Job directives Directive Description Default --job-name=... A descriptive name for the job Script name --output=... --error=... --workdir=... Path to the file where standard output is redirected. Special placeholders for job id ( %j ) and the execution node ( %N ) Path to the file where standard error is redirected. Special placeholders for job id ( %j ) and the execution node ( %N ) Working directory of the job. The output and error files can be defined relative to this directory. slurm-%j.out output value submitting dir --qos=... Quality of service (queue) where the job is to be submitted normal* --time=... --mail-type=... Wall clock limit of the job (not cpu time limit!) Format: m, m:s, h:m:s, d-h, d-h:m or d-h:m:s Notify user by when certain event types occur. Valid type values are BEGIN, END, FAIL, REQUEUE, and ALL qos default disabled --mail-user=... address to send the submit user --hold Submit the job in held state. It won t run until released with scontrol release <jobid> not used Slide 11
12 Submitting a job: sbatch sbatch: Submits a job to the system. Job is configured: including the directives in the job script using the same directives as command line options The job to be submitted can be specified: As an argument of sbatch If no script is passed as an argument, sbatch will read the job from standard input $> sbatch hello.sh Submitted batch job $> cat hello out Hello world! $> The corresponding job id will be returned if successful, or an error if the job could not be submitted Slide 12
13 Submitting a job from cron Slurm jobs take the environment from the submission session Submitting from cron will cause the jobs to run with a very limited environment and will most likely fail Use a crontab line similar to: $> * * * $HOME/cronrun sbatch $HOME/cronjob Where the script cronrun is: #!/bin/ksh # cronrun script. ~/.profile. ~/.kshrc $@ #!/bin/bash # cronrun script. ~/.bash_profile $@ #!/bin/csh # cronrun script. ~/.login $@ Slide 13
14 Checking the queue: squeue squeue: displays some information about the jobs currently running or waiting By default it shows all jobs from all users, but some filtering options are possible: -u <comma separated list of users> -q <comma separated list of QoSs> -n <comma separated list of job names> -j <comma separated list of job ids> -t <comma separated list of job states> $> squeue -u $USER JOBID NAME USER QOS STATE TIME TIMELIMIT NODELIST(REASON) helloworld usxa express RUNNING 0:08 5:00 ecgb07 Slide 14
15 Canceling a job: scancel scancel: Cancels the specified job(s) $> sbatch hello.sh Submitted batch job $> scancel $> scancel scancel: error: Kill job error on job id : Invalid job id specified $> sbatch hello.sh Submitted batch job $> scancel -in hello.sh Cancel job_id= name=hello.sh partition=batch [y/n]? y $> sbatch hello.sh Submitted batch job $> scancel -i v scancel: auth plugin for Munge ( loaded Cancel job_id= name=hello.sh partition=batch [y/n]? y scancel: Terminating job A job can be cancelled either if it is running or still waiting on the queue A running job will be killed, and message will be appended on the error file: slurmd[ecgb07]: *** JOB CANCELLED AT T17:08:29 *** Slide 15
16 Canceling a job: scancel options The most common usage of scancel is: $> scancel <jobid1> <jobid2> <jobid3> More advanced options: Option Description -n <jobname> Cancel all the jobs with the specified job name -t <state> Cancel all the jobs that are in the specified state (PENDING/RUNNING) -q <qos> Cancel only jobs on the specified QoS -u $USER Cancel ALL the jobs of the current user. Use carefully! -i Interactive option: ask for confirmation before cancelling jobs -v Verbose option. It will show what is being done Note: An ordinary user can only cancel their own jobs Slide 16
17 Practical 1: Basic job submission Practicals must be run on ecgate, so make sure you log in there first! $> ssh ecgate $> cd $SCRATCH $> tar xvzf ~trx/intro/batch_ecgate_practicals.tar.gz $> cd batch_ecgate_practicals/basic 1. Have a look at the script env.sh 2. Submit the job and check whether it is running What QoS is it using? What is the time limit of the job? 3. Where did the output of the job go? Have a look at the output 4. Submit the job again and then once it starts cancel it 5. Check the output Slide 17
18 Practical 1: Basic job setup Can you modify the previous job so it 1. runs in the express QoS, with a wall clock limit of 5 minutes? 2. uses the subdirectory work/ as the working directory? 3. sends the a) output to the file work/env_out_<jobid>.out? b) error to work/env_out_<jobid>.err? 4. sends you an when the job starts? Try your job after the modifications and check if they are correct You can do the modifications one by one or all at once Slide 18
19 Why doesn t my job start? Check the last column of the squeue output for a hint $> squeue -j JOBID NAME USER QOS STATE TIME TIMELIMIT NODELIST(REASON) sbatch usxa long PENDING 0: :00:00 (QOSResourceLimit) Reason Priority Resources JobUserHeld QOSResourceLimit Description There are other jobs with more priority No free resources are available The job is held. Release with scontrol release <jobid> You have reached a limit in the number of jobs you can submit to a QoS My job is PENDING because of a QOSResourceLimit. How do I check my limits? Slide 19
20 Checking limits and general usage: sqos sqos: Utility to have an overview of the different QoSs where the user have access, including usage and limits This utility is ECMWF specific (not part of a standard Slurm installation) $> sqos QoS Prio Max Wall Total Jobs User Jobs Max CPUS Max Mem express :00:00 11 / / MB normal :00:00 23 / / MB long :00:00 7 / 32 4 / MB large :00:00 0 / 8 0 / MB timecrit :00:00 0 / 96 0 / MB Total: 43 Jobs, 41 RUNNING, 2 PENDING Account Def QoS Running Jobs Submitted Jobs *ectrain normal 15 / / 1000 User trx: 17 Jobs, 15 RUNNING, 2 PENDING Slide 20
21 More details on current jobs and nodes scontrol: view and modify Slurm jobs and node configuration $> scontrol show job JobId=24789 Name=test_slurm_csh UserId=us2(1666) GroupId=gb(3070) Priority=3000 Account=ecus QOS=normal JobState=COMPLETED Reason=None Dependency=(null) Requeue=0 Restarts=0 BatchFlag=1 ExitCode=0:0 RunTime=00:01:25 TimeLimit=00:10:00 TimeMin=N/A SubmitTime= T08:55:34 EligibleTime= T08:55:34 StartTime= T08:55:34 EndTime= T08:56:59 PreemptTime=None SuspendTime=None SecsPreSuspend=0 Partition=batch AllocNode:Sid=ecgb05:36790 ReqNodeList=(null) ExcNodeList=(null) NodeList=ecgb05 BatchHost=ecgb05 NumNodes=1 NumCPUs=1 CPUs/Task=1 ReqS:C:T=*:*:* MinCPUsNode=1 MinMemoryCPU=1900M MinTmpDiskNode=0 Features=(null) Gres=(null) Reservation=(null) Shared=1 Contiguous=0 Licenses=(null) Network=(null) Command=/home/ms/gb/us2/slurm_csh.job WorkDir=/scratch/ms/gb/us2/csh Comment=Output=/scratch/ms/gb/us2/csh/slurm_24789.out;Error=/scratch/ms/gb/us2/csh/slurm_24789.out; sinfo: View information about node status $> sinfo PARTITION AVAIL TIMELIMIT NODES STATE NODELIST batch* up infinite 1 drain ecgb11 batch* up infinite 6 alloc ecgb[04-05,07-10] test up infinite 1 idle ecgb06 Slide 21
22 More details on current jobs and nodes sview: GUI to see the queue and node status Slide 22
23 Access to the Slurm accounting DB: sacct sacct: View present and past job information $> sacct -X JobID JobName QOS State ExitCode Elapsed NodeList test.sh normal COMPLETED 0:0 00:00:13 ecgb test.sh normal COMPLETED 0:0 00:01:10 ecgb test.sh normal COMPLETED 0:0 00:00:47 ecgb test.sh normal COMPLETED 0:0 00:01:32 ecgb test.sh normal COMPLETED 0:0 00:02:19 ecgb test.sh normal COMPLETED 0:0 00:00:45 ecgb test.sh normal RUNNING 0:0 00:02:35 ecgb test.sh normal RUNNING 0:0 00:02:35 ecgb test.sh normal CANCELLED+ 0:0 00:01:24 ecgb test.sh normal RUNNING 0:0 00:02:35 ecgb test.sh normal COMPLETED 0:0 00:00:40 ecgb test.sh normal RUNNING 0:0 00:02:35 ecgb test.sh normal COMPLETED 0:0 00:00:40 ecgb test.sh normal RUNNING 0:0 00:02:33 ecgb helloworld normal FAILED 1:0 00:00:01 ecgb test.sh normal CANCELLED+ 0:0 00:00:33 ecgb test.sh normal RUNNING 0:0 00:01:39 ecgb test.sh express RUNNING 0:0 00:01:23 ecgb test.sh express RUNNING 0:0 00:01:23 ecgb test.sh long RUNNING 0:0 00:01:19 ecgb04 Slide 23
24 Access to the Slurm accounting DB: sacct options By default, sacct will return information about your jobs that started today Option Description -j <jobid> Show the job with that jobid -u <user> Show jobs for the specified user. Use option a for all users -E <endtime> Show jobs eligible before that date and time -S <starttime> Show jobs eligible after that date and time -s <statelist> Show jobs on the states (comma-separated) given during the time period. Valids states are: CANCELLED, COMPLETED, FAILED, NODE_FAIL, RUNNING, PENDING, TIMEOUT -q <qos> Show jobs only for the qos selected -o <outformat> Format option. Comma-separated names of fields to display -e Show the different columns to be used for the o option -X Hide the job step information, showing the allocation only Slide 24
25 What happened to my job: job_forensics job_forensics: Custom ECMWF utility to dump forensic information about a job $> job_forensics DB Information: Job: JobID: JobName:sbatch User:trx UID:414 Group:ectrain GID:1400 Account:ectrain QOS:long Priority:2000 Partition:batch NCPUS:32 NNodes:1 NodeList:ecgb09 State:COMPLETED Timelimit:7-00:00:00 Submit: T16:19:06 Eligible: T16:19:06 Start: T16:19:06 End: T16:20:07 Elapsed:00:01:01 CPUTime:00:32:32 UserCPU:00: SystemCPU:00: TotalCPU:00: DerivedExitCode:0:0 ExitCode:0:0 Output:/home/ectrain/trx/slurm out Error:/home/ectrain/trx/slurm out... Main step: JobID: batch JobName:batch NCPUS:1 CPUTime:00:01:01 AveRSS:1796K MaxRSS:1796K MaxRSSNode:ecgb09 MaxRSSTask:0 Controller Logs: [ T16:19:06+00:00] _slurm_rpc_submit_batch_job JobId= usec= ecgb09 log (main): [ T16:19:07+00:00] Launching batch job for UID 414 [ T16:20:07+00:00] [ ] sending REQUEST_COMPLETE_BATCH_SCRIPT, error:0 [ T16:20:07+00:00] [ ] done with job Slide 25
26 Practical 2: reviewing past runs How would you retrieve the list of jobs that you ran today? $> sacct retrieve the list of all the jobs that were cancelled today by user trx? $> sacct ask for the submit, start and end times for a job of your choice? $> sacct find out the output an error paths for a job of your choice? $> sacct Slide 26
27 Practical 3: Fixing broken jobs $> cd $SCRATCH/batch_ecgate_practicals/broken What is wrong in job1? Can you fix it? What is wrong in job2? Can you fix it? What is wrong in job3? Can you fix it? Slide 27
28 Bonus: Migrating from LoadLeveler You can submit a LL job to Slurm, but the LL directives will be ignored! Translation required: manually or using ll2slurm $> ll2slurm -h usage: ll2slurm [-h] [-i INSCRIPT] [-o OUTSCRIPT] [-q] [-f] Job translator from LoadLeveler to Slurm optional arguments: -h, --help show this help message and exit -i INSCRIPT, --inscript INSCRIPT Input script. By default reads stdin -o OUTSCRIPT, --outscript OUTSCRIPT Output translated script. By default writes to stdout -q, --quiet Do not produce warning or error messages on stderr -f, --force Overwrite the output file if it exists Not all the LoadLeveler keywords can be translated. Some manual additions might be required! You may play with the example: $SCRATCH/batch_ecgate_practicals/loadleveler Slide 28
29 Migration cheatsheet (I) User Commands LoadLeveler SLURM Job submission llsubmit [script] sbatch [script] Job cancel llcancel [job_id] scancel [job_id] Job status llq [ -j job_id] squeue [job_id] Queue list llq squeue Environment Variables LoadLeveler SLURM Job ID $LOADL_STEP_ID $SLURM_JOBID Working Dir $LOADL_STEP_INITDIR pwd command Node List $LOADL_PROCESSOR_LIST $SLURM_JOB_NODELIST Slide 29
30 Migration cheatsheet (II) Job Configuration LoadLeveler SLURM Script directive #SBATCH Job Name job_name=[name] --job-name=[name] Queue class=[queue] --qos=[queue] Wall Clock Limit wall_clock_limit=[hh:mm:ss] --time=[min] --time=[days-hh:mm:ss] Std Output File output=[file] --output=[file] Std Error File error=[file] --error=[file] Working Directory initialdir=[dir_name] --workdir=[dir_name] Copy Environment environment=copy_all --export=[all NONE] --export=[variables] Notification notification= --mail-type=[events] Address notify_user=[address] --mail-user=[address] Slide 30
31 Additional Info Ecgate job examples: SLURM website and documentation: Questions? Slide 31
SLURM Workload Manager
SLURM Workload Manager What is SLURM? SLURM (Simple Linux Utility for Resource Management) is the native scheduler software that runs on ASTI's HPC cluster. Free and open-source job scheduler for the Linux
SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education www.accre.vanderbilt.
SLURM: Resource Management and Job Scheduling Software Advanced Computing Center for Research and Education www.accre.vanderbilt.edu Simple Linux Utility for Resource Management But it s also a job scheduler!
SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education www.accre.vanderbilt.
SLURM: Resource Management and Job Scheduling Software Advanced Computing Center for Research and Education www.accre.vanderbilt.edu Simple Linux Utility for Resource Management But it s also a job scheduler!
Introduction to Running Computations on the High Performance Clusters at the Center for Computational Research
! Introduction to Running Computations on the High Performance Clusters at the Center for Computational Research! Cynthia Cornelius! Center for Computational Research University at Buffalo, SUNY! cdc at
Until now: tl;dr: - submit a job to the scheduler
Until now: - access the cluster copy data to/from the cluster create parallel software compile code and use optimized libraries how to run the software on the full cluster tl;dr: - submit a job to the
General Overview. Slurm Training15. Alfred Gil & Jordi Blasco (HPCNow!)
Slurm Training15 Agenda 1 2 3 About Slurm Key Features of Slurm Extending Slurm Resource Management Daemons Job/step allocation 4 5 SMP MPI Parametric Job monitoring Accounting Scheduling Administration
Miami University RedHawk Cluster Working with batch jobs on the Cluster
Miami University RedHawk Cluster Working with batch jobs on the Cluster The RedHawk cluster is a general purpose research computing resource available to support the research community at Miami University.
Introduction to the SGE/OGS batch-queuing system
Grid Computing Competence Center Introduction to the SGE/OGS batch-queuing system Riccardo Murri Grid Computing Competence Center, Organisch-Chemisches Institut, University of Zurich Oct. 6, 2011 The basic
To connect to the cluster, simply use a SSH or SFTP client to connect to:
RIT Computer Engineering Cluster The RIT Computer Engineering cluster contains 12 computers for parallel programming using MPI. One computer, cluster-head.ce.rit.edu, serves as the master controller or
Biowulf2 Training Session
Biowulf2 Training Session 9 July 2015 Slides at: h,p://hpc.nih.gov/docs/b2training.pdf HPC@NIH website: h,p://hpc.nih.gov System hardware overview What s new/different The batch system & subminng jobs
An introduction to compute resources in Biostatistics. Chris Scheller [email protected]
An introduction to compute resources in Biostatistics Chris Scheller [email protected] 1. Resources 1. Hardware 2. Account Allocation 3. Storage 4. Software 2. Usage 1. Environment Modules 2. Tools 3.
Job Scheduler Daemon Configuration Guide
Job Scheduler Daemon Configuration Guide A component of Mark Dickinsons Unix Job Scheduler This manual covers the server daemon component of Mark Dickinsons unix Job Scheduler. This manual is for version
Introduction to Programming and Computing for Scientists
Oxana Smirnova (Lund University) Programming for Scientists Tutorial 7b 1 / 48 Introduction to Programming and Computing for Scientists Oxana Smirnova Lund University Tutorial 7b: Grid certificates and
Introduction to parallel computing and UPPMAX
Introduction to parallel computing and UPPMAX Intro part of course in Parallel Image Analysis Elias Rudberg [email protected] March 22, 2011 Parallel computing Parallel computing is becoming increasingly
Quick Tutorial for Portable Batch System (PBS)
Quick Tutorial for Portable Batch System (PBS) The Portable Batch System (PBS) system is designed to manage the distribution of batch jobs and interactive sessions across the available nodes in the cluster.
Matlab on a Supercomputer
Matlab on a Supercomputer Shelley L. Knuth Research Computing April 9, 2015 Outline Description of Matlab and supercomputing Interactive Matlab jobs Non-interactive Matlab jobs Parallel Computing Slides
LoadLeveler Overview. January 30-31, 2012. IBM Storage & Technology Group. IBM HPC Developer Education @ TIFR, Mumbai
IBM HPC Developer Education @ TIFR, Mumbai IBM Storage & Technology Group LoadLeveler Overview January 30-31, 2012 Pidad D'Souza ([email protected]) IBM, System & Technology Group 2009 IBM Corporation
Batch Job Analysis to Improve the Success Rate in HPC
Batch Job Analysis to Improve the Success Rate in HPC 1 JunWeon Yoon, 2 TaeYoung Hong, 3 ChanYeol Park, 4 HeonChang Yu 1, First Author KISTI and Korea University, [email protected] 2,3, KISTI,[email protected],[email protected]
Ecaccess Gateway, Tools and ectrans
Ecaccess Gateway, Tools and ectrans Dominique Lucas User Support [email protected] 18/02/2015 1 Content ECaccess concepts Ecaccess tools Ectrans Tutorial 18/02/2015 2 ECaccess - Concepts ECaccess provides
PBS Tutorial. Fangrui Ma Universit of Nebraska-Lincoln. October 26th, 2007
PBS Tutorial Fangrui Ma Universit of Nebraska-Lincoln October 26th, 2007 Abstract In this tutorial we gave a brief introduction to using PBS Pro. We gave examples on how to write control script, and submit
Managing GPUs by Slurm. Massimo Benini HPC Advisory Council Switzerland Conference March 31 - April 3, 2014 Lugano
Managing GPUs by Slurm Massimo Benini HPC Advisory Council Switzerland Conference March 31 - April 3, 2014 Lugano Agenda General Slurm introduction Slurm@CSCS Generic Resource Scheduling for GPUs Resource
Using NeSI HPC Resources. NeSI Computational Science Team ([email protected])
NeSI Computational Science Team ([email protected]) Outline 1 About Us About NeSI Our Facilities 2 Using the Cluster Suitable Work What to expect Parallel speedup Data Getting to the Login Node 3 Submitting
Introduction to HPC Workshop. Center for e-research ([email protected])
Center for e-research ([email protected]) Outline 1 About Us About CER and NeSI The CS Team Our Facilities 2 Key Concepts What is a Cluster Parallel Programming Shared Memory Distributed Memory 3 Using
Tutorial-4a: Parallel (multi-cpu) Computing
HTTP://WWW.HEP.LU.SE/COURSES/MNXB01 Introduction to Programming and Computing for Scientists (2015 HT) Tutorial-4a: Parallel (multi-cpu) Computing Balazs Konya (Lund University) Programming for Scientists
Cloud Bursting with SLURM and Bright Cluster Manager. Martijn de Vries CTO
Cloud Bursting with SLURM and Bright Cluster Manager Martijn de Vries CTO Architecture CMDaemon 2 Management Interfaces Graphical User Interface (GUI) Offers administrator full cluster control Standalone
Job scheduler details
Job scheduler details Advanced Computing Center for Research & Education (ACCRE) Job scheduler details 1 / 25 Outline 1 Batch queue system overview 2 Torque and Moab 3 Submitting jobs (ACCRE) Job scheduler
The Asterope compute cluster
The Asterope compute cluster ÅA has a small cluster named asterope.abo.fi with 8 compute nodes Each node has 2 Intel Xeon X5650 processors (6-core) with a total of 24 GB RAM 2 NVIDIA Tesla M2050 GPGPU
Running applications on the Cray XC30 4/12/2015
Running applications on the Cray XC30 4/12/2015 1 Running on compute nodes By default, users do not log in and run applications on the compute nodes directly. Instead they launch jobs on compute nodes
NorduGrid ARC Tutorial
NorduGrid ARC Tutorial / Arto Teräs and Olli Tourunen 2006-03-23 Slide 1(34) NorduGrid ARC Tutorial Arto Teräs and Olli Tourunen CSC, Espoo, Finland March 23
Using the Yale HPC Clusters
Using the Yale HPC Clusters Stephen Weston Robert Bjornson Yale Center for Research Computing Yale University Oct 2015 To get help Send an email to: [email protected] Read documentation at: http://research.computing.yale.edu/hpc-support
Beginners Shell Scripting for Batch Jobs
Beginners Shell Scripting for Batch Jobs Evan Bollig and Geoffrey Womeldorff Before we begin... Everyone please visit this page for example scripts and grab a crib sheet from the front http://www.scs.fsu.edu/~bollig/techseries
Manual. Version: 1.0.0
Manual Version: 1.0.0 Table of Contents I. INTRODUCTION... 3 II. INSTALLATION... 5 a. System Requirements... 5 b. Installation... 5 c. Configure PayPal IPN... 5 d. Cron Setup... 6 e. Upload Email Logo...
Grid Engine Users Guide. 2011.11p1 Edition
Grid Engine Users Guide 2011.11p1 Edition Grid Engine Users Guide : 2011.11p1 Edition Published Nov 01 2012 Copyright 2012 University of California and Scalable Systems This document is subject to the
ECaccess User Guide. User Support. Operations Department. Version 4.0. October 2011
ECaccess User Guide User Support Operations Department Version 4.0 October 2011 c Copyright 2011 European Centre for Medium-Range Weather Forecasts Shinfield Park, Reading, RG2 9AX, United Kingdom Literary
SGE Roll: Users Guide. Version @VERSION@ Edition
SGE Roll: Users Guide Version @VERSION@ Edition SGE Roll: Users Guide : Version @VERSION@ Edition Published Aug 2006 Copyright 2006 UC Regents, Scalable Systems Table of Contents Preface...i 1. Requirements...1
Introduction to Sun Grid Engine (SGE)
Introduction to Sun Grid Engine (SGE) What is SGE? Sun Grid Engine (SGE) is an open source community effort to facilitate the adoption of distributed computing solutions. Sponsored by Sun Microsystems
Hodor and Bran - Job Scheduling and PBS Scripts
Hodor and Bran - Job Scheduling and PBS Scripts UND Computational Research Center Now that you have your program compiled and your input file ready for processing, it s time to run your job on the cluster.
Script-controlled Backups in SAP HANA Database
Script-controlled Backups in SAP HANA Database Documentation of a template shell script for controlling backup execution Richard Bremer, SAP Customer Solution Adoption (CSA) Data Backups in SAP HANA Database
NYUAD HPC Center Running Jobs
NYUAD HPC Center Running Jobs 1 Overview... Error! Bookmark not defined. 1.1 General List... Error! Bookmark not defined. 1.2 Compilers... Error! Bookmark not defined. 2 Loading Software... Error! Bookmark
Linux command line. An introduction to the Linux command line for genomics. Susan Fairley
Linux command line An introduction to the Linux command line for genomics Susan Fairley Aims Introduce the command line Provide an awareness of basic functionality Illustrate with some examples Provide
Introduction to Running Hadoop on the High Performance Clusters at the Center for Computational Research
Introduction to Running Hadoop on the High Performance Clusters at the Center for Computational Research Cynthia Cornelius Center for Computational Research University at Buffalo, SUNY 701 Ellicott St
The RWTH Compute Cluster Environment
The RWTH Compute Cluster Environment Tim Cramer 11.03.2013 Source: D. Both, Bull GmbH Rechen- und Kommunikationszentrum (RZ) How to login Frontends cluster.rz.rwth-aachen.de cluster-x.rz.rwth-aachen.de
Using Parallel Computing to Run Multiple Jobs
Beowulf Training Using Parallel Computing to Run Multiple Jobs Jeff Linderoth August 5, 2003 August 5, 2003 Beowulf Training Running Multiple Jobs Slide 1 Outline Introduction to Scheduling Software The
HPC-Nutzer Informationsaustausch. The Workload Management System LSF
HPC-Nutzer Informationsaustausch The Workload Management System LSF Content Cluster facts Job submission esub messages Scheduling strategies Tools and security Future plans 2 von 10 Some facts about the
Installing IBM Websphere Application Server 7 and 8 on OS4 Enterprise Linux
Installing IBM Websphere Application Server 7 and 8 on OS4 Enterprise Linux By the OS4 Documentation Team Prepared by Roberto J Dohnert Copyright 2013, PC/OpenSystems LLC This whitepaper describes how
Unix the Bare Minimum
Unix the Bare Minimum Norman Matloff September 27, 2005 c 2001-2005, N.S. Matloff Contents 1 Purpose 2 2 Shells 2 3 Files and Directories 4 3.1 Creating Directories.......................................
Grid Engine Basics. Table of Contents. Grid Engine Basics Version 1. (Formerly: Sun Grid Engine)
Grid Engine Basics (Formerly: Sun Grid Engine) Table of Contents Table of Contents Document Text Style Associations Prerequisites Terminology What is the Grid Engine (SGE)? Loading the SGE Module on Turing
Kiko> A personal job scheduler
Kiko> A personal job scheduler V1.2 Carlos allende prieto october 2009 kiko> is a light-weight tool to manage non-interactive tasks on personal computers. It can improve your system s throughput significantly
Linux für bwgrid. Sabine Richling, Heinz Kredel. Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim. 27.
Linux für bwgrid Sabine Richling, Heinz Kredel Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim 27. June 2011 Richling/Kredel (URZ/RUM) Linux für bwgrid FS 2011 1 / 33 Introduction
CrontabFile Converter
JobScheduler - Job Execution and Scheduling System CrontabFile Converter Users Manual July 2014 July 2014 CrontabFile Converter page: 1 CrontabFile Converter - Contact Information Contact Information Software-
Compute Cluster Documentation
Drucksachenkategorie Compute Cluster Documentation Compiled by Martin Geier (DLR FA-STM) Authors Michael Schäfer (INIT GmbH) Martin Geier (DLR FA-STM) Date 22.05.2015 Table of contents Table of contents...
CS10110 Introduction to personal computer equipment
CS10110 Introduction to personal computer equipment PRACTICAL 4 : Process, Task and Application Management In this practical you will: Use Unix shell commands to find out about the processes the operating
Fair Scheduler. Table of contents
Table of contents 1 Purpose... 2 2 Introduction... 2 3 Installation... 3 4 Configuration...3 4.1 Scheduler Parameters in mapred-site.xml...4 4.2 Allocation File (fair-scheduler.xml)... 6 4.3 Access Control
Scheduling in SAS 9.4 Second Edition
Scheduling in SAS 9.4 Second Edition SAS Documentation The correct bibliographic citation for this manual is as follows: SAS Institute Inc. 2015. Scheduling in SAS 9.4, Second Edition. Cary, NC: SAS Institute
Scheduling in SAS 9.3
Scheduling in SAS 9.3 SAS Documentation The correct bibliographic citation for this manual is as follows: SAS Institute Inc 2011. Scheduling in SAS 9.3. Cary, NC: SAS Institute Inc. Scheduling in SAS 9.3
Thirty Useful Unix Commands
Leaflet U5 Thirty Useful Unix Commands Last revised April 1997 This leaflet contains basic information on thirty of the most frequently used Unix Commands. It is intended for Unix beginners who need a
SendMIME Pro Installation & Users Guide
www.sendmime.com SendMIME Pro Installation & Users Guide Copyright 2002 SendMIME Software, All Rights Reserved. 6 Greer Street, Stittsville, Ontario Canada K2S 1H8 Phone: 613-831-4023 System Requirements
PaperCut Payment Gateway Module - RBS WorldPay Quick Start Guide
PaperCut Payment Gateway Module - RBS WorldPay Quick Start Guide This guide is designed to supplement the Payment Gateway Module documentation and provides a guide to installing, setting up and testing
NASA Workflow Tool. User Guide. September 29, 2010
NASA Workflow Tool User Guide September 29, 2010 NASA Workflow Tool User Guide 1. Overview 2. Getting Started Preparing the Environment 3. Using the NED Client Common Terminology Workflow Configuration
Gentran Integration Suite. Archiving and Purging. Version 4.2
Gentran Integration Suite Archiving and Purging Version 4.2 Copyright 2007 Sterling Commerce, Inc. All rights reserved. Additional copyright information is located on the Gentran Integration Suite Documentation
Command Line Interface User Guide for Intel Server Management Software
Command Line Interface User Guide for Intel Server Management Software Legal Information Information in this document is provided in connection with Intel products. No license, express or implied, by estoppel
Beam Loss Monitor Software Guide
Version 1.0 Beam Loss Monitor Software Guide James Leaver 5th December 2008 1 Operating Procedure In order to run the Beam Loss Monitor software, the user should complete the following steps: 1. Open a
IBM Redistribute Big SQL v4.x Storage Paths IBM. Redistribute Big SQL v4.x Storage Paths
Redistribute Big SQL v4.x Storage Paths THE GOAL The Big SQL temporary tablespace is used during high volume queries to spill sorts or intermediate data to disk. To improve I/O performance for these queries,
Working with Docker on Microsoft Azure
Working with Docker on Microsoft Azure Lab Overview In this lab you will create a Docker enabled virtual machine from the Azure Marketplace. You will then go through basic Docker commands. After that,
End User Setup and Handling
on IM and Presence Service, page 1 Authorization Policy Setup On IM and Presence Service, page 1 Bulk Rename User Contact IDs, page 4 Bulk Export User Contact Lists, page 5 Bulk Export Non-Presence Contact
Configuring MailArchiva with Insight Server
Copyright 2009 Bynari Inc., All rights reserved. No part of this publication may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopy, recording, or any
Apparo Fast Edit. Excel data import via email 1 / 19
Apparo Fast Edit Excel data import via email 1 / 19 1 2 3 4 5 Definition 3 Benefits at a glance 3 Example 4 3.1 Use Case 4 3.2 How users experience this feature 4 Email ImportBusiness Case 6 4.1 Creating
Installing and running COMSOL on a Linux cluster
Installing and running COMSOL on a Linux cluster Introduction This quick guide explains how to install and operate COMSOL Multiphysics 5.0 on a Linux cluster. It is a complement to the COMSOL Installation
Automating FTP with the CP 443-1 IT
Automating FTP with the CP 443-1 IT Contents Page Introduction 2 FTP Basics with the SIMATIC NET CP 443-1 IT 3 CONFIGURATION 3 FTP SERVICES 6 FTP Server with the SIMATIC NET CP 443-1 IT 9 OVERVIEW 9 CONFIGURATION
Rocoto. HWRF Python Scripts Training Miami, FL November 19, 2015
Rocoto HWRF Python Scripts Training Miami, FL November 19, 2015 Outline Introduction to Rocoto How it works Overview and description of XML Effectively using Rocoto (run, boot, stat, check, rewind, logs)
System Administration Guide
www.novell.com/documentation System Administration Guide Data Synchronizer 1.2 August 22, 2012 Legal Notices Novell, Inc. makes no representations or warranties with respect to the contents or use of this
Batch Scripts for RA & Mio
Batch Scripts for RA & Mio Timothy H. Kaiser, Ph.D. [email protected] 1 Jobs are Run via a Batch System Ra and Mio are shared resources Purpose: Give fair access to all users Have control over where jobs
Agenda. Using HPC Wales 2
Using HPC Wales Agenda Infrastructure : An Overview of our Infrastructure Logging in : Command Line Interface and File Transfer Linux Basics : Commands and Text Editors Using Modules : Managing Software
High-Performance Reservoir Risk Assessment (Jacta Cluster)
High-Performance Reservoir Risk Assessment (Jacta Cluster) SKUA-GOCAD 2013.1 Paradigm 2011.3 With Epos 4.1 Data Management Configuration Guide 2008 2013 Paradigm Ltd. or its affiliates and subsidiaries.
Streamline Computing Linux Cluster User Training. ( Nottingham University)
1 Streamline Computing Linux Cluster User Training ( Nottingham University) 3 User Training Agenda System Overview System Access Description of Cluster Environment Code Development Job Schedulers Running
Slurm Workload Manager Architecture, Configuration and Use
Slurm Workload Manager Architecture, Configuration and Use Morris Jette [email protected] Goals Learn the basics of SLURM's architecture, daemons and commands Learn how to use a basic set of commands Learn
SFTP SHELL SCRIPT USER GUIDE
SFTP SHELL SCRIPT USER GUIDE FCA US INFORMATION & COMMUNICATION TECHNOLOGY MANAGEMENT Overview The EBMX SFTP shell scripts provide a parameter driven workflow to place les on the EBMX servers and queue
Job Scheduling with Moab Cluster Suite
Job Scheduling with Moab Cluster Suite IBM High Performance Computing February 2010 Y. Joanna Wong, Ph.D. [email protected] 2/22/2010 Workload Manager Torque Source: Adaptive Computing 2 Some terminology..
Ra - Batch Scripts. Timothy H. Kaiser, Ph.D. [email protected]
Ra - Batch Scripts Timothy H. Kaiser, Ph.D. [email protected] Jobs on Ra are Run via a Batch System Ra is a shared resource Purpose: Give fair access to all users Have control over where jobs are run Set
How to Run Parallel Jobs Efficiently
How to Run Parallel Jobs Efficiently Shao-Ching Huang High Performance Computing Group UCLA Institute for Digital Research and Education May 9, 2013 1 The big picture: running parallel jobs on Hoffman2
Smart Card Authentication Client. Administrator's Guide
Smart Card Authentication Client Administrator's Guide April 2013 www.lexmark.com Contents 2 Contents Overview...3 Configuring Smart Card Authentication Client...4 Configuring printer settings for use
UMass High Performance Computing Center
.. UMass High Performance Computing Center University of Massachusetts Medical School October, 2014 2 / 32. Challenges of Genomic Data It is getting easier and cheaper to produce bigger genomic data every
SWsoft Plesk 8.2 for Linux/Unix Backup and Restore Utilities. Administrator's Guide
SWsoft Plesk 8.2 for Linux/Unix Backup and Restore Utilities Administrator's Guide 2 Copyright Notice ISBN: N/A SWsoft. 13755 Sunrise Valley Drive Suite 325 Herndon VA 20171 USA Phone: +1 (703) 815 5670
PUBLIC Model Manager User Guide
SAP Predictive Analytics 2.4 2015-11-23 PUBLIC Content 1 Introduction....4 2 Concepts....5 2.1 Roles....5 2.2 Rights....6 2.3 Schedules....7 2.4 Tasks.... 7 3....8 3.1 My Model Manager....8 Overview....
one Managing your PBX Administrator ACCESSING YOUR PBX ACCOUNT CHECKING ACCOUNT ACTIVITY
one Managing your PBX Administrator ACCESSING YOUR PBX ACCOUNT Navigate to https://portal.priorityonenet.com/ and log in to the PriorityOne portal account. If you would like your web browser to keep you
Cisco Networking Academy Program Curriculum Scope & Sequence. Fundamentals of UNIX version 2.0 (July, 2002)
Cisco Networking Academy Program Curriculum Scope & Sequence Fundamentals of UNIX version 2.0 (July, 2002) Course Description: Fundamentals of UNIX teaches you how to use the UNIX operating system and
753 Broad Street Phone: 706-312-3535 Suite 200 Fax: 706-868-8655 Augusta, GA 30901-5518. Copyrights
Ipswitch, Inc. Web: www.imailserver.com 753 Broad Street Phone: 706-312-3535 Suite 200 Fax: 706-868-8655 Augusta, GA 30901-5518 Copyrights 1995-2011 Ipswitch, Inc. All rights reserved. IMail Collaboration
Unix Sampler. PEOPLE whoami id who
Unix Sampler PEOPLE whoami id who finger username hostname grep pattern /etc/passwd Learn about yourself. See who is logged on Find out about the person who has an account called username on this host
Work Environment. David Tur HPC Expert. HPC Users Training September, 18th 2015
Work Environment David Tur HPC Expert HPC Users Training September, 18th 2015 1. Atlas Cluster: Accessing and using resources 2. Software Overview 3. Job Scheduler 1. Accessing Resources DIPC technicians
CPSC 2800 Linux Hands-on Lab #7 on Linux Utilities. Project 7-1
CPSC 2800 Linux Hands-on Lab #7 on Linux Utilities Project 7-1 In this project you use the df command to determine usage of the file systems on your hard drive. Log into user account for this and the following
Setting up PostgreSQL
Setting up PostgreSQL 1 Introduction to PostgreSQL PostgreSQL is an object-relational database management system based on POSTGRES, which was developed at the University of California at Berkeley. PostgreSQL
? Index. Introduction. 1 of 38 About the QMS Network Print Monitor for Windows NT
1 of 38 About the QMS Network for Windows NT System Requirements" Installing the " Using the " Troubleshooting Operations" Introduction The NT Print Spooler (both workstation and server versions) controls
Informatica Corporation Proactive Monitoring for PowerCenter Operations Version 3.0 Release Notes May 2014
Contents Informatica Corporation Proactive Monitoring for PowerCenter Operations Version 3.0 Release Notes May 2014 Copyright (c) 2012-2014 Informatica Corporation. All rights reserved. Installation...
EVALUATION ONLY. WA2088 WebSphere Application Server 8.5 Administration on Windows. Student Labs. Web Age Solutions Inc.
WA2088 WebSphere Application Server 8.5 Administration on Windows Student Labs Web Age Solutions Inc. Copyright 2013 Web Age Solutions Inc. 1 Table of Contents Directory Paths Used in Labs...3 Lab Notes...4
