The SUN ONE Grid Engine BATCH SYSTEM
|
|
|
- Gilbert Adams
- 10 years ago
- Views:
Transcription
1 The SUN ONE Grid Engine BATCH SYSTEM Juan Luis Chaves Sanabria Centro Nacional de Cálculo Científico (CeCalCULA) Latin American School in HPC on Linux Cluster October 27 November What is SGE? Is a cluster resource management software Acceptsjobssubmittedby usersand schedules them for execution on the cluster based upon resource management policies (who gets how much resources when) Jobs are distributed in a way that optimizes uniform workload across the cluster
2 Who develop SGE? SGE is developed by Sun Microsystems SUN adquired Gridware. Developer of Distributed Resource Management (DRM) software (July 2000) SUN release SGE as a free downloable binary for solaris and linux OS to facilitate deployment of compute farms. Source code is available. Open source project to enable the Grid Computing Model. SGE 5.3 supported platforms Compaq Tru64 Unix 5.0, 5.1 Hewlett Packard HP-UX 10.20, IBM AIX 4.3.X Linux x86, kernel 2.4, glibc 2.2 Linux Alpha/AXP, kernel 2.2, glibc 2.2 SGI IRIX SUN Solaris (sparc) 2.6, 7, 8, 9 32-bit SUN Solaris (sparc) 2.6, 7, 8, 9 64-bit Sun Solaris (x86) 8
3 How the System Operates? SGE accepts jobs requests for computer resources (requeriment profile by each job) Jobs requests are located in a holding area until they can be executed When are ready to be executed, the request is forwarded to the adequate execution(s) device(s) SGE manage the execution of the request Logs the record of their execution when it s finalized SGE Components Hosts: Master (sge_qmaster y sge_schedd): control all the SGE components and the overall cluster activity Execution (sge_execd): authorized to execute jobs through SGE Administration: designated to carry out any kind of administrative task for the SGE system Submit: for submitting (qsub) and controlling (qstat, qdel, qhold, qrls,... ) batch jobs
4 SGE Components (2) Queues: A queue is a container for a class of jobs (Batch/Parallel/Interactive/Checkpoint) allowed to execute on a particular host concurrently Commands applied to a queue affect all jobs associated with this. SGE Components (3) Queues (2): Properties: name: queue s name hostname: machine host of the queue processors: in a multiprocessor system are the processors to which queue has access qtype: type of jobs permited to run in this queue (Interactive, Batch, Parallel, Checkpointing) slots: the numbers of jobs that can run concurrently in that queue
5 SGE Components (4) Queues (3): Properties (2): owner_lists: queue s owners user_lists: users o grups ids of those who may access the queue xuser_lists: userso grupsidsofthosewhomay not access the queue complex_list: indicate the complexs associated with the queue complex values: assigns capacities as provided for this queue for certain complex attributes SGE Components (5) Complex: Set of features (resources) associated with a queue, a hosts, or the entire cluster that are known to SGE. Cell: Each loosely separated SGE cluster, with a different configuration and master machine. The SGE_CELL environment variable permit discriminate among clusters
6 SGE funcionality Is controlled by four daemons: sge_qmaster: control all the cluster s management and scheduling activities Receive scheduling decisions from sge_schedd Requets actions from sge_execd on the execution hosts Mantain tables about cluster status sge_shadowd: daemon used if exist a host backup (shadow master host) for the functionality of sge_qmaster SGE functionality (2) sge_schedd: mantain an up to date view of the cluster s status with the data provided by the sge_qmaster daemon. It : Decide which jobs are forwarded to which queues Comunicate these decisions to the sge_qmaster, who initiates the appropriate actions
7 SGE funcionality (3) sge_execd: is responsible for the queues on its host and for the execution of the jobs in this queues. It send information to the master host (sge_qmaster) about jobs status or load on its host. sge_commd: all the daemons communicates among them through the communication daemons (one per host) SGE functionality (4) Master Host sge_qmaster sge_schedd q2 q3 sge_execd sge_commd sge_commd sge_commd sge_execd q1 switch sge_commd sge_execd q4 q5
8 Using SGE Depend of the user type executing the SGE command. SGE define four types of users: Managers: Have full capabilities to manipulate SGE Operators: Can execute all the commands like managers, with the exception of making configuration changes to the SGE Owners: Are defined by queue and can manipulate the owned queues or jobs within them. Users: Only can manage the owned jobs and only can use queues or parallel environments where are authorized Using SGE (2) Command Manager Operator Owner User qacct qalter qconf No system setup changes Shown only Shown only qdel qhold qhost qlogin
9 Using SGE (3) Command Manager Operator Owner User qmod qmon qrls No system setup changes Own jobs and owned queues only No configuration changes No configuration changes qselect qsh qstat Submitting Jobs Prerequisites ensure that in your.[t]cshrc or. bashrc no commands are executed that need a terminal (tty) bash, sh or ksh tty s if [ $? = 0 ]; then stty erase ^H fi csh or tcsh tty s if ( $status = 0 ) then stty erase ^H endif
10 Submitting Jobs (2) Prerequisites (2) ensure that in your.[t]cshrc or.bashrc you set executable search path and other SGE environmental conditions csh or tcsh: source <sge_root_dir>/default/common/settings.csh bash, sh or ksh:. <sge_root_dir>/default/common/settings.sh Submitting Jobs (3) specify what script should be executed qsub cwd job_script -cwd: run the job from the current working directory. (Default: $HOME) in the simplest case the job script contains one line, the name of the executable various examples in <sge_root_dir>/examples/jobs/ many options are available for qsub man qsub
11 Submitting Jobs (4) Example of a script file #!/bin/csh WORKDIR=/tmp/scratch/$USER DATADIR=$HOME/data mkdir -p $WORKDIR cp $DATADIR/input_data $WORKDIR cd $WORKDIR executable < input_data > out_executable cp out_executable $DATADIR rm rf $WORKDIR Submitting Jobs (5) Output and Error redirection: Default standard output filename: <Job_name>.o<Job_id> Can by changed with the o option Default standard error filename: <Job_name>.e<Job_id> Can by changed with the e option Active SGE comments in script files: Per default are identified by #$
12 Submitting Jobs (6) Array Jobs: Are parametrized executions of the same script SGE view them as an array of independent tasks joined into a single job. task_id is the array job task index number Each task can use the environment variable $SGE_TASK_ID to retrieve their own task index number and use it to access input data sets arranged for this task_id Submitting Jobs (7) Array Jobs (2): Example: qsub l h_cpu=0:30:0 t 2-10:2 script.sh input.data Default standard output filename: <Job_name>.o<Job_id>.<Task_id> Default standard error filename: <Job_name>.e<Job_id>.<Task_id> Can be monitored and controlled as a total or by individual or subset of tasks
13 Submitting Jobs (8) Interactive Jobs: Are executed on interactive queues Three ways are available: qlogin: start a telnet-like sesion on a host choosed by SGE qrsh: Is like rsh or rlogin UNIX commands qsh: Is an xterm that is brought up with the display set corresponding to the setting of the DISPLAY environment variable. If this variable is not set, the xterm is directed to the 0.0 screen of the X server on the host from which the interactive job was submitted. DISPLAY can be set with the -display option. Monitoring and Controlling Jobs qstat: show job/queue status Whithout arguments show running/pending jobs -j show detailed information on running/pending jobs -f show submitted jobs and full listing of all queues qhost: show job/host status Whithout arguments show all execution host and their configuration -q show detailed information on queues at each host
14 Monitoring and Controlling Jobs (2) qdel: cancel jobs submitted through SGE qdel <job_id> qmod: suspend/unsuspend running jobs qmod s <job_id> (suspend) qmod us <job_id> (unsuspend) qhold: holds back pending jobs from execution qrls: releases jobs from holds previously assigned to them Parallel Jobs Are submitted to run on parallel environments Parallel environments are procedures to accomplish with requeriments needed to run a specific parallel application One parallel environment by each class or type of parallel application configured into the cluster
15 Parallel Jobs (2) qconf ap <parallel environment name> create a new parallel environment qconf spl list all defined parallel environments qconf sp <parallel environment name> show detailed information on the specified parallel environtment name Parallel Jobs (3) Parallel environment example: $ qconf -sp mpich pe_name mpich queue_list all slots 8 user_lists NONE xuser_lists NONE start_proc_args $pe_hostfile /usr/local/sge/mpi/startmpi.sh -catch_rsh stop_proc_args /usr/local/sge/mpi/stopmpi.sh allocation_rule $round_robin control_slaves TRUE job_is_first_task FALSE
16 Parallel Jobs (4) Script example: #!/bin/csh # # (c) 2002 Sun Microsystems, Inc. Use is subject to license terms. # # our name #$ -N MPI_calc_PI_Job # # pe request #$ -pe mpich 2-6 # #$ -v MPIR_HOME=/usr/local/mpich # # needs in # $NSLOTS # the number of tasks to be used # $TMPDIR/machines # a valid machine file to be passed to mpirun # echo "Got $NSLOTS slots." # $MPIR_HOME/bin/mpirun -np $NSLOTS -machinefile $TMPDIR/machines $HOME/MPI/cpi Checkpointing SGE support two class of checkpointing: User level checkpointing Operating system level checkpointing Checkpointing environments must be defined by each type of application with this support When a checkpointing job is launched this must be indicated using the ckpt option of the qsub command
17 Checkpointing (2) Checkpointing environments are defined in configuration files: Define the operations to: initiating a checkpoint generation migrate a checkpoint job to another host restart of a checkpointed application As well as the list of queues which are eligible for a check-pointing method. Checkpointing (3) Checkpoint environment file format: ckpt_name <name> interface user defined or os provided. ckpt_command command to initiate the checkpoint. migr_command command used during a migration of a checkpointing job from one host to another. restart_command command used to restart a previously checkpointed application. clean_command command used to cleanup after a checkpointed application has finished. ckpt_dir where checkpoint file should be stored. queue_list all or comma separated list of queues signal Unix signal to be sent to a job to initiate a checkpoint generation when when generate the checkpoints: s (shutdown the node) m (periodically, at the min_cpu_interval interval defined by the queue) x (when the job gets suspended) r (job will be rescheduled (not checkpointed))
18 SGE Administration All administration activities on SGE are commited through the qmon command Basically: qconf a<h q s > <associated arguments> qconf d<h q e conf s > <associated arguments> qconf m<q conf > <associated arguments> qconf s<h s sel conf > <associated arguments> QMON: the SGE GUI
Grid Engine. Application Integration
Grid Engine Application Integration Getting Stuff Done. Batch Interactive - Terminal Interactive - X11/GUI Licensed Applications Parallel Jobs DRMAA Batch Jobs Most common What is run: Shell Scripts Binaries
Introduction to Sun Grid Engine (SGE)
Introduction to Sun Grid Engine (SGE) What is SGE? Sun Grid Engine (SGE) is an open source community effort to facilitate the adoption of distributed computing solutions. Sponsored by Sun Microsystems
Introduction to Sun Grid Engine 5.3
CHAPTER 1 Introduction to Sun Grid Engine 5.3 This chapter provides background information about the Sun Grid Engine 5.3 system that is useful to users and administrators alike. In addition to a description
Grid Engine experience in Finis Terrae, large Itanium cluster supercomputer. Pablo Rey Mayo Systems Technician, Galicia Supercomputing Centre (CESGA)
Grid Engine experience in Finis Terrae, large Itanium cluster supercomputer Pablo Rey Mayo Systems Technician, Galicia Supercomputing Centre (CESGA) Agenda Introducing CESGA Finis Terrae Architecture Grid
Grid Engine Basics. Table of Contents. Grid Engine Basics Version 1. (Formerly: Sun Grid Engine)
Grid Engine Basics (Formerly: Sun Grid Engine) Table of Contents Table of Contents Document Text Style Associations Prerequisites Terminology What is the Grid Engine (SGE)? Loading the SGE Module on Turing
User s Guide. Introduction
CHAPTER 3 User s Guide Introduction Sun Grid Engine (Computing in Distributed Networked Environments) is a load management tool for heterogeneous, distributed computing environments. Sun Grid Engine provides
Notes on the SNOW/Rmpi R packages with OpenMPI and Sun Grid Engine
Notes on the SNOW/Rmpi R packages with OpenMPI and Sun Grid Engine Last updated: 6/2/2008 4:43PM EDT We informally discuss the basic set up of the R Rmpi and SNOW packages with OpenMPI and the Sun Grid
SGE Roll: Users Guide. Version @VERSION@ Edition
SGE Roll: Users Guide Version @VERSION@ Edition SGE Roll: Users Guide : Version @VERSION@ Edition Published Aug 2006 Copyright 2006 UC Regents, Scalable Systems Table of Contents Preface...i 1. Requirements...1
Running ANSYS Fluent Under SGE
Running ANSYS Fluent Under SGE ANSYS, Inc. Southpointe 275 Technology Drive Canonsburg, PA 15317 [email protected] http://www.ansys.com (T) 724-746-3304 (F) 724-514-9494 Release 15.0 November 2013 ANSYS,
Grid Engine Users Guide. 2011.11p1 Edition
Grid Engine Users Guide 2011.11p1 Edition Grid Engine Users Guide : 2011.11p1 Edition Published Nov 01 2012 Copyright 2012 University of California and Scalable Systems This document is subject to the
Cluster Computing With R
Cluster Computing With R Stowers Institute for Medical Research R/Bioconductor Discussion Group Earl F. Glynn Scientific Programmer 18 December 2007 1 Cluster Computing With R Accessing Linux Boxes from
Oracle Grid Engine. User Guide Release 6.2 Update 7 E21976-02
Oracle Grid Engine User Guide Release 6.2 Update 7 E21976-02 February 2012 Oracle Grid Engine User Guide, Release 6.2 Update 7 E21976-02 Copyright 2000, 2012, Oracle and/or its affiliates. All rights reserved.
Grid Engine Training Introduction
Grid Engine Training Jordi Blasco ([email protected]) 26-03-2012 Agenda 1 How it works? 2 History Current status future About the Grid Engine version of this training Documentation 3 Grid Engine internals
SMRT Analysis Software Installation (v2.3.0)
SMRT Analysis Software Installation (v2.3.0) Introduction This document describes the basic requirements for installing SMRT Analysis v2.3.0 on a customer system. SMRT Analysis is designed to be installed
Grid 101. Grid 101. Josh Hegie. [email protected] http://hpc.unr.edu
Grid 101 Josh Hegie [email protected] http://hpc.unr.edu Accessing the Grid Outline 1 Accessing the Grid 2 Working on the Grid 3 Submitting Jobs with SGE 4 Compiling 5 MPI 6 Questions? Accessing the Grid Logging
Efficient cluster computing
Efficient cluster computing Introduction to the Sun Grid Engine (SGE) queuing system Markus Rampp (RZG, MIGenAS) MPI for Evolutionary Anthropology Leipzig, Feb. 16, 2007 Outline Introduction Basic concepts:
Streamline Computing Linux Cluster User Training. ( Nottingham University)
1 Streamline Computing Linux Cluster User Training ( Nottingham University) 3 User Training Agenda System Overview System Access Description of Cluster Environment Code Development Job Schedulers Running
GRID Computing: CAS Style
CS4CC3 Advanced Operating Systems Architectures Laboratory 7 GRID Computing: CAS Style campus trunk C.I.S. router "birkhoff" server The CAS Grid Computer 100BT ethernet node 1 "gigabyte" Ethernet switch
Oracle Grid Engine. Administration Guide Release 6.2 Update 7 E21978-01
Oracle Grid Engine Administration Guide Release 6.2 Update 7 E21978-01 August 2011 Oracle Grid Engine Administration Guide, Release 6.2 Update 7 E21978-01 Copyright 2000, 2011, Oracle and/or its affiliates.
Installing and running COMSOL on a Linux cluster
Installing and running COMSOL on a Linux cluster Introduction This quick guide explains how to install and operate COMSOL Multiphysics 5.0 on a Linux cluster. It is a complement to the COMSOL Installation
Grid Engine 6. Troubleshooting. BioTeam Inc. [email protected]
Grid Engine 6 Troubleshooting BioTeam Inc. [email protected] Grid Engine Troubleshooting There are two core problem types Job Level Cluster seems OK, example scripts work fine Some user jobs/apps fail Cluster
How To Use A Job Management System With Sun Hpc Cluster Tools
A Comparison of Job Management Systems in Supporting HPC ClusterTools Presentation for SUPerG Vancouver, Fall 2000 Chansup Byun and Christopher Duncan HES Engineering-HPC, Sun Microsystems, Inc. Stephanie
1.0. User Manual For HPC Cluster at GIKI. Volume. Ghulam Ishaq Khan Institute of Engineering Sciences & Technology
Volume 1.0 FACULTY OF CUMPUTER SCIENCE & ENGINEERING Ghulam Ishaq Khan Institute of Engineering Sciences & Technology User Manual For HPC Cluster at GIKI Designed and prepared by Faculty of Computer Science
PBS Tutorial. Fangrui Ma Universit of Nebraska-Lincoln. October 26th, 2007
PBS Tutorial Fangrui Ma Universit of Nebraska-Lincoln October 26th, 2007 Abstract In this tutorial we gave a brief introduction to using PBS Pro. We gave examples on how to write control script, and submit
Quick Tutorial for Portable Batch System (PBS)
Quick Tutorial for Portable Batch System (PBS) The Portable Batch System (PBS) system is designed to manage the distribution of batch jobs and interactive sessions across the available nodes in the cluster.
Enigma, Sun Grid Engine (SGE), and the Joint High Performance Computing Exchange (JHPCE) Cluster
Enigma, Sun Grid Engine (SGE), and the Joint High Performance Computing Exchange (JHPCE) Cluster http://www.biostat.jhsph.edu/bit/sge_lecture.ppt.pdf Marvin Newhouse Fernando J. Pineda The JHPCE staff:
Sun Grid Engine 5.2.3 Manual
Sun Grid Engine 5.2.3 Manual Sun Microsystems, Inc. 901 San Antonio Road Palo Alto, CA 94303-4900 U.S.A. 650-960-1300 Part No. 816-2077-10 July 2001 Copyright 2001 Sun Microsystems, Inc., 901 San Antonio
High Performance Computing
High Performance Computing at Stellenbosch University Gerhard Venter Outline 1 Background 2 Clusters 3 SU History 4 SU Cluster 5 Using the Cluster 6 Examples What is High Performance Computing? Wikipedia
Submitting Jobs to the Sun Grid Engine. CiCS Dept The University of Sheffield. Email [email protected] [email protected].
Submitting Jobs to the Sun Grid Engine CiCS Dept The University of Sheffield Email [email protected] [email protected] October 2012 Topics Covered Introducing the grid and batch concepts.
High Performance Computing Facility Specifications, Policies and Usage. Supercomputer Project. Bibliotheca Alexandrina
High Performance Computing Facility Specifications, Policies and Usage Supercomputer Project Bibliotheca Alexandrina Bibliotheca Alexandrina 1/16 Topics Specifications Overview Site Policies Intel Compilers
Sun ONE Grid Engine, Enterprise Edition Administration and User s Guide
Sun ONE Grid Engine, Enterprise Edition Administration and User s Guide Sun Microsystems, Inc. 4150 Network Circle Santa Clara, CA 95054 U.S.A. 650-960-1300 Part No. 816-4739-11 October 2002, Revision
Beyond Windows: Using the Linux Servers and the Grid
Beyond Windows: Using the Linux Servers and the Grid Topics Linux Overview How to Login & Remote Access Passwords Staying Up-To-Date Network Drives Server List The Grid Useful Commands Linux Overview Linux
LSKA 2010 Survey Report Job Scheduler
LSKA 2010 Survey Report Job Scheduler Graduate Institute of Communication Engineering {r98942067, r98942112}@ntu.edu.tw March 31, 2010 1. Motivation Recently, the computing becomes much more complex. However,
Grid Engine Administration. Overview
Grid Engine Administration Overview This module covers Grid Problem Types How it works Distributed Resource Management Grid Engine 6 Variants Grid Engine Scheduling Grid Engine 6 Architecture Grid Problem
MPI / ClusterTools Update and Plans
HPC Technical Training Seminar July 7, 2008 October 26, 2007 2 nd HLRS Parallel Tools Workshop Sun HPC ClusterTools 7+: A Binary Distribution of Open MPI MPI / ClusterTools Update and Plans Len Wisniewski
Hodor and Bran - Job Scheduling and PBS Scripts
Hodor and Bran - Job Scheduling and PBS Scripts UND Computational Research Center Now that you have your program compiled and your input file ready for processing, it s time to run your job on the cluster.
Cluster@WU User s Manual
Cluster@WU User s Manual Stefan Theußl Martin Pacala September 29, 2014 1 Introduction and scope At the WU Wirtschaftsuniversität Wien the Research Institute for Computational Methods (Forschungsinstitut
CycleServer Grid Engine Support Install Guide. version 1.25
CycleServer Grid Engine Support Install Guide version 1.25 Contents CycleServer Grid Engine Guide 1 Administration 1 Requirements 1 Installation 1 Monitoring Additional OGS/SGE/etc Clusters 3 Monitoring
LAE 4.6.0 Enterprise Server Installation Guide
LAE 4.6.0 Enterprise Server Installation Guide 2013 Lavastorm Analytics, Inc. Rev 01/2013 Contents Introduction... 3 Installing the LAE Server on UNIX... 3 Pre-Installation Steps... 3 1. Third-Party Software...
Scheduling in SAS 9.3
Scheduling in SAS 9.3 SAS Documentation The correct bibliographic citation for this manual is as follows: SAS Institute Inc 2011. Scheduling in SAS 9.3. Cary, NC: SAS Institute Inc. Scheduling in SAS 9.3
High Performance Computing with Sun Grid Engine on the HPSCC cluster. Fernando J. Pineda
High Performance Computing with Sun Grid Engine on the HPSCC cluster Fernando J. Pineda HPSCC High Performance Scientific Computing Center (HPSCC) " The Johns Hopkins Service Center in the Dept. of Biostatistics
Introduction to Grid Engine
Introduction to Grid Engine Workbook Edition 8 January 2011 Document reference: 3609-2011 Introduction to Grid Engine for ECDF Users Workbook Introduction to Grid Engine for ECDF Users Author: Brian Fletcher,
IBM WebSphere Application Server Version 7.0
IBM WebSphere Application Server Version 7.0 Centralized Installation Manager for IBM WebSphere Application Server Network Deployment Version 7.0 Note: Before using this information, be sure to read the
INF-110. GPFS Installation
INF-110 GPFS Installation Overview Plan the installation Before installing any software, it is important to plan the GPFS installation by choosing the hardware, deciding which kind of disk connectivity
Release Notes for Open Grid Scheduler/Grid Engine. Version: Grid Engine 2011.11
Release Notes for Open Grid Scheduler/Grid Engine Version: Grid Engine 2011.11 New Features Berkeley DB Spooling Directory Can Be Located on NFS The Berkeley DB spooling framework has been enhanced such
Introduction to the SGE/OGS batch-queuing system
Grid Computing Competence Center Introduction to the SGE/OGS batch-queuing system Riccardo Murri Grid Computing Competence Center, Organisch-Chemisches Institut, University of Zurich Oct. 6, 2011 The basic
Oracle Grid Engine. Installation and Upgrade Guide Release 6.2 Update 7 E21973-02
Oracle Grid Engine Installation and Upgrade Guide Release 6.2 Update 7 E21973-02 February 2012 Oracle Grid Engine Installation and Upgrade Guide, Release 6.2 Update 7 E21973-02 Copyright 2000, 2012, Oracle
Manual for using Super Computing Resources
Manual for using Super Computing Resources Super Computing Research and Education Centre at Research Centre for Modeling and Simulation National University of Science and Technology H-12 Campus, Islamabad
Using Parallel Computing to Run Multiple Jobs
Beowulf Training Using Parallel Computing to Run Multiple Jobs Jeff Linderoth August 5, 2003 August 5, 2003 Beowulf Training Running Multiple Jobs Slide 1 Outline Introduction to Scheduling Software The
Sun Grid Engine, a new scheduler for EGEE middleware
Sun Grid Engine, a new scheduler for EGEE middleware G. Borges 1, M. David 1, J. Gomes 1, J. Lopez 2, P. Rey 2, A. Simon 2, C. Fernandez 2, D. Kant 3, K. M. Sephton 4 1 Laboratório de Instrumentação em
Sun Grid Engine, a new scheduler for EGEE
Sun Grid Engine, a new scheduler for EGEE G. Borges, M. David, J. Gomes, J. Lopez, P. Rey, A. Simon, C. Fernandez, D. Kant, K. M. Sephton IBERGRID Conference Santiago de Compostela, Spain 14, 15, 16 May
Sun Grid Engine Update
Sun Grid Engine Update SGE Workshop 2007, Regensburg September 10-12, 2007 Andy Schwierskott Sun Microsystems Copyright Sun Microsystems What is Grid Computing? The network is the computer > Distributed
Linux command line. An introduction to the Linux command line for genomics. Susan Fairley
Linux command line An introduction to the Linux command line for genomics Susan Fairley Aims Introduce the command line Provide an awareness of basic functionality Illustrate with some examples Provide
159.735. Final Report. Cluster Scheduling. Submitted by: Priti Lohani 04244354
159.735 Final Report Cluster Scheduling Submitted by: Priti Lohani 04244354 1 Table of contents: 159.735... 1 Final Report... 1 Cluster Scheduling... 1 Table of contents:... 2 1. Introduction:... 3 1.1
Running applications on the Cray XC30 4/12/2015
Running applications on the Cray XC30 4/12/2015 1 Running on compute nodes By default, users do not log in and run applications on the compute nodes directly. Instead they launch jobs on compute nodes
How to Run Parallel Jobs Efficiently
How to Run Parallel Jobs Efficiently Shao-Ching Huang High Performance Computing Group UCLA Institute for Digital Research and Education May 9, 2013 1 The big picture: running parallel jobs on Hoffman2
Batch Systems. provide a mechanism for submitting, launching, and tracking jobs on a shared resource
PBS INTERNALS PBS & TORQUE PBS (Portable Batch System)-software system for managing system resources on workstations, SMP systems, MPPs and vector computers. It was based on Network Queuing System (NQS)
Cisco Setting Up PIX Syslog
Table of Contents Setting Up PIX Syslog...1 Introduction...1 Before You Begin...1 Conventions...1 Prerequisites...1 Components Used...1 How Syslog Works...2 Logging Facility...2 Levels...2 Configuring
Miami University RedHawk Cluster Working with batch jobs on the Cluster
Miami University RedHawk Cluster Working with batch jobs on the Cluster The RedHawk cluster is a general purpose research computing resource available to support the research community at Miami University.
KISTI Supercomputer TACHYON Scheduling scheme & Sun Grid Engine
KISTI Supercomputer TACHYON Scheduling scheme & Sun Grid Engine 슈퍼컴퓨팅인프라지원실 윤 준 원 ([email protected]) 2014.07.15 Scheduling (batch job processing) Distributed resource management Features of job schedulers
Ra - Batch Scripts. Timothy H. Kaiser, Ph.D. [email protected]
Ra - Batch Scripts Timothy H. Kaiser, Ph.D. [email protected] Jobs on Ra are Run via a Batch System Ra is a shared resource Purpose: Give fair access to all users Have control over where jobs are run Set
Batch Job Analysis to Improve the Success Rate in HPC
Batch Job Analysis to Improve the Success Rate in HPC 1 JunWeon Yoon, 2 TaeYoung Hong, 3 ChanYeol Park, 4 HeonChang Yu 1, First Author KISTI and Korea University, [email protected] 2,3, KISTI,[email protected],[email protected]
TABLE OF CONTENTS OVERVIEW SYSTEM REQUIREMENTS - SAP FOR ORACLE IDATAAGENT GETTING STARTED - DEPLOYING ON WINDOWS
Page 1 of 44 Quick Start - SAP for Oracle idataagent TABLE OF CONTENTS OVERVIEW Introduction Key Features Full Range of Backup and Recovery Options SnapProtect Backup Command Line Support Backup and Recovery
Batch Scripts for RA & Mio
Batch Scripts for RA & Mio Timothy H. Kaiser, Ph.D. [email protected] 1 Jobs are Run via a Batch System Ra and Mio are shared resources Purpose: Give fair access to all users Have control over where jobs
Features - SRM UNIX File System Agent
Page 1 of 45 Features - SRM UNIX File System Agent TABLE OF CONTENTS OVERVIEW SYSTEM REQUIREMENTS - SRM UNIX FILE SYSTEM AGENT INSTALLATION Install the SRM Unix File System Agent Install the SRM Unix File
An Introduction to High Performance Computing in the Department
An Introduction to High Performance Computing in the Department Ashley Ford & Chris Jewell Department of Statistics University of Warwick October 30, 2012 1 Some Background 2 How is Buster used? 3 Software
Configuring LocalDirector Syslog
Configuring LocalDirector Syslog Document ID: 22178 LocalDirector is now End of Sale. Refer to the Cisco LocalDirector 400 Series bulletins for more information. Contents Introduction Before You Begin
High-Performance Reservoir Risk Assessment (Jacta Cluster)
High-Performance Reservoir Risk Assessment (Jacta Cluster) SKUA-GOCAD 2013.1 Paradigm 2011.3 With Epos 4.1 Data Management Configuration Guide 2008 2013 Paradigm Ltd. or its affiliates and subsidiaries.
Job Scheduling with Moab Cluster Suite
Job Scheduling with Moab Cluster Suite IBM High Performance Computing February 2010 Y. Joanna Wong, Ph.D. [email protected] 2/22/2010 Workload Manager Torque Source: Adaptive Computing 2 Some terminology..
Wolfr am Lightweight Grid M TM anager USER GUIDE
Wolfram Lightweight Grid TM Manager USER GUIDE For use with Wolfram Mathematica 7.0 and later. For the latest updates and corrections to this manual: visit reference.wolfram.com For information on additional
SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education www.accre.vanderbilt.
SLURM: Resource Management and Job Scheduling Software Advanced Computing Center for Research and Education www.accre.vanderbilt.edu Simple Linux Utility for Resource Management But it s also a job scheduler!
Sun N1 Grid Engine 6.1 Release Notes
Sun N1 Grid Engine 6.1 Release Notes Sun Microsystems, Inc. 4150 Network Circle Santa Clara, CA 95054 U.S.A. Part No: 820 0700 13 May 2007 (Updated August 2008) Copyright 2007 Sun Microsystems, Inc. 4150
Configuration of High Performance Computing for Medical Imaging and Processing. SunGridEngine 6.2u5
Configuration of High Performance Computing for Medical Imaging and Processing SunGridEngine 6.2u5 A manual guide for installing, configuring and using the cluster. Mohammad Naquiddin Abd Razak Summer
Tutorial: Using WestGrid. Drew Leske Compute Canada/WestGrid Site Lead University of Victoria
Tutorial: Using WestGrid Drew Leske Compute Canada/WestGrid Site Lead University of Victoria Fall 2013 Seminar Series Date Speaker Topic 23 September Lindsay Sill Introduction to WestGrid 9 October Drew
Parallel Debugging with DDT
Parallel Debugging with DDT Nate Woody 3/10/2009 www.cac.cornell.edu 1 Debugging Debugging is a methodical process of finding and reducing the number of bugs, or defects, in a computer program or a piece
Sun ONE Grid Engine 5.3 Release Notes
Sun ONE Grid Engine 5.3 Release Notes Sun Microsystems, Inc. 4150 Network Circle Santa Clara, CA 95054 U.S.A. 650-960-1300 Part No. 816-5077-11 October 2002, Revision 01 Send comments about this document
Running on Blue Gene/Q at Argonne Leadership Computing Facility (ALCF)
Running on Blue Gene/Q at Argonne Leadership Computing Facility (ALCF) ALCF Resources: Machines & Storage Mira (Production) IBM Blue Gene/Q 49,152 nodes / 786,432 cores 768 TB of memory Peak flop rate:
IGEL Universal Management. Installation Guide
IGEL Universal Management Installation Guide Important Information Copyright This publication is protected under international copyright laws, with all rights reserved. No part of this manual, including
Using Symantec NetBackup with Symantec Security Information Manager 4.5
Using Symantec NetBackup with Symantec Security Information Manager 4.5 Using Symantec NetBackup with Symantec Security Information Manager Legal Notice Copyright 2007 Symantec Corporation. All rights
MATLAB Distributed Computing Server System Administrator s Guide. R2013b
MATLAB Distributed Computing Server System Administrator s Guide R2013b How to Contact MathWorks www.mathworks.com Web comp.soft-sys.matlab Newsgroup www.mathworks.com/contact_ts.html Technical Support
Scheduling in SAS 9.4 Second Edition
Scheduling in SAS 9.4 Second Edition SAS Documentation The correct bibliographic citation for this manual is as follows: SAS Institute Inc. 2015. Scheduling in SAS 9.4, Second Edition. Cary, NC: SAS Institute
HP OpenView Storage Data Protector
HP OpenView Storage Data Protector Backup and Restore of High Availability Cluster Multi-Processing for AIX 5.3 Version 1.0 Backup and Restore of HACMP Cluster 1 1. Introduction... 3 2. Overview of HACMP
Opswise Managed File Transfer 5.2.0 for UNIX Quick Start Guide
Opswise Managed File Transfer 5.2.0 for UNIX Quick Start Guide 2015 by Stonebranch, Inc. All Rights Reserved. Opswise Managed File Transfer 5.2.0 for UNIX Quick Start Guide Objective System Requirements
Ahsay Replication Server v5.5. Administrator s Guide. Ahsay TM Online Backup - Development Department
Ahsay Replication Server v5.5 Administrator s Guide Ahsay TM Online Backup - Development Department October 9, 2009 Copyright Notice Ahsay Systems Corporation Limited 2008. All rights reserved. Author:
Job scheduler details
Job scheduler details Advanced Computing Center for Research & Education (ACCRE) Job scheduler details 1 / 25 Outline 1 Batch queue system overview 2 Torque and Moab 3 Submitting jobs (ACCRE) Job scheduler
HPC system startup manual (version 1.30)
HPC system startup manual (version 1.30) Document change log Issue Date Change 1 12/1/2012 New document 2 10/22/2013 Added the information of supported OS 3 10/22/2013 Changed the example 1 for data download
Architecture and Mode of Operation
Open Source Scheduler Architecture and Mode of Operation http://jobscheduler.sourceforge.net Contents Components Platforms & Databases Architecture Configuration Deployment Distributed Processing Security
MATLAB Distributed Computing Server System Administrator's Guide
MATLAB Distributed Computing Server System Administrator's Guide R2015b How to Contact MathWorks Latest news: www.mathworks.com Sales and services: www.mathworks.com/sales_and_services User community:
