An introduction to compute resources in Biostatistics. Chris Scheller schelcj@umich.edu



Similar documents
SLURM Workload Manager

The Asterope compute cluster

General Overview. Slurm Training15. Alfred Gil & Jordi Blasco (HPCNow!)

Introduction to Running Computations on the High Performance Clusters at the Center for Computational Research

SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education

Until now: tl;dr: - submit a job to the scheduler

SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education

Matlab on a Supercomputer

Biowulf2 Training Session

Getting Started with HPC

Berkeley Research Computing. Town Hall Meeting Savio Overview

Introduction to Supercomputing with Janus

Introduction to parallel computing and UPPMAX

An introduction to Fyrkat

Tutorial-4a: Parallel (multi-cpu) Computing

Parallel Processing using the LOTUS cluster

To connect to the cluster, simply use a SSH or SFTP client to connect to:

Managing GPUs by Slurm. Massimo Benini HPC Advisory Council Switzerland Conference March 31 - April 3, 2014 Lugano

Introduction to Running Hadoop on the High Performance Clusters at the Center for Computational Research

Linux für bwgrid. Sabine Richling, Heinz Kredel. Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim. 27.

How to Run Parallel Jobs Efficiently

Submitting batch jobs Slurm on ecgate. Xavi Abellan User Support Section

Grid 101. Grid 101. Josh Hegie.

icer Bioinformatics Support Fall 2011

NEC HPC-Linux-Cluster

Grid Engine Basics. Table of Contents. Grid Engine Basics Version 1. (Formerly: Sun Grid Engine)

An Introduction to High Performance Computing in the Department

Introduction to Linux and Cluster Basics for the CCR General Computing Cluster

Requesting Nodes, Processors, and Tasks in Moab

Using the Yale HPC Clusters

The RWTH Compute Cluster Environment

Background and introduction Using the cluster Summary. The DMSC datacenter. Lars Melwyn Jensen. Niels Bohr Institute University of Copenhagen

Juropa. Batch Usage Introduction. May 2014 Chrysovalantis Paschoulas

How to Use NoMachine 4.4

Using Google Compute Engine

JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert

Bright Cluster Manager 5.2. User Manual. Revision: Date: Fri, 30 Nov 2012

A High Performance Computing Scheduling and Resource Management Primer

HPC-Nutzer Informationsaustausch. The Workload Management System LSF

Job Scheduling Explained More than you ever want to know about how jobs get scheduled on WestGrid systems...

Batch Usage on JURECA Introduction to Slurm. Nov 2015 Chrysovalantis Paschoulas

Work Environment. David Tur HPC Expert. HPC Users Training September, 18th 2015

Running on Blue Gene/Q at Argonne Leadership Computing Facility (ALCF)

CHEOPS Cologne High Efficient Operating Platform for Science Brief Instructions

Chapter 2: Getting Started

PSE Molekulardynamik

CNAG User s Guide. Barcelona Supercomputing Center Copyright c 2015 BSC-CNS December 18, Introduction 2

Visualization Cluster Getting Started

Parallel Debugging with DDT

Hodor and Bran - Job Scheduling and PBS Scripts

bwgrid Treff MA/HD Sabine Richling, Heinz Kredel Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim 29.

Running applications on the Cray XC30 4/12/2015

Martinos Center Compute Clusters

Slurm Workload Manager Architecture, Configuration and Use

A Crash course to (The) Bighouse

8/15/2014. Best (and more) General Information. Staying Informed. Staying Informed. Staying Informed-System Status

Cloud Bursting with SLURM and Bright Cluster Manager. Martijn de Vries CTO

The CNMS Computer Cluster

Agenda. Using HPC Wales 2

The Information Technology Solution. Denis Foueillassar TELEDOS project coordinator

OLCF Best Practices. Bill Renaud OLCF User Assistance Group

InterWorx Clustering Guide. by InterWorx LLC

Performance Monitor. Intellicus Web-based Reporting Suite Version 4.5. Enterprise Professional Smart Developer Smart Viewer

User s Manual

HPC at IU Overview. Abhinav Thota Research Technologies Indiana University

Linux VPS with cpanel. Getting Started Guide

ABAQUS High Performance Computing Environment at Nokia

Remote & Collaborative Visualization. Texas Advanced Compu1ng Center

Using the Windows Cluster

Using NeSI HPC Resources. NeSI Computational Science Team

Using WestGrid. Patrick Mann, Manager, Technical Operations Jan.15, 2014

Creating a DUO MFA Service in AWS

bwgrid Treff MA/HD Sabine Richling, Heinz Kredel Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim 20.

Informationsaustausch für Nutzer des Aachener HPC Clusters

Computing in High- Energy-Physics: How Virtualization meets the Grid

Advanced PBS Workflow Example Bill Brouwer 05/01/12 Research Computing and Cyberinfrastructure Unit, PSU

RWTH GPU Cluster. Sandra Wienke November Rechen- und Kommunikationszentrum (RZ) Fotos: Christian Iwainsky

OLCF Best Practices (and More) Bill Renaud OLCF User Assistance Group

HTCondor at the RAL Tier-1

Slurm License Management

Job Scheduling on a Large UV Chad Vizino SGI User Group Conference May Pittsburgh Supercomputing Center

High Performance Computing Facility Specifications, Policies and Usage. Supercomputer Project. Bibliotheca Alexandrina

Beyond Windows: Using the Linux Servers and the Grid

Miami University RedHawk Cluster Working with batch jobs on the Cluster

Using Parallel Computing to Run Multiple Jobs

Using the Yale HPC Clusters

Simulation of batch scheduling using real production-ready software tools

Compute Cluster Documentation

Quick Tutorial for Portable Batch System (PBS)

Solution for private cloud computing

Debugging with TotalView

Introduction to HPC Workshop. Center for e-research

Overview of HPC Resources at Vanderbilt

Batch Scripts for RA & Mio

CPSC 2800 Linux Hands-on Lab #7 on Linux Utilities. Project 7-1

bwgrid Treff MA/HD Sabine Richling, Heinz Kredel Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim 24.

Introduction to Computing Facilities

Streamline Computing Linux Cluster User Training. ( Nottingham University)

Transcription:

An introduction to compute resources in Biostatistics Chris Scheller schelcj@umich.edu

1. Resources 1. Hardware 2. Account Allocation 3. Storage 4. Software 2. Usage 1. Environment Modules 2. Tools 3. Compiles 3. SLURM 1. Scheduler 2. Submitting Jobs 3. Queue management 4. Summary 1. Other Compute Resources 2. More information

Hardware 19 nodes, 264 cores 32 GB of memory per node gigabit networking Nodes running Ubuntu-server (a debian based distribution) uses the SLURM Resource manager and job scheduler login node: biostat-login.sph.umich.edu

Account Allocation Cluster accounts are available at cost to all biostat faculty, students and researchers. Account requests have to come from your faculty advisor. Your login name is your umich uniqname and the password is your SPH AFS password.

Storage /home 50GB quota per user Only for current jobs Backed up nightly /tmp not shared across nodes can server as scratch space for jobs not backed up /afs SPH AFS cell only only available on login nodes

Software Defaults OpenMPI R gcc/g++ Available SAS Coming soon MATLAB OpenMP Standard tools are also available make, awk, sed perl, python vim, nano

Usage Software modules module list list loaded modules module load <modulename> load <modulename> for use module avail show all available modules module avail <modulename> show all versions of <modulename> available module rm <modulename> unload a currently loaded module

Introduction to SLURM SLURM is the Simple Linux Utility for Resource Management Slurm manages the resources of the cluster ie; cores, nodes Slurm also is the job scheduler Access to all compute nodes goes through the slurm job scheduler

The Scheduler Current scheduler limitations Limited number of jobs in the queue in a running or pending state Maximum number of running jobs Max walltime of any running job Upcoming changes The longer a job is in the queue the higher its priority Recent usage: Accounts with less usage over the past week will have a higher priority than those with a lot of usage.

Job Queue Jobs are ran on a First In First Out basis Any holes in resource usage will be backfilled by the scheduler with smaller jobs that would have sat waiting. Reservations can be placed on nodes for users and or groups

Submitting jobs There are two primary means of submitting jobs to the cluster. srun submit a single job via the command line srun --time= <command_to_run> time is a required option srun --time=02:00 R CMD BATCH script.r srun --time=00:05 --nodes=2 hostname srun --time=1-0 --license=sas script.sas srun --time=28-0 --cpus_per_task=4 a.out sbatch submit batch jobs from a batch file. A batch file is a short text file with the commands to run your job along with commands for the scheduler. sbatch <batch_file>

$ cat job.txt #!/bin/sh #SBATCH --mail-type=all #SBATCH --mail-user=uniqname@umich.edu #SBATCH --time=1-12:00 #SBATCH --nodes=2 #SBATCH --job-name=test_job R CMD BATCH script.r $sbatch./job.txt Submitted batch job 123456 $ cat slurm-123456.out # STDOUT/STDERR are written to this file

Queue Management squeue view the contents of the job queue scancel kill a queued job regardless of state sinfo view information about cluster nodes and partitions

Summary Other compute resources unfunded student compute servers. Not part of the cluster, jobs are run locally. compute0.sph.umich.edu compute4.sph.umich.edu http://www.sph.umich.edu/computing/unix/machines.html FLUX University wide cluster ran by the Center for Advanced Computing. Currently has 2000 cores available Allocations have a cost billed per core per month http://cac.engin.umich.edu/

More Information To request an account have your advisor email sph-biostathelp@umich.edu with your uniqname. Primary cluster documentation https://webservices.itcs.umich.edu/mediawiki/biostatistics/index.php New user introduction page https://webservices.itcs.umich.edu/mediawiki/biostatistics/index.php/cluster/newuserintro