8/15/2014. Best Practices @OLCF (and more) General Information. Staying Informed. Staying Informed. Staying Informed-System Status



Similar documents
OLCF Best Practices (and More) Bill Renaud OLCF User Assistance Group

OLCF Best Practices. Bill Renaud OLCF User Assistance Group

Work Environment. David Tur HPC Expert. HPC Users Training September, 18th 2015

Using WestGrid. Patrick Mann, Manager, Technical Operations Jan.15, 2014

Research Technologies Data Storage for HPC

Introduction to Linux and Cluster Basics for the CCR General Computing Cluster

SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education

High Performance Computing at the Oak Ridge Leadership Computing Facility

SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education

High Performance Computing Facility Specifications, Policies and Usage. Supercomputer Project. Bibliotheca Alexandrina

The RWTH Compute Cluster Environment

An Introduction to High Performance Computing in the Department

Introduction to Running Hadoop on the High Performance Clusters at the Center for Computational Research

Running applications on the Cray XC30 4/12/2015

Introduction to Running Computations on the High Performance Clusters at the Center for Computational Research

NEC HPC-Linux-Cluster

Job Scheduling Explained More than you ever want to know about how jobs get scheduled on WestGrid systems...

Tutorial: Using WestGrid. Drew Leske Compute Canada/WestGrid Site Lead University of Victoria

Job Scheduling with Moab Cluster Suite

Martinos Center Compute Clusters

HPC at IU Overview. Abhinav Thota Research Technologies Indiana University

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

The Moab Scheduler. Dan Mazur, McGill HPC Aug 23, 2013

Parallel Debugging with DDT

Linux für bwgrid. Sabine Richling, Heinz Kredel. Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim. 27.

Biowulf2 Training Session

Hodor and Bran - Job Scheduling and PBS Scripts

Introduction to Supercomputing with Janus

LOCKSS on LINUX. Installation Manual and the OpenBSD Transition 02/17/2011

User s Manual

RA MPI Compilers Debuggers Profiling. March 25, 2009

Streamline Computing Linux Cluster User Training. ( Nottingham University)

SLURM Workload Manager

Agenda. Using HPC Wales 2

Installation Manual v2.0.0

Miami University RedHawk Cluster Working with batch jobs on the Cluster

Rev 7 06-OCT Site Manager Installation Guide

Using the Yale HPC Clusters

A Crash course to (The) Bighouse

PBS + Maui Scheduler

WebSphere Business Monitor

How To Limit Volume In Bacula

Cleaning Up Your Outlook Mailbox and Keeping It That Way ;-) Mailbox Cleanup. Quicklinks >>

Incremental Backup Script. Jason Healy, Director of Networks and Systems

Using NeSI HPC Resources. NeSI Computational Science Team

Parallel Processing using the LOTUS cluster

The CNMS Computer Cluster

SysPatrol - Server Security Monitor

RWTH GPU Cluster. Sandra Wienke November Rechen- und Kommunikationszentrum (RZ) Fotos: Christian Iwainsky

Job scheduler details

An introduction to compute resources in Biostatistics. Chris Scheller

SQL Server Replication Guide

Optimization tools. 1) Improving Overall I/O

Metalogix SharePoint Backup. Advanced Installation Guide. Publication Date: August 24, 2015

Getting Started with HPC

11. Configuring the Database Archiving Mode.

Informationsaustausch für Nutzer des Aachener HPC Clusters

Avira AntiVir MailGate 3.2 Release Notes

Deploying LoGS to analyze console logs on an IBM JS20

LOCKSS on LINUX. CentOS6 Installation Manual 08/22/2013

JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert

UNISOL SysAdmin. SysAdmin helps systems administrators manage their UNIX systems and networks more effectively.

Database Maintenance Guide

Survey of Filesystems for Embedded Linux. Presented by Gene Sally CELF

Features of AnyShare

PBS Tutorial. Fangrui Ma Universit of Nebraska-Lincoln. October 26th, 2007

Section 1 Using the Support Center

INF-110. GPFS Installation

Batch Scripts for RA & Mio

Advanced PBS Workflow Example Bill Brouwer 05/01/12 Research Computing and Cyberinfrastructure Unit, PSU

Using Parallel Computing to Run Multiple Jobs

How To Backup Your Computer With A File Copy Engine

The Asterope compute cluster

Enterprise Manager Performance Tips

SQL Server Database Administrator s Guide

Windchill ProjectLink Curriculum Guide

User Guide. Version 3.2. Copyright Snow Software AB. All rights reserved.

Windows HPC 2008 Cluster Launch

Gladinet Cloud Backup V3.0 User Guide

Application Note. UDO Archive Appliance and C2C Archive One

Charter Business Desktop Security Administrator's Guide

Running on Blue Gene/Q at Argonne Leadership Computing Facility (ALCF)

WhatsUp Gold v16.3 Installation and Configuration Guide

Matlab on a Supercomputer

PGP Portable Quick Start Guide Version 10.2

Running a Workflow on a PowerCenter Grid

Web Server (Step 1) Processes request and sends query to SQL server via ADO/OLEDB. Web Server (Step 2) Creates HTML page dynamically from record set

LISTSERV in a High-Availability Environment DRAFT Revised

Spectrum Technology Platform. Version 9.0. Spectrum Spatial Administration Guide

NETWRIX USER ACTIVITY VIDEO REPORTER

Acronis Backup & Recovery 10 Server for Linux. Installation Guide

Archiving Troubleshooting Guide

IUCLID 5 Guidance and support. Installation Guide Distributed Version. Linux - Apache Tomcat - PostgreSQL

Transcription:

Best Practices @OLCF (and more) Bill Renaud OLCF User Support General Information This presentation covers some helpful information for users of OLCF Staying informed Aspects of system usage that may differ from your past experience Common error messages Common questions Quick tips & best practices This is by no means an all-inclusive presentation Feel free to ask questions ORNL is managed by UT-Battelle for the US Department of Energy 2 Best Practices @OLCF (& more) Staying Informed Staying Informed OLCF provides multiple layers of user notifications about system status and downtimes OLCF Weekly Update Status indicators on olcf.ornl.gov Opt-in email lists (subscription management currently unavailable) Twitter (@OLCFStatus) StatusCast (coming soon) A summary of these items can also be found at http://www.olcf.ornl.gov/kb_articles/communications-to-users/ 4 Best Practices @OLCF (& more) Staying Informed-Weekly Update Sent weekly (Thu/Fri) Contains Announcements about upcoming training Announcements about system changes Planned outages All OLCF users should receive this email Let us know if you re not receiving it Logs from monitoring software parsed to make educated guess on system status Status is sent to multiple destinations Websites Twitter Email lists Fairly accurate, but still a fully automated process Possibility of both false positives and false negative We do take some measures to mitigate this 5 Best Practices @OLCF (& more) 6 Best Practices @OLCF (& more) 1

(via www.olcf.ornl.gov) (via Twitter) @OLCFStatus 7 Best Practices @OLCF (& more) 8 Best Practices @OLCF (& more) (via Opt-In Email Lists) We also send status up/down notices via email These area available on an opt-in basis Online sign up (coming soon) at http://www.olcf.ornl.gov/kb_articles/system-notification-lists/ Sign up now by emailing help@olcf.ornl.gov Lists aren t just for status; you may want to sign up (via Opt-in Email Lists) http://users.nccs.gov/subscribe 9 Best Practices @OLCF (& more) 10 Best Practices @OLCF (& more) (via StatusCast) Users have told us via our survey that they d like more status information online Work is ongoing on a new tool to provide that information Current status Outages (upcoming, current, and historical) Announcements Stay tuned to the OLCF Weekly Update for more information System Differences: Compiling & Finding Software 11 Best Practices @OLCF (& more) 2

Finding Software Software Updates Some is part of the default environment Basic commands Text editors Other software typically managed via modules Software is actually installed in /sw module avail, module load, etc. More information is available on the OLCF website Compilers, MPI and GPU libraries, etc. are also provided via modules (but are not in /sw) Cray-provided software (compilers, MPI, CUDA, etc.) goes through testing before we receive it Usually not available to us immediately after vendor releases it OLCF provided software is moving to a model of updates at certain intervals Not all minor versions will be installed We ll provide build recipes in case you want/need a minor version 13 Best Practices @OLCF (& more) 14 Best Practices @OLCF (& more) Software Installation Compiling for the XK7 and XC30 You are free to install software in your directories Subject to export control, license agreements, etc. If you think software would be of general interest, you might ask us to install it for general use Via email to help@olcf.ornl.gov or the request form at http://www.olcf.ornl.gov/support/software/software-request/ Requests are reviewed by our software council May differ (greatly) from your previous experience Mix of?-asyncpe (craype) & PrgEnv-? modules PrgEnv loads compiler, math, MPI, etc. modules craype/asyncpe loads compiler wrapper scripts Regardless of actual compiler (PGI, Intel, etc), invoke with cc, CC, or ftn MPI, math, & scientific libraries automatically linked No lmpi, -lblas, etc. Can be challenging for some build processes 15 Best Practices @OLCF (& more) 16 Best Practices @OLCF (& more) Compiling for the XK7 and XC30 Running Jobs You are actually cross-compiling Processors/instructions sets differ between login & compute nodes It s very important to remember this utilities like configure don t expect this Compiling for batch/login nodes is occasionally necessary See 7.2 of the Titan User Guide for examples https://www.olcf.ornl.gov/support/system-user-guides/titan-user-guide/ Batch job information is on the web site Note that XK7 queue policy favors large jobs Jobs are charged based on what you make unavailable to others, not what you use Requests for high priority/quick turnaround are considered 17 Best Practices @OLCF (& more) 18 Best Practices @OLCF (& more) 3

Debugging/Optimization Common Runtime Error Messages Several tools available: DDT, Vampir, ScoreP, CrayPAT Some documentation is available on the OLCF website; you can also contact User Assistance 19 Best Practices @OLCF (& more) Common Runtime Errors Common Runtime Errors (2) Illegal Instruction A code compiled for compute nodes was executed on login/batch/service nodes (Remember, you re cross-compiling) request exceeds max nodes alloc The aprun request requires more nodes than the job requested -or- Request was correct, but at launch time some nodes were discovered to be down (see potential fix later in this presentation) relocation truncated to fit: R_X86_64_PC32 Program is using too much static memory Limit on Titan is 2GB Solution: use dynamic memory allocation as much as possible 21 Best Practices @OLCF (& more) 22 Best Practices @OLCF (& more) Common Runtime Errors (3) Common Questions aprun: [NID 94]Exec a.out failed: chdir /autofs/na1_home/user1 No such file or directory Compute nodes only mount Lustre directories When you launch aprun, you must be in a directory visible to compute nodes (in this example, it was launched from $HOME) Any files used by the processes on compute nodes must also be in Lustre The executable itself need not be in Lustre 23 Best Practices @OLCF (& more) 4

Why isn t my job running? Why isn t my job running? (2) Can be a number of reasons Insufficient nodes available Pending downtime Queue policy Dependency on another job not met Use checkjob <jobid> to diagnose Use v for verbose mode Reason for job not running is usually near the end Use showres to see upcoming reservations/outages Also shows running jobs look carefully Insufficient resources available NOTE: job cannot run (insufficient available procs: 3264 available) Unresolved depencency NOTE: job cannot run (dependency 1851375 jobcomplete not met) Queue policy issue NOTE: job violates constraints for partition titan (job 1854474 violates active HARD MAXJOB limit of 2 for qos smallmaxjobs user partition ALL (Req: 1 InUse: 2)) BLOCK MSG: job 1854474 violates active HARD MAXJOB limit of 2 for qos smallmaxjobs user partition ALL (Req: 1 InUse: 2) (recorded at last scheduling iteration) 25 Best Practices @OLCF (& more) 26 Best Practices @OLCF (& more) Why isn t my job running? (3) When will my job start? Pending maintenance period/reservation This is example output from showres Look for alphanumeric ReservationIDs (those without letters are likely currently running jobs) Note the full system reservation starting in 15 hours am I asking for more than 15 hours in my job? ReservationID Type S Start End Duration N/P StartTime DTNOutage.127 User - 15:05:13 1:03:05:13 12:00:00 11/176 Tue Jan 28 08:00:00 PM.128 User - 15:05:13 1:03:05:13 12:00:00 18688/299008 Tue Jan 28 08:00:00 PM_login.129 User - 15:05:13 1:03:05:13 12:00:00 16/2048 Tue Jan 28 08:00:00 Can be difficult to tell since jobs submitted in the future can affect it showstart <jobid> is not always reliable showq output is sorted by priority (can be helpful) mdiag p gives detailed priority info Shows boosts based on job size Shows overallocation penalties 27 Best Practices @OLCF (& more) 28 Best Practices @OLCF (& more) What is my project ID and what is the project allocation? Use showproj and showusage to list projects and usage, respectively Both commands have a help option (run with h) $ showproj User1 is a member of the following project(s) on titan: stf007 $ showusage titan usage for the project's current allocation period: Project Totals user1 Project Allocation Usage Remaining Usage stf007 7350000 11138 7338862 12 What happens when my project overruns its allocation? Jobs will run with a lower priority to facilitate some degree of fairshare with projects that have allocation remaining 30 day penalty for slightly over (usage 100-125% of allocation) 365 day penalty for usage >125% of allocation Projects are NOT disabled for going over allocation 29 Best Practices @OLCF (& more) 30 Best Practices @OLCF (& more) 5

My project lost X hours due to system issues can I get a refund? Since we don t disable projects for going over allocation, we don t issue refunds (per se) If many jobs were affected, we can delay the priority reduction This is the same thing as a refund but is much easier to manage Can I get a priority boost/higher walltime limit/purge exemption/etc? Perhaps Make request via https://www.olcf.ornl.gov/support/documents-forms/ (Forms to Request Changes to Computers, Jobs or Accounts) Reviewed by Resource Utilization Council Please make requests well in advance to allow for review If requesting job priority, make sure you submit the job they often run more quickly than you expect 31 Best Practices @OLCF (& more) 32 Best Practices @OLCF (& more) Are my files backed up? NFS directories: Yes, to a limited extent $ cd ~/SomeDir/SomeOtherDir $ ls file1 file2 $ rm file1 $ ls file2 $ cp ~/.yesterday/somedir/someotherdir/file1. $ ls file1 file2 Where do I archive data? Is there a mass storage system available? Yes, HPSS Accessed via hsi and htar See the Storage @ OLCF presentation for details https://www.olcf.ornl.gov/wp-content/uploads/2013/02/storage_at_olcf_2-mg.pdf Lustre: No HPSS: No (You might use it as a backup, but HPSS itself is not backed up. It s a good idea to have another level of backup at another site.) 33 Best Practices @OLCF (& more) 34 Best Practices @OLCF (& more) How do I transfer my data? Can I access my data when Titan is down? System user guides discuss data transfer Data user guide: https://www.olcf.ornl.gov/computing-resources/data-management/datamanagement-user-guide/ Data transfer nodes (dtn.ccs.ornl.gov) are the preferred place for internal and external data transfer Often remain available during system outages All users should have access Where do I store data? Can be complicated NFS, Lustre, HPSS User directories, Project directories See the Data Management Policy https://www.olcf.ornl.gov/kb_articles/data-management-policy 35 Best Practices @OLCF (& more) 36 Best Practices @OLCF (& more) 6

How do I check my disk usage? Where do I find documentation? For home directories, use the quota command $ quota quota: error while getting quota from master:/cm/shared for user1 (id 98765): Connection refused quota: error while getting quota from master:/home for user1 (id 98765): Connection refused Disk quotas for user user1(uid 98765): Filesystem blocks quota limit grace files quota limit grace nccsfiler3.ccs.ornl.gov:/vol/home1 8 10485760 10485760 2 4294967295 4294967295 nccsfiler4.ccs.ornl.gov:/vol/home2 8669108 15728640 15728640 51766 4294967295 4294967295 For lustre directories, use lustredu $ lustredu $MEMBERWORK/stf007 Last Collected Date Size File Count Directory 2014-01-10 12:30:08 31.21 MB 9 /lustre/atlas1/stf007/scratch/user1 http://www.olcf.ornl.gov User Guides available at https://www.olcf.ornl.gov/support/system-user-guides/ http://docs.cray.com (CrayDocs) http://docs.nvidia.com (Nvidia hosted documentation) 37 Best Practices @OLCF (& more) 38 Best Practices @OLCF (& more) Quick Tips, Best Practices, and Other Useful Information Quick Tip-Dealing With the Lustre Purge Problem: Needed files were deleted by the purge Several fixes Pulling from HPSS every time: inefficient Multiple scripts (data in/not in place): cumbersome touch: not ideal Conditional transfer: good #!/bin/bash... if [[! a $MEMBERWORK/stf007/some_important_file ]]; then hsi -q get /home/user1/data/some_important_file fi aprun -n 4096./a.out... 40 Best Practices @OLCF (& more) Quick Tip-Dealing With Failed Nodes Best Practices: Data Storage Sometimes nodes fail when your job starts Since node allocation has already happened, the batch system can t replace them You can work around this by requesting extra nodes #!/bin/bash #PBS lnodes=104... APRUN_RETURN_VALUE=1 while [[ $APRUN_RETURN_VALUE ne 0 ]]; do aprun n1600./a.out APRUN_RETURN_VALUE=$? done... HPSS is the proper location for long-term sotrage Project areas offer a common area for files shared among project members, but is not long-term storage Monitor disk usage Should be backed up May be purged 41 Best Practices @OLCF (& more) 42 Best Practices @OLCF (& more) 7

Best Practices: Lustre Filesystem Best Practices: Using HPSS Codes may need revisions to prevent large metadata operations Compiling in Lustre can be slow There is a purge, but please delete files when no longer needed HPSS prefers a few large files rather than many small files (htar is your friend) Running many simultaneous transfers can be problematic Planned upgrades (larger disk cache) should make a notable difference in performance 43 Best Practices @OLCF (& more) 44 Best Practices @OLCF (& more) Best Practices: User Support Finally Email is often useful if there are long/complicated error messages Send as many error messages as possible Or place them in a file & direct us to it When sending codes, create a.tar file & direct us to it Include all files necessary to run More efficient than sending via email Send new issues in new tickets, not replies to old ones We re here to help you Questions/comments/etc. can be sent to the OLCF User Assistance Center Staffed 9AM 5PM US Eastern Time, exclusive of ORNL holidays help@olcf.ornl.gov (865) 241-6536 45 Best Practices @OLCF (& more) 46 Best Practices @OLCF (& more) 8