Manual for using Super Computing Resources



Similar documents
1.0. User Manual For HPC Cluster at GIKI. Volume. Ghulam Ishaq Khan Institute of Engineering Sciences & Technology

Grid 101. Grid 101. Josh Hegie.

NEC HPC-Linux-Cluster

Streamline Computing Linux Cluster User Training. ( Nottingham University)

The Asterope compute cluster

An Introduction to High Performance Computing in the Department

Introduction to Linux and Cluster Basics for the CCR General Computing Cluster

User s Manual

Grid Engine Basics. Table of Contents. Grid Engine Basics Version 1. (Formerly: Sun Grid Engine)

The CNMS Computer Cluster

High Performance Computing Facility Specifications, Policies and Usage. Supercomputer Project. Bibliotheca Alexandrina

Running on Blue Gene/Q at Argonne Leadership Computing Facility (ALCF)

CycleServer Grid Engine Support Install Guide. version 1.25

Introduction to Running Hadoop on the High Performance Clusters at the Center for Computational Research

Notes on the SNOW/Rmpi R packages with OpenMPI and Sun Grid Engine

Cloud Implementation using OpenNebula

GRID Computing: CAS Style

How To Run A Tompouce Cluster On An Ipra (Inria) (Sun) 2 (Sun Geserade) (Sun-Ge) 2/5.2 (

Introduction to Running Computations on the High Performance Clusters at the Center for Computational Research

Introduction to Sun Grid Engine (SGE)

Getting Started with HPC

High-Performance Reservoir Risk Assessment (Jacta Cluster)

Installing and running COMSOL on a Linux cluster

HPCC USER S GUIDE. Version 1.2 July IITS (Research Support) Singapore Management University. IITS, Singapore Management University Page 1 of 35

How to Backup XenServer VM with VirtualIQ

Deploying Cloudera CDH (Cloudera Distribution Including Apache Hadoop) with Emulex OneConnect OCe14000 Network Adapters

SGE Roll: Users Guide. Version Edition

Single Node Hadoop Cluster Setup

1 Bull, 2011 Bull Extreme Computing

Setup Cisco Call Manager on VMware

SYMANTEC BACKUPEXEC2010 WITH StorTrends

Secure Shell. The Protocol

Oracle EXAM - 1Z Oracle Weblogic Server 11g: System Administration I. Buy Full Product.

Using WestGrid. Patrick Mann, Manager, Technical Operations Jan.15, 2014

High Performance Computing

Deploy Apache Hadoop with Emulex OneConnect OCe14000 Ethernet Network Adapters

13.1 Backup virtual machines running on VMware ESXi / ESX Server

HP reference configuration for entry-level SAS Grid Manager solutions

How To Install Storegrid Server On Linux On A Microsoft Ubuntu 7.5 (Amd64) Or Ubuntu (Amd86) (Amd77) (Orchestra) (For Ubuntu) (Permanent) (Powerpoint

OpenMP & MPI CISC 879. Tristan Vanderbruggen & John Cavazos Dept of Computer & Information Sciences University of Delaware

Tutorial: Using WestGrid. Drew Leske Compute Canada/WestGrid Site Lead University of Victoria

Configuration of High Performance Computing for Medical Imaging and Processing. SunGridEngine 6.2u5

Neptune. A Domain Specific Language for Deploying HPC Software on Cloud Platforms. Chris Bunch Navraj Chohan Chandra Krintz Khawaja Shams

IBM Smart Cloud guide started

installation administration and monitoring of beowulf clusters using open source tools

Cluster Implementation and Management; Scheduling

INF-110. GPFS Installation

TS-800. Configuring SSH Client Software in UNIX and Windows Environments for Use with the SFTP Access Method in SAS 9.2, SAS 9.3, and SAS 9.

Adobe Marketing Cloud Using FTP and sftp with the Adobe Marketing Cloud

HP StorageWorks EBS Solutions guide for VMware Consolidated Backup

Cluster Computing at HRI

Hodor and Bran - Job Scheduling and PBS Scripts

Testing New Applications In The DMZ Using VMware ESX. Ivan Dell Era Software Engineer IBM

Estonian Scientific Computing Infrastructure (ETAIS)

Signiant Agent installation

HP StorageWorks Modular Smart Array 1000 Small Business SAN Kit Hardware and Software Demonstration

HPC Cluster Decisions and ANSYS Configuration Best Practices. Diana Collier Lead Systems Support Specialist Houston UGM May 2014

Table of Contents. Online backup Manager User s Guide

Connecting to the School of Computing Servers and Transferring Files

A National Computing Grid: FGI

1z0-102 Q&A. DEMO Version

Easy Setup Guide 1&1 CLOUD SERVER. Creating Backups. for Linux

Beyond Windows: Using the Linux Servers and the Grid

Improved LS-DYNA Performance on Sun Servers

Using Parallel Computing to Run Multiple Jobs

LOCKSS on LINUX. CentOS6 Installation Manual 08/22/2013

Grid Engine Users Guide p1 Edition

LBNC and IBM Corporation Document: LBNC-Install.doc Date: Path: D:\Doc\EPFL\LNBC\LBNC-Install.doc Version: V1.0

FileCruiser Backup & Restoring Guide

WinSCP PuTTY as an alternative to F-Secure July 11, 2006

New High-performance computing cluster: PAULI. Sascha Frick Institute for Physical Chemistry

Building Clusters for Gromacs and other HPC applications

High Performance Computing Cluster Quick Reference User Guide

Clusters: Mainstream Technology for CAE

What the student will need:

JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert

Security Configuration Guide P/N Rev A05

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Using Red Hat Network Satellite Server to Manage Dell PowerEdge Servers

Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers

How to Run Parallel Jobs Efficiently

VTrak SATA RAID Storage System

Set Up Panorama. Palo Alto Networks. Panorama Administrator s Guide Version 6.0. Copyright Palo Alto Networks

Building a Private Cloud Cloud Infrastructure Using Opensource

Linux für bwgrid. Sabine Richling, Heinz Kredel. Universitätsrechenzentrum Heidelberg Rechenzentrum Universität Mannheim. 27.

To connect to the cluster, simply use a SSH or SFTP client to connect to:

Extending Remote Desktop for Large Installations. Distributed Package Installs

Remote & Collaborative Visualization. Texas Advanced Compu1ng Center

Remote Access to Unix Machines

SSH and Basic Commands

Instructions for Accessing the Advanced Computing Facility Supercomputing Cluster at the University of Kansas

Install and configure SSH server

Transcription:

Manual for using Super Computing Resources Super Computing Research and Education Centre at Research Centre for Modeling and Simulation National University of Science and Technology H-12 Campus, Islamabad 1

Table of Contents 1. Computational Resources 03 1.1 Specifications of Super Computer 03 1.1.1 Snapshot of Data Center 04 1.1.2 Rack Diagram of Super Computer 05 1.1.3 Locations of Compilers 06 1.1.4 Locations of Runtimes 06 1.2 Storage Area Network (SAN) Setup 07 1.3 Network Diagram 08 2. Submitting Processes to Super Computer 09 2.1 Steps to Login through SSH 09 2.1.1 Linux Client 09 2.1.2 Windows Client 10 2.2 Steps for Password-less Authentications 11 2.3 Steps for Submitting Jobs on the Cluster 12 2.3.1 Running Parallel MPI Jobs 12 2.3.2 Running Parallel MPI Scripts 13 3. Contact Information 14 3.1 Contact Persons 14 3.2 Contact Address 14 2

1. Computational Resources RCMS Super Computer is installed in state of art data center with 80 KVA of UPS backup and 12 ton precision cooling system. The data center is protected by FM-200 based Automatic Fire Detection and Suppression System and manual fire extinguishers. CCTV Cameras and Access Control systems are being procured for effective surveillance of data center. Specifications of Super Computer are given below:- 1.1 Specifications of Super Computer The Super Computer is comprised of 32 Intel Xeon based machines and each one of them is connected to Nvidia Tesla S1070 (each of which contains 4 GPU s). All nodes are connected by40gbps QDR InfiniBand Interconnect for internal communication. A high-performance and reliable SAN storage is linked to Servers, accessible by all computational nodes. Table 1shows the detailed specification of RCMS Super Computer. Cluster Name afrit.rcms.nust.edu.pk Brand HP ProLiant DL380 G6 Servers/ HP ProLiant DL160se G6 Server Total Processors 272 Intel Xeon Total Nodes 34 Total Memory 1.312 TB Operating System Redhat Enterprise Linux 5.6 Interconnects InfiniBand Switch Storage HP P2000 SAN Storage 22TB capacity, SAN Switches, Host Bus Adapters (HBA s), Fiber Channel Switch with RAID Controllers Graphic Processing Unit 32 x NVidia Tesla S1070 (each system contains 4 GPU s) Table 1: Specification of RCMS Super Computer 3

1.1.1 Snapshot of Data Center Figure 1: SnapshotofRCMSSuper Computer Cluster Nodes afrit.rcms.nust.edu.pk Compute-0-13 Compute-0-24 Compute-0-3 Compute-0-14 Compute-0-25 Compute-0-4 Compute-0-15 Compute-0-26 Compute-0-5 Compute-0-16 Compute-0-27 Compute-0-6 Compute-0-17 Compute-0-28 Compute-0-7 Compute-0-18 Compute-0-29 Compute-0-8 Compute-0-19 Compute-0-30 Compute-0-9 Compute-0-20 Compute-0-31 Compute-0-10 Compute-0-21 Compute-0-32 Compute-0-11 Compute-0-22 Compute-0-33 Compute-0-12 Compute-0-23 Compute-0-35 4

1.1.2 Rack Diagram of Super Computer Figure 2: Logical Rack DiagramofRCMSCluster 5

1.1.3 Locations of Compilers Name Command Location Make utility make /usr/bin/make GNU C compiler gcc /usr/bin/gcc GNU C++ compiler g++ /usr/bin/g++ GNU F77 Compiler g77 /usr/bin/g77 MPI C compiler mpicc /usr/mpi/intel/openmpi-1.4.3/bin/mpicc MPI C++ compiler mpic++ /usr/mpi/intel/openmpi-1.4.3/bin/mpic++ MPI Fortran 77 Compiler mpif77 /usr/mpi/intel/openmpi-1.4.3/bin/mpif77 Java Compiler javac /usr/java/latest/bin/javac Ant Utility ant /opt/rocks/bin/ant C compiler cc /usr/bin/cc F77 Compiler f77 /usr/bin/f77 GFortran Compiler gfortran /usr/bin/gfortran Fortran95 Compiler f95 /usr/bin/f95 UPC Compiler upcc /share/apps/upc/upc-installation/upcc 1.1.4 Locations of Runtimes: Name Command Location MPI Runtime mpirun /usr/mpi/intel/openmpi-1.4.3/bin/mpirun Java Virtual Machine java /usr/java/latest/bin/java UPC Runtime upcrun /share/apps/upc/upc-installation/upcrun 6

1.2 SAN Setup Total 22 TB of SAN Storage is available for storing users data. Two SAN Switches are installed in UBB Rack. 8 x 8 Gb transceiver installed in each of the SAN Switch. Total 48 slots are available which are occupied with 450GB each. The system is configured on RAID -1 one unit and RAID-5 on 4 units, each containing 16 drives. One online spare drive is marked in each of the disk enclosure for high availability. In case of drive failure the online spare drive will take over and data will be re-created depending on the RAID level. Each unit is presented to Storage Node, whose hostname is u2. NFS Server daemon is installed on u2. NFS share has been created in order to assign storage to other nodes on network. The Storage is managed using an application called Storage Management Utility. Figure 3: SAN Storage 7

1.3 Network Diagram for Super Computer Figure 4: Network diagram of Super Computing Resources. 8

2. Submitting Processes to Super Computer 2.1 Steps to Login using SSH 2.1.1 Linux Client a) On the shell prompt type: $ ssh -p 2299 username@111.68.97.5 Where username is the login name assigned by System Administrator b) The following text will be displayed: You will have to type yes as shown and press enter c) After pressing enter system will prompt for password. You may enter your password in order to login. Please note that your password will not be displayed for security reasons. 9

2.1.2 Windows Client: a) An SSH Client will be required in order to login using SSH. You may download PuTTY from http://www.chiark.greenend.org.uk/~sgtatham/putty/download.html b) In Host Name field of putty write IP Address: 111.68.97.5 c) Select SSH as connection type and enter port: 2299. c) Then click open. e) On clicking Open you will be asked to enter your username and password. Write your username and password followed by an Enter to login to Super Computer. Please note that your password will not be displayed for security reasons. 10

2.2 Password-less Authentication 2.2.1 The private / public key pair will be required in order to authenticate you on the target machine. To generate this key pair, type the following in console: $ ssh-keygen t rsa 2.2.2 Shell will ask the following questions: Enter file in which to save the key (/export/home2/rcms/usman/.ssh/id_rsa): (Press enter here) Enter passphrase: (Press enter here or give passphrase) Enter same passphrase again: (Press enter here or give passphrase) 2.2.3 After generating the public key, copy the public key to a file named authorized_keys, as shown below: $ cd ~ $ cat./.ssh/id_rsa.pub >>./.ssh/ authorized_keys 11

2.3 Running Scripts on Cluster 2.3.1 Running Sequential Scripts: a) Create a shell Script #!/bin/bash #myscript.sh # # **************************** # Following is the title of the Sequential script. #$ -N SEQ_PRO # The output and errors of the programmes will be written in # SEQ_PRO.OPRO_ID and SEQ_PRO.EPRO_ID respectively. #Path to the executable file. /usr/bin/myscript # End of Script b) Set execute permissions to the script $ chmod 755 myscript.sh c) Now submit your script as follows:- $ qsub -V myscript.sh d) In order to see the status of script type the following command followed by enter: $ qstat e) In order to del your running script type the following command: $ qdel 19 19 is the ID of your script 12

2.3.2 Running Parallel MPI Scripts a) Create a shell script for parallel processing #!/bin/bash # mpi_script.sh # ********************** # Following is the name of MPI Script # $ -N MPI_PRO # Following will be the output and error files # MPI_PRO.OPRO_ID and MPI_PRO.EPRO_ID respectively. # $ -pe mpi 16 #This will assign 16 cores to the script # $NSLOTS # No of cores allocated by the sun grid engine # machines # a machine file containing names of all available nodes echo "Allocated $NSLOTS slots." mpirun -np $NSLOTS -mca ras gridengine --hostfile machines mpi_script.sh # End of Script b) Set execute permissions to the script $ chmod 755 mpi_script.sh c) Now submit your script as follows:- $ qsub -V mpi_script.sh d) In order to see the status of script type the following command followed by enter: $ qstat e) In order to del your running script type the following command: $ qdel 19 19 is the ID of your script 13

3. Contact Information: 3.1 Contact Persons: In case of any inquiry or assistance feel free to contact the following persons: S. No Designation Name Contact No Email Address 1 Director ScREC Engr. Taufique-ur-Rehman 051-90855730 taufique.rehman@rcms.nust.edu.pk 2 Faculty Member Mr. Tariq Saeed 051-90855731 tariq@rcms.nust.edu.pk 3 System Administrator 4 Assistant System Administrator Engr. Muhammad Usman 051-90855717 usman@rcms.nust.edu.pk Mr. Shahzad Shoukat 051-90855714 shahzad@rcms.nust.edu.pk 3.2 Contact Address: Supper Computing Lab, 1 st Floor, Acad1 Block, Research Center of Modeling and Simulation (RCMS) National University of Sciences and Technology, Sector H 12, Islamabad Pakistan. Special Thanks to: Mr. Hammad Siddiqi, System Administrator High Performance Computing Lab, School of Electrical Engineering and Computer Science, National University of Sciences and Technology, H-12, Islamabad http://hpc.seecs.nust.edu.pk/ 14