CNAG User s Guide. Barcelona Supercomputing Center Copyright c 2015 BSC-CNS December 18, Introduction 2
|
|
|
- Bertina Rich
- 10 years ago
- Views:
Transcription
1 CNAG User s Guide Barcelona Supercomputing Center Copyright c 2015 BSC-CNS December 18, 2015 Contents 1 Introduction 2 2 System Overview 2 3 Connecting to CNAG cluster Transferring files Graphical applications File Systems NFS Local Hard Drive Lustre Running Jobs Classes Submitting jobs Exclusive nodes for groups Software Environment C Compilers FORTRAN Compilers Modules Environment Debuggers Software in CNAG cluster Getting help Frequently Asked Questions (FAQ) Appendices SSH Transferring files Using X
2 1 Introduction This user s guide for the CNAG cluster is intended to provide the minimum amount of information needed by a new user of this system. As such, it assumes that the user is familiar with many of the standard aspects of supercomputing such as the Unix operating system. Here you can find most of the information you need to use our computing resources and the technical documentation about the machine. Please read carefully this document and if any doubt arises do not hesitate to contact us (Getting help (chapter 7)). 2 System Overview The CNAG cluster comprises 100 compute nodes, 2 login nodes and 20 HiMem nodes. The operative system is a Linux RedHat release 6.2. Every compute node (and both logins) has two processors Xeon Quad Core at 2.93 GHz with 48 GB of RAM, that is 8 cores with 5,8GB per core, and 166 GB of local disk storage. Each HiMem node has two processors Intel Xeon E at 2.60 GHz with 128 GB of RAM, 16 cores with 7,8 GB/core, and 394 GB of local disk storage. Every node has access to a couple of Lustre parallel filesystems, /project (1130T) and /scratch (873T), providing a total of 2 PB of disk storage. The networks that interconnect the CNAG cluster are: Infiniband Network: High bandwidth network used by parallel applications communications (MPI) available in all compute nodes except in the HiMem ones. Gigabit Network: Ethernet network used by the Lustre Filesystem and the communications between the HiMem nodes. 3 Connecting to CNAG cluster First thing you should know is your username and password. Once you have a login and its associated password you can get into the cluster through one of the following login nodes: cn501 ( ) cn502 ( ) You must use Secure Shell (ssh) tools to login into or transfer files to the cluster. We do not accept incoming connections from protocols such as telnet, ftp, rlogin, rcp, or rsh commands. To get more information about the Secure Shell version supported and how to get ssh for your system (including Windows systems) see Appendices (chapter 8). Here you have two examples of logging into the cluster from a UNIX environment: > ssh -l username cn501 username s password: Last login: Thu Oct 08 15:07: from XXXX username@cn501:~> or alternatively > ssh -l username username s password: Last login: Thu Oct 08 15:07: from XXXX username@cn501:~> These addresses are only accesible inside CRG s network, comprising CNAG s offices and CRG s VPN connections. As can be seen, you will be prompted for your password before access is granted. If you are on a Windows system, you need to download and install a Secure Shell client to perform the connection to the machine (See Appendices (chapter 8) for more information). Most of these applications are graphical and you will have to fill some information in some of the fields offered. In the field Host name or Remote Host name you will need to introduce the URL of 2
3 the login node. After entering your username and password as requested by the Secure Shell client you should be logged in. The first time that you connect to the cluster Secure Shell exchanges some information to establish the communication. Then it requires that you accept of the RSA key of the remote host (this is just an unique identifier for the server). You must answer yes to accept the key and connect. If you select no, the connection will be refused and terminated. If you cannot get access to the system after following this procedure, first consult Appendices (chapter 8) for extended information about Secure Shell and, if the problem persists, you can contact us (see Getting Help (chapter 7)). Once inside the machine you will be presented with a UNIX shell prompt and you will normally be in your home directory ($HOME). If you are new to UNIX, you will have to learn the basics before doing anything useful. You can jump from one login node to the other but it is not possible to jump to a computing node. Outgoing connections of any kind are only possible from the login nodes. 3.1 Transferring files The easiest way to tranfer files to/from the cluster is with the use of scp and sftp commands. You should execute them from your local machine. Here there are some examples of each of these tools transferring files to the cluster: localsystem$ scp localfile username@ : username s password: localsystem$ sftp username@ username s password: sftp> put localfile These are the ways to retrieve files from CNAG cluster to your local machine: localsystem$ scp username@ :remotefile localdir username s password: localsystem$ sftp username@ username s password: sftp> get remotefile On a Windows system, most of the Secure Shell clients come with a tool to make secure copies or secure ftp s. There are several tools that accomplishes the requirements, please refer to the Appendices (chapter 8) for the most common ones and examples of use. 3.2 Graphical applications The only way to use graphical applications on the cluster is tunneling all the graphical traffic of the application you opened to your local machine through the Secure Shell connection you established to connect to the cluster. You will need to have an Xserver running on your local machine to be able to show the graphical information. Most of the UNIX flavors have an X server installed by default. In a Windows environment, you will probably need to download and install some type of X server emulator (see Appendices (chapter 8)). The second step in order to be able to execute graphical applications is to enable in your Secure Shell connection the forwarding of the graphical information through the secure channel created. This is normally done adding the -X flag to your normal ssh command used to connect to the cluster. Here you have an example: localsystem$ ssh -X -l username username s password: Last login: Fri Apr 23 15:07: from XXXX username@cn501:~> 3
4 For Windows systems you will have to enable the X11 forwarding in your SSH client. That option normally resides on the Tunneling or Connection menu of the client configuration window. (See Appendices (chapter 8) for further details). 4 File Systems IMPORTANT: It is your responsibility as a user of the CNAG cluster to backup all your critical data. We only guarantee a daily backup of user data under /home and of the qseq and fastq files from finished runs of the sequencers. Each user has several areas of disk space to store files and some of them may have size or time limits. Please read carefully this section to know about the usage policy of each of these filesystems. There are 3 different types of storage available inside a node: NFS: NFS is a distributed file system protocol which allows the user access to /home and /apps from all nodes. Lustre: Lustre is a parallel distributed file system which can be accessed from all the nodes. Local hard drive: Every node has an internal hard drive accessible only from the node itself. 4.1 NFS /home and /apps are mounted in all nodes through NFS from one of the servers. /apps contains all the software installed in the cluster for all users by support team. If you need an application/library or different version which can not be found there you can ask us for its installation (see Getting Help (chapter 7)). You should indicate the name and version of the desired application/ library, as well as the web page where it can be downloaded (more information about the available software at Software Environment (chapter 6)). /home contains the personal directory of the users and is the directory where you start when logging into the cluster. Every user has its own home directory to store code, small files or applications which only they will use. As this is NOT a parallel file system (only 1 server) it is not recommended to use it as the working directory or output directory of production runs. Using it that way is likely to hurt your jobs performance and may harm other user s experience while using the cluster. You should use your personal space in Lustre for anything other than small tests. An incremental backup of these directories is performed daily. 4.2 Local Hard Drive Every node has a local hard drive that can be used as a local scratch space to store temporary files during the execution of a job. This space is mounted over /scratch_tmp directory. The amount of space within the /scratch_tmp filesystem is about 160G in the compute nodes and 375G in the HiMem. All data stored in these local hard drives is only accessible from the node itself. In order to use the local hard drive from your jobs or commands, you should refer to it through the environment variable $TMPDIR. 4.3 Lustre Lustre is a parallel distributed file system which provides a high performance and scalable file system. There are two Lustre filesystems available in CNAG cluster from all nodes: /project: This space is intended to store data that needs to be shared between the users of the same group or project. /scratch: This space is intended to store the temporary files of your jobs during its execution. /project has a directory for each group of users while /scratch has a directory per user. A quota per user is active in both Lustre file systems. You can check the limits of your quota with the command: lquota 4
5 You should foresee the quantity of space that your jobs will generate and must make certain you have enough quota available to meet your jobs needs. You can reduce your quota by archiving and compressing your data and by deleting old, intermediate or no longer useful data. Whenever you are using the following commands anywhere in the cluster you should add these options to use the node s hard drive for temporary files: sort -T $TMPDIR java -Djava.io.tmpdir=$TMPDIR Tips & tricks It is a good practice to keep the filesystems as clean as possible. We recommend you remove any temporary files created during a jobs execution at the end of the job or, if that s not possible, to store temporary data on specific paths so it s easy to find and delete afterwards. If the temporary data does not need to be available after the jobs end you may consider using the nodes local hard drive to store it (it doesn t count towards your quota and will be deleted automatically after the job ends). If you try to write data into Lustre when you are near your quota the system slows down for all users that access the filesystem. Use the lquota command to monitor your quota usage before submitting the executions to avoid this situation. If a directory holds a great number of files (more than 1024) its access will be slowed down. This can severely impact performance in a sudden way. It is preferrable to store as few files as possible or to distribute them among several subdirectories before assuming the performance penalty. Another performance problem is caused by reading and writing many files at the same time in the same directory, e.g., when several similar jobs start executing at the same time or temporary intermediate files are created and destroyed in quick succession (sort command). In that case it s preferred that you create a per-job directory to store the output files, avoiding also the huge number of files in a directory at the same time, or to use the nodes local hard drive. Lustre FileSystem commands The usual filesystem utilities don t interact well on Lustre filesystems and may cause some performance problems. Lustre provides its own version of some of those tools to efficiently access the filesystem without hurting performance. Whenever a wide filesystem operation needs to be carried out you should use these commands instead of their common counterparts. Check the commands documentation for usage and information. lfs find lfs ls lfs cp 5 Running Jobs Slurm is the utility used at CNAG Cluster for batch processing support, so all jobs must be run through it. This document provides information for getting started with job execution at CNAG Cluster. There are two main partitions in the cluster: projects: 25% of the machine. Only the high priority jobs for the sequencing machines and those shorter than 6 hours are run in this partition. limited: 75% of the machine. The standard jobs are run in this partition. Furthermore, there are other partitions like debug for testing purposes, the special partition for the development groups and the himem partition. These last two partitions allow to run jobs shorter than 3 hours. 5
6 5.1 Classes The user s default class is normal, although there are lowprio and highprio classes, as explained below: lowprio: This class is for low priority jobs, which means that the job will only run if there are available nodes. normal: This is the default class for every submitted job. highprio: This class is suitable for high priority jobs, which will be executed before normal or low priority jobs. mysql: This class is only for jobs using the mysql server running on the cluster. The limits for these queues are the following: Table 1: Joblimits by queue Class Maximum Wall Clock Maximum Number of Jobs lowprio 7 days Unlimited normal 1 day Unlimited highprio 1 day Unlimited msyql 1 day 10 jobs 5.2 Submitting jobs A job is the execution unit for the SLURM. We have created wrappers to make easier the adaptation to the batch system. A job is defined by a text file containing a set of directives describing the job, and the commands to execute. SLURM wrapper commands These are the basic directives to submit jobs: mnsubmit <job_script> submits a job script to the queue system (see below for job script directives). You can also submit a job that depends on the previous execution of another job: mnsubmit -dep afterok:<job_id>:<job_id> <job_script> The <job_script> will be executed after the correct completion of the jobs with id <job_id> mnq shows all the jobs submitted. mncancel <job_id> remove the job from the queue system, canceling the execution of the processes, if they were already running. mnhold <job_id> Hold the job and prevent it from starting. mnhold accepts a list of jobs separated by blanks. mnhold --release <job_id> Release the job that was held. 6
7 mnsh It allocates a shell with a 8 hour limit for testing and debug purposes. mnsh -x It allocates an interactive shell in the same console (without X forwarding). sacct Show basic information of your running or completed jobs after the 00:00 hours of the same day. Option -S allows to change this to the specified date. Also, a lot more of information can be obtained with long option. sacct also provides information about the memory used on your jobs. If you want to know the maximum amount of memory per process that any of your jobs used, the commands to run would be: sacct -o jobid,maxrss sacct -j o maxrss Sacct is a very powerful command with a lot of different options. Please refer to the man page for further information. (man sacct). Job directives A job must contain a series of directives to inform the batch system about the characteristics of the job. These directives appear as comments in the job script, with the following syntax: directive = value Additionally, the job script may contain a set of commands to execute. If not, an external script must be provided with the executable directive. Here you may find the most common directives: class = class_name With the class you can specify the priority of your job as is explained at Section 5.1. partition = himem The partition where the job is to be submitted. Let this field empty unless you need to use special partitions such as himem or development. wall_clock_limit = HH:MM:SS The limit of wall clock time. This is a mandatoryfield and you must set it to a value greater than the real execution time for your application and smaller than the time limits granted to the user. Notice that your job will be killed after the elapsed period initialdir = pathname The working directory of your job (i.e. where the job will run). If not specified, it is the current working directory at the time the job was submitted. error = file The name of the file to collect the stderr output of the job. output = file The name of the file to collect the standard output (stdout) of the job. 7
8 total_tasks = number The number of processes to start. Optionally, you can specify how many threads each process would open with the directive: cpus_per_task = number The number of cpus assigned to the job will be the total_tasks number * cpus_per_task number. tasks_per_node = number The number of tasks assigned to a node. memory = n The amount of memory per cpu requested in MB. If it is not set, it will take 5,8GB per cpu requested. requeue = 1 With this directive, slurm will requeue the job if it died due to a node fail. scratch = n Slurm guarantees that when the job starts there are at least n MB free in the local scratch of the allocated nodes. There are also a few SLURM environment variables you can use in your scripts: Table 2: SLURM environment variables Variable SLURM_JOBID SLURM_NPROCS SLURM_NNODES SLURM_PROCID SLURM_NODEID SLURM_LOCALID Meaning Specifies the job ID of the executing job Specifies the total number of processes in the job Is the actual number of nodes assigned to run your job Specifies the MPI rank (or relative process ID) for the current process. The range is from 0-(SLURM_NPROCS 1) Specifies relative node ID of the current job. The range is from 0-(SLURM_NNODES 1) Specifies the node-local task ID for the process within a job Examples Example for a sequential job : #!/bin/bash job_name = test_serial initialdir =. output = serial_%j.out error = serial_%j.err total_tasks = 1 wall_clock_limit = 00:02:00./serial_binary > serial.out The job would be submitted using: > mnsubmit ptest.cmd 8
9 Examples for a parallel job : #!/bin/bash job_name = test_parallel initialdir =. output = mpi_%j.out error = mpi_%j.err total_tasks = 8 wall_clock_limit = 00:02:00 srun./parallel_binary > parallel.output 5.3 Exclusive nodes for groups The CNAG users belonging to the production or development groups are able to submit jobs to special nodes that are reserved for their exclusive use. production The node cn100 is exclusive for interactive use of the production group. This node can be used for testing or debugging purposes. The command for getting resources in the node is the following: mnsh -a production development Each development group s team has a set of 3 dedicated nodes and 1 exclusive node. The dedicated nodes allow running any job shorter than 3 hours by any user of the cluster, but the members of the group have the maximum priority for running any kind of job there (it s only used by other jobs when no group user is executing). To submit a job to a dedicated node of your team, you must include the following directives in your job script: partition = development account = TEAM features = TEAM The exclusive node allows members of the group only to submit their jobs there, either for executions or to use it as an interactive node for testing or debugging purposes. You must add these directives to your jobscript to submit it to the exclusive node: partition = debug account = TEAM features = ExTEAM Or you can get the node in interactive mode with the command: mnsh -a TEAM Table 3: Shorthand identifiers for teams as must be selected in jobscripts Team name Annotation & Assembly Algorithm Development Functional Bioinformatics Comparative Genomics Structural Genomics Statistical Genomics Shorthand (TEAM) AA AD FB CG StruG StatG 6 Software Environment All software and numerical libraries available at the cluster can be found at /apps/. If you need something that is not there please contact us to get it installed (see Getting Help (chapter 7)). 9
10 6.1 C Compilers In the cluster you can find these C/C++ compilers : icc /icpc -> Intel C/C++ Compilers % man icc % man icpc gcc /g++ -> GNU Compilers for C/C++ % man gcc % man g++ All invocations of the C or C++ compilers follow these suffix conventions for input files:.c,.cc,.cpp, or.cxx -> C++ source file..c -> C source file.i -> preprocessed C source file.so -> shared object file.o -> object file for ld command.s -> assembler source file By default, the preprocessor is run on both C and C++ source files. These are the default sizes of the standard C/C++ datatypes on the machine Table 4: Default datatype sizes on the machine Type bool (c++ only) 1 char 1 wchar_t 4 short 2 int 4 long 8 float 4 double 8 long double 16 Length (bytes) Distributed Memory Parallelism To compile MPI programs it is recommended to use the following handy wrappers: mpicc, mpicxx for C and C++ source code. You need to choose the Parallel environment first: module load openmpi /module load impi /module load poe. These wrappers will include all the necessary libraries to build MPI applications without having to specify all the details by hand. % mpicc a.c -o a.exe % mpicxx a.c -o a.exe Shared Memory Parallelism OpenMP directives are fully supported by the Intel C and C++ compilers. To use it, the flag -openmp must be added to the compile line. % icc -openmp -o exename filename.c % icpc -openmp -o exename filename.c You can also mix MPI + OPENMP code using -openmp with the mpi wrappers mentioned above. 10
11 Automatic Parallelization The Intel C and C++ compilers are able to automatically parallelize simple loop constructs, using the option -parallel : % icc -parallel a.c 6.2 FORTRAN Compilers In the cluster you can find these compilers : ifort -> Intel Fortran Compilers % man ifort gfortran -> GNU Compilers for FORTRAN % man gfortran By default, the compilers expect all FORTRAN source files to have the extension.f, and all FORTRAN source files that require preprocessing to have the extension.f. The same applies to FORTRAN 90 source files with extensions.f90 and.f90. Distributed Memory Parallelism In order to use MPI, again you can use the wrappers mpif77 or mpif90 depending on the source code type. You can always man mpif77 to see a detailed list of options to configure the wrappers, ie: change the default compiler. % mpif77 a.f -o a.exe Shared Memory Parallelism OpenMP directives are fully supported by the Intel Fortran compiler when the option -openmp is set: % ifort -openmp Automatic Parallelization The Intel Fortran compiler will attempt to automatically parallelize simple loop constructs using the option -parallel : % ifort -parallel 6.3 Modules Environment The Environment Modules package ( provides a dynamic modification of a user s environment via modulefiles. Each modulefile contains the information needed to configure the shell for an application or a compilation. Modules can be loaded and unloaded dynamically, in a clean fashion. All popular shells are supported, including bash, ksh, zsh, sh, csh, tcsh, as well as some scripting languages such as perl. Installed software packages are divided into five categories: Environment: modulefiles dedicated to prepare the environment, for example, get all necessary variables to use openmpi to compile or run programs Tools: useful tools which can be used at any time (php, perl,... ) 11
12 Applications: High Performance Computers programs (GROMACS,... ) Libraries: Those are tipycally loaded at a compilation time, they load into the environment the correct compiler and linker flags (FFTW, LAPACK,... ) Compilers: Compiler suites available for the system (intel, gcc,... ) Modules tool usage Modules can be invoked in two ways: by name alone or by name and version. Invoking them by name implies loading the default module version. This is usually the most recent version that has been tested to be stable (recommended) or the only version available. % module load intel Invoking by version loads the version specified of the application. As of this writing, the previous command and the following one load the same module. % module load intel/ The most important commands for modules are these: module list shows all the loaded modules module avail shows all the modules the user is able to load module purge removes all the loaded modules module load <modulename> loads the necessary environment variables for the selected modulefile (PATH, MANPATH, LD_LIBRARY_PATH... ) module unload <modulename> removes all environment changes made by module load command module switch <oldmodule> <newmodule> unloads the first module (oldmodule) and loads the second module (newmodule) You can run module help any time to check the command s usage and options or check the module(1) manpage for further information. 6.4 Debuggers GDB (GNU DEBUGGER) /usr/bin/gdb DDD (/apps/ddd) /apps/ddd/3.3.12/bin/ddd 6.5 Software in CNAG cluster All software and numerical libraries available at CNAG cluster can be found at /apps. If you need something that is not available there contact us to get it installed (see Getting Help (chapter 7)). 7 Getting help BSC provides excellent consulting assistance for the cluster users. User support consultants are available during normal business hours, Monday to Friday, 09 a.m. to 18 p.m. (CEST time). User questions and support are handled at: [email protected] If you need assistance, please send an stating the nature of the issue, the date and time the issue occurred, and any other relevant information such as output files or error messages. Please contact BSC if you have any questions or comments regarding policies or procedures. Our mailing address is: 12
13 Barcelona Supercomputing Center Centro Nacional de Supercomputación C/ Jordi Girona, 31, Edificio Capilla Barcelona 7.1 Frequently Asked Questions (FAQ) You can check the answers to most common questions at BSC s Support Knowledge Center 1. There you will find online and updated versions of our documentation, including this guide, and a listing with deeper answers to the most common questions we receive as well as advanced specific questions unfit for a general-purpose user guide. 8 Appendices 8.1 SSH SSH is a program that enables secure logins over an insecure network. It encrypts all the data passing both ways, so that if it is intercepted it cannot be read. It also replaces the old an insecure tools like telnet, rlogin, rcp, ftp,etc. SSH is a client-server software. Both machines must have ssh installed for it to work. We have already installed a ssh server in our machines. You must have installed an ssh client in your local machine. SSH is available without charge for almost all versions of UNIX (including Linux and MacOS X). For UNIX and derivatives, we recommend using the OpenSSH client, downloadable from and for Windows users we recommend using Putty, a free SSH client that can be downloaded from Otherwise, any client compatible with SSH version 2 can be used. This section describes installing, configuring and using the client on Windows machines. No matter your client, you will need to specify the following information: Select SSH as default protocol Select port 22 Specify the remote machine and username For example with putty client: This is the first window that you will see at putty startup. Once finished, press the Open button. If it is your first connection to the machine, your will get a Warning telling you that the host key from the server is unknown, and will ask you if you are agree to cache the new host key, press Yes. IMPORTANT: If you see this warning another time and you haven t modified or reinstalled the ssh client, please do not log in, and contact us as soon as possible (see Getting Help (chapter 7)). Finally, a new window will appear asking for your login and password: 8.2 Transferring files To transfer files to or from the cluster you need a secure ftp (sftp) o secure copy (scp) client. There are several different clients, but as previously mentioned, we recommend using of Putty clients for transferring files: psftp and pscp. You can find it at the same web page as Putty ( 2 ). Some other possible tools for users requiring graphical file transfers could be: WinSCP: Freeware Sftp and Scp client for Windows ( ) SSH: Not free. ( ) 13
14 Figure 1: Putty client Figure 2: Putty certificate security alert Figure 3: Cluster login 14
15 Figure 4: WinSCP authentication Using WinSCP WinSCP is the common software used at CNAG for file transferring via SFTP. The program first asks for the machine IP ( /16), and then for the username and password as you can see below Once you are aunthenticated, you will be able to copy files from your local machine to CNAG, and from CNAG to your local machine. Simply you can drag n drop files from the left panel (your local machine) to the right panel (CNAG) and viceversa, as you can see below: Figure 5: WinSCP transfer window 8.3 Using X11 In order to start remote X applications you need and X-Server running in your local machine. Here is a list of most common X-servers for windows:
16 Cygwin/X: X-Win32 : WinaXe : XconnectPro : Exceed : The only Open Source X-server listed here is Cygwin/X, you need to pay for the others. Once the X-Server is running run putty with X11 forwarding enabled: Figure 6: Putty X11 configuration 16
Introduction to Linux and Cluster Basics for the CCR General Computing Cluster
Introduction to Linux and Cluster Basics for the CCR General Computing Cluster Cynthia Cornelius Center for Computational Research University at Buffalo, SUNY 701 Ellicott St Buffalo, NY 14203 Phone: 716-881-8959
The Asterope compute cluster
The Asterope compute cluster ÅA has a small cluster named asterope.abo.fi with 8 compute nodes Each node has 2 Intel Xeon X5650 processors (6-core) with a total of 24 GB RAM 2 NVIDIA Tesla M2050 GPGPU
An Introduction to High Performance Computing in the Department
An Introduction to High Performance Computing in the Department Ashley Ford & Chris Jewell Department of Statistics University of Warwick October 30, 2012 1 Some Background 2 How is Buster used? 3 Software
To connect to the cluster, simply use a SSH or SFTP client to connect to:
RIT Computer Engineering Cluster The RIT Computer Engineering cluster contains 12 computers for parallel programming using MPI. One computer, cluster-head.ce.rit.edu, serves as the master controller or
Streamline Computing Linux Cluster User Training. ( Nottingham University)
1 Streamline Computing Linux Cluster User Training ( Nottingham University) 3 User Training Agenda System Overview System Access Description of Cluster Environment Code Development Job Schedulers Running
CASHNet Secure File Transfer Instructions
CASHNet Secure File Transfer Instructions Copyright 2009, 2010 Higher One Payments, Inc. CASHNet, CASHNet Business Office, CASHNet Commerce Center, CASHNet SMARTPAY and all related logos and designs are
Secure Shell. The Protocol
Usually referred to as ssh The name is used for both the program and the protocol ssh is an extremely versatile network program data encryption and compression terminal access to remote host file transfer
Introduction to Running Computations on the High Performance Clusters at the Center for Computational Research
! Introduction to Running Computations on the High Performance Clusters at the Center for Computational Research! Cynthia Cornelius! Center for Computational Research University at Buffalo, SUNY! cdc at
Introduction to Running Hadoop on the High Performance Clusters at the Center for Computational Research
Introduction to Running Hadoop on the High Performance Clusters at the Center for Computational Research Cynthia Cornelius Center for Computational Research University at Buffalo, SUNY 701 Ellicott St
Manual for using Super Computing Resources
Manual for using Super Computing Resources Super Computing Research and Education Centre at Research Centre for Modeling and Simulation National University of Science and Technology H-12 Campus, Islamabad
NEC HPC-Linux-Cluster
NEC HPC-Linux-Cluster Hardware configuration: 4 Front-end servers: each with SandyBridge-EP processors: 16 cores per node 128 GB memory 134 compute nodes: 112 nodes with SandyBridge-EP processors (16 cores
24/08/2004. Introductory User Guide
24/08/2004 Introductory User Guide CSAR Introductory User Guide Introduction This material is designed to provide new users with all the information they need to access and use the SGI systems provided
Introduction to ACENET Accelerating Discovery with Computational Research May, 2015
Introduction to ACENET Accelerating Discovery with Computational Research May, 2015 What is ACENET? What is ACENET? Shared regional resource for... high-performance computing (HPC) remote collaboration
File Transfers. Contents
A File Transfers Contents Overview..................................................... A-2................................... A-2 General Switch Software Download Rules..................... A-3 Using
Overview. Remote access and file transfer. SSH clients by platform. Logging in remotely
Remote access and file transfer Overview Remote logins to Bio-Linux with ssh Running software from another machine Logging in from another machine Getting files on and off Bio-Linux Transferring files
Getting Started with HPC
Getting Started with HPC An Introduction to the Minerva High Performance Computing Resource 17 Sep 2013 Outline of Topics Introduction HPC Accounts Logging onto the HPC Clusters Common Linux Commands Storage
Easy Setup Guide 1&1 CLOUD SERVER. Creating Backups. for Linux
Easy Setup Guide 1&1 CLOUD SERVER Creating Backups for Linux Legal notice 1&1 Internet Inc. 701 Lee Road, Suite 300 Chesterbrook, PA 19087 USA www.1and1.com [email protected] August 2015 Copyright 2015 1&1
Cluster@WU User s Manual
Cluster@WU User s Manual Stefan Theußl Martin Pacala September 29, 2014 1 Introduction and scope At the WU Wirtschaftsuniversität Wien the Research Institute for Computational Methods (Forschungsinstitut
High Performance Computing Facility Specifications, Policies and Usage. Supercomputer Project. Bibliotheca Alexandrina
High Performance Computing Facility Specifications, Policies and Usage Supercomputer Project Bibliotheca Alexandrina Bibliotheca Alexandrina 1/16 Topics Specifications Overview Site Policies Intel Compilers
Using NeSI HPC Resources. NeSI Computational Science Team ([email protected])
NeSI Computational Science Team ([email protected]) Outline 1 About Us About NeSI Our Facilities 2 Using the Cluster Suitable Work What to expect Parallel speedup Data Getting to the Login Node 3 Submitting
WinSCP PuTTY as an alternative to F-Secure July 11, 2006
WinSCP PuTTY as an alternative to F-Secure July 11, 2006 Brief Summary of this Document F-Secure SSH Client 5.4 Build 34 is currently the Berkeley Lab s standard SSH client. It consists of three integrated
Using WestGrid. Patrick Mann, Manager, Technical Operations Jan.15, 2014
Using WestGrid Patrick Mann, Manager, Technical Operations Jan.15, 2014 Winter 2014 Seminar Series Date Speaker Topic 5 February Gino DiLabio Molecular Modelling Using HPC and Gaussian 26 February Jonathan
Cisco Networking Academy Program Curriculum Scope & Sequence. Fundamentals of UNIX version 2.0 (July, 2002)
Cisco Networking Academy Program Curriculum Scope & Sequence Fundamentals of UNIX version 2.0 (July, 2002) Course Description: Fundamentals of UNIX teaches you how to use the UNIX operating system and
High Performance Computing Cluster Quick Reference User Guide
High Performance Computing Cluster Quick Reference User Guide Base Operating System: Redhat(TM) / Scientific Linux 5.5 with Alces HPC Software Stack Copyright 2011 Alces Software Ltd All Rights Reserved
SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education www.accre.vanderbilt.
SLURM: Resource Management and Job Scheduling Software Advanced Computing Center for Research and Education www.accre.vanderbilt.edu Simple Linux Utility for Resource Management But it s also a job scheduler!
JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert
Mitglied der Helmholtz-Gemeinschaft JUROPA Linux Cluster An Overview 19 May 2014 Ulrich Detert JuRoPA JuRoPA Jülich Research on Petaflop Architectures Bull, Sun, ParTec, Intel, Mellanox, Novell, FZJ JUROPA
The CNMS Computer Cluster
The CNMS Computer Cluster This page describes the CNMS Computational Cluster, how to access it, and how to use it. Introduction (2014) The latest block of the CNMS Cluster (2010) Previous blocks of the
PuTTY/Cygwin Tutorial. By Ben Meister Written for CS 23, Winter 2007
PuTTY/Cygwin Tutorial By Ben Meister Written for CS 23, Winter 2007 This tutorial will show you how to set up and use PuTTY to connect to CS Department computers using SSH, and how to install and use the
Introduction to HPC Workshop. Center for e-research ([email protected])
Center for e-research ([email protected]) Outline 1 About Us About CER and NeSI The CS Team Our Facilities 2 Key Concepts What is a Cluster Parallel Programming Shared Memory Distributed Memory 3 Using
Grid 101. Grid 101. Josh Hegie. [email protected] http://hpc.unr.edu
Grid 101 Josh Hegie [email protected] http://hpc.unr.edu Accessing the Grid Outline 1 Accessing the Grid 2 Working on the Grid 3 Submitting Jobs with SGE 4 Compiling 5 MPI 6 Questions? Accessing the Grid Logging
Fundamentals of UNIX Lab 16.2.6 Networking Commands (Estimated time: 45 min.)
Fundamentals of UNIX Lab 16.2.6 Networking Commands (Estimated time: 45 min.) Objectives: Develop an understanding of UNIX and TCP/IP networking commands Ping another TCP/IP host Use traceroute to check
Introduction to Supercomputing with Janus
Introduction to Supercomputing with Janus Shelley Knuth [email protected] Peter Ruprecht [email protected] www.rc.colorado.edu Outline Who is CU Research Computing? What is a supercomputer?
Parallel Processing using the LOTUS cluster
Parallel Processing using the LOTUS cluster Alison Pamment / Cristina del Cano Novales JASMIN/CEMS Workshop February 2015 Overview Parallelising data analysis LOTUS HPC Cluster Job submission on LOTUS
CONNECTING TO DEPARTMENT OF COMPUTER SCIENCE SERVERS BOTH FROM ON AND OFF CAMPUS USING TUNNELING, PuTTY, AND VNC Client Utilities
CONNECTING TO DEPARTMENT OF COMPUTER SCIENCE SERVERS BOTH FROM ON AND OFF CAMPUS USING TUNNELING, PuTTY, AND VNC Client Utilities DNS name: turing.cs.montclair.edu -This server is the Departmental Server
Navigating the Rescue Mode for Linux
Navigating the Rescue Mode for Linux SUPPORT GUIDE DEDICATED SERVERS ABOUT THIS GUIDE This document will take you through the process of booting your Linux server into rescue mode to identify and fix the
Linux VPS with cpanel. Getting Started Guide
Linux VPS with cpanel Getting Started Guide First Edition October 2010 Table of Contents Introduction...1 cpanel Documentation...1 Accessing your Server...2 cpanel Users...2 WHM Interface...3 cpanel Interface...3
HPCC - Hrothgar Getting Started User Guide
HPCC - Hrothgar Getting Started User Guide Transfer files High Performance Computing Center Texas Tech University HPCC - Hrothgar 2 Table of Contents Transferring files... 3 1.1 Transferring files using
SLURM Workload Manager
SLURM Workload Manager What is SLURM? SLURM (Simple Linux Utility for Resource Management) is the native scheduler software that runs on ASTI's HPC cluster. Free and open-source job scheduler for the Linux
Online Backup Client User Manual Mac OS
Online Backup Client User Manual Mac OS 1. Product Information Product: Online Backup Client for Mac OS X Version: 4.1.7 1.1 System Requirements Operating System Mac OS X Leopard (10.5.0 and higher) (PPC
Online Backup Client User Manual Mac OS
Online Backup Client User Manual Mac OS 1. Product Information Product: Online Backup Client for Mac OS X Version: 4.1.7 1.1 System Requirements Operating System Mac OS X Leopard (10.5.0 and higher) (PPC
WHY USE ILLUMIN8 MARKETING FOR HOSTING YOUR WEB SITE?
QUESTIONS OR COMMENTS? email [email protected] or call 608-796-1990 WHY USE ILLUMIN8 MARKETING FOR HOSTING YOUR WEB SITE? Yes, there are cheaper (even free) alternatives out there. What we will
The RWTH Compute Cluster Environment
The RWTH Compute Cluster Environment Tim Cramer 11.03.2013 Source: D. Both, Bull GmbH Rechen- und Kommunikationszentrum (RZ) How to login Frontends cluster.rz.rwth-aachen.de cluster-x.rz.rwth-aachen.de
RecoveryVault Express Client User Manual
For Linux distributions Software version 4.1.7 Version 2.0 Disclaimer This document is compiled with the greatest possible care. However, errors might have been introduced caused by human mistakes or by
How To Run Linux On Windows 7 (For A Non-Privileged User) On A Windows 7 Computer (For Non-Patty) On Your Computer (Windows) On An Unix Computer (Unix) On Windows) On The Same
Working from Home Tools and Technologies for Improving Your Programming Environment Daniel J. Hood [email protected] Overview Tools for MS Windows (and others OSs) SSH SCP X Windows Tunneling X Connections
Agenda. Using HPC Wales 2
Using HPC Wales Agenda Infrastructure : An Overview of our Infrastructure Logging in : Command Line Interface and File Transfer Linux Basics : Commands and Text Editors Using Modules : Managing Software
SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education www.accre.vanderbilt.
SLURM: Resource Management and Job Scheduling Software Advanced Computing Center for Research and Education www.accre.vanderbilt.edu Simple Linux Utility for Resource Management But it s also a job scheduler!
1. Product Information
ORIXCLOUD BACKUP CLIENT USER MANUAL LINUX 1. Product Information Product: Orixcloud Backup Client for Linux Version: 4.1.7 1.1 System Requirements Linux (RedHat, SuSE, Debian and Debian based systems such
Instructions for Accessing the Advanced Computing Facility Supercomputing Cluster at the University of Kansas
ACF Supercomputer Access Instructions 1 Instructions for Accessing the Advanced Computing Facility Supercomputing Cluster at the University of Kansas ACF Supercomputer Access Instructions 2 Contents Instructions
1.0. User Manual For HPC Cluster at GIKI. Volume. Ghulam Ishaq Khan Institute of Engineering Sciences & Technology
Volume 1.0 FACULTY OF CUMPUTER SCIENCE & ENGINEERING Ghulam Ishaq Khan Institute of Engineering Sciences & Technology User Manual For HPC Cluster at GIKI Designed and prepared by Faculty of Computer Science
Online Backup Client User Manual Linux
Online Backup Client User Manual Linux 1. Product Information Product: Online Backup Client for Linux Version: 4.1.7 1.1 System Requirements Operating System Linux (RedHat, SuSE, Debian and Debian based
File Protection using rsync. Setup guide
File Protection using rsync Setup guide Contents 1. Introduction... 2 Documentation... 2 Licensing... 2 Overview... 2 2. Rsync technology... 3 Terminology... 3 Implementation... 3 3. Rsync data hosts...
TREK HOSC PAYLOAD ETHERNET GATEWAY (HPEG) USER GUIDE
TREK HOSC PAYLOAD ETHERNET GATEWAY (HPEG) USER GUIDE April 2016 Approved for Public Release; Distribution is Unlimited. TABLE OF CONTENTS PARAGRAPH PAGE 1 Welcome... 1 1.1 Getting Started... 1 1.2 System
Online Backup Linux Client User Manual
Online Backup Linux Client User Manual Software version 4.0.x For Linux distributions August 2011 Version 1.0 Disclaimer This document is compiled with the greatest possible care. However, errors might
Oracle EXAM - 1Z0-102. Oracle Weblogic Server 11g: System Administration I. Buy Full Product. http://www.examskey.com/1z0-102.html
Oracle EXAM - 1Z0-102 Oracle Weblogic Server 11g: System Administration I Buy Full Product http://www.examskey.com/1z0-102.html Examskey Oracle 1Z0-102 exam demo product is here for you to test the quality
Online Backup Client User Manual
For Linux distributions Software version 4.1.7 Version 2.0 Disclaimer This document is compiled with the greatest possible care. However, errors might have been introduced caused by human mistakes or by
SSH and Basic Commands
SSH and Basic Commands In this tutorial we'll introduce you to SSH - a tool that allows you to send remote commands to your Web server - and show you some simple UNIX commands to help you manage your website.
13.1 Backup virtual machines running on VMware ESXi / ESX Server
13 Backup / Restore VMware Virtual Machines Tomahawk Pro This chapter describes how to backup and restore virtual machines running on VMware ESX, ESXi Server or VMware Server 2.0. 13.1 Backup virtual machines
Introweb Remote Backup Client for Mac OS X User Manual. Version 3.20
Introweb Remote Backup Client for Mac OS X User Manual Version 3.20 1. Contents 1. Contents...2 2. Product Information...4 3. Benefits...4 4. Features...5 5. System Requirements...6 6. Setup...7 6.1. Setup
1z0-102 Q&A. DEMO Version
Oracle Weblogic Server 11g: System Administration Q&A DEMO Version Copyright (c) 2013 Chinatag LLC. All rights reserved. Important Note Please Read Carefully For demonstration purpose only, this free version
Online Backup Client User Manual
Online Backup Client User Manual Software version 3.21 For Linux distributions January 2011 Version 2.0 Disclaimer This document is compiled with the greatest possible care. However, errors might have
2 Advanced Session... Properties 3 Session profile... wizard. 5 Application... preferences. 3 ASCII / Binary... Transfer
Contents I Table of Contents Foreword 0 Part I SecEx Overview 3 1 What is SecEx...? 3 2 Quick start... 4 Part II Configuring SecEx 5 1 Session Profiles... 5 2 Advanced Session... Properties 6 3 Session
Domain Name. Domain Registrar. Web Site cpanel. www..com. www..net. www..biz URL: Username: Password: www.mydomain.com/cpanel. Username: Password:
314 Main Street Onalaska, WI 54650 608-787-8887 Why Use The Idea Center for Hosting Your Web Site? Yes, there are cheaper (even free) alternatives out there. What we will offer you is a reliable service,
CPSC 226 Lab Nine Fall 2015
CPSC 226 Lab Nine Fall 2015 Directions. Our overall lab goal is to learn how to use BBB/Debian as a typical Linux/ARM embedded environment, program in a traditional Linux C programming environment, and
Hodor and Bran - Job Scheduling and PBS Scripts
Hodor and Bran - Job Scheduling and PBS Scripts UND Computational Research Center Now that you have your program compiled and your input file ready for processing, it s time to run your job on the cluster.
OroTimesheet 7 Installation Guide
Installation Guide Copyright 1996-2011 OroLogic Inc. http://www.orologic.com Revision 7.00 Contents I Contents Installation Guide 2 Introduction 2 Installing OroTimesheet 2 Installing OroTimesheet in stand-alone
Online Backup Client User Manual
For Mac OS X Software version 4.1.7 Version 2.2 Disclaimer This document is compiled with the greatest possible care. However, errors might have been introduced caused by human mistakes or by other means.
Miami University RedHawk Cluster Connecting to the Cluster Using Windows
Miami University RedHawk Cluster Connecting to the Cluster Using Windows The RedHawk cluster is a general purpose research computing resource available to support the research community at Miami University.
Introduction... 1. Connecting Via FTP... 4. Where do I upload my website?... 4. What to call your home page?... 5. Troubleshooting FTP...
This guide is designed to show you the different ways of uploading your site using ftp, including the basic principles of understanding and troubleshooting ftp issues. P a g e 0 Introduction... 1 When
Using the Windows Cluster
Using the Windows Cluster Christian Terboven [email protected] aachen.de Center for Computing and Communication RWTH Aachen University Windows HPC 2008 (II) September 17, RWTH Aachen Agenda o Windows Cluster
File Transfer with Secure FTP
File Transfer with Secure FTP The Princeton Review values the security of all client s information and would like to encourage all FTP file transfers to and from our clients be conducted utilizing a Secure
File transfer clients manual File Delivery Services
File transfer clients manual File Delivery Services Publisher Post CH Ltd Information Technology Webergutstrasse 12 CH-3030 Berne (Zollikofen) Contact Post CH Ltd Information Technology Webergutstrasse
Getting Started With Your Virtual Dedicated Server. Getting Started Guide
Getting Started Guide Getting Started With Your Virtual Dedicated Server Setting up and hosting a domain on your Linux Virtual Dedicated Server using cpanel. Getting Started with Your Virtual Dedicated
DiskPulse DISK CHANGE MONITOR
DiskPulse DISK CHANGE MONITOR User Manual Version 7.9 Oct 2015 www.diskpulse.com [email protected] 1 1 DiskPulse Overview...3 2 DiskPulse Product Versions...5 3 Using Desktop Product Version...6 3.1 Product
Windows HPC 2008 Cluster Launch
Windows HPC 2008 Cluster Launch Regionales Rechenzentrum Erlangen (RRZE) Johannes Habich [email protected] Launch overview Small presentation and basic introduction Questions and answers Hands-On
4PSA Total Backup 3.0.0. User's Guide. for Plesk 10.0.0 and newer versions
4PSA Total Backup 3.0.0 for Plesk 10.0.0 and newer versions User's Guide For more information about 4PSA Total Backup, check: http://www.4psa.com Copyright 2009-2011 4PSA. User's Guide Manual Version 84359.5
Setting up PostgreSQL
Setting up PostgreSQL 1 Introduction to PostgreSQL PostgreSQL is an object-relational database management system based on POSTGRES, which was developed at the University of California at Berkeley. PostgreSQL
virtualization.info Review Center SWsoft Virtuozzo 3.5.1 (for Windows) // 02.26.06
virtualization.info Review Center SWsoft Virtuozzo 3.5.1 (for Windows) // 02.26.06 SWsoft Virtuozzo 3.5.1 (for Windows) Review 2 Summary 0. Introduction 1. Installation 2. VPSs creation and modification
A SHORT INTRODUCTION TO BITNAMI WITH CLOUD & HEAT. Version 1.12 2014-07-01
A SHORT INTRODUCTION TO BITNAMI WITH CLOUD & HEAT Version 1.12 2014-07-01 PAGE _ 2 TABLE OF CONTENTS 1. Introduction.... 3 2. Logging in to Cloud&Heat Dashboard... 4 2.1 Overview of Cloud&Heat Dashboard....
Data Movement and Storage. Drew Dolgert and previous contributors
Data Movement and Storage Drew Dolgert and previous contributors Data Intensive Computing Location Viewing Manipulation Storage Movement Sharing Interpretation $HOME $WORK $SCRATCH 72 is a Lot, Right?
LOCKSS on LINUX. CentOS6 Installation Manual 08/22/2013
LOCKSS on LINUX CentOS6 Installation Manual 08/22/2013 1 Table of Contents Overview... 3 LOCKSS Hardware... 5 Installation Checklist... 6 BIOS Settings... 9 Installation... 10 Firewall Configuration...
OLCF Best Practices. Bill Renaud OLCF User Assistance Group
OLCF Best Practices Bill Renaud OLCF User Assistance Group Overview This presentation covers some helpful information for users of OLCF Staying informed Some aspects of system usage that may differ from
File Transfer Examples. Running commands on other computers and transferring files between computers
Running commands on other computers and transferring files between computers 1 1 Remote Login Login to remote computer and run programs on that computer Once logged in to remote computer, everything you
AG MacOS Standalone Array Client Administration Guide
AG MacOS Standalone Array Client Administration Guide Copyright Statement Copyright Statement Copyright 2015 Array Networks, Inc., 1371 McCarthy Blvd, Milpitas, California 95035, USA. All rights reserved.
How To Set Up A Backupassist For An Raspberry Netbook With A Data Host On A Nsync Server On A Usb 2 (Qnap) On A Netbook (Qnet) On An Usb 2 On A Cdnap (
WHITEPAPER BackupAssist Version 5.1 www.backupassist.com Cortex I.T. Labs 2001-2008 2 Contents Introduction... 3 Hardware Setup Instructions... 3 QNAP TS-409... 3 Netgear ReadyNas NV+... 5 Drobo rev1...
Connectivity using ssh, rsync & vsftpd
Connectivity using ssh, rsync & vsftpd A Presentation for the 2005 Linux Server Boot Camp by David Brown David has 15 years of systems development experience with EDS, and has been writing Linux based
LifeSize Control Installation Guide
LifeSize Control Installation Guide April 2005 Part Number 132-00001-001, Version 1.0 Copyright Notice Copyright 2005 LifeSize Communications. All rights reserved. LifeSize Communications has made every
Quick Start Guide. Cerberus FTP is distributed in Canada through C&C Software. Visit us today at www.ccsoftware.ca!
Quick Start Guide Cerberus FTP is distributed in Canada through C&C Software. Visit us today at www.ccsoftware.ca! How to Setup a File Server with Cerberus FTP Server FTP and SSH SFTP are application protocols
Uploading and Downloading Files to the HIS Server. Application Requirements
Uploading and Downloading Files to the HIS Server Your PHLive environment and How To document screenshots may not always match due to the dynamic nature of HIS. If any modification to the system impacts
Parallels Plesk Panel 11 for your Linux server
Getting Started Guide Parallels Plesk Panel 11 for your Linux server Getting Started Guide Page 1 Getting Started Guide: Parallels Plesk Panel 11, Linux Server Version 1.1 (11.1.2012) Copyright 2012. All
Tutorial: Using WestGrid. Drew Leske Compute Canada/WestGrid Site Lead University of Victoria
Tutorial: Using WestGrid Drew Leske Compute Canada/WestGrid Site Lead University of Victoria Fall 2013 Seminar Series Date Speaker Topic 23 September Lindsay Sill Introduction to WestGrid 9 October Drew
Debugging with TotalView
Tim Cramer 17.03.2015 IT Center der RWTH Aachen University Why to use a Debugger? If your program goes haywire, you may... ( wand (... buy a magic... read the source code again and again and...... enrich
EVault for Data Protection Manager. Course 361 Protecting Linux and UNIX with EVault
EVault for Data Protection Manager Course 361 Protecting Linux and UNIX with EVault Table of Contents Objectives... 3 Scenario... 3 Estimated Time to Complete This Lab... 3 Requirements for This Lab...
Using the Yale HPC Clusters
Using the Yale HPC Clusters Stephen Weston Robert Bjornson Yale Center for Research Computing Yale University Oct 2015 To get help Send an email to: [email protected] Read documentation at: http://research.computing.yale.edu/hpc-support
Simple. Control Panel. for your Linux Server. Getting Started Guide. Simple Control Panel // Linux Server
Getting Started Guide Simple Control Panel for your Linux Server Getting Started Guide Page 1 Getting Started Guide: Simple Control Panel, Linux Server Version 2.1 (02.01.10) Copyright 2010. All rights
TANDBERG MANAGEMENT SUITE 10.0
TANDBERG MANAGEMENT SUITE 10.0 Installation Manual Getting Started D12786 Rev.16 This document is not to be reproduced in whole or in part without permission in writing from: Contents INTRODUCTION 3 REQUIREMENTS
GRID Computing: CAS Style
CS4CC3 Advanced Operating Systems Architectures Laboratory 7 GRID Computing: CAS Style campus trunk C.I.S. router "birkhoff" server The CAS Grid Computer 100BT ethernet node 1 "gigabyte" Ethernet switch
