LANL Computing Environment for PSAAP Partners



Similar documents
LANL Computing Environment for PSAAP Partners

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Using WestGrid. Patrick Mann, Manager, Technical Operations Jan.15, 2014

Getting Started with HPC

The Asterope compute cluster

Parallel Processing using the LOTUS cluster

MPI / ClusterTools Update and Plans

1 DCSC/AU: HUGE. DeIC Sekretariat /RB. Bilag 1. DeIC (DCSC) Scientific Computing Installations

Work Environment. David Tur HPC Expert. HPC Users Training September, 18th 2015

Introduction to Running Computations on the High Performance Clusters at the Center for Computational Research

Tutorial: Using WestGrid. Drew Leske Compute Canada/WestGrid Site Lead University of Victoria

Introduction to Supercomputing with Janus

JUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert

Caltech Center for Advanced Computing Research System Guide: MRI2 Cluster (zwicky) January 2014

Introduction to Linux and Cluster Basics for the CCR General Computing Cluster

Cloud Computing through Virtualization and HPC technologies

Streamline Computing Linux Cluster User Training. ( Nottingham University)

A High Performance Computing Scheduling and Resource Management Primer

Sun Constellation System: The Open Petascale Computing Architecture

SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education

Hack the Gibson. John Fitzpatrick Luke Jennings. Exploiting Supercomputers. 44Con Edition September Public EXTERNAL

SUN HPC SOFTWARE CLUSTERING MADE EASY

Job Scheduling with Moab Cluster Suite

Designed for Maximum Accelerator Performance

Introduction to ACENET Accelerating Discovery with Computational Research May, 2015

RWTH GPU Cluster. Sandra Wienke November Rechen- und Kommunikationszentrum (RZ) Fotos: Christian Iwainsky

Sun in HPC. Update for IDC HPC User Forum Tucson, AZ, Sept 2008

High Performance Computing Infrastructure at DESY

Overview of HPC systems and software available within

SRNWP Workshop. HP Solutions and Activities in Climate & Weather Research. Michael Riedmann European Performance Center

An Introduction to High Performance Computing in the Department

System Software for High Performance Computing. Joe Izraelevitz

NEC HPC-Linux-Cluster

CNR-INFM DEMOCRITOS and SISSA elab Trieste

High Performance Computing in CST STUDIO SUITE

SLURM: Resource Management and Job Scheduling Software. Advanced Computing Center for Research and Education

Fujitsu HPC Cluster Suite

High Performance Computing in Aachen

icer Bioinformatics Support Fall 2011

HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

The RWTH Compute Cluster Environment

Linux Cluster Computing An Administrator s Perspective

CAS2K5. Jim Tuccillo

Introduction to Running Hadoop on the High Performance Clusters at the Center for Computational Research

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

ABAQUS High Performance Computing Environment at Nokia

Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar

A GPU COMPUTING PLATFORM (SAGA) AND A CFD CODE ON GPU FOR AEROSPACE APPLICATIONS

Using the Yale HPC Clusters

OpenMP Programming on ScaleMP

High Performance Computing at the Oak Ridge Leadership Computing Facility

Cloud Computing for Control Systems CERN Openlab Summer Student Program 9/9/2011 ARSALAAN AHMED SHAIKH

IBM Platform Computing : infrastructure management for HPC solutions on OpenPOWER Jing Li, Software Development Manager IBM

Ra - Batch Scripts. Timothy H. Kaiser, Ph.D. tkaiser@mines.edu

Windows HPC 2008 Cluster Launch

The CNMS Computer Cluster

Building Clusters for Gromacs and other HPC applications

HPC Software Requirements to Support an HPC Cluster Supercomputer

Recommended hardware system configurations for ANSYS users

Smarter Cluster Supercomputing from the Supercomputer Experts

SLURM Workload Manager

- An Essential Building Block for Stable and Reliable Compute Clusters

Cloud Bursting with SLURM and Bright Cluster Manager. Martijn de Vries CTO

CORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER

The Maui High Performance Computing Center Department of Defense Supercomputing Resource Center (MHPCC DSRC) Hadoop Implementation on Riptide - -

The GRID and the Linux Farm at the RCF

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance


PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute

IBM System x SAP HANA

Supercomputing Status und Trends (Conference Report) Peter Wegner

Cluster Implementation and Management; Scheduling

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC Denver

Technical Computing Suite Job Management Software

FLOW-3D Performance Benchmark and Profiling. September 2012

Performance Comparison of SQL based Big Data Analytics with Lustre and HDFS file systems

Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband

ECLIPSE Best Practices Performance, Productivity, Efficiency. March 2009

Using NeSI HPC Resources. NeSI Computational Science Team

Parallel Programming Survey

Provisioning and Resource Management at Large Scale (Kadeploy and OAR)

Background and introduction Using the cluster Summary. The DMSC datacenter. Lars Melwyn Jensen. Niels Bohr Institute University of Copenhagen

Stanford HPC Conference. Panasas Storage System Integration into a Cluster

DeIC Watson Agreement - hvad betyder den for DeIC medlemmerne

Batch Scripts for RA & Mio

Energy-aware job scheduler for highperformance

Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer

Advanced MPI. Hybrid programming, profiling and debugging of MPI applications. Hristo Iliev RZ. Rechen- und Kommunikationszentrum (RZ)

MOSIX: High performance Linux farm

PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN

Transcription:

LANL Computing Environment for PSAAP Partners Robert Cunningham rtc@lanl.gov HPC Systems Group (HPC-3) October 2009

LANL Resources Available To Alliance Users Coyote is old, has a tiny allocation Accounts available for porting, data transfer, small scale runs Bproc Linux, difficult port, LSF scheduler 2,600 Opterons with IB, 13.5 TF, max of 510 procs/job Lobo is current workhorse -- TLCC platform Simple Linux OS, Moab job scheduler, Panasas filesystem 4,352 compute CPUs; max 2,144 procs/job Small Roadrunner platform available: Cerrillos 1,440 hybrid pairs (AMD/cell), max 360 procs (x2) / job

LANL Lobo Cluster Standard DoE Lab cookie-cutter from Appro, Intl. 75% allocation for ASC, primarily for PSAAP 2 Connected Units (CUs) combined include: 4,352 x AMD Opteron @ 2.2 GHz cores on compute nodes Voltaire InfiniBand interconnect 8.7 TeraBytes RAM Theoretical peak of ~17.4 teraflops

LANL Cerrillos Cluster Hybrid architecture: Opteron+Cell, from IBM. 25% allocation for ASC, primarily for PSAAP 2 Connected Units (CUs) combined include: 1,440 x AMD Opteron @ 1.8 GHz cores on compute nodes 1,440 Cell Broadband Engines Voltaire InfiniBand interconnect 11.8 TeraBytes RAM Theoretical peak of ~152 teraflops

LANL HPC Environment Obtain an account, acquire cryptocard (foreign nationals start early!) Access HPC platforms in Open Collaborative (Turquoise) Network via firewall: ssh wtrw.lanl.gov After connecting, ssh into front-end: lo1-fe, lo2-fe; ce-fe1, ce-fe2 Moab + [Slurm Torque] schedules nodes, batch or interactive msub scriptname -or - msub -I (hopefully) checkjob, showstate, showq, mjobctl mpirun to run parallel jobs across nodes Fairshare scheduling to deliver pre-determined allocation Unique security model no connecting (ssh) between platforms No kerberos tickets; keep cryptocard handy scp or sftp using File Transfer Agents (FTAs), turq-fta1.lanl.gov and turqfta2.lanl.gov

LANL HPC Environment Usage model Front-end: text editing, job script set-up, pathname arrangements, compilations, linking, and job launching Compute nodes: run applications I/O nodes: out to local disk, no direct user access Possible intra-network File Transfer Agent (FTA) arrangement in future Modulefiles to establish compilers, libraries, tools in $PATH, $LD_LIBRARY_PATH Compilers: Intel, PGI, Pathscale. MPI: OpenMPI, Mvapich Tools: TotalView, IDL, python, perl, Subversion, Atlas

LANL Turquoise HPC Storage Tiny home directories, not shared between s (security) Larger NFS-based workspace in /usr/projects/proj_name Big parallel, globally-accessible filesystem: /scratchn Cross-mounted to all HPC nodes in Turquoise ~800 TB total space Fast for parallel I/O, slower than NFS for serial transfers No automated back-ups Purged weekly, 30-days or older! Archival storage available via Tivoli (TSM), NFS-mounted as /archive/proj_name/moniker File Transfers in/out of Turquoise are S L O W, we will be addressing this in the next fiscal year, may need your help.

Parallel Scalable Back Bone (PaScalBB) Disk Poor Clients Job Queue Scalable OBFS Enterprise Global FS File System Gateways BackBone Object Archive Diskless Clients Relieve the compute nodes Multiple s sharing large, global namespace parallel I/O sub Includes Cerrillos/Lobo/Coyote Network is combination of HPC Interconnect + commodity networking bridge Panasas is storage vendor I/O through a set of fileserver nodes over IB; nodes serve as interconnect<->gige routers.

3 LANL Web Sites You Can't Live Without User Docs http://bear.lanl.gov/drupal/?q=tlcc_home (or) https://bear.lanl.gov/drupal (or call 505-665-4444 option 3, consult@lanl.gov) Calendar http://ccp.lanl.gov HPC Training http://int.lanl.gov/projects/asci/training

HPC Accounts Don t forget Photo Op!

Questions?