Power Monitoring At NCSA ISL and Blue Waters

Similar documents
Power Efficiency Metrics for the Top500. Shoaib Kamil and John Shalf CRD/NERSC Lawrence Berkeley National Lab

High Performance Computing in CST STUDIO SUITE

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

GPU System Architecture. Alan Gray EPCC The University of Edinburgh

NVIDIA Tesla K20-K20X GPU Accelerators Benchmarks Application Performance Technical Brief

Performance of HPC Applications on the Amazon Web Services Cloud

Energy Efficient High Performance Computing Power Measurement Methodology. (version 1.2RC2)

A Smart Investment for Flexible, Modular and Scalable Blade Architecture Designed for High-Performance Computing.

Datacenter Design Issues

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

A PERFORMANCE COMPARISON USING HPC BENCHMARKS: WINDOWS HPC SERVER 2008 AND RED HAT ENTERPRISE LINUX 5

CORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

Bytes and BTUs: Holistic Approaches to Data Center Energy Efficiency. Steve Hammond NREL

Recent Advances in HPC for Structural Mechanics Simulations

The Green Index: A Metric for Evaluating System-Wide Energy Efficiency in HPC Systems

TSUBAME-KFC : a Modern Liquid Submersion Cooling Prototype Towards Exascale

Comparing the performance of the Landmark Nexus reservoir simulator on HP servers

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/ CAE Associates

How To Compare Amazon Ec2 To A Supercomputer For Scientific Applications

Building Clusters for Gromacs and other HPC applications

Designed for Maximum Accelerator Performance

HPC Update: Engagement Model

Getting Started with HPC

High-Performance Computing and Big Data Challenge

SGI High Performance Computing

Cooling and thermal efficiently in

Cray Gemini Interconnect. Technical University of Munich Parallel Programming Class of SS14 Denys Sobchyshak

Error Characterization of Petascale Machines: A study of the error logs from Blue Waters Anno Accademico

Mississippi State University High Performance Computing Collaboratory Brief Overview. Trey Breckenridge Director, HPC

Kriterien für ein PetaFlop System

White Paper on Consolidation Ratios for VDI implementations

HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect

LS DYNA Performance Benchmarks and Profiling. January 2009

HP ProLiant SL270s Gen8 Server. Evaluation Report

GPU Clusters for HPC. Volodymyr Kindratenko Innovative Systems Laboratory

Trends in High-Performance Computing for Power Grid Applications

Hyperscale. The new frontier for HPC. Philippe Trautmann. HPC/POD Sales Manager EMEA March 13th, 2011

Accelerating From Cluster to Cloud: Overview of RDMA on Windows HPC. Wenhao Wu Program Manager Windows HPC team

Copyright 2011 Penguin Computing, Inc. All rights reserved. Interactive Preclinical Analytics via GPU Cloud Platform

InfiniBand Strengthens Leadership as the High-Speed Interconnect Of Choice

Deploying and managing a Visualization Onera

Building a Top500-class Supercomputing Cluster at LNS-BUAP

HPC Cluster Decisions and ANSYS Configuration Best Practices. Diana Collier Lead Systems Support Specialist Houston UGM May 2014

Power Efficiency Comparison: Cisco UCS 5108 Blade Server Chassis and IBM FlexSystem Enterprise Chassis

SRNWP Workshop. HP Solutions and Activities in Climate & Weather Research. Michael Riedmann European Performance Center

Clusters: Mainstream Technology for CAE

GTC Presentation March 19, Copyright 2012 Penguin Computing, Inc. All rights reserved

Green Data Center and Virtualization Reducing Power by Up to 50% Pages 10

Green HPC - Dynamic Power Management in HPC

Pedraforca: ARM + GPU prototype

Using the Intel Xeon Phi (with the Stampede Supercomputer) ISC 13 Tutorial

Amazon EC2 XenApp Scalability Analysis


Accelerating Simulation & Analysis with Hybrid GPU Parallelization and Cloud Computing

Vertical Scaling of Oracle 10g Performance on Red Hat Enterprise Linux 5 on Intel Xeon Based Servers. Version 1.0

CUBIX ACCEL-APP SYSTEMS Linux2U Rackmount Elite

TOOLS AND TIPS FOR MANAGING A GPU CLUSTER. Adam DeConinck HPC Systems Engineer, NVIDIA

Overview of HPC systems and software available within

PRIMERGY server-based High Performance Computing solutions

High Performance Oracle RAC Clusters A study of SSD SAN storage A Datapipe White Paper

SUN ORACLE EXADATA STORAGE SERVER

Journée Mésochallenges 2015 SysFera and ROMEO Make Large-Scale CFD Simulations Only 3 Clicks Away

Smarter Cluster Supercomputing from the Supercomputer Experts

HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK

Desktop Consolidation. Stéphane Verdy, CTO Devon IT

Solve your IT energy crisis WITH An energy SMArT SoluTIon FroM Dell

A-CLASS The rack-level supercomputer platform with hot-water cooling

The CNMS Computer Cluster

Challenges and Opportunities for Exscale Resource Management and How Today's Petascale Systems are Guiding the Way

Manual for using Super Computing Resources

PCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters. from One Stop Systems (OSS)

The benefits of data center temperature monitoring

Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

ECLIPSE Performance Benchmarks and Profiling. January 2009

Sun Constellation System: The Open Petascale Computing Architecture

Big Data Challenges In Leadership Computing

Data Sheet FUJITSU Server PRIMERGY CX400 M1 Multi-Node Server Enclosure

Smart Campus Management with Cloud Services

HPC Growing Pains. Lessons learned from building a Top500 supercomputer

Numerical Calculation of Laminar Flame Propagation with Parallelism Assignment ZERO, CS 267, UC Berkeley, Spring 2015

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

HP Workstations graphics card options

CAS2K5. Jim Tuccillo

StruxureWare TM Center Expert. Data

Icepak High-Performance Computing at Rockwell Automation: Benefits and Benchmarks

Power Efficiency Comparison: Cisco UCS 5108 Blade Server Chassis and Dell PowerEdge M1000e Blade Enclosure

Calculating power requirements for HP ProLiant rack-mounted systems

HP Workstations graphics card options

Optimization of energy consumption in HPC centers: Energy Efficiency Project of Castile and Leon Supercomputing Center - FCSCL

Parallel Programming Survey

1 DCSC/AU: HUGE. DeIC Sekretariat /RB. Bilag 1. DeIC (DCSC) Scientific Computing Installations

Performance Comparison of Fujitsu PRIMERGY and PRIMEPOWER Servers

Finite Elements Infinite Possibilities. Virtual Simulation and High-Performance Computing

BC43: Virtualization and the Green Factor. Ed Harnish

High-Performance Computing Clusters

Smarter Cluster Supercomputing from the Supercomputer Experts

Power efficiency and power management in HP ProLiant servers

THE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC

Transcription:

Power Monitoring At NCSA ISL and Blue Waters Salishan 2013 Conference Craig P Steffen Blue Waters Science and Engineering Applications Suppport Group csteffen@ncsa.illinois.edu

NCSA: Measuring Power Use and Power Effectiveness of User Applications using Sustained System Performance Metrics Use a real user application It is important to use a real user application that makes realistic demands on the system Measure power of whole machine As close as possible to data-center/machine boundary (where the facility charges the owner) Buy (create) and install power monitoring Make the data available in a useful form to anyone who can use it Measure whole application (wall clock) run time Scheduler deals with wall clock time (that s the time other applications can t be running) User allocations charged wall clock time 2

Overview: Power Monitor System Progression 32 node Innovative Systems Lab Accelerator Cluster system Power monitoring on single node 128 node ISL/ECE EcoG system Power monitoring on 8 nodes block 25,712 compute-node Cray XE/XK Blue Waters system Whole-system power monitoring per building transformer and per computational cabinet 3

Stage1: Innovative Systems Lab Accelerator Cluster (AC) (2008?-2012) 32 nodes HP xw9400 workstation 2216 AMD Opteron 2.4 GHz dual socket dual-core Infiniband QDR Each node: Tesla S1070 1U GPU Computing Server 1.3 GHz Tesla T10 processors 4x4 GB GDDR3 SDRAM 4

Commercial Power Meters (~2010) price readout monitors Kill-a-watt $50 LCD display 120 VAC PowerSight PS3000 $2450 asynch 100-250 VAC ElitePro Recording Poly-Phase Power Meter $965 asynch 120 VAC Watts Up Smart Circuit 20 $194.95 Web-only 120 VAC 5

AC Power Hardware Monitor Conclusion: Nothing Suitable Exists; Make One (monitor single node) V1: Kill-a-watt based Xbee Tweet a watt Wireless transmitter Voltage and current V2 and V3: Arduino based power monitors 1 Arduino Duemilanove per chassis 1 Manutech MN-220 20A AC power sensing transformer per measured channel Arduino forms RMS value of AC voltage current (5 times per second) 6

price readout monitors Power Meter: Final Result Kill-a-watt $50 LCD display 120 VAC PowerSight PS3000 $2450 asynch 100-250 VAC ElitePro Recording Poly- Phase Power Meter $965 asynch 120 VAC Watts Up Smart Circuit 20 $194.95 Web-only 120 VAC Our Arduino-based monitor ~$100 (2 channels) USB text (python) base config: 120/208/250 VAC 7

Power data Power Data Harvesting During Job AC Head/Login Node Database Server Python monitor script Power data Power monitor Database USB 208 VAC Power Monitor Monitored Compute node PCI-E NVIDIA Tesla S1070 8

AC Power Monitor Software: Tied Into Job Software, Exports Data Automatically Use qsub to request powermon feature Blue = GPU power Red = CPU node power Prolog script creates link to power graph (left) and raw power data file Power graph available during the run; graph and link to data works afterward (GPU has separate power supply) 9

Application Speedup Summary (small or single-node versions of apps) Application Raw GPU speedup (wall clock time) Speedup scaled by (GPU+node)/node power ratio NAMD 6 2.8 VMD 26 10.5 QMCPack 62 23 MILC 20 8 10

Stage 2: EcoG Student-built GPU Cluster (2010-2012) ECE Independent study course in cluster building to create high performance GPU-based architecture Maps to GPU math capabilities Frequent but not constant nodeto-node updates Likely target apps: Molecular dynamics Fluid dynamics HPL for testing 128 nodes Tesla 2050 GPUs primary computing element; single modest CPU per node Single-socket motherboard Each node: Intel Core i3 2.93 GHz CPU 4 GB RAM main memory (smal to lower power footprint) 1 two-port QDR infiniband High-performance GPUs, lower power CPUs NFS root file system (stateless nodes) 11

EcoG Power Monitoring: modified rack PDU Re-used rack-mounted PDU 2 voltage probes for 208V power legs 2 clamp-on current probes for current measurement Probes secured INSIDE enclosure Asynchronous readout via text file 12

Green 500 Run Rules and EcoG Submission (HPL known quantity for entire system) Green 500 run rules as of fall 2011: Fraction: ANY sub-fraction can be measured and scaled up Measurement position: anywhere Time: at least 20% of run in the middle 80% Subsystems: Compute nodes (only) required NCSA s EcoG (3 page) Submission + technical report 8 of 128 compute nodes Power measured upstream of node (entire chassis) We decided to average power from whole 80% of run HPL is not an application but it is a valid system stress test 13

Sources of Uncertainty in Green 500 Results Subsampling allows small numbers of elements to skew the average (up or down) Measuring power inside the system leaves out efficiency of AC/DC or DC/DC conversion efficiency (10 or 20%) Not requiring whole run allows lower power section of run to be used (~3% effect) Leaving out subsystems (network, head nodes, storage, cooling etc.) artificially lowers power cost of system 14

NCSA has been working with EEHPCWG to create a new power monitoring and reporting specification Energy Efficient HPC Working group: http://eehpcwg.lbl.gov Creating a specification for wholesystem power measurement for application efficiency characterization Hope to convince Top-500, Green- 500, and Green Grid to adopt this standard for power part of submissions Also to drive specifications for machines and comparisons Compare power measurement and performance measurement (general measurement of value) Three quality levels Level 1 current Green-500 Level 3 = current best possible Requires 100% of system Power measured upstream of AC/DC conversion OR loss is accounted for 100% of parallel run used in average power calculation ALL participating subsystems required to be MEASURED Metering devices must be integrating total energy meters 15

Stage 3: Cray XE6/XK7 Blue Waters (2012--) 3D Torus Gemini network topology 24x23x24 Mix of XE and XK Cray nodes (XK in contiguous block) 22640 XE compute nodes Each node 2-die Interlagos processor 16 Bulldozer cores 64 GB of RAM 3072 XK compute nodes Each node 1-die Interlagos processor 8 Bulldozer cores 1 Kepler K20X GPU 32 GB of RAM 3 large Sonexion Lustre file systems (usable capacities): /home (2.2 PB) /project (2.2 PB) /scratch (21.6 PB) 16

Power supply monitor Cray cabinet L 1 277 / 480 VAC Power Transformer Blue Waters Power Delivery L1 controller monitors and reports power from within (each) cabinet Each transformer output monitored and logged 17

airflow Blue Waters Cooling: Flow rates and Temperatures Monitored Evap chillers pump Heat exchanger refrigerant blades pump valve XDP fan flow control valve Campus chilled water 18

Integrated Systems Console Tool being developed as system visualization/monitoring/analysis tool for Blue Waters admins Selected features are also available on the Blue Waters Portal for users All this information in one place allows triage of cooling problems All data together 19

Selected Applications on Blue Waters During Acceptance four large-scale science applications: VPIC, PPM, QMCPACK and SPECFEM3DGLOBE sustained performance >1PF on Blue Waters Weather Research & Forecasting (WRF) run on Blue Waters is the largest WRF simulation ever documented Simulating hurricane Sandy Submitted as paper to SC 2013 These applications are part of the NCSA Blue Waters Sustained Petascale Performance (SPP) suite and represent valid scientific workloads. 20

(sizes are in MPI ranks) App NSF nonpetascale SPP full system (>1.0 PF/s) SPP Interlagos SPP Kepler K20X NSF petascale turbulence (PSDNS) 360,000 nwchem 80000, 8000 specfem3d 693,600 173400, 21600 vpic 180,224 131072, 73728, 8192 ppm 85,668 33024, 32250, 2112 milc 316,416 milc 2048 milc 65856, 8232 wrf 512 wrf 72960, 4920 qmcpack 90,000 153600, 76800, 38400 700 namd (100M) 25,650 namd 20000, 768 768 chroma 768 gamess 1536 paratec 512 Full system: XE: 724,480 XK: 49,152 Composite SPP 1.3 PF/s 21

Friendly User Period Power Usage Friendly user Period: users running real production workload Average Power February: 9.46 MW Average Power March: 9.71 MW 22

Good measurements + effective delivery = ability to understand system behavior Real effectiveness is in terms of performance per watt Real performance in terms of wall clock time Real power in terms of full system power Get the data to those who can use it 23

Thanks and References Thanks to: National Science Foundation State of Illinois Microsoft IBM Linux Technology Center References Enos, J.; Steffen, C.; Fullop, J.; Showerman, M.; Guochun Shi; Esler, K.; Kindratenko, V.; Stone, J.E.; Phillips, J.C., "Quantifying the impact of GPUs on performance and energy efficiency in HPC clusters," Green Computing Conference, 2010 International, vol., no., pp.317,324, 15-18 Aug. 2010 doi: 10.1109/GREENCOMP.2010.5598297 http://www.ncsa.illinois.edu/news/stories/pfapps/ http://www.nersc.gov/research-and-development/performance-and-monitoringtools/sustained-system-performance-ssp-benchmark/ 24