Energy-aware job scheduler for highperformance



Similar documents
Dynamic resource management for energy saving in the cloud computing environment

Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging

How To Test For Performance And Scalability On A Server With A Multi-Core Computer (For A Large Server)

OPTIMAL MULTI SERVER CONFIGURATION FOR PROFIT MAXIMIZATION IN CLOUD COMPUTING

Configuring a U170 Shared Computing Environment

ECLIPSE Performance Benchmarks and Profiling. January 2009

EPOBF: ENERGY EFFICIENT ALLOCATION OF VIRTUAL MACHINES IN HIGH PERFORMANCE COMPUTING CLOUD

Ready Time Observations

Performance Testing of a Cloud Service

System Requirements Table of contents

LS DYNA Performance Benchmarks and Profiling. January 2009

Chapter 2: Getting Started

High Performance Computing in CST STUDIO SUITE

Energy Constrained Resource Scheduling for Cloud Environment

Energy-aware Memory Management through Database Buffer Control

An Energy-aware Multi-start Local Search Metaheuristic for Scheduling VMs within the OpenNebula Cloud Distribution

Tableau Server 7.0 scalability

DELL. Virtual Desktop Infrastructure Study END-TO-END COMPUTING. Dell Enterprise Solutions Engineering

Dynamic Resource allocation in Cloud

Monitoring Databases on VMware

LSKA 2010 Survey Report Job Scheduler

Run-time Resource Management in SOA Virtualized Environments. Danilo Ardagna, Raffaela Mirandola, Marco Trubian, Li Zhang

EMC Business Continuity for Microsoft SQL Server Enabled by SQL DB Mirroring Celerra Unified Storage Platforms Using iscsi

Clusters: Mainstream Technology for CAE

Power Efficiency Comparison: Cisco UCS 5108 Blade Server Chassis and IBM FlexSystem Enterprise Chassis

LCMON Network Traffic Analysis

Towards Energy Efficient Query Processing in Database Management System

Figure 1A: Dell server and accessories Figure 1B: HP server and accessories Figure 1C: IBM server and accessories

HP reference configuration for entry-level SAS Grid Manager solutions

ECLIPSE Best Practices Performance, Productivity, Efficiency. March 2009

Cloud Computing through Virtualization and HPC technologies

NetIQ Privileged User Manager

CHAPTER 4 PERFORMANCE ANALYSIS OF CDN IN ACADEMICS

Oracle Applications Release 10.7 NCA Network Performance for the Enterprise. An Oracle White Paper January 1998

Enabling Technologies for Distributed Computing

Web Application s Performance Testing

CPU Performance. Lecture 8 CAP

Oracle Database Reliability, Performance and scalability on Intel Xeon platforms Mitch Shults, Intel Corporation October 2011

Where IT perceptions are reality. Test Report. OCe14000 Performance. Featuring Emulex OCe14102 Network Adapters Emulex XE100 Offload Engine

SP Apps Performance test Test report. 2012/10 Mai Au

Recommended hardware system configurations for ANSYS users

Arrow ECS sp. z o.o. Oracle Partner Academy training environment with Oracle Virtualization. Oracle Partner HUB

Dragon Medical Enterprise Network Edition Technical Note: Requirements for DMENE Networks with virtual servers

PRIMERGY server-based High Performance Computing solutions

Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays

How System Settings Impact PCIe SSD Performance

Overlapping Data Transfer With Application Execution on Clusters

Kingston Technology. KingstonConsult WHD Local. October 2014

Using Fuzzy Logic Control to Provide Intelligent Traffic Management Service for High-Speed Networks ABSTRACT:

Performance And Scalability In Oracle9i And SQL Server 2000

Power-Aware High-Performance Scientific Computing

High Performance Oracle RAC Clusters A study of SSD SAN storage A Datapipe White Paper

Performance Tuning and Optimizing SQL Databases 2016

Introduction to IBM tools to manage energy consumption

Power Management in Cloud Computing using Green Algorithm. -Kushal Mehta COP 6087 University of Central Florida

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

FLOW-3D Performance Benchmark and Profiling. September 2012

Power Efficiency Metrics for the Top500. Shoaib Kamil and John Shalf CRD/NERSC Lawrence Berkeley National Lab

McAfee Enterprise Mobility Management Performance and Scalability Guide

Power Efficiency Comparison: Cisco UCS 5108 Blade Server Chassis and Dell PowerEdge M1000e Blade Enclosure

HUAWEI Tecal E6000 Blade Server

1. Simulation of load balancing in a cloud computing environment using OMNET

Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking

Improving Scalability for Citrix Presentation Server

Enabling Technologies for Distributed and Cloud Computing

Microsoft Dynamics CRM 2011 Guide to features and requirements

International Journal of Computer & Organization Trends Volume20 Number1 May 2015

Heterogeneous Workload Consolidation for Efficient Management of Data Centers in Cloud Computing

A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures

HP ProLiant DL585 G5 earns #1 virtualization performance record on VMmark Benchmark

HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect

How To Compare Two Servers For A Test On A Poweredge R710 And Poweredge G5P (Poweredge) (Power Edge) (Dell) Poweredge Poweredge And Powerpowerpoweredge (Powerpower) G5I (

SERVER CLUSTERING TECHNOLOGY & CONCEPT

HPC performance applications on Virtual Clusters

Characterizing Task Usage Shapes in Google s Compute Clusters

System Software for High Performance Computing. Joe Izraelevitz

Microsoft SharePoint Server 2010

WHITE PAPER. ClusterWorX 2.1 from Linux NetworX. Cluster Management Solution C ONTENTS INTRODUCTION

International Journal of Advance Research in Computer Science and Management Studies

Multifaceted Resource Management for Dealing with Heterogeneous Workloads in Virtualized Data Centers

Module: Sharepoint Administrator

Best Practices. Server: Power Benchmark

Power consumption & management: Windows Vista versus Windows XP

Transcription:

Energy-aware job scheduler for highperformance computing 7.9.2011 Olli Mämmelä (VTT), Mikko Majanen (VTT), Robert Basmadjian (University of Passau), Hermann De Meer (University of Passau), André Giesler (Jülich Supercomputing Centre), Willi Homberg (Jülich Supercomputing Centre), olli.mammela@vtt.fi

2 Outline Introduction HPC energy-aware scheduler Evaluation with simulation model Evaluation with real-world testbed Conclusions

3 Introduction Energy-awareness has become a major topic nowadays ICT as a whole is estimated to cover 2% of world s carbon dioxide emissions HPC is no exception: growing demand for higher performance increases total power consumption Research in energy-aware HPC Energy-efficient hardware Dynamic Voltage and Frequency Scaling (DVFS) technique Shutting down HW components at low system utilization Power capping and thermal management This work presents an energy-aware job scheduler for HPC

4 HPC energy-aware scheduler HPC cluster consists of a resource management system (RMS) and several compute nodes Users submit jobs to the queue(s) inside the RMS Job scheduler is responsible for scheduling decisions Several algorithms available for job scheduling Energy-aware scheduler supports three commonly used scheduling algorithms with energy-saving features

5 HPC energy-aware scheduler FIFO When a job is completed, resources are checked for the first queue item If not enough resources, all jobs have to wait Energy-aware FIFO (E-FIFO) Go through the queue until the first job cannot be started Check estimated start time of the 1st job in the queue based on the available resources and currently running jobs If estimated start time is more than T seconds all idle nodes are powered off Back Front

6 HPC energy-aware scheduler Backfilling (first fit and best fit) Functions like FIFO, but when there are not enough resources for the execution of the first job in the queue, the rest of the queue is checked for jobs that can be executed Execution should not cause any delay for the first job Backfill First Fit (BFF): first job that meets the resource and time constraints is chosen Backfill Best Fit (BBF): all potential backfill jobs are searched and the selection is made based on certain criteria In this work BBF uses these criteria to select the best job 1. Nodes 2. Cores 3. Memory Energy-aware backfilling (E-BFF and E-BBF) Same methods for energy savings as in FIFO Idle nodes are powered off if the estimated start time of the first job in the queue is more than T seconds Backfilling has less opportunities to turn off idle nodes than FIFO

7 Simulation model HPC simulation model implemented with OMNeT++ and the INET Framework Models for clients, data centre, servers, and the RMS Network topology consists of three backbone routers and a gateway router Clients send job requests to the data centre

8 Simulation model Data centre module consists of servers, the RMS and a router between them RMS handles incoming job requests and schedules the jobs to the servers RMS also sends power off / power on actions when needed Servers receive jobs from the RMS and execute the jobs

9 Simulation model RMS, servers, and clients derived from StandardHost module of INET Framework Transport, network, physical layer protocols already available Functionalities developed as an application layer program

10 Simulation model Application models also include models of the server components and their power consumption models Details of server CPUs, cores, memory, fans, etc. are defined Power consumption models Processor, memory, hard disk, network interface card, mainboard, fan, power supply unit Models were derived by performing various observations with physical equipment and specific benchmark programs

11 Simulation parameters Parameter Number of clients 20 Number of servers 32 Value Number of job requests 20 * 20 = 400 Job cores 1, 2 or 4 Job core load uniform(30, 99) Job memory uniform(100 MB, 2 GB) Job wall time uniform(600 s, 86400 s) Job nodes Number of simulation runs 10 uniform(1, 5), uniform(1, 10) and uniform(1, 32)

12 Server parameters Parameter Value Number of CPUs 2 Cores per CPU 2 Core frequency 2.4 GHz RAM size 4 * 2 GB = 8GB RAM vendor Kingston RAM type DDR2 800 MHz, unbuffered

13 Energy savings Comparing standard scheduling algorithms to their energy-aware versions Highest energy saving of 16 % with E-FIFO (1-32 nodes) Other savings approx. 6-10 % Savings are highly dependent on system utilization

14 Energy consumption (J), 1-10 nodes FIFO is the most energy consuming Backfilling itself can decrease energy consumption 1.3 % BFF vs FIFO 2.8 % BBF vs FIFO Energy-aware backfill best fit (E-BBF) consumes least amount of energy E-BBF saves 9.1 % energy compared to FIFO

15 Energy consumption (J), 1-32 nodes FIFO consumes again most energy Compared to FIFO, E-BBF can reduce energy consumption by 33 % Savings by standard backfilling is approximately 23 % compared to FIFO

16 Average simulation duration (s) 1-10 nodes requirements At highest 0.62 % increase (BBF vs E-BBF) 1-32 nodes requirements 2.32 % increase at highest (BFF vs E-BFF)

17 Average wait time (s) 1-10 nodes 1.2 % increase at highest (BBF vs E-BBF) 1-32 nodes 0.81 % increase at highest (BBF vs E-BBF)

18 Testbed configuration Energy-aware scheduler was also implemented in Juggle cluster at Jülich Supercomputing Centre Testing environment simulated typical usage of a supercomputer by using a workload generator Several benchmarks programs were used in the tests, more details in the paper Default scheduler of the testbed was Torque RMS Power was measured by Raritan device in intervals of three seconds Strategy for power savings was to place idle nodes in low-power standby state if no jobs which could make use of them Standby mode consumes 50 W less power than idle state/mode

19 Juggle testbed parameters Parameter Value Number of nodes 4 CPUs per node 2 Cores per CPU 2 Core frequency 2.4 GHz CPU architecture AMD Opteron F2216 Operating System Linux CPU Idle Power 95 W RAM size 4 * 8 * 1 GB = 32 GB RAM vendor Kingston RAM type DDR2 667 MHz, unbuffered

20 Testbed results Elapsed time Energy consumed Avg. power consumed Torque Scheduler E-BFF 2049 s 2062 s 1600 kj 1500 kj 781 W 729 W An energy saving of 6.3% was achieved Elapsed time increases 0.63%

21 Conclusions Developed energy-aware scheduler can be applied to HPC data centres without any changes any hardware With the simulation energy savings of 6-16 % were achieved with energy-aware scheduling strategies compared to standard scheduling algorithms Choice of a job scheduling algorithm can have an effect on the energy consumption Testbed experiments also showed energy savings without a large increase in completion time Simulation and testbed experiments showed similar results, which means that the simulation is able to model real-world environment accurately

22 Future work Apply DVFS technique when appropriate Explore different variaties of the backfill best fit algorithm with regards to energy Try out different low power states, such as standby or hybernated Expand the work to include multiple data centres in a federated site scenario

23 More information olli.mammela@vtt.fi This work was supported by the EU FP7 project FIT4Green www.fit4green.eu

24 VTT creates business from technology