A Practical Method for Estimating Performance Degradation on Multicore Processors, and its Application to HPC Workloads.

Similar documents
Optimizing Shared Resource Contention in HPC Clusters

Power Efficiency Metrics for the Top500. Shoaib Kamil and John Shalf CRD/NERSC Lawrence Berkeley National Lab

Distributed forests for MapReduce-based machine learning

A Holistic Model of the Energy-Efficiency of Hypervisors

Studying Auto Insurance Data

ORACLE OPS CENTER: PROVISIONING AND PATCH AUTOMATION PACK

Classification On The Clouds Using MapReduce

FACT: a Framework for Adaptive Contention-aware Thread migrations

!"!!"#$$%&'()*+$(,%!"#$%$&'()*""%(+,'-*&./#-$&'(-&(0*".$#-$1"(2&."3$'45"

Data Mining: STATISTICA

A Property & Casualty Insurance Predictive Modeling Process in SAS

ACCELERATING COMMERCIAL LINEAR DYNAMIC AND NONLINEAR IMPLICIT FEA SOFTWARE THROUGH HIGH- PERFORMANCE COMPUTING

Table of Contents. June 2010

Data Mining Part 5. Prediction

Addressing Shared Resource Contention in Multicore Processors via Scheduling

Data Mining Part 5. Prediction

In-Situ Bitmaps Generation and Efficient Data Analysis based on Bitmaps. Yu Su, Yi Wang, Gagan Agrawal The Ohio State University

Easily Identify Your Best Customers

High Performance Computing for Operation Research

KNIME TUTORIAL. Anna Monreale KDD-Lab, University of Pisa

Decision Trees from large Databases: SLIQ

How System Settings Impact PCIe SSD Performance

Data mining techniques: decision trees

Automatic Workload Management in Clusters Managed by CloudStack

Applied Data Mining Analysis: A Step-by-Step Introduction Using Real-World Data Sets

Abstract: Motivation: Description of proposal:

Oracle Enterprise Manager 13c Cloud Control

Virtualization. Clothing the Wolf in Wool. Wednesday, April 17, 13

Application of Predictive Analytics for Better Alignment of Business and IT

Azure Machine Learning, SQL Data Mining and R

Analysis of WEKA Data Mining Algorithm REPTree, Simple Cart and RandomTree for Classification of Indian News

Knowledge Discovery and Data Mining. Bootstrap review. Bagging Important Concepts. Notes. Lecture 19 - Bagging. Tom Kelsey. Notes

Model Combination. 24 Novembre 2009

CI6227: Data Mining. Lesson 11b: Ensemble Learning. Data Analytics Department, Institute for Infocomm Research, A*STAR, Singapore.

Characterizing Task Usage Shapes in Google s Compute Clusters

The Data Center as a Grid Load Stabilizer

The Impact of Memory Subsystem Resource Sharing on Datacenter Applications. Lingia Tang Jason Mars Neil Vachharajani Robert Hundt Mary Lou Soffa

Methodology for predicting the energy consumption of SPMD application on virtualized environments *

Data Mining Techniques for Prognosis in Pancreatic Cancer

Comparison of Data Mining Techniques used for Financial Data Analysis

Applied Multivariate Analysis - Big data analytics

An Overview and Evaluation of Decision Tree Methodology

NoSQL Failover Characteristics: Aerospike, Cassandra, Couchbase, MongoDB

testo dello schema Secondo livello Terzo livello Quarto livello Quinto livello

Windows Server 2008 R2 Hyper V. Public FAQ

Towards energy-aware scheduling in data centers using machine learning

COC131 Data Mining - Clustering

A Comparison of Variable Selection Techniques for Credit Scoring

Autonomous Resource Sharing for Multi-Threaded Workloads in Virtualized Servers

Classification and Prediction

Performance Evaluation of NAS Parallel Benchmarks on Intel Xeon Phi

Final Project Report

The Gurobi Optimizer

Nine Common Types of Data Mining Techniques Used in Predictive Analytics

THE HYBRID CART-LOGIT MODEL IN CLASSIFICATION AND DATA MINING. Dan Steinberg and N. Scott Cardell

Practical Data Science with Azure Machine Learning, SQL Data Mining, and R

Dell Virtualization Solution for Microsoft SQL Server 2012 using PowerEdge R820

Data Mining III: Numeric Estimation

PERFORMANCE ENHANCEMENTS IN TreeAge Pro 2014 R1.0

Virtualization. Types of Interfaces

Predicting Student Performance by Using Data Mining Methods for Classification

Welcome to the IBM Education Assistant module for Tivoli Storage Manager version 6.2 Hyper-V backups. hyper_v_backups.ppt.

Data Mining Practical Machine Learning Tools and Techniques

Statistical Process Control (SPC) Training Guide

USTC Course for students entering Clemson F2013 Equivalent Clemson Course Counts for Clemson MS Core Area. CPSC 822 Case Study in Operating Systems

Clusters: Mainstream Technology for CAE

Tableau Server 7.0 scalability

Building an energy dashboard. Energy measurement and visualization in current HPC systems

Scalable Data Analysis in R. Lee E. Edlefsen Chief Scientist UserR! 2011

Data quality in Accounting Information Systems

Red Hat Enterprprise Linux - Renewals DETAILS SUPPORTED ARCHITECTURE

Ensemble Data Mining Methods

Best Practices. Server: Power Benchmark

Multi-Objective Job Placement in Clusters

Lecture 10: Regression Trees

TPCalc : a throughput calculator for computer architecture studies

Best Practices for Hadoop Data Analysis with Tableau

A Regression Approach for Forecasting Vendor Revenue in Telecommunication Industries

Data Mining Classification: Decision Trees

Data Mining from A to Z: Better Insights, New Opportunities WHITE PAPER

BOOSTING - A METHOD FOR IMPROVING THE ACCURACY OF PREDICTIVE MODEL

Chapter 12 Bagging and Random Forests

AV-24 Advanced Analytics for Predictive Maintenance

Web Document Clustering

CSC 177 Fall 2014 Team Project Final Report

A Property and Casualty Insurance Predictive Modeling Process in SAS

White Paper. Redefine Your Analytics Journey With Self-Service Data Discovery and Interactive Predictive Analytics

Classification of Bad Accounts in Credit Card Industry

Comparison of Windows IaaS Environments

Data Mining. Nonlinear Classification

SERVER CLUSTERING TECHNOLOGY & CONCEPT

MONITORING power consumption of a microprocessor

Ensemble Learning of Colorectal Cancer Survival Rates

Social Media Mining. Data Mining Essentials

Classification/Decision Trees (II)

Using multiple models: Bagging, Boosting, Ensembles, Forests

COMP 598 Applied Machine Learning Lecture 21: Parallelization methods for large-scale machine learning! Big Data by the numbers

Marketing Strategies for Retail Customers Based on Predictive Behavior Models

CONTINUOUSLY IMPROVE TO OBTAIN THE ISO 50001

International Journal of Computer Trends and Technology (IJCTT) volume 4 Issue 8 August 2013

Transcription:

A Practical Method for Estimating Performance Degradation on Multicore Processors, and its Application to HPC Workloads Tyler Dwyer, Alexandra Fedorova, Sergey Blagodurov, Mark Roth, Fabien Gaud, Jian Pei 2012 Sameer Wadgaonkar Department of Computer & Information Sciences University of Delaware

Motivation Performance Degradation :- When Multiple programs are run on a modern multicore processor they compete for shared resources. Performance degradation is how much slower does each program run as compared to when run individually on the same system. Performance Degradation can be measured as high as 200%. This causes loss in time and power consumption.

Introduction In this paper the Authors have proposed: 1. A methodology for modeling performance degradation on multicore systems using machine learning. 2. Evaluating strengths and limitations of the resulting model. 3. Designing a confidence predictor that signals when the model is unable to produce an accurate estimate. 4. Demonstrating how the model can be applied to improve performance fidelity to save energy in HPC setting.

Model Testing Platforms: Two systems were used for building and testing the model, Intel and AMD. The models were built using exactly the same procedure on both the systems. The System parameters were as follows

Model Decision tree Learning was used to develop the model. The Nodes of the Decision tree are the attributes with their individual threshold values. Performance Degradation was calculated using the following formula The authors perform the above operation on all instances in the dataset. After the above procedure the authors had 340 attributes per core = 1360 attributes from the event counters.(intel) Weka, a machine learning tool was used for attribute selection. Correlation based feature subset attribute selection(cfssubset) was used within Weka. After attribute selection the number of attributes were reduced to 19 per core from the original 340 attributes per core.(intel)

Model List of Attributes selected after applying attribute selection for the Intel System.

Model The authors have used all modeling procedures available on Weka and compared each of them. After evaluation of all the models present in Weka the Authors choose REPTree as it yielded the highest accuracy. Regression tree mode was used instead of classification tree. The authors also used bagging to lower the error rate further.

Model Root of the Decision tree for Intel System. The number under the attribute is the value used for branching

Results Difference between the actual and predicted degradation for Best, Median and Worst Predicted co-schedules for each primary benchmark. The right most chart shows coscheduled when we apply the confidence predictor.

Results Baseline Cluster Scheduling Policies: Best-fit and Min-collocation. Best-fit allocates the process of the same job on all available cores on the node using additional nodes if needed, but if single job does not fill the cores, it fills them with processes from another job. Min-collocation attempts to schedule no more than one job per node, as long as there are unused nodes available. The Balanced Scheduler is based on the model described above. Job allocation across Nodes

Results Performance and energy consumption Experiment 1: Improved Performance Fidelity

Results Performance and energy consumption Experiment 2: Improved Power Efficiency

Conclusion The study was aimed to investigate the effectiveness of machine learning in modeling contention-induced performance degradation. The proposed model could be run on live workload without a prior knowledge of the applications or the need to run them in isolation. The model accurately estimates degradation within 16% of its true value. The confidence predictor will successfully estimate when the model is likely to produce an inaccurate estimate and reduce the maximum error.

Questions?