Lecture 24: WSC, Datacenters. Topics: network-on-chip wrap-up, warehouse-scale computing and datacenters (Sections 6.1-6.7)



Similar documents
Lecture 23: Interconnection Networks. Topics: communication latency, centralized and decentralized switches (Appendix E)

Concepts Introduced in Chapter 6. Warehouse-Scale Computers. Important Design Factors for WSCs. Programming Models for WSCs

Chapter 7: Distributed Systems: Warehouse-Scale Computing. Fall 2011 Jussi Kangasharju

Last time. Data Center as a Computer. Today. Data Center Construction (and management)

Topological Properties

Hadoop. MPDL-Frühstück 9. Dezember 2013 MPDL INTERN

Hadoop Parallel Data Processing

Interconnection Networks Programmierung Paralleler und Verteilter Systeme (PPV)

COMP 422, Lecture 3: Physical Organization & Communication Costs in Parallel Machines (Sections 2.4 & 2.5 of textbook)

Hadoop: A Framework for Data- Intensive Distributed Computing. CS561-Spring 2012 WPI, Mohamed Y. Eltabakh

COM 444 Cloud Computing

Welcome to the unit of Hadoop Fundamentals on Hadoop architecture. I will begin with a terminology review and then cover the major components

Cloud Computing Data Centers

Hadoop Cluster Applications

Big Data With Hadoop

Energy Efficient MapReduce

RAID. RAID 0 No redundancy ( AID?) Just stripe data over multiple disks But it does improve performance. Chapter 6 Storage and Other I/O Topics 29

CIT 668: System Architecture

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software

Ch. 4 - Topics of Discussion

Parallel Processing of cluster by Map Reduce

Green Computing: Datacentres

CSE-E5430 Scalable Cloud Computing Lecture 2

BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB

Resource Efficient Computing for Warehouse-scale Datacenters

Data Mining for Data Cloud and Compute Cloud

Benchmarking Cassandra on Violin

RAMCloud and the Low- Latency Datacenter. John Ousterhout Stanford University

System Interconnect Architectures. Goals and Analysis. Network Properties and Routing. Terminology - 2. Terminology - 1

Interconnection Networks. Interconnection Networks. Interconnection networks are used everywhere!

Scheduling in the Cloud

Hadoop Architecture. Part 1

Architectures for Big Data Analytics A database perspective

Scalable Cloud Computing Solutions for Next Generation Sequencing Data

Design of Warehouse Scale Computers (WSC)

Green Computing: Datacentres

Hyper Node Torus: A New Interconnection Network for High Speed Packet Processors

Parallel Programming Map-Reduce. Needless to Say, We Need Machine Learning for Big Data

Jeffrey D. Ullman slides. MapReduce for data intensive computing

Parallel Databases. Parallel Architectures. Parallelism Terminology 1/4/2015. Increase performance by performing operations in parallel

Cloud Computing at Google. Architecture

Components: Interconnect Page 1 of 18

Interconnection Network Design

Lecture 7: Data Center Networks"

COST MINIMIZATION OF RUNNING MAPREDUCE ACROSS GEOGRAPHICALLY DISTRIBUTED DATA CENTERS

Apache Hadoop. Alexandru Costan

Leveraging BlobSeer to boost up the deployment and execution of Hadoop applications in Nimbus cloud environments on Grid 5000

Analysis and Research of Cloud Computing System to Comparison of Several Cloud Computing Platforms

GraySort and MinuteSort at Yahoo on Hadoop 0.23

Map Reduce / Hadoop / HDFS

Chapter 7. Using Hadoop Cluster and MapReduce

How To Scale Out Of A Nosql Database

Radware ADC-VX Solution. The Agility of Virtual; The Predictability of Physical

CURTAIL THE EXPENDITURE OF BIG DATA PROCESSING USING MIXED INTEGER NON-LINEAR PROGRAMMING

Modern Data Centers: Creating a New Value Proposition

Infrastructures for big data

Introduction to Cloud Computing

Big Data and Apache Hadoop s MapReduce

CS 91: Cloud Systems & Datacenter Networks

ebay Storage, From Good to Great

HadoopTM Analytics DDN

Developing MapReduce Programs

Interconnection Network

A Cost-Evaluation of MapReduce Applications in the Cloud

NoSQL for SQL Professionals William McKnight

DISTRIBUTED SYSTEMS AND CLOUD COMPUTING. A Comparative Study

Evaluation of NoSQL databases for large-scale decentralized microblogging

Big Data Technology Map-Reduce Motivation: Indexing in Search Engines

Paolo Costa

Improving MapReduce Performance in Heterogeneous Environments

Maximizing Hadoop Performance and Storage Capacity with AltraHD TM

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

Lecture 2 Parallel Programming Platforms

IMPROVED FAIR SCHEDULING ALGORITHM FOR TASKTRACKER IN HADOOP MAP-REDUCE

Parallel Computing. Benson Muite. benson.

Big Data Analytics with MapReduce VL Implementierung von Datenbanksystemen 05-Feb-13

CS2510 Computer Operating Systems

CS2510 Computer Operating Systems

EX ECUT IV E ST RAT EG Y BR IE F. Server Design for Microsoft s Cloud Infrastructure. Cloud. Resources

Internet Scale Storage Microsoft Storage Community

A Novel Cloud Based Elastic Framework for Big Data Preprocessing

The Total Cost of (Non) Ownership of a NoSQL Database Cloud Service

LARGE-SCALE DATA STORAGE APPLICATIONS

Data-Intensive Computing with Map-Reduce and Hadoop

Overview : Computer Networking. The Big Switch, Redux. Data Center Overview. Networking in the DC. L-24 Data Center Networking

How To Handle Big Data With A Data Scientist

Data Warehousing and Analytics Infrastructure at Facebook. Ashish Thusoo & Dhruba Borthakur athusoo,dhruba@facebook.com

Facebook: Cassandra. Smruti R. Sarangi. Department of Computer Science Indian Institute of Technology New Delhi, India. Overview Design Evaluation

Management & Analysis of Big Data in Zenith Team

Hypertable Architecture Overview

Cloud computing doesn t yet have a

Input / Ouput devices. I/O Chapter 8. Goals & Constraints. Measures of Performance. Anatomy of a Disk Drive. Introduction - 8.1

Data Centers and Cloud Computing

CS246: Mining Massive Datasets Jure Leskovec, Stanford University.

Datacenters and Cloud Computing. Jia Rao Assistant Professor in CS

Google File System. Web and scalability

Weekly Report. Hadoop Introduction. submitted By Anurag Sharma. Department of Computer Science and Engineering. Indian Institute of Technology Bombay

From GWS to MapReduce: Google s Cloud Technology in the Early Days

Apache HBase. Crazy dances on the elephant back

A High-availability and Fault-tolerant Distributed Data Management Platform for Smart Grid Applications

Transcription:

Lecture 24: WSC, Datacenters Topics: network-on-chip wrap-up, warehouse-scale computing and datacenters (Sections 6.1-6.7) 1

Topology Examples Grid Torus Hypercube Criteria 64 nodes Performance Bisection bandwidth Bus Ring 2Dtorus 6-cube Fully connected 1 2 16 32 1024 Cost Ports/switch 3 5 7 64 Total links 1 128 192 256 2080 2

k-ary d-cube Consider a k-ary d-cube: a d-dimension array with k elements in each dimension, there are links between elements that differ in one dimension by 1 (mod k) Number of nodes N = k d (with no wraparound) Number of switches : Switch degree : Number of links : Pins per node : N 2d + 1 Nd 2wd Avg. routing distance: Diameter : Bisection bandwidth : Switch complexity : Should we minimize or maximize dimension? d(k-1)/2 d(k-1) 2wk d-1 (2d + 1) 2 3

Warehouse-Scale Computer (WSC) 100K+ servers in one WSC ~$150M overall cost Requests from millions of users (Google, Facebook, etc.) Cloud Computing: a model where users can rent compute and storage within a WSC, there s an associated service-level agreement (SLA) Datacenter: a collection of WSCs in a single building, possibly belonging to different clients and using different hardware/architecture 4

Workloads Typically, software developed in-house MapReduce, BigTable, etc. MapReduce: embarrassingly parallel operations performed on very large datasets, e.g., organize data into clusters, aggregate a count over several documents Hadoop is an open-source implementation of the MapReduce framework; makes it easy for users to write MapReduce programs without worrying about low-level task/data management 5

MapReduce Application-writer provides Map and Reduce functions that operate on key-value pairs Each map function operates on a collection of records; a record is (say) a webpage or a facebook user profile The records are in the file system and scattered across several servers; thousands of map functions are spawned to work on all records in parallel The Reduce function aggregates and sorts the results produced by the Mappers, also performed in parallel 6

Word Count Histogram Example 7

MR Framework Duties Replicate data for fault tolerance Detect failed threads and re-start threads Handle variability in thread response times Use of MR within Google has been growing every year: Aug 04 Sep 09 Number of MR jobs has increased 100x+ Data being processed has increased 100x+ Number of servers per job has increased 3x 8

WSC Hierarchy A rack can hold 48 1U servers (1U is 1.75 inches high and is the maximum height for a server unit) A rack switch is used for communication within and out of a rack; an array switch connects an array of racks Latency grows if data is fetched from remote DRAM or disk (300us vs. 0.1us for DRAM and 12ms vs. 10ms for disk ) Bandwidth within a rack is much higher than between arrays; hence, software must be aware of data placement and locality 9

Power Delivery and Efficiency Figure 6.9 Power distribution and where losses occur. Note that the best improvement is 11%. (From Hamilton [2010].) Source: H&P Textbook Copyright 2011, Elsevier Inc. All rights Reserved. 10

PUE Metric and Power Breakdown PUE = Total facility power / IT equipment power (power utilization effectiveness) It is greater than 1; ranges from 1.33 to 3.03, median of 1.69 The cooling power is roughly half the power used by servers Within a server (circa 2007), the power distribution is as follows: Processors (33%), DRAM memory (30%), Disks (10%), Networking (5%), Miscellaneous (22%) 11

CapEx and OpEx Capital expenditure: infrastructure costs for the building, power delivery, cooling, and servers Operational expenditure: the monthly bill for energy, failures, personnel, etc. CapEx can be amortized into a monthly estimate by assuming that the facilities will last 10 years, server parts will last 3 years, and networking parts will last 4 12

CapEx/OpEx Case Study 8 MW facility : facility cost: $88M, server/networking cost: $79M Monthly expense: $3.8M. Breakdown: Servers 53% (amortized CapEx) Networking 8% (amortized CapEx) Power/cooling infrastructure 20% (amortized CapEx) Other infrastructure 4% (amortized CapEx) Monthly power bill 13% (true OpEx) Monthly personnel salaries 2% (true OpEx) 13

Improving Energy Efficiency An unloaded server dissipates a large amount of power Ideally, we want energy-proportional computing, but in reality, servers are not energy-proportional Can approach energy-proportionality by turning on a few servers that are heavily utilized See figures on next two slides for power/utilization profile of a server and a utilization profile of servers in a WSC 14

Power/Utilization Profile Source: H&P textbook. Copyright 2011, Elsevier Inc. All rights Reserved. 15

Server Utilization Profile Source: H&P textbook. Copyright 2011, Elsevier Inc. All rights Reserved. Figure 6.3 Average CPU utilization of more than 5000 servers during a 6-month period at Google. Servers are rarely completely idle or fully utilized, in-stead operating most of the time at between 10% and 50% of their maximum utilization. (From Figure 1 in Barroso and Hölzle [2007].) The column the third from the right in Figure 6.4 calculates percentages plus or minus 5% to come up with the weightings; thus, 1.2% for the 90% row means that 1.2% of servers were between 85% and 95% utilized. 16

Other Metrics Performance does matter, especially latency An analysis of the Bing search engine shows that if a 200ms delay is introduced in the response, the next click by the user is delayed by 500ms; so a poor response time amplifies the user s non-productivity Reliability (MTTF) and Availability (MTTF/MTTF+MTTR) are very important, given the large scale A server with MTTF of 25 years (amazing!) : 50K servers would lead to 5 server failures a day; Similarly, annual disk failure rate is 2-10% 1 disk failure every hour 17

Important Problems Reducing power in power-down states Maximizing utilization Reducing cost with virtualization Reducing data movement Building a low-power low-cost processor Building a low-power low-cost hi-bw memory Low-power low-cost on-demand reliability 18

Title Bullet 19