Delay Scheduling. A Simple Technique for Achieving Locality and Fairness in Cluster Scheduling
|
|
|
- Geoffrey Harvey
- 9 years ago
- Views:
Transcription
1 Delay Scheduling A Simple Technique for Achieving Locality and Fairness in Cluster Scheduling Matei Zaharia, Dhruba Borthakur *, Joydeep Sen Sarma *, Khaled Elmeleegy +, Scott Shenker, Ion Stoica UC Berkeley, * Facebook Inc, + Yahoo! Research
2 Motivation MapReduce / Hadoop originally designed for high throughput batch processing Today s workload is far more diverse: Many users want to share a cluster Engineering, marketing, business intelligence, etc Vast majority of jobs are short Ad- hoc queries, sampling, periodic reports Response time is critical Interactive queries, deadline- driven reports How can we efticiently share MapReduce clusters between users?
3 Example: Hadoop at Facebook 600- node, 2 PB data warehouse, growing at 15 TB/day Applications: data mining, spam detection, ads 200 users (half non- engineers) 7500 MapReduce jobs / day CDF 100% 80% 60% 40% 20% 0% Median: 84s Job Running Time (s)
4 Approaches to Sharing Hadoop default scheduler (FIFO) Problem: short jobs get stuck behind long ones Separate clusters Problem 1: poor utilization Problem 2: costly data replication Full replication across clusters nearly infeasible at Facebook/Yahoo! scale Partial replication prevents cross- dataset queries
5 Our Work Hadoop Fair Scheduler Fine- grained sharing at level of map & reduce tasks Predictable response times and user isolation Main challenge: data locality For efticiency, must run tasks near their input data Strictly following any job queuing policy hurts locality: job picked by policy may not have data on free nodes Solution: delay scheduling Relax queuing policy for limited time to achieve locality
6 The Problem Master Job 1 Job 2 Scheduling order Task 2 Task 5 Task 3 Task 1 Task 7 Task 4 Slave Slave Slave Slave Slave Slave File 1: File 2:
7 The Problem Master Job 2 Job 1 Scheduling order Task 2 Task 5 Task 31 Task 1 Task 72 Slave Slave Slave Slave Slave Task 43 Slave File 1: File 2: Problem: Fair decision hurts locality Especially bad for jobs with small input Tiles
8 Locality vs. Job Size at Facebook Data Locality in Production at Facebook Percent Local Map Tasks 100% 80% 58% of jobs 60% 40% 20% 0% Job Size (Number of Input Blocks) Node Locality Rack Locality
9 Special Instance: Sticky Slots Under fair sharing, locality can be poor even when all jobs have large input Tiles Problem: jobs get stuck in the same set of task slots When one a task in job j Tinishes, the slot it was running in is given back to j, because j is below its share Bad because data Tiles are spread out across all nodes Master Task Task Task Task Slave Slave Slave Slave Fair Running Job Share Tasks Job Job 2 2 2
10 Special Instance: Sticky Slots Under fair sharing, locality can be poor even when all jobs have large input Tiles Problem: jobs get stuck in the same set of task slots When one a task in job j Tinishes, the slot it was running in is given back to j, because j is below its share Bad because data Tiles are spread out across all nodes % Local Map Tasks Data Locality vs. Number of Concurrent Jobs 100% 80% 60% 40% 20% 0% 5 Jobs 10 Jobs 20 Jobs 50 Jobs
11 Solution: Delay Scheduling Relax queuing policy to make jobs wait for a limited time if they cannot launch local tasks Result: Very short wait time (1-5s) is enough to get nearly 100% locality
12 Delay Scheduling Example Wait! Master Job 2 Job 1 Scheduling order Task 2 Task 51 Task 38 Task 13 Task 72 Slave Slave Slave Slave Slave Task 46 Slave File 1: File 2: Idea: Wait a short time to get data- local scheduling opportunities
13 Delay Scheduling Details Scan jobs in order given by queuing policy, picking Tirst that is permitted to launch a task Jobs must wait before being permitted to launch non- local tasks If wait < T 1, only allow node- local tasks If T 1 < wait < T 2, also allow rack- local If wait > T 2, also allow off- rack Increase a job s time waited when it is skipped
14 Analysis When is it worth it to wait, and for how long? Waiting worth it if E(wait time) < cost to run non-locally Simple model: Data- local scheduling opportunities arrive following Poisson process Expected response time gain from delay scheduling: E(gain) = (1 e - w/t )(d t) Wait amount Expected time to get local scheduling opportunity Delay from running non- locally Optimal wait time is inainity
15 Analysis Continued E(gain) = (1 e - w/t )(d t) Wait amount Expected time to get local scheduling opportunity Delay from running non- locally Typical values of t and d: t = (avg task length) / (Tile replication tasks per node) d (avg task length) 3 6
16 More Analysis What if some nodes are occupied by long tasks? Suppose we have: Fraction φ of active tasks are long R replicas of each block L task slots per node For a given data block, P(all replicas blocked by long tasks) = φ RL With R = 3 and L = 6, this is less than 2% if φ < 0.8
17 Evaluation Macrobenchmark IO- heavy workload CPU- heavy workload Mixed workload Microbenchmarks Sticky slots Small jobs Hierarchical fair scheduling Sensitivity analysis Scheduler overhead
18 Macrobenchmark 100- node EC2 cluster, 4 cores/node Job submission schedule based on job sizes and inter- arrival times at Facebook 100 jobs grouped into 9 bins of sizes Three workloads: IO- heavy, CPU- heavy, mixed Three schedulers: FIFO Fair sharing Fair + delay scheduling (wait time = 5s)
19 Results for IO-Heavy Workload Job Response Times Up to 5x % 0.8 CDF CDF CDF FIFO Fair Fair + DS FIFO Fair Fair + DS FIFO 0.2 Fair Fair + DS Time (s) Time (s) Time (s) Small Jobs (1-10 input blocks) Medium Jobs ( input blocks) Large Jobs (4800 input blocks)
20 Results for IO-Heavy Workload Percent Local Maps Speedup from Delay Scheduling 100% 80% 60% 40% 20% 0% 80% 60% 40% 20% 0% FIFO Fair Fair + Delay Sched Bin Bin
21 Sticky Slots Microbenchmark 5-50 jobs on EC node cluster 4 cores / node Percent Local Maps 100% 80% 60% 40% 20% 0% 5 Jobs 10 Jobs 20 Jobs 50 Jobs 5s delay scheduling Benchmark Running Time (minutes) Without Delay Scheduling With Delay Scheduling 5 Jobs 10 Jobs 20 Jobs 50 Jobs 2x Without Delay Scheduling With Delay Scheduling
22 Conclusions Delay scheduling works well under two conditions: SufTicient fraction of tasks are short relative to jobs There are many locations where a task can run efticiently Blocks replicated across nodes, multiple tasks/node Generalizes beyond MapReduce and fair sharing Applies to any queuing policy that gives ordered job list Used in Hadoop Fair Scheduler to implement hierarchical policy Applies to placement needs other than data locality
23 Thanks! This work is open source, packaged with Hadoop as the Hadoop Fair Scheduler Old version without delay scheduling in use at Facebook since fall 2008 Delay scheduling will appear in Hadoop 0.21 My
24 Related Work Quincy (SOSP 09) Locality aware fair scheduler for Dryad Expresses scheduling problem as min- cost Tlow Kills tasks to reassign nodes when jobs change One task slot per machine HPC scheduling Inelastic jobs (cannot scale up/down over time) Data locality generally not considered Grid computing Focused on common interfaces for accessing compute resources in multiple administrative domains
25 Delay Scheduling Details when there is a free task slot on node n: sort jobs according to queuing policy for j in jobs: if j has node- local task t on n: j.level := 0; j.wait := 0; return t else if j has rack- local task t on n and (j.level 1 or j.wait T 1 ): j.level := 1; j.wait := 0; return t else if j.level = 2 or (j.level = 1 and j.wait T 2 ) or (j.level = 0 and j.wait T 1 + T 2 ): j.level := 2; j.wait := 0; return t else: j.wait += time since last scheduling decision
26 Sensitivity Analysis Percent Local Maps 100% 80% 60% 40% 20% 0% 98% 99% 100% 100% 80% 68% 11% 5% No Delay 1s Delay 5s Delay 10s Delay 4- Map Jobs 12- Map Jobs
27 Hadoop Terminology Job = entire MapReduce computation; consists of map & reduce tasks Task = unit of work that is part of a job Slot = location on a slave where a task can run Block Map Input File Block Map Reduce Reduce Output File Block Map
28 Analysis Required waiting time is a fraction of the average task length as long as there are many locations where a task can run Ex: with 3 replicas / block and 8 tasks / node, wait 1/24 th of average task length Non- locality decreases exponentially with waiting time
29 Analysis Suppose we have: M machines L slots/machine T second average task length Job with data on fraction f of nodes Let D = max # of scheduling opportunities skipped P(non- local) = (1- f) D waiting time = D * T/(ML) Decreases exponentially with D Fraction of mean task length Decreases with # slots/node
30 Small Jobs Experiment min benchmark consisting of several hundred IO- heavy jobs Job sizes: 3, 10 and 100 map tasks Environment: 100- node 0.4 cluster on 4 racks; 5 map slots per node Normalized Running Time Job Size Default Scheduler Node Loc. Rack Loc. Node Loc. Rack Loc. 3 maps 2% 50% 10 maps 37% 98% 100 maps 84% 99% 3 Maps 10 Maps 100 Maps Job Size (Number of Maps) With Delay Sched. 75% 94% 99% 99% 94% 99% No Delay Scheduling With Delay Scheduling
Job Scheduling for MapReduce
UC Berkeley Job Scheduling for MapReduce Matei Zaharia, Dhruba Borthakur *, Joydeep Sen Sarma *, Scott Shenker, Ion Stoica RAD Lab, * Facebook Inc 1 Motivation Hadoop was designed for large batch jobs
Delay Scheduling: A Simple Technique for Achieving Locality and Fairness in Cluster Scheduling
Delay Scheduling: A Simple Technique for Achieving Locality and Fairness in Cluster Scheduling Matei Zaharia University of California, Berkeley [email protected] Khaled Elmeleegy Yahoo! Research [email protected]
IMPROVED FAIR SCHEDULING ALGORITHM FOR TASKTRACKER IN HADOOP MAP-REDUCE
IMPROVED FAIR SCHEDULING ALGORITHM FOR TASKTRACKER IN HADOOP MAP-REDUCE Mr. Santhosh S 1, Mr. Hemanth Kumar G 2 1 PG Scholor, 2 Asst. Professor, Dept. Of Computer Science & Engg, NMAMIT, (India) ABSTRACT
Mesos: A Platform for Fine- Grained Resource Sharing in Data Centers (II)
UC BERKELEY Mesos: A Platform for Fine- Grained Resource Sharing in Data Centers (II) Anthony D. Joseph LASER Summer School September 2013 My Talks at LASER 2013 1. AMP Lab introduction 2. The Datacenter
Matchmaking: A New MapReduce Scheduling Technique
Matchmaking: A New MapReduce Scheduling Technique Chen He Ying Lu David Swanson Department of Computer Science and Engineering University of Nebraska-Lincoln Lincoln, U.S. {che,ylu,dswanson}@cse.unl.edu
Improving MapReduce Performance in Heterogeneous Environments
UC Berkeley Improving MapReduce Performance in Heterogeneous Environments Matei Zaharia, Andy Konwinski, Anthony Joseph, Randy Katz, Ion Stoica University of California at Berkeley Motivation 1. MapReduce
Survey on Scheduling Algorithm in MapReduce Framework
Survey on Scheduling Algorithm in MapReduce Framework Pravin P. Nimbalkar 1, Devendra P.Gadekar 2 1,2 Department of Computer Engineering, JSPM s Imperial College of Engineering and Research, Pune, India
Big Data Analysis and Its Scheduling Policy Hadoop
IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 1, Ver. IV (Jan Feb. 2015), PP 36-40 www.iosrjournals.org Big Data Analysis and Its Scheduling Policy
Apache Hadoop. Alexandru Costan
1 Apache Hadoop Alexandru Costan Big Data Landscape No one-size-fits-all solution: SQL, NoSQL, MapReduce, No standard, except Hadoop 2 Outline What is Hadoop? Who uses it? Architecture HDFS MapReduce Open
GraySort on Apache Spark by Databricks
GraySort on Apache Spark by Databricks Reynold Xin, Parviz Deyhim, Ali Ghodsi, Xiangrui Meng, Matei Zaharia Databricks Inc. Apache Spark Sorting in Spark Overview Sorting Within a Partition Range Partitioner
An Adaptive Scheduling Algorithm for Dynamic Heterogeneous Hadoop Systems
An Adaptive Scheduling Algorithm for Dynamic Heterogeneous Hadoop Systems Aysan Rasooli, Douglas G. Down Department of Computing and Software McMaster University {rasooa, downd}@mcmaster.ca Abstract The
Introduction to Big Data! with Apache Spark" UC#BERKELEY#
Introduction to Big Data! with Apache Spark" UC#BERKELEY# This Lecture" The Big Data Problem" Hardware for Big Data" Distributing Work" Handling Failures and Slow Machines" Map Reduce and Complex Jobs"
Computing Load Aware and Long-View Load Balancing for Cluster Storage Systems
215 IEEE International Conference on Big Data (Big Data) Computing Load Aware and Long-View Load Balancing for Cluster Storage Systems Guoxin Liu and Haiying Shen and Haoyu Wang Department of Electrical
Network-Aware Scheduling of MapReduce Framework on Distributed Clusters over High Speed Networks
Network-Aware Scheduling of MapReduce Framework on Distributed Clusters over High Speed Networks Praveenkumar Kondikoppa, Chui-Hui Chiu, Cheng Cui, Lin Xue and Seung-Jong Park Department of Computer Science,
Job Scheduling with the Fair and Capacity Schedulers
Job Scheduling with the Fair and Capacity Schedulers Matei Zaharia UC Berkeley Wednesday, June 10, 2009 Santa Clara Marriott Motivation» Provide fast response times to small jobs in a shared Hadoop cluster»
Survey on Job Schedulers in Hadoop Cluster
IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661, p- ISSN: 2278-8727Volume 15, Issue 1 (Sep. - Oct. 2013), PP 46-50 Bincy P Andrews 1, Binu A 2 1 (Rajagiri School of Engineering and Technology,
Scheduling Algorithms in MapReduce Distributed Mind
Scheduling Algorithms in MapReduce Distributed Mind Karthik Kotian, Jason A Smith, Ye Zhang Schedule Overview of topic (review) Hypothesis Research paper 1 Research paper 2 Research paper 3 Project software
Deadline-based Workload Management for MapReduce Environments: Pieces of the Performance Puzzle
Deadline-based Workload Management for MapReduce Environments: Pieces of the Performance Puzzle Abhishek Verma University of Illinois at Urbana-Champaign, IL, US. Ludmila Cherkasova Hewlett-Packard Labs
Petabyte Scale Data at Facebook. Dhruba Borthakur, Engineer at Facebook, SIGMOD, New York, June 2013
Petabyte Scale Data at Facebook Dhruba Borthakur, Engineer at Facebook, SIGMOD, New York, June 2013 Agenda 1 Types of Data 2 Data Model and API for Facebook Graph Data 3 SLTP (Semi-OLTP) and Analytics
Large scale processing using Hadoop. Ján Vaňo
Large scale processing using Hadoop Ján Vaňo What is Hadoop? Software platform that lets one easily write and run applications that process vast amounts of data Includes: MapReduce offline computing engine
Petabyte Scale Data at Facebook. Dhruba Borthakur, Engineer at Facebook, UC Berkeley, Nov 2012
Petabyte Scale Data at Facebook Dhruba Borthakur, Engineer at Facebook, UC Berkeley, Nov 2012 Agenda 1 Types of Data 2 Data Model and API for Facebook Graph Data 3 SLTP (Semi-OLTP) and Analytics data 4
How To Make A Cluster Of Workable, Efficient, And Efficient With A Distributed Scheduler
: A Platform for Fine-Grained Resource Sharing in the Data Center Benjamin Hindman, Andy Konwinski, Matei Zaharia, Ali Ghodsi, Anthony D. Joseph, Randy Katz, Scott Shenker, Ion Stoica University of California,
A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM
A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM Sneha D.Borkar 1, Prof.Chaitali S.Surtakar 2 Student of B.E., Information Technology, J.D.I.E.T, [email protected] Assistant Professor, Information
Extending Hadoop beyond MapReduce
Extending Hadoop beyond MapReduce Mahadev Konar Co-Founder @mahadevkonar (@hortonworks) Page 1 Bio Apache Hadoop since 2006 - committer and PMC member Developed and supported Map Reduce @Yahoo! - Core
Evaluating Task Scheduling in Hadoop-based Cloud Systems
2013 IEEE International Conference on Big Data Evaluating Task Scheduling in Hadoop-based Cloud Systems Shengyuan Liu, Jungang Xu College of Computer and Control Engineering University of Chinese Academy
BBM467 Data Intensive ApplicaAons
Hace7epe Üniversitesi Bilgisayar Mühendisliği Bölümü BBM467 Data Intensive ApplicaAons Dr. Fuat Akal [email protected] Problem How do you scale up applicaaons? Run jobs processing 100 s of terabytes
Hadoop & its Usage at Facebook
Hadoop & its Usage at Facebook Dhruba Borthakur Project Lead, Hadoop Distributed File System [email protected] Presented at the The Israeli Association of Grid Technologies July 15, 2009 Outline Architecture
Research on Job Scheduling Algorithm in Hadoop
Journal of Computational Information Systems 7: 6 () 5769-5775 Available at http://www.jofcis.com Research on Job Scheduling Algorithm in Hadoop Yang XIA, Lei WANG, Qiang ZHAO, Gongxuan ZHANG School of
Hadoop Distributed File System. Dhruba Borthakur Apache Hadoop Project Management Committee [email protected] [email protected]
Hadoop Distributed File System Dhruba Borthakur Apache Hadoop Project Management Committee [email protected] [email protected] Hadoop, Why? Need to process huge datasets on large clusters of computers
Take An Internal Look at Hadoop. Hairong Kuang Grid Team, Yahoo! Inc [email protected]
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc [email protected] What s Hadoop Framework for running applications on large clusters of commodity hardware Scale: petabytes of data
Data Warehousing and Analytics Infrastructure at Facebook. Ashish Thusoo & Dhruba Borthakur athusoo,[email protected]
Data Warehousing and Analytics Infrastructure at Facebook Ashish Thusoo & Dhruba Borthakur athusoo,[email protected] Overview Challenges in a Fast Growing & Dynamic Environment Data Flow Architecture,
Hadoop & its Usage at Facebook
Hadoop & its Usage at Facebook Dhruba Borthakur Project Lead, Hadoop Distributed File System [email protected] Presented at the Storage Developer Conference, Santa Clara September 15, 2009 Outline Introduction
CSE-E5430 Scalable Cloud Computing Lecture 2
CSE-E5430 Scalable Cloud Computing Lecture 2 Keijo Heljanko Department of Computer Science School of Science Aalto University [email protected] 14.9-2015 1/36 Google MapReduce A scalable batch processing
Application Development. A Paradigm Shift
Application Development for the Cloud: A Paradigm Shift Ramesh Rangachar Intelsat t 2012 by Intelsat. t Published by The Aerospace Corporation with permission. New 2007 Template - 1 Motivation for the
Heterogeneity-Aware Resource Allocation and Scheduling in the Cloud
Heterogeneity-Aware Resource Allocation and Scheduling in the Cloud Gunho Lee, Byung-Gon Chun, Randy H. Katz University of California, Berkeley, Yahoo! Research Abstract Data analytics are key applications
GraySort and MinuteSort at Yahoo on Hadoop 0.23
GraySort and at Yahoo on Hadoop.23 Thomas Graves Yahoo! May, 213 The Apache Hadoop[1] software library is an open source framework that allows for the distributed processing of large data sets across clusters
Data Management in the Cloud
Data Management in the Cloud Ryan Stern [email protected] : Advanced Topics in Distributed Systems Department of Computer Science Colorado State University Outline Today Microsoft Cloud SQL Server
Adaptive Task Scheduling for Multi Job MapReduce
Adaptive Task Scheduling for MultiJob MapReduce Environments Jordà Polo, David de Nadal, David Carrera, Yolanda Becerra, Vicenç Beltran, Jordi Torres and Eduard Ayguadé Barcelona Supercomputing Center
Towards an Optimized Big Data Processing System
Towards an Optimized Big Data Processing System The Doctoral Symposium of the IEEE/ACM CCGrid 2013 Delft, The Netherlands Bogdan Ghiţ, Alexandru Iosup, and Dick Epema Parallel and Distributed Systems Group
Snapshots in Hadoop Distributed File System
Snapshots in Hadoop Distributed File System Sameer Agarwal UC Berkeley Dhruba Borthakur Facebook Inc. Ion Stoica UC Berkeley Abstract The ability to take snapshots is an essential functionality of any
Introduction to Apache YARN Schedulers & Queues
Introduction to Apache YARN Schedulers & Queues In a nutshell, YARN was designed to address the many limitations (performance/scalability) embedded into Hadoop version 1 (MapReduce & HDFS). Some of the
Hadoop: A Framework for Data- Intensive Distributed Computing. CS561-Spring 2012 WPI, Mohamed Y. Eltabakh
1 Hadoop: A Framework for Data- Intensive Distributed Computing CS561-Spring 2012 WPI, Mohamed Y. Eltabakh 2 What is Hadoop? Hadoop is a software framework for distributed processing of large datasets
Dominant Resource Fairness: Fair Allocation of Multiple Resource Types
Dominant Resource Fairness: Fair Allocation of Multiple Resource Types Ali Ghodsi Matei Zaharia Benjamin Hindman Andrew Konwinski Scott Shenker Ion Stoica Electrical Engineering and Computer Sciences University
Cloud Management: Knowing is Half The Battle
Cloud Management: Knowing is Half The Battle Raouf BOUTABA David R. Cheriton School of Computer Science University of Waterloo Joint work with Qi Zhang, Faten Zhani (University of Waterloo) and Joseph
Hadoop and Map-Reduce. Swati Gore
Hadoop and Map-Reduce Swati Gore Contents Why Hadoop? Hadoop Overview Hadoop Architecture Working Description Fault Tolerance Limitations Why Map-Reduce not MPI Distributed sort Why Hadoop? Existing Data
Efficient Data Replication Scheme based on Hadoop Distributed File System
, pp. 177-186 http://dx.doi.org/10.14257/ijseia.2015.9.12.16 Efficient Data Replication Scheme based on Hadoop Distributed File System Jungha Lee 1, Jaehwa Chung 2 and Daewon Lee 3* 1 Division of Supercomputing,
CloudRank-D:A Benchmark Suite for Private Cloud Systems
CloudRank-D:A Benchmark Suite for Private Cloud Systems Jing Quan Institute of Computing Technology, Chinese Academy of Sciences and University of Science and Technology of China HVC tutorial in conjunction
Tachyon: Reliable File Sharing at Memory- Speed Across Cluster Frameworks
Tachyon: Reliable File Sharing at Memory- Speed Across Cluster Frameworks Haoyuan Li UC Berkeley Outline Motivation System Design Evaluation Results Release Status Future Directions Outline Motivation
MapReduce Online. Tyson Condie, Neil Conway, Peter Alvaro, Joseph Hellerstein, Khaled Elmeleegy, Russell Sears. Neeraj Ganapathy
MapReduce Online Tyson Condie, Neil Conway, Peter Alvaro, Joseph Hellerstein, Khaled Elmeleegy, Russell Sears Neeraj Ganapathy Outline Hadoop Architecture Pipelined MapReduce Online Aggregation Continuous
Task Scheduling in Hadoop
Task Scheduling in Hadoop Sagar Mamdapure Munira Ginwala Neha Papat SAE,Kondhwa SAE,Kondhwa SAE,Kondhwa Abstract Hadoop is widely used for storing large datasets and processing them efficiently under distributed
Hadoop implementation of MapReduce computational model. Ján Vaňo
Hadoop implementation of MapReduce computational model Ján Vaňo What is MapReduce? A computational model published in a paper by Google in 2004 Based on distributed computation Complements Google s distributed
YARN Apache Hadoop Next Generation Compute Platform
YARN Apache Hadoop Next Generation Compute Platform Bikas Saha @bikassaha Hortonworks Inc. 2013 Page 1 Apache Hadoop & YARN Apache Hadoop De facto Big Data open source platform Running for about 5 years
Hadoop and its Usage at Facebook. Dhruba Borthakur [email protected], June 22 rd, 2009
Hadoop and its Usage at Facebook Dhruba Borthakur [email protected], June 22 rd, 2009 Who Am I? Hadoop Developer Core contributor since Hadoop s infancy Focussed on Hadoop Distributed File System Facebook
Hadoop Parallel Data Processing
MapReduce and Implementation Hadoop Parallel Data Processing Kai Shen A programming interface (two stage Map and Reduce) and system support such that: the interface is easy to program, and suitable for
Cloud Computing using MapReduce, Hadoop, Spark
Cloud Computing using MapReduce, Hadoop, Spark Benjamin Hindman [email protected] Why this talk? At some point, you ll have enough data to run your parallel algorithms on multiple computers SPMD (e.g.,
Open source Google-style large scale data analysis with Hadoop
Open source Google-style large scale data analysis with Hadoop Ioannis Konstantinou Email: [email protected] Web: http://www.cslab.ntua.gr/~ikons Computing Systems Laboratory School of Electrical
CSE-E5430 Scalable Cloud Computing Lecture 11
CSE-E5430 Scalable Cloud Computing Lecture 11 Keijo Heljanko Department of Computer Science School of Science Aalto University [email protected] 30.11-2015 1/24 Distributed Coordination Systems Consensus
Spark. Fast, Interactive, Language- Integrated Cluster Computing
Spark Fast, Interactive, Language- Integrated Cluster Computing Matei Zaharia, Mosharaf Chowdhury, Tathagata Das, Ankur Dave, Justin Ma, Murphy McCauley, Michael Franklin, Scott Shenker, Ion Stoica UC
Hadoop Fair Scheduler Design Document
Hadoop Fair Scheduler Design Document October 18, 2010 Contents 1 Introduction 2 2 Fair Scheduler Goals 2 3 Scheduler Features 2 3.1 Pools........................................ 2 3.2 Minimum Shares.................................
Scalable Cloud Computing Solutions for Next Generation Sequencing Data
Scalable Cloud Computing Solutions for Next Generation Sequencing Data Matti Niemenmaa 1, Aleksi Kallio 2, André Schumacher 1, Petri Klemelä 2, Eija Korpelainen 2, and Keijo Heljanko 1 1 Department of
Design and Evolution of the Apache Hadoop File System(HDFS)
Design and Evolution of the Apache Hadoop File System(HDFS) Dhruba Borthakur Engineer@Facebook Committer@Apache HDFS SDC, Sept 19 2011 Outline Introduction Yet another file-system, why? Goals of Hadoop
Shareability and Locality Aware Scheduling Algorithm in Hadoop for Mobile Cloud Computing
Shareability and Locality Aware Scheduling Algorithm in Hadoop for Mobile Cloud Computing Hsin-Wen Wei 1,2, Che-Wei Hsu 2, Tin-Yu Wu 3, Wei-Tsong Lee 1 1 Department of Electrical Engineering, Tamkang University
Benchmarking Hadoop & HBase on Violin
Technical White Paper Report Technical Report Benchmarking Hadoop & HBase on Violin Harnessing Big Data Analytics at the Speed of Memory Version 1.0 Abstract The purpose of benchmarking is to show advantages
Hadoop Distributed File System. Dhruba Borthakur Apache Hadoop Project Management Committee [email protected] June 3 rd, 2008
Hadoop Distributed File System Dhruba Borthakur Apache Hadoop Project Management Committee [email protected] June 3 rd, 2008 Who Am I? Hadoop Developer Core contributor since Hadoop s infancy Focussed
MAPREDUCE [1] is proposed by Google in 2004 and
IEEE TRANSACTIONS ON COMPUTERS 1 Improving MapReduce Performance Using Smart Speculative Execution Strategy Qi Chen, Cheng Liu, and Zhen Xiao, Senior Member, IEEE Abstract MapReduce is a widely used parallel
Cura: A Cost-optimized Model for MapReduce in a Cloud
: A -optimized Model for MapReduce in a Cloud Balaji Palanisamy Aameek Singh Ling Liu Bryan Langston College of Computing, Georgia Tech IBM Research - Almaden {balaji, lingliu}@cc.gatech.edu, {aameek.singh,
CapacityScheduler Guide
Table of contents 1 Purpose... 2 2 Overview... 2 3 Features...2 4 Installation... 3 5 Configuration...4 5.1 Using the CapacityScheduler...4 5.2 Setting up queues...4 5.3 Queue properties... 4 5.4 Resource
From GWS to MapReduce: Google s Cloud Technology in the Early Days
Large-Scale Distributed Systems From GWS to MapReduce: Google s Cloud Technology in the Early Days Part II: MapReduce in a Datacenter COMP6511A Spring 2014 HKUST Lin Gu [email protected] MapReduce/Hadoop
DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters
IEEE TRANSACTIONS ON CLOUD COMPUTING 1 DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Shanjiang Tang, Bu-Sung Lee, Bingsheng He Abstract MapReduce is a popular computing
Nexus: A Common Substrate for Cluster Computing
Nexus: A Common Substrate for Cluster Computing Benjamin Hindman, Andy Konwinski, Matei Zaharia, Ali Ghodsi, Anthony D. Joseph, Scott Shenker, Ion Stoica University of California, Berkeley {benh,andyk,matei,alig,adj,shenker,stoica}@cs.berkeley.edu
Quantcast Petabyte Storage at Half Price with QFS!
9-131 Quantcast Petabyte Storage at Half Price with QFS Presented by Silvius Rus, Director, Big Data Platforms September 2013 Quantcast File System (QFS) A high performance alternative to the Hadoop Distributed
COSHH: A Classification and Optimization based Scheduler for Heterogeneous Hadoop Systems
COSHH: A Classification and Optimization based Scheduler for Heterogeneous Hadoop Systems Aysan Rasooli a, Douglas G. Down a a Department of Computing and Software, McMaster University, L8S 4K1, Canada
Hadoop: Distributed Data Processing. Amr Awadallah Founder/CTO, Cloudera, Inc. ACM Data Mining SIG Thursday, January 25 th, 2010
Hadoop: Distributed Data Processing Amr Awadallah Founder/CTO, Cloudera, Inc. ACM Data Mining SIG Thursday, January 25 th, 2010 Outline Scaling for Large Data Processing What is Hadoop? HDFS and MapReduce
DATA MINING WITH HADOOP AND HIVE Introduction to Architecture
DATA MINING WITH HADOOP AND HIVE Introduction to Architecture Dr. Wlodek Zadrozny (Most slides come from Prof. Akella s class in 2014) 2015-2025. Reproduction or usage prohibited without permission of
NoSQL and Hadoop Technologies On Oracle Cloud
NoSQL and Hadoop Technologies On Oracle Cloud Vatika Sharma 1, Meenu Dave 2 1 M.Tech. Scholar, Department of CSE, Jagan Nath University, Jaipur, India 2 Assistant Professor, Department of CSE, Jagan Nath
Sparrow: Scalable Scheduling for Sub-Second Parallel Jobs
Sparrow: Scalable Scheduling for Sub-Second Parallel Jobs Kay Ousterhout Patrick Wendell Matei Zaharia Ion Stoica Electrical Engineering and Computer Sciences University of California at Berkeley Technical
Optimizing Shared Resource Contention in HPC Clusters
Optimizing Shared Resource Contention in HPC Clusters Sergey Blagodurov Simon Fraser University Alexandra Fedorova Simon Fraser University Abstract Contention for shared resources in HPC clusters occurs
Capacity Planning Process Estimating the load Initial configuration
Capacity Planning Any data warehouse solution will grow over time, sometimes quite dramatically. It is essential that the components of the solution (hardware, software, and database) are capable of supporting
Four Orders of Magnitude: Running Large Scale Accumulo Clusters. Aaron Cordova Accumulo Summit, June 2014
Four Orders of Magnitude: Running Large Scale Accumulo Clusters Aaron Cordova Accumulo Summit, June 2014 Scale, Security, Schema Scale to scale 1 - (vt) to change the size of something let s scale the
FP-Hadoop: Efficient Execution of Parallel Jobs Over Skewed Data
FP-Hadoop: Efficient Execution of Parallel Jobs Over Skewed Data Miguel Liroz-Gistau, Reza Akbarinia, Patrick Valduriez To cite this version: Miguel Liroz-Gistau, Reza Akbarinia, Patrick Valduriez. FP-Hadoop:
