Data Warehousing and Analytics Infrastructure at Facebook. Ashish Thusoo & Dhruba Borthakur athusoo,dhruba@facebook.com



Similar documents
Facebook s Petabyte Scale Data Warehouse using Hive and Hadoop

America s Most Wanted a metric to detect persistently faulty machines in Hadoop

Apache Hadoop FileSystem and its Usage in Facebook

Hadoop & its Usage at Facebook

Hadoop and its Usage at Facebook. Dhruba Borthakur June 22 rd, 2009

Hadoop Architecture and its Usage at Facebook

Hadoop & its Usage at Facebook

Petabyte Scale Data at Facebook. Dhruba Borthakur, Engineer at Facebook, SIGMOD, New York, June 2013

Hadoop and Hive Development at Facebook. Dhruba Borthakur Zheng Shao {dhruba, Presented at Hadoop World, New York October 2, 2009

Hadoop Distributed File System. Dhruba Borthakur Apache Hadoop Project Management Committee

Design and Evolution of the Apache Hadoop File System(HDFS)

An Industrial Perspective on the Hadoop Ecosystem. Eldar Khalilov Pavel Valov

CSE-E5430 Scalable Cloud Computing Lecture 2

Apache Hadoop FileSystem Internals

Petabyte Scale Data at Facebook. Dhruba Borthakur, Engineer at Facebook, UC Berkeley, Nov 2012

Petabyte Scale Data at Facebook. Dhruba Borthakur, Engineer at Facebook, XLDB Conference at Stanford University, Sept 2012

Data Warehousing and Analytics Infrastructure at Facebook

HIVE. Data Warehousing & Analytics on Hadoop. Joydeep Sen Sarma, Ashish Thusoo Facebook Data Team

Big Data With Hadoop

Realtime Apache Hadoop at Facebook. Jonathan Gray & Dhruba Borthakur June 14, 2011 at SIGMOD, Athens

So What s the Big Deal?

BookKeeper. Flavio Junqueira Yahoo! Research, Barcelona. Hadoop in China 2011

Open source software framework designed for storage and processing of large scale data on clusters of commodity hardware

Distributed File System. MCSN N. Tonellotto Complements of Distributed Enabling Platforms

Open source Google-style large scale data analysis with Hadoop

Big Data. Facebook Wall Data using Graph API. Presented by: Prashant Patel Jaykrushna Patel

Hadoop: Embracing future hardware

Welcome to the unit of Hadoop Fundamentals on Hadoop architecture. I will begin with a terminology review and then cover the major components

DATA MINING WITH HADOOP AND HIVE Introduction to Architecture

Managing Big Data with Hadoop & Vertica. A look at integration between the Cloudera distribution for Hadoop and the Vertica Analytic Database

Hadoop Distributed File System. Dhruba Borthakur June, 2007

Take An Internal Look at Hadoop. Hairong Kuang Grid Team, Yahoo! Inc

Journal of science STUDY ON REPLICA MANAGEMENT AND HIGH AVAILABILITY IN HADOOP DISTRIBUTED FILE SYSTEM (HDFS)

Hadoop: A Framework for Data- Intensive Distributed Computing. CS561-Spring 2012 WPI, Mohamed Y. Eltabakh

Data Warehouse Overview. Namit Jain

Big Data Analytics Nokia

Alternatives to HIVE SQL in Hadoop File Structure

Real-time Analytics at Facebook: Data Freeway and Puma. Zheng Shao 12/2/2011

An Oracle White Paper June High Performance Connectors for Load and Access of Data from Hadoop to Oracle Database

Hadoop Distributed File System. Dhruba Borthakur Apache Hadoop Project Management Committee June 3 rd, 2008

Architectural patterns for building real time applications with Apache HBase. Andrew Purtell Committer and PMC, Apache HBase

Chase Wu New Jersey Ins0tute of Technology

BIG DATA CAN DRIVE THE BUSINESS AND IT TO EVOLVE AND ADAPT RALPH KIMBALL BUSSUM 2014

Hadoop implementation of MapReduce computational model. Ján Vaňo

Overview. Big Data in Apache Hadoop. - HDFS - MapReduce in Hadoop - YARN. Big Data Management and Analytics

<Insert Picture Here> Oracle and/or Hadoop And what you need to know

Well packaged sets of preinstalled, integrated, and optimized software on select hardware in the form of engineered systems and appliances

How To Scale Out Of A Nosql Database

Data-Intensive Programming. Timo Aaltonen Department of Pervasive Computing

Oracle Big Data SQL Technical Update

2009 Oracle Corporation 1

Using distributed technologies to analyze Big Data

Luncheon Webinar Series May 13, 2013

Lecture 5: GFS & HDFS! Claudia Hauff (Web Information Systems)! ti2736b-ewi@tudelft.nl

Hadoop Architecture. Part 1

the missing log collector Treasure Data, Inc. Muga Nishizawa

RCFile: A Fast and Space-efficient Data Placement Structure in MapReduce-based Warehouse Systems CLOUD COMPUTING GROUP - LITAO DENG

A very short Intro to Hadoop

Hadoop MapReduce and Spark. Giorgio Pedrazzi, CINECA-SCAI School of Data Analytics and Visualisation Milan, 10/06/2015

Chapter 7. Using Hadoop Cluster and MapReduce

BIG DATA TRENDS AND TECHNOLOGIES

Performance Comparison of SQL based Big Data Analytics with Lustre and HDFS file systems

Jeffrey D. Ullman slides. MapReduce for data intensive computing

CitusDB Architecture for Real-Time Big Data

Networking in the Hadoop Cluster

BigData. An Overview of Several Approaches. David Mera 16/12/2013. Masaryk University Brno, Czech Republic

Open source large scale distributed data management with Google s MapReduce and Bigtable

CDH AND BUSINESS CONTINUITY:

Hadoop Scalability at Facebook. Dmytro Molkov YaC, Moscow, September 19, 2011

Einsatzfelder von IBM PureData Systems und Ihre Vorteile.

Scalable Cloud Computing Solutions for Next Generation Sequencing Data

INTRODUCTION TO APACHE HADOOP MATTHIAS BRÄGER CERN GS-ASE

Apache Hadoop: Past, Present, and Future

MySQL and Hadoop: Big Data Integration. Shubhangi Garg & Neha Kumari MySQL Engineering

CSE 590: Special Topics Course ( Supercomputing ) Lecture 10 ( MapReduce& Hadoop)

Big Data Primer. 1 Why Big Data? Alex Sverdlov alex@theparticle.com

IBM Netezza High Capacity Appliance

Big Data and Market Surveillance. April 28, 2014

Hadoop at Yahoo! Owen O Malley Yahoo!, Grid Team owen@yahoo-inc.com

Hadoop Distributed File System. Jordan Prosch, Matt Kipps

Hadoop IST 734 SS CHUNG

BIG DATA What it is and how to use?

Tips and Tricks for Using Oracle TimesTen In-Memory Database in the Application Tier

<Insert Picture Here> Big Data

BIG DATA TECHNOLOGY. Hadoop Ecosystem

Data-Intensive Computing with Map-Reduce and Hadoop

Parquet. Columnar storage for the people

Oracle s Big Data solutions. Roger Wullschleger. <Insert Picture Here>

Large scale processing using Hadoop. Ján Vaňo

Hadoop Distributed File System. T Seminar On Multimedia Eero Kurkela

Hadoop & Spark Using Amazon EMR

The Hadoop Distributed File System

THE ATLAS DISTRIBUTED DATA MANAGEMENT SYSTEM & DATABASES

BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB

How Cisco IT Built Big Data Platform to Transform Data Management

MapReduce and Hadoop. Aaron Birkland Cornell Center for Advanced Computing. January 2012

Big Data Analytics with MapReduce VL Implementierung von Datenbanksystemen 05-Feb-13

Inge Os Sales Consulting Manager Oracle Norway

Transcription:

Data Warehousing and Analytics Infrastructure at Facebook Ashish Thusoo & Dhruba Borthakur athusoo,dhruba@facebook.com

Overview Challenges in a Fast Growing & Dynamic Environment Data Flow Architecture, Applications & Middleware Backend Infrastructure Challenges & Usage Statistics Conclusion

Key Challenge (I) Data, data and more data Recent Growth of Data for Analysis at Facebook (TB of data)

Key Challenge (II) Accessibility Simplify data access and use prevalent and familiar tools Make data available as soon as possible

Key Challenge (III) Flexibility Support for different data formats Support for user defined computations and extensions Ability to evolve with data

Technologies Hadoop Scalable Storage and Compute Grid Hive, cohive & HiPal Extensible & Flexible Data Warehousing Framework the provides SQL like Query Language on Hadoop Scribe-HDFS Scalable Log Collection Infrastructure Nectar Flexible Instrumentation Framework Databee & Chronos Dependency based Scheduling Infrastructure

Data Flow Architecture at Facebook Web Servers Scribe- Hadoop Cluster Hive replicadon SILVER Hive- Hadoop Cluster Oracle RAC PLATINUM Hive- Hadoop Cluster Federated MySQL

Usage Types of Applications: Reporting Eg: Daily/Weekly aggregations of impression/click counts Measures of user engagement Microstrategy reports Ad hoc Analysis Eg: how many group admins broken down by state/country Machine Learning (Assembling training data) Ad Optimization Eg: User Engagement as a function of user attributes Index Generation etc.

Analysis and Data Organization > 99% of analysis through Hive on Hadoop Hive: Easy to use: Familiar SQL interface with Data as Tables and Columns) Easy to extend: Can embed map/reduce user programs in the data flow Support for user defined functions Flexible: Supports user defined data formats and storage formats Support user defined types Interoperable: JDBC, ODBC and thrift interfaces for integration with BI tools

Application Instrumentation (10,000 feet) 1. Setup new scribe category Class MyEvent : NectarAppEvent 2. Copiers to copy data into warehouse 3. Automatic creation of tables in Hive _current and _r tables 4. Automatic detection of schema changes MyEvent->log() 5. Near real time event statistics dashboard Nectar Framework Services

Data Discovery (10,000 feet) 1. Tag based search on Table Metadata 2. Crowd sourcing to generate tag information Data Discovery 3. Ability to ask questions from expert users that are identified by analyzing usage logs 4. Lineage information of data shown for browsing and data discovery cohive Services in HiPal

Batch Jobs (10,000 feet) Builtin Operators for Common Tasks Chaining Transformations Specification for DAG of Transforms Smart retries and checkpointing on failures Chronos Scheduler Monitoring, Alerting & Statistics Data Availability Statistics and Alerts Resource Utilization History DataBee & Chronos Services

Hadoop & Hive Cluster @ Facebook Hadoop/Hive cluster 16000 cores Raw Storage capacity ~ 21PB 8 cores + 12 TB per node 32 GB RAM per node Two level network topology 1 Gbit/sec from node to rack switch 10 Gbit/sec to top level rack switch 2 clusters One for adhoc users (SILVER cluster) One for strict SLA jobs (PLATINUM cluster)

Hive & Hadoop Usage @ Facebook Statistics per day: 800TB of I/O per day 10K 25K Hadoop jobs per day Hive simplifies Hadoop: New engineers go though a Hive training session Analysts (non-engineers) use Hadoop through Hive Most of jobs are Hive Jobs

Data Collection and Real time Availability Challenge: Large Volume of data to be queryable as soon as it is produced: Logs produce 100 TB per day uncompressed Store in Scribe-HDFS clusters co-located with web-tier Query-able within 15 seconds (via tail f <filename>) Overlapping HDFS clusters to avoid Single Point of Failure Two separate NameNodes Two DataNode instances on same physical machine Gzip compressed, uploaded into central warehouse Query-able within 15 minutes (via Hive external tables)

Scribe-HDFS: Near Real Time Log Collection Scribe An Open Source Technology created at Facebook for Log Collection Routes log messages from source (web machines) to destination storage Persistently buffers data on intermediate nodes to deal with failures Scribe-HDFS A highly scalable routing service + a highly scalable storage service Enables us to scale with commodity nodes no more expensive filers

Scribe-HDFS: 101 Scribed HDFS Data Node Scribed <category, msgs> Scribed Append to /staging/<category>/<file> HDFS Data Node Scribed Scribed HDFS Data Node Scribe-HDFS

Warehouse is Business Critical Challenge: Remove all single points of failure (SPOF) Hadoop NameNode was a SPOF Hadoop AvatarNode Active-passive Hot Standby pair of NameNodes Failover time for 20 PB file system having 65 million files is 10 seconds Work-in progress to support active-active AvatarNode

Warehouse Utilization and Workload Compute Map-Reduce cluster is CPU bound Peak usage of 95% CPU utilization Peak network usage is 70% (network is not a bottleneck) 70% tasks find data on local rack We do not compress map outputs Storage HDFS cluster is capacity bound 75% storage full (current size is 21 PB, 2000 nodes) Disk bandwidth used is 20% of capacity (IO is not a bottleneck) All Hive tables compressed using gzip

Hive: Optimizing Resource Utilization Joins: Joins try to reduce the number of map/reduce jobs needed. Memory efficient joins by streaming largest tables. Map Joins User specified small tables stored in hash tables on the mapper No reducer needed Aggregations: Map side partial aggregations Hash-based aggregates Serialized key/values in hash tables 90% speed improvement on Query SELECT count(1) FROM t; Load balancing for data skew

Storage Costs Challenge: disk is cheap, but the cost of PetaBytes of disk is huge! HDFS-RAID Default policy for HDFS is to replicate data 3 times Generate parity block from a stripe of blocks from a file HDFS-RAID replicates data 2 times and keeps parity blocks Saves around 25% disk space Columnar Compression Hive RCFile format organizes content by column Saves around 15-25% of disk space

Single Warehouse for Adhoc and Periodic jobs Challenge: support adhoc and periodic jobs in a single instance of the warehouse New job arrives every 10 seconds 90% jobs are small and finish within a few minutes Users run lots of experiments to mine data Batch processing for large hourly/daily/weekly reporting 99% uptime Hadoop Map-Reduce FairShare Scheduler Each user gets an equal share of the cluster Optimize latency for small jobs Optimize cluster utilization and fair-share for larger jobs Resource aware scheduling (CPU and memory resources only)

Isolation for Business Critical Pipelines Challenge: 100% uptime for a select set of pipeline Separate a small storage & compute cluster PLATINUM cluster Small subset of data from production warehouse No adhoc queries High bar for graduating a pipeline from production warehouse to PLATINUM warehouse Poor man s Disaster Recovery Solution (DR) 99.9% uptime SLA Hive replication Scribe copies data into PLATINUM Copier looks at Hive query logs to replicate data from PLATINUM to SILVER cluster

Conclusion Scalable, Accessible & Flexible infrastructure a MUST Use & Contribute to Open Source to keep the costs low Hadoop Hive Scribe-HDFS Always a challenge to keep up with a dynamically growing & changing environment