THE ATLAS DISTRIBUTED DATA MANAGEMENT SYSTEM & DATABASES

Similar documents
ATLAS Data Management Accounting with Hadoop Pig and HBase

How To Scale Out Of A Nosql Database

Hadoop IST 734 SS CHUNG

Oracle s Big Data solutions. Roger Wullschleger. <Insert Picture Here>

A Performance Analysis of Distributed Indexing using Terrier

CSE-E5430 Scalable Cloud Computing Lecture 2

Introduction to Hadoop. New York Oracle User Group Vikas Sawhney

Chapter 11 Map-Reduce, Hadoop, HDFS, Hbase, MongoDB, Apache HIVE, and Related

Scaling Up 2 CSE 6242 / CX Duen Horng (Polo) Chau Georgia Tech. HBase, Hive

THE HADOOP DISTRIBUTED FILE SYSTEM

HBase A Comprehensive Introduction. James Chin, Zikai Wang Monday, March 14, 2011 CS 227 (Topics in Database Management) CIT 367

Hadoop & its Usage at Facebook

Apache Hadoop. Alexandru Costan

Oracle Big Data SQL Technical Update

Performance and Scalability Overview

Performance Comparison of SQL based Big Data Analytics with Lustre and HDFS file systems

Hadoop Distributed File System. Dhruba Borthakur Apache Hadoop Project Management Committee

Data-Intensive Programming. Timo Aaltonen Department of Pervasive Computing

Introduction to Hadoop HDFS and Ecosystems. Slides credits: Cloudera Academic Partners Program & Prof. De Liu, MSBA 6330 Harvesting Big Data

Hadoop. Apache Hadoop is an open-source software framework for storage and large scale processing of data-sets on clusters of commodity hardware.

Hadoop and Map-Reduce. Swati Gore

BIG DATA TRENDS AND TECHNOLOGIES

Introduction to Apache Cassandra

How To Use Big Data For Telco (For A Telco)

MapReduce with Apache Hadoop Analysing Big Data

Big Data Analytics - Accelerated. stream-horizon.com

Lecture 10: HBase! Claudia Hauff (Web Information Systems)!

Building Scalable Big Data Infrastructure Using Open Source Software. Sam William

Hadoop. Sunday, November 25, 12

Hadoop 只 支 援 用 Java 開 發 嘛? Is Hadoop only support Java? 總 不 能 全 部 都 重 新 設 計 吧? 如 何 與 舊 系 統 相 容? Can Hadoop work with existing software?

An Approach to Implement Map Reduce with NoSQL Databases

Prepared By : Manoj Kumar Joshi & Vikas Sawhney

DATA MINING WITH HADOOP AND HIVE Introduction to Architecture

NoSQL Data Base Basics

Cloudera Certified Developer for Apache Hadoop

Department of Computer Science University of Cyprus EPL646 Advanced Topics in Databases. Lecture 14

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software

A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM

Cost-Effective Business Intelligence with Red Hat and Open Source

Take An Internal Look at Hadoop. Hairong Kuang Grid Team, Yahoo! Inc

INTRODUCTION TO APACHE HADOOP MATTHIAS BRÄGER CERN GS-ASE

Data Warehousing and Analytics Infrastructure at Facebook. Ashish Thusoo & Dhruba Borthakur athusoo,dhruba@facebook.com

Infomatics. Big-Data and Hadoop Developer Training with Oracle WDP

Comparison of the Frontier Distributed Database Caching System with NoSQL Databases

HDFS Under the Hood. Sanjay Radia. Grid Computing, Hadoop Yahoo Inc.

Introduction to Big data. Why Big data? Case Studies. Introduction to Hadoop. Understanding Features of Hadoop. Hadoop Architecture.

Accelerating and Simplifying Apache

Performance and Scalability Overview

Big Data With Hadoop

Xiaoming Gao Hui Li Thilina Gunarathne

An Oracle White Paper November Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics

Can High-Performance Interconnects Benefit Memcached and Hadoop?

Can the Elephants Handle the NoSQL Onslaught?

Realtime Apache Hadoop at Facebook. Jonathan Gray & Dhruba Borthakur June 14, 2011 at SIGMOD, Athens

Hadoop: A Framework for Data- Intensive Distributed Computing. CS561-Spring 2012 WPI, Mohamed Y. Eltabakh

Big Data? Definition # 1: Big Data Definition Forrester Research

Getting Started with Hadoop. Raanan Dagan Paul Tibaldi

Scalable Architecture on Amazon AWS Cloud

Dominik Wagenknecht Accenture

Complete Java Classes Hadoop Syllabus Contact No:

Big Data and Analytics: A Conceptual Overview. Mike Park Erik Hoel

Hadoop & its Usage at Facebook

Hadoop Ecosystem B Y R A H I M A.

Database Scalability and Oracle 12c

Hadoop & Spark Using Amazon EMR

Implement Hadoop jobs to extract business value from large and varied data sets

HDFS. Hadoop Distributed File System

Big Data Management and Security

BIG DATA What it is and how to use?

BIG DATA CAN DRIVE THE BUSINESS AND IT TO EVOLVE AND ADAPT RALPH KIMBALL BUSSUM 2014

Big Fast Data Hadoop acceleration with Flash. June 2013

Open source large scale distributed data management with Google s MapReduce and Bigtable

Navigating the Big Data infrastructure layer Helena Schwenk

News and trends in Data Warehouse Automation, Big Data and BI. Johan Hendrickx & Dirk Vermeiren

Hypertable Goes Realtime at Baidu. Yang Dong Sherlock Yang(

Using MySQL for Big Data Advantage Integrate for Insight Sastry Vedantam

A Brief Outline on Bigdata Hadoop

MADOCA II Data Logging System Using NoSQL Database for SPring-8

So What s the Big Deal?

Overview. Big Data in Apache Hadoop. - HDFS - MapReduce in Hadoop - YARN. Big Data Management and Analytics

Analytics March 2015 White paper. Why NoSQL? Your database options in the new non-relational world

Real Time Big Data Processing

Hadoop implementation of MapReduce computational model. Ján Vaňo

NoSQL for SQL Professionals William McKnight

Hadoop: Distributed Data Processing. Amr Awadallah Founder/CTO, Cloudera, Inc. ACM Data Mining SIG Thursday, January 25 th, 2010

Managing Big Data with Hadoop & Vertica. A look at integration between the Cloudera distribution for Hadoop and the Vertica Analytic Database

Practical Cassandra. Vitalii

Big Data Explained. An introduction to Big Data Science.

Big Data & QlikView. Democratizing Big Data Analytics. David Freriks Principal Solution Architect

An Industrial Perspective on the Hadoop Ecosystem. Eldar Khalilov Pavel Valov

Hadoop MapReduce over Lustre* High Performance Data Division Omkar Kulkarni April 16, 2013

Scaling Up HBase, Hive, Pegasus

Workshop on Hadoop with Big Data

A survey of big data architectures for handling massive data

BIG DATA TECHNOLOGY. Hadoop Ecosystem

Large scale processing using Hadoop. Ján Vaňo

Using distributed technologies to analyze Big Data

Transcription:

THE ATLAS DISTRIBUTED DATA MANAGEMENT SYSTEM & DATABASES Vincent Garonne, Mario Lassnig, Martin Barisits, Thomas Beermann, Ralph Vigne, Cedric Serfon Vincent.Garonne@cern.ch ph-adp-ddm-lab@cern.ch XLDB 2013 Workshop Europe

Overview 2 Distributed Data Management (DDM) Overview Architecture Relational Database Management Systems (RDBMS) & DDM Use cases & Experiences NoSQL & DDM Use cases & (operational) Experiences Present & Future: Rucio Conclusions

DDM Background 3 The Distributed Data Management project manages ATLAS data on the grid The current system is Don Quijote 2 (DQ2) 144 Petabytes 600k datasets 450 million files 800 active users 130 sites + History Data Taking Total Grid Usage (PB) Long Shutdown 1 (LS1)

DDM Implementation: DQ2 4 DQ2 In production since 2004 2007-2011: Many updates RDBMS Oracle Critical dependency Proven technology Expertise@cern Great for enforcing data integrity Tool of choice for Online transaction processing applications (OLTP)

Oracle & DDM 5 Constant optimization over the years has allowed us to scale avg. 300 Hz, physical write 7 MB/s, physical read 30 MB/s Improvement with I/O, concurrency, front-end deployment, time partitioning, table space management, denormalization for performance, etc. Developing and maintaining an high availability service with Oracle requires some good knowledge and expertise Bind variables, index hint, oracle optimizer, pl/sql, execution plan, etc.

Concepts 6 RDBMS Vertical scalability ( scale up ) Few powerful nodes Shared state Explicit partitioning Resistant hardware ACID Implicit queries (WHAT) NoSQL (Structured storage) Horizontal scalability ( scale out ) Lots of interconnected low cost nodes Shared nothing architecture Implicit partitioning Reliability in software BASE Explicit data pipeline (HOW)

NoSQL & DDM 7 Data warehousing use cases & applications relevant for NoSQL Lot of data No transactions and relaxed consistency Multi dimensional queries Three technologies evaluated: MongoDB, Cassandra & Hadoop Many more available, but these were chosen with the following things in mind Large community available and widely installed In production in several larger companies with respectable data sizes Potential commercial support 12 node cluster located in CERN IT data center to evaluate technologies 96 CPU cores (Intel Xeon, 2.27GHz, 8/node) 288 GB RAM (24/node) 24 TB space (1 TB SATA each, 2/node) 1 GigE network

Data Models Structured Storage :: Technologies :: Data Models 8 7! Explicit row-key! Native datatypes! Everything indexable! Implicit row-keys! Data is byte streams! Column Families group row-keys! Implicit row-key! Data is byte streams! Row-keys group Column Families! Row-keys are sorted

Structured Storage :: Technologies :: Data Bases Architectures 9! Master/Slave! Smart client implements failover! Write-ahead log! Limited MapReduce! interleaved! bound to single thread! Keyed binary storage! Indexes! Table locking! Replica sets! Explicit partitioning! No single point of failure! ring of nodes! forwarding of requests! Write-ahead log! No MapReduce! can use Hadoop! No file storage! Bloom filter! Row locking! Snapshotting! Implicit partitioning! No single point of failure! multiple masters! Write-ahead log! MapReduce! File storage! Data on HDFS! Can be used as a source and sink within Hadoop! Bloom filter! Row locking! HDFS-backed redundancy! Implicit partitioning

Technology Selection Structured Storage :: Technology Selection 10 Installation/ Configuration MongoDB Cassandra Hadoop/HBase Download, unpack, run Download, unpack, configure, run Distribution, Complex config Buffered read 256 250 000/sec 180 000/sec 150 000/sec Random read 256 20 000/sec 20 000/sec 20 000/sec Relaxed write 256 10 000/sec 19 000/sec 9 000/sec Durable Write 256 2 500/sec 9 000/sec 6 000/sec Analytics Limited MapReduce Hadoop MapReduce MapReduce, Pig, Hive Durability support Full Full Full Native API Binary JSON Java Java Generic API None Thrift Thrift, REST

rage Structured Technology: :: Technologies Storage :: :: Technology Selection 11 mework for distributed data processing Installation/ Configuration MongoDB Cassandra Hadoop/HBase Download, unpack, run abase like MongoDB or Cassandra ents Download, unpack, configure, run Distribution, Complex config Buffered read 256 250 000/sec 180 000/sec 150 000/sec Random read 256 20 000/sec 20 000/sec 20 000/sec uted filesystem distributed processing of large data sets uted data base for structured storage Analytics ntend and warehouse w language for parallel execution oordination service Relaxed write 256 10 000/sec 19 000/sec 9 000/sec Durable Write 256 2 500/sec 9 000/sec 6 000/sec Limited MapReduce Hadoop MapReduce MapReduce, Pig, Hive Durability support Full Full Full Native API Binary JSON Java Java Generic API None Thrift Thrift, REST Hadoop is a framework for distributed data processing (not only a database) with many components: HDFS (distributed filesystem), MapReduce (distributed processing of large data sets), HBase (distributed data base for structured storage), Hive(SQL frontend), Pig: dataflow language for parallel execution,

12 Use Cases : Log File Aggregation 12 Use cases :: Log file aggregation! HDFS is mounted as a POSIX filesystem via FUSE HDFS! is mounted Daily copies as a of POSIX all the ATLAS filesystem DDM via log FUSE files are aggregated in a single place Daily! copies 8 months of of all logs the ATLAS accumulated, DDM log already files are using aggregated 3 TB of space in a single on HDFS place! 8 months Python of MapReduce logs accumulated jobs analyse ~ 3 TB of the space log on files HDFS Python! MapReduce Streaming API: jobs read analyse from stdin, the log write files to stdout! Streaming Processing API: the read data from takes stdin, about write to 70 stdout minutes Processing! Average the data IO at takes 70MB/s about 70 minutes Average! Potential IO at 70MB/s for 15% performance increase if re-written in pure Java Potential for 15% performance increase if re-written in pure Java " Better read patterns and reducing temporary network usage n Better read patterns and reducing temporary network usag Apache Apache Apache write log file FUSE HDFS MapReduce Apache

Use Cases : Trace Mining 13 Client interaction with ATLAS DDM generates traces E.g., downloading a dataset/file from a remote site Lots of information (25 attributes), time-based One month of traces uncompressed 80GB, compressed 25GB n Can be mapreduced in under 2 minutes Implemented in HBase as distributed atomic counters Previously developed in Cassandra At various granularities (minutes, hours, days) Size of HBase tables negligible Average rate at 300 insertions/s Migrated from Cassandra within 2 days Almost the same column-based data model Get extra Hadoop benefits for free (mature ecosystem with many tools) The single Cassandra benefit, HA, was implemented in Hadoop

Use cases :: DQ2Share Use Cases : DQ2Share 14! HTTP cache for for dataset/file downloads! Downloads via via ATLAS DDM DDM tools tools HDFS, to HDFS, serves serves via Apache via Apache! Get all the the features of of HDFS HDFS for for free, free, i.e., i.e., one one large large reliable reliable disk pool disk pool

! (come and see the poster) 15 Use Cases : Accounting Use cases :: Accounting LAS 6 data Break contents down usage of ATLAS data contents data! queries Break Historical down usage free-form of ATLAS meta data queries contents ta10*, datatype=esd, location=cern*} {site, nbfiles, bytes} := {project=data10*, datatype=esd, location=cern*}!! Historical free-form meta data queries mmaries s about 8 minutes {site, nbfiles, bytes} := {project=data10*, datatype=esd, location=cern*} Non-relational periodic summaries! Non-relational periodic summaries A full accounting run takes about 8 minutes MapReduce jobs B of output data DFS! A full accounting run takes about 8 minutes Oracle " Pig data pipeline pipeline creates creates MapReduce MapReduce jobs jobs 7 GB of input data, 100 MB of output data " 7 GB of input data, 100 MB of output data Pig publish Apache periodic snapshot HDFS Pig publish Apache

Operational experiences: Hadoop 16 Cloudera distribution Tests and packages the Hadoop ecosystem Straightforward installation via Puppet/YUM But the configuration was... not so obvious with many parameters, extensive documentation, but bad default performance Disk failure is common and cannot be ignored Data centre annual disk replacement rate up to 13% (Google & CMU, 2011) Within one year we had: n 5 disk failures (20% failure rate!) Out of which 3 happened at the same time n 1 Mainboard failure together with the disk failure, but another node Worst case scenario experienced up to now: 4 nodes out of 12 dead within a few minutes Hadoop Reported erroneous nodes, blacklisted them and re-synced the remaining ones No manual intervention necessary Nothing was lost (but one node is still causing trouble, RAID controller?, cluster not affected)

The Next DDM Version: Rucio 17 DQ2 will simply not continue to scale Order(s) of magnitude more data with significant higher trigger rates and event pileup (after LS1) Extensibility issues Computing model and middleware changes Rucio is the new major version, anticipated for 2014, to ensure system scalability, remove unused concepts and support new use cases Rucio exploits commonalities between experiments (not just LHC) and other data intensive sciences (e.g., Cloud/Big data computing) Rucio & Databases Relational database management system n SQLAlchemy Object Relational Mapper(ORM) supporting Oracle, sqlite, mysql, n Use cases: Real-time data and transactional consistency n Test with representative data volume is ongoing n Evaluation of synchronization strategies between DQ2 and Rucio Non relational structured storage (Hadoop) n Use cases: Search functionality, realtime stats, monitoring, meta-data, complex analytical reports over large volumes

Conclusions 18 ATLAS Distributed Data Management delivered a working system to the collaboration in time for LHC data taking relying heavily on Oracle NoSQL: DDM use cases well covered Hadoop proved to be the correct choice: Stable reliable fast easy to work with We see Hadoop complementary to RDBMS, not as a replacement DDM team is currently developing and (stress-)testing a new DDM version called Rucio anticipated for 2014 Using both SQL and NoSQL

Thanks! 19 http://rucio.cern.ch