The Flash Effect on Big Data

Similar documents
Solid State Storage in the Evolution of the Data Center

Benchmarking Cassandra on Violin

BIG DATA APPLIANCES. July 23, TDWI. R Sathyanarayana. Enterprise Information Management & Analytics Practice EMC Consulting

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software

2010 Ingres Corporation. Interactive BI for Large Data Volumes Silicon India BI Conference, 2011, Mumbai Vivek Bhatnagar, Ingres Corporation

System Architecture. In-Memory Database

Evaluating NoSQL for Enterprise Applications. Dirk Bartels VP Strategy & Marketing

Big Data Technologies Compared June 2014

Main Memory Data Warehouses

Scaling from Datacenter to Client

Architecting for Big Data Analytics and Beyond: A New Framework for Business Intelligence and Data Warehousing

Big Data & QlikView. Democratizing Big Data Analytics. David Freriks Principal Solution Architect

Flash for Databases. September 22, 2015 Peter Zaitsev Percona

Database Performance with In-Memory Solutions

News and trends in Data Warehouse Automation, Big Data and BI. Johan Hendrickx & Dirk Vermeiren

Performance and Scalability Overview

Accelerating I/O- Intensive Applications in IT Infrastructure with Innodisk FlexiArray Flash Appliance. Alex Ho, Product Manager Innodisk Corporation

Using distributed technologies to analyze Big Data

SSD Performance Tips: Avoid The Write Cliff

SALSA Flash-Optimized Software-Defined Storage

The Data Placement Challenge

Benchmarking Hadoop & HBase on Violin

Hadoop and Relational Database The Best of Both Worlds for Analytics Greg Battas Hewlett Packard

Lecture Data Warehouse Systems

The Evolution of Microsoft SQL Server: The right time for Violin flash Memory Arrays

Hypertable Goes Realtime at Baidu. Yang Dong Sherlock Yang(

Future Prospects of Scalable Cloud Computing

An Integrated Analytics & Big Data Infrastructure September 21, 2012 Robert Stackowiak, Vice President Data Systems Architecture Oracle Enterprise

Accelerating Real Time Big Data Applications. PRESENTATION TITLE GOES HERE Bob Hansen

Performance and Scalability Overview

Data Center Storage Solutions

Maximum performance, minimal risk for data warehousing

How SSDs Fit in Different Data Center Applications

Driving Peak Performance IBM Corporation

Data Center Solutions

Hurtownie Danych i Business Intelligence: Big Data

Big Fast Data Hadoop acceleration with Flash. June 2013

Accelerating Cassandra Workloads using SanDisk Solid State Drives

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

Oracle Big Data SQL Technical Update

How To Scale Out Of A Nosql Database

ebay Storage, From Good to Great

Investor Presentation. Second Quarter 2015

Minimize cost and risk for data warehousing

Reference Architecture, Requirements, Gaps, Roles

An Overview of Flash Storage for Databases

Data Center Solutions

Peninsula Strategy. Creating Strategy and Implementing Change

Big Data Analytics - Accelerated. stream-horizon.com

Flash Controller Architecture for All Flash Arrays

Well packaged sets of preinstalled, integrated, and optimized software on select hardware in the form of engineered systems and appliances

IS IN-MEMORY COMPUTING MAKING THE MOVE TO PRIME TIME?

Hyperscale Use Cases for Scaling Out with Flash. David Olszewski

HYPER-CONVERGED INFRASTRUCTURE STRATEGIES

Making Sense ofnosql A GUIDE FOR MANAGERS AND THE REST OF US DAN MCCREARY MANNING ANN KELLY. Shelter Island

Emerging storage and HPC technologies to accelerate big data analytics Jerome Gaysse JG Consulting

IBM Data Retrieval Technologies: RDBMS, BLU, IBM Netezza, and Hadoop

SOLID STATE DRIVES AND PARALLEL STORAGE

SAP HANA PLATFORM Top Ten Questions for Choosing In-Memory Databases. Start Here

Best Practices for Optimizing SQL Server Database Performance with the LSI WarpDrive Acceleration Card

Dell Microsoft Business Intelligence and Data Warehousing Reference Configuration Performance Results Phase III

How To Handle Big Data With A Data Scientist

Big Data and Its Impact on the Data Warehousing Architecture

Flash Memory Arrays Enabling the Virtualized Data Center. July 2010

Bringing Greater Efficiency to the Enterprise Arena

Big Data and Data Science: Behind the Buzz Words

Cloud Scale Distributed Data Storage. Jürmo Mehine

Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers

Integrated Grid Solutions. and Greenplum

Enterprise Operational SQL on Hadoop Trafodion Overview

Introducing Oracle Exalytics In-Memory Machine

Structured Data Storage

Speeding Up Cloud/Server Applications Using Flash Memory

Can the Elephants Handle the NoSQL Onslaught?

Tap into Hadoop and Other No SQL Sources

Accelerating Server Storage Performance on Lenovo ThinkServer

Essentials Guide CONSIDERATIONS FOR SELECTING ALL-FLASH STORAGE ARRAYS

Solid State Technology What s New?

2009 Oracle Corporation 1

Calsoft Webinar - Debunking QA myths for Flash- Based Arrays

Tiber Solutions. Understanding the Current & Future Landscape of BI and Data Storage. Jim Hadley

Big Data Architectures. Tom Cahill, Vice President Worldwide Channels, Jaspersoft

The What, Why and How of the Pure Storage Enterprise Flash Array

The Flash Transformed Data Center & the Unlimited Future of Flash John Scaramuzzo Sr. Vice President & General Manager, Enterprise Storage Solutions

BIG DATA CAN DRIVE THE BUSINESS AND IT TO EVOLVE AND ADAPT RALPH KIMBALL BUSSUM 2014

Big Data Technology ดร.ช ชาต หฤไชยะศ กด. Choochart Haruechaiyasak, Ph.D.

Getting the Most Out of Flash Storage

Inge Os Sales Consulting Manager Oracle Norway

HP ProLiant DL580 Gen8 and HP LE PCIe Workload WHITE PAPER Accelerator 90TB Microsoft SQL Server Data Warehouse Fast Track Reference Architecture

Can Flash help you ride the Big Data Wave? Steve Fingerhut Vice President, Marketing Enterprise Storage Solutions Corporation

DataStax Enterprise Reference Architecture

Overview: X5 Generation Database Machines

Evaluation Report: Database Acceleration with HP 3PAR StoreServ 7450 All-flash Storage

Flash 101. Violin Memory Switzerland. Violin Memory Inc. Proprietary 1

Accelerating Datacenter Server Workloads with NVDIMMs

Ramesh Bhashyam Teradata Fellow Teradata Corporation

Performance Management in Big Data Applica6ons. Michael Kopp, Technology

IAF Business Intelligence Solutions Make the Most of Your Business Intelligence. White Paper November 2002

Hadoop and Data Warehouse Friends, Enemies or Profiteers? What about Real Time?

Oracle Aware Flash: Maximizing Performance and Availability for your Database

Transcription:

The Flash Effect on Big Data Bruce Moxon CTO, Systems and Solutions Santa Clara, CA 1

Discussion Outline Defining Big Data Big Data s HPC Legacy and Hyperscale Data Warehousing and Analytics NoSQL Solid State Storage Key Features Big Data acceleration scenarios Hadoop, Cassandra, Splunk Future Santa Clara, CA 2

What is Big Data? How Big is Big? Big data is high volume, high velocity, and/or high variety information assets that require new forms of processing to enable enhanced decision making, insight discovery and process optimization. - Gartner (based on Doug Laney s 3V s) http://en.wikipedia.org/wiki/big_data New (evolved) Technologies, Approaches, Tools Scale-out computing Application-specific storage (SDS) In-memory databases Columnar databases MapReduce and Distributed File Systems NoSQL (k-v, graph, document), NewSQL Complex Event Processing A significant (dominant) Open Source model. Santa Clara, CA 3

Scale-out, HPC, and Hyperscale Data Centers Web Browsers Client Internet / Intranet Load Balancers Web / App Servers Service Application Services Caching (e.g. memcached) Resource Analytics Master DBs Slave DBs Santa Clara, CA 4

Why Open Compute Matters for more than Facebook Traditional Infrastructure Lifecycle Hyperscale Infrastructure Lifecycle Santa Clara, CA 5

Traditional Data Warehousing Process (DW/BI) Santa Clara, CA 6

Operational vs Analytical Schema Organization of data should reflect the activities (processes) that use it Operational (OLTP) Optimized for tracking and status reporting (e.g. LIMS, sample tracking) Transactional queries Short, UPDATE intensive Well defined, indexed Few records per query High volume Exploits RI, triggers, stored procedures No results Normalized Schema (3NF) Analytical (OLAP) Optimized for large-scale analysis and decision support applications Analytical queries Long, SELECT intensive Ad hoc, table scans Many records (scans), many tables (joins) Relatively low volume scalar functions, aggregation, extended functions Results used elsewhere (BI reporting, visualization, ) De-normalized Schema Santa Clara, CA 7

Denormalization, Columnar Stores, Star Schema, and Cubes Efficient Columnar Representations Santa Clara, CA 8

OLAP, Multi-dimensional Databases (MDDs), and Cubes An approach to multi-resolution, N-dimensional data analysis (OLAP) central fact table(s) peripheral dimension tables Optimized for quick response to slice and dice queries, drillup / drilldown, pivot pre-aggregations (rollups) highly indexed special schema (star, snowflake) read mostly / batch update model browsable dimensions filtering of facts by dimension attributes Efficient data representations sparse data, time series Library Organ Donor Geneid Pathology Expr Level L01833 brain 103391 g121131 normal 0.00214 L01833 brain 103391 g124122 normal 0.00133 L01833 brain 103391 g089913 normal 0.00212 L01011 lung 110122 g113314 cancerous 0.00413 Gene (Protein) Expression MDD or RDBMS implementations Tissue Time Santa Clara, CA 9

An Evolution of Data Warehousing / Analytics Platforms Teradata (1983) First DSS appliance based on MPP architecture Netezza (2000); acquired by IBM (2010) Hardware accelerated function shipping (FPGA-based AMPP architecture) Greenplum (2003); acquired by EMC (2010) Shared nothing implementation of Postgres on commodity server MPP architecture Columnar stores and MapReduce support added Vertica (2005); acquired by HP (2011) Columnar data stores (scan performance, high compression factors) ParAccel (2005); Columnar data stores, commodity servers Aster Data (2005); acquired by Teradata (2011) MapReduce-based implementation, SQL-MapReduce NoSQL Data Warehouses HBase, Cassandra, Gen 1 Scaleout MPP Gen 2 Custom HW (x) Gen 3 Commodity servers (scaleout) Gen 3.5 Columnar stores Gen 4 MapReduce and Derivatives, NoSQL Santa Clara, CA 10

The New Analytics Infrastructure The Internet of Things Big Data Ecosystem Key Implications Data Warehouse, Operational Data Store, Data Mart distinction blurring Historically batch-oriented processes (ETL, summary/aggregation) are becoming more real-time Indexing, Query (SQL) Background MapReduce http://www.funambol.com/documents/funambol_internetofthings_may10.pdf http://stevenimmons.org/wp-content/uploads/2012/02/bigdataecosystem.png Santa Clara, CA 11

Time-value of Data http://blog.voltdb.com/big-data-value-continuum/ Santa Clara, CA 12

NoSQL NoSQL and NewSQL Highly scalable, (generally) looser consistency models than traditional RDBMSs (CAP Theorem; MVCC, ACID, and BASE) Column: HBase, Cassandra, Accumulo Document: MongoDB, Couchbase, MarkLogic Key-value: Dynamo, Riak, Redis, LevelDB, BerkeleyDB, MemcacheDB Graph: Neo4J, Objectivity Infinite Graph NewSQL New class of high performance databases supporting the relational model and SQL (transactional, realtime ETL and analytics) NuoDB, VoltDB, Memsql, Pivotal SQLFire http://en.wikipedia.org/wiki/nosql http://nosql-database.org/ http://www.marklogic.com/blog/acid-base-and-nosql/ Santa Clara, CA 13

Solid State Storage Fundamentals http://en.wikipedia.org/wiki/write_amplification#basic_ssd_operation Everything I Needed to Know I learned at FMS Data is read/written in pages (typically 4-8 KB) Data is *erased* in multi-page blocks (e.g., 128) Data can only be written (programmed) into a previously erased block (no overwrite ) Background garbage collection (GC) copies valid pages to squeeze out deleted pages and make room for new data OS/FS integration (TRIM) Additional write amplification can occur in support of wear leveling and GC Flash memory can only be programmed/erased (P/E) a limited number of times (Endurance) Performance degrades over device lifetime Overprovisioning is usually employed to afford more flexibility in background tasks such as wear leveling and garbage collection. It reduces write amplification, in turn extending the life of the device Background tasks affect performance, especially at peak device speeds (latency) Implementations vary widely Santa Clara, CA 14

Log Structured File Systems First Described in a paper by Mendel Rosenblum and John K. Ousterhout http://www.cs.berkeley.edu/~brewer/cs262/lfs.pdf Basic principles (Solid State implementations) Extra space or over-provisioning is added to the capacity of the drive Writes are always written to new logical blocks (LBs); Updates invalidate old LBs Garbage Collection periodically purges the invalidated LBs by re-writing still valid LBs to fresh Erase Blocks and erasing the cleared Erase Blocks Effectively transforms random writes into sequential writes The percentage of transformation is dependent on the amount of over-provisioning FTL embodies the virtual-to-physical (V2P) mapping Host Random Writes Santa Clara, CA 15

Solid State Performance Characteristics (General) HDD (SAS) Sequential Random Read 200 MB/s 200 IOPS Write 200 MB/s 200 IOPS Rand Response 3-8x SSD / PCIe Sequential 4K Random Read.5 1.5 GB/s 60-200K IOPS Write.5 1.3 GB/s 50-250K IOPS Read Write HDD 8 ms 0.5 ms* SSD 60 us 20 us 100-1000x General Performance Characteristics. YMMV, depending on Device architecture Interface (SATA, SAS, PCIe) 2-5x performance range Transfer sizes QDs (concurrency) Cost Differential $0.50 - $1.50 / GB SAS HDD $2 - $12 / GB SSD/PCIe (< $1) Sweet Spot High Random IOPS (esp. Read) Low Latency Santa Clara, CA 16

Key Solid State Storage Characteristics Outstanding random IO relative to HDD Random write performance through log-structured filesystem approaches; corresponding amplification / endurance considerations Moderate increase in sequential IO relative to HDD $-per-mb/s parity; Capacity premium Significant latency advantage (10x or more) Much higher concurrency support (64-128+ per device) without significant latency impact Santa Clara, CA 17

Flash for MapReduce (?) Investigations and Conclusions From http://hadoopblog.blogspot.com/2012/05/hadoop-and-solid-state-drives.html Basic map-reduce is large block sequential and doesn t benefit much (not economically) from SSDs HDFS+HBase Tested HDFS with DRAM-based data (~ infinitely fast disk) to determine basic HDFS capability: 92K IOPS and 95% CPU Tested HBase with DRAM-based table (~ infintely fast disk) to determine basic HBase capability: 35K ops/s at 45% CPU; heavy lock contention and context switching My conclusion is that database and storage technologies would need to be developed from scratch if we want to utilize the full potential of Solid State Devices Ongoing Work: BucketCache victim cache (HBASE-7404); Tiered HFile Storage (HBASE-6572); Heterogeneous Storage (HBASE-2832) Santa Clara, CA 18

Flash for Splunk From http://blogs.splunk.com/2012/05/10/quantifying-the-benefits-of-splunk-with-ssds/ Guidance Ingest and indexing are (sequential)-write-intensive activities; these will benefit only modestly from use of SSDs Searches: Dense searches typically used for aggregation will mostly use large, sequential reads ( scan operations or range queries ) Sparse searches including cross-series queries depend on random reads (both indexes and data) and will make good use of SSD s (200x demonstrated for pure point queries) Bloom Filters may also benefit (Splunk 4.3) Santa Clara, CA 19

Flash for Cassandra Cassandra Writes (1.2) Santa Clara, CA 20

Flash for Cassandra Cassandra Reads (1.2) 1.2: Compression offsets and Bloom filter off-heap (capacity scaling) Santa Clara, CA 21

Flash for Cassandra http://datastax.com/documentation/cassandra/1.2/webhelp/index.html?pagename=docs&version=1.2&file=dml#cassandra/architecture/architectureplanninghardware_c.html There are a number of cassandra.yaml settings specifically for Cassandra tuning for SSDs, mostly taking into consideration the increased concurrency and performance afforded by SSDs. The.yaml file includes suggested values. YMMV concurrent_reads and concurrent_writes multithreaded_compaction and compaction_throughput_mb_per_sec (disable throttling) commitlog_sync_period_in_ms There may be other not-so-obvious settings depending on workload (ingest rate, amount of indexing, query types and concurrency, ). E.g. compaction frequency/type, which can generate more SSTables (which SSD s should handle well), more or less aggressive Bloom Filters, etc. With HDD s, the tendency is to defer flushes (so there may be less data to write); with SSD s, more frequent flushes and more SSTables files may yield better performance (higher concurrencies, which SSD s accommodate). Santa Clara, CA 22

Flash and Big Data: Future There are some things on the horizon that can fundamentally change the evolving architectures High thread count processors (Intel Xeon Phi) PCIe and NVMe Non-spinning disk abstractions on Flash Fusion-io iomemory SDK stec FOS/kvs NVM Programming TWG Storage Class Memory (MRAM, PCM, resistive RAM, ) Santa Clara, CA 23

Flash 3.0 Flash technology presents new opportunities for application optimizations in the data path Random access media Near-memory latency at near-disk cost These will be optimally delivered through new device abstractions better-matched to device characteristics Memory abstractions Object abstractions Flash Object Store (FOS/KVS) Natural extension to Storage Class Memory Library File System OS Buffer Cache Block Device Application Library Flash-optimized Device Flash Management Santa Clara, CA 24

Summary A range of Big Data stores and tools are emerging based on scale-out commodity server implementations; many are open source These tools provide new approaches and extensions to traditional Data Warehouse/Business Intelligence problems Many of these stores/tools optimize for sequential writes for ingest and indexing (and logging); this approach is Flash friendly (limiting write amplification), but doesn t drastically improve ingest performance Focused queries (point queries) can benefit greatly from the increased random read performance and higher concurrency support of Flash-based devices (both index and data reads) Next generation NVMe devices promise significant sequential IO improvement to boost sequential read/write performance (ingest and scan queries) Architectural enhancements (e.g. victim caches) and new IO architectures are needed to fully leverage Flash Storage Class Memories will present the next opportunity with a new set of architectural opportunities and challenges Santa Clara, CA 25