Breaking News! Big Data is Solved. What Is In-Memory Computing and What Does It Mean to U.S. Leaders? EXECUTIVE WHITE PAPER



Similar documents
What Is In-Memory Computing and What Does It Mean to U.S. Leaders? EXECUTIVE WHITE PAPER

From Big Data to Actionable Insight

Dell* In-Memory Appliance for Cloudera* Enterprise

A Next-Generation Analytics Ecosystem for Big Data. Colin White, BI Research September 2012 Sponsored by ParAccel

SAP and Hortonworks Reference Architecture

SAP HANA SAP s In-Memory Database. Dr. Martin Kittel, SAP HANA Development January 16, 2013

News and trends in Data Warehouse Automation, Big Data and BI. Johan Hendrickx & Dirk Vermeiren

In-memory computing with SAP HANA

Hadoop: Strengths and Limitations in National Security Missions

Architecting for Big Data Analytics and Beyond: A New Framework for Business Intelligence and Data Warehousing

Big Data Performance Growth on the Rise

Parallel Data Warehouse

Introducing Oracle Exalytics In-Memory Machine

Decoding the Big Data Deluge a Virtual Approach. Dan Luongo, Global Lead, Field Solution Engineering Data Virtualization Business Unit, Cisco

Why DBMSs Matter More than Ever in the Big Data Era

ENABLING REAL-TIME BUSINESS WITH SAP HANA IN THE CLOUD

Real-Time Big Data Analytics SAP HANA with the Intel Distribution for Apache Hadoop software

Focus on the business, not the business of data warehousing!

Interactive data analytics drive insights

Big Data Are You Ready? Thomas Kyte

Infrastructure Matters: POWER8 vs. Xeon x86

Einsatzfelder von IBM PureData Systems und Ihre Vorteile.

BIG DATA TRENDS AND TECHNOLOGIES

Understanding the Value of In-Memory in the IT Landscape

Oracle Database - Engineered for Innovation. Sedat Zencirci Teknoloji Satış Danışmanlığı Direktörü Türkiye ve Orta Asya

The Future of Data Management

Gain Contextual Awareness for a Smarter Digital Enterprise with SAP HANA Vora

Elasticsearch on Cisco Unified Computing System: Optimizing your UCS infrastructure for Elasticsearch s analytics software stack

Why Big Data in the Cloud?

SAP HANA FAQ. A dozen answers to the top questions IT pros typically have about SAP HANA

SAP HANA PLATFORM Top Ten Questions for Choosing In-Memory Databases. Start Here

IBM Netezza High Capacity Appliance

A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM

SAP Big Data and Cloud Application Development. Mark Mumy Director, Enterprise Architecture and Big Data

IS IN-MEMORY COMPUTING MAKING THE MOVE TO PRIME TIME?

WHITEPAPER. A Technical Perspective on the Talena Data Availability Management Solution

In-Memory Analytics for Big Data

Beyond Web Application Log Analysis using Apache TM Hadoop. A Whitepaper by Orzota, Inc.

BIG DATA APPLIANCES. July 23, TDWI. R Sathyanarayana. Enterprise Information Management & Analytics Practice EMC Consulting

HADOOP SOLUTION USING EMC ISILON AND CLOUDERA ENTERPRISE Efficient, Flexible In-Place Hadoop Analytics

1 Performance Moves to the Forefront for Data Warehouse Initiatives. 2 Real-Time Data Gets Real

Using an In-Memory Data Grid for Near Real-Time Data Analysis

Database Performance with In-Memory Solutions

Native Connectivity to Big Data Sources in MicroStrategy 10. Presented by: Raja Ganapathy

Big Data. Value, use cases and architectures. Petar Torre Lead Architect Service Provider Group. Dubrovnik, Croatia, South East Europe May, 2013

5 Keys to Unlocking the Big Data Analytics Puzzle. Anurag Tandon Director, Product Marketing March 26, 2014

IBM Analytics. Just the facts: Four critical concepts for planning the logical data warehouse

Business Performance without limits how in memory. computing changes everything

How To Handle Big Data With A Data Scientist

The Brave New World of Power BI and Hybrid Cloud

Why EMC for SAP HANA. EMC is the #1 Storage Vendor for SAP (IDC Storage User Demand Study, Fall 2011)

Dell s SAP HANA Appliance

Oracle Big Data Building A Big Data Management System

Oracle Big Data Discovery The Visual Face of Hadoop

Accenture and SAP: Delivering Visual Data Discovery Solutions for Agility and Trust at Scale

BIG DATA: FROM HYPE TO REALITY. Leandro Ruiz Presales Partner for C&LA Teradata

<Insert Picture Here> Big Data

An Oracle White Paper November Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics

Three Open Blueprints For Big Data Success

2015 Ironside Group, Inc. 2

Information Architecture

SAP HANA - an inflection point

Innovative technology for big data analytics

How to use Big Data in Industry 4.0 implementations. LAURI ILISON, PhD Head of Big Data and Machine Learning

Big Data & QlikView. Democratizing Big Data Analytics. David Freriks Principal Solution Architect

NoSQL for SQL Professionals William McKnight

IBM DB2 Near-Line Storage Solution for SAP NetWeaver BW

Well packaged sets of preinstalled, integrated, and optimized software on select hardware in the form of engineered systems and appliances

Safe Harbor Statement

Big Data and Your Data Warehouse Philip Russom

BigMemory and Hadoop: Powering the Real-time Intelligent Enterprise

Apache Hadoop: The Big Data Refinery

Streaming Big Data Performance Benchmark. for

IBM Data Retrieval Technologies: RDBMS, BLU, IBM Netezza, and Hadoop

SAP Solution Brief SAP Technology SAP HANA. SAP HANA An In-Memory Data Platform for Real-Time Business

High Performance IT Insights. Building the Foundation for Big Data

How Companies are! Using Spark

Safe Harbor Statement

Cisco Business Intelligence Appliance for SAP

Hadoop Architecture. Part 1

Colgate-Palmolive selects SAP HANA to improve the speed of business analytics with IBM and SAP

Unisys ClearPath Forward Fabric Based Platform to Power the Weather Enterprise

Actian Vector in Hadoop

WA2192 Introduction to Big Data and NoSQL EVALUATION ONLY

Traditional BI vs. Business Data Lake A comparison

Protecting Big Data Data Protection Solutions for the Business Data Lake

Advanced Big Data Analytics with R and Hadoop

IBM System x reference architecture solutions for big data

Big Data Analytics. with EMC Greenplum and Hadoop. Big Data Analytics. Ofir Manor Pre Sales Technical Architect EMC Greenplum

Converged, Real-time Analytics Enabling Faster Decision Making and New Business Opportunities

Chukwa, Hadoop subproject, 37, 131 Cloud enabled big data, 4 Codd s 12 rules, 1 Column-oriented databases, 18, 52 Compression pattern, 83 84

While a number of technologies fall under the Big Data label, Hadoop is the Big Data mascot.

Transcription:

Breaking News! Big Data is Solved. What Is In-Memory Computing and What Does It Mean to U.S. Leaders? EXECUTIVE WHITE PAPER

There is a revolution happening in information technology, and it s not just the newest app for your smartphone. There is a sea change occurring in how enterprises can manage their data, breaking a technology paradigm that has been in place since the midpoint of the last century. The potential benefits include radically better computing performance at less cost in US national security missions. Without much risk of contradiction, it can be said the requirements to ingest and store the various types of Big Data have been addressed by both demand-driven, commercial-off-the-shelf (COTS) software manufacturers and by the voluntary software development efforts of participants in the free and open source (FOSS) consortia. However, business and government leaders may not be satisfied that their investments in Big Data solutions have produced meaningful improvements in their mission and business processes. The sticking point has been that the visualization, analysis, and exploitation of Big Data pools to date have been largely batch-oriented, not interactive; awkward, not agile; and reliant on having very highly skilled personnel. In this paper we discuss a new paradigm, In-Memory Computing, and how it can help you fully exploit the value of Big Data and leverage the investment already made in your data infrastructure. To understand how this new paradigm can deliver value to your organization, let s put our nerd on and review how we reached this revolution in data management. A WALK DOWN MEMORY LANE Since 1944, mainstream computer design has been based on the architecture created by pioneering computer scientist John Von Neumann. His architecture is based on a control unit taking chunks of data from external memory into main memory, where it is operated upon by the logic processor. The control unit takes in the data that is needed for the program to operate in manageable chunks; the logic processor performs its operations on each chunk of data; and then the system fetches the next chunk of data for processing; and so on, over and over again, until the processing is completed. The essence of the Von Neumann architecture was to bring the data to the logic, given the limitations that existed at the time in logic processor speed and main memory capacity. Figure 1: The Von Neumann Architecture SAP National Security Services TM SAP NS2 TM 2

A few years ago, the Hadoop Distributed File System (HDFS or simply Hadoop ) was developed at Yahoo, with inspiration from a paper published by Google. 1 Hadoop has a fundamental difference in its design. It splits up very large files of data into small chunks spread across hundreds or thousands of computers; then it splits up and sends the business logic to be processed to each of the CPUs and memories on all of those computers. This is what makes Hadoop work well on very large Big Data files; the processing logic that the data scientist writes (called a MapReduce program) is run simultaneously on all of the small chunks of the big file. So a fundamental paradigm shift occurs when Hadoop is used to process data. Instead of bringing the data to the logic processor in small chunks, Hadoop distributes and brings the logic to the data. Figure 2: The Hadoop Distributed File System Architecture SOFTWARE BUILT TO LEVERAGE THE NEW CAPABILITIES OF PROCESSORS AND MEMORY Most recently, SAP has pioneered a new approach to computing that is another fundamental paradigm shift. It is called the High Performance Analytical Appliance, or SAP HANA TM, and it leverages recent advancements in both logic processors and memory chips to deliver a different architecture from the Von Neumann machine, and from Hadoop. 2 Co-innovated with Intel Corporation 3, the architecture of SAP HANA makes it possible to radically accelerate Big Data analyses in the enterprise, including data stored in the Hadoop Distributed File System. 1 Doug Cutting and others at Yahoo wrote Hadoop during 2007-08, based on Google s Big Table architecture. See F. Chang, J. Dean, et. al, Bigtable: A Distributed Storage System for Structured Data, Proceedings of the 7th Symposium on Operating Systems Design and Implementation, 2006. 2 SAP HANA-certified servers based on the Intel chipset are available from Cisco, Dell, Fujitsu, Hitachi, HP, and IBM. 3 Co-innovation with Intel allows the SAP in-memory solution to leverage HANA-specific instruction sets in the Xeon Ivory Bridge and Hayworth chipsets. See the Intel white paper, Analyzing Business as it Happens, Version 1.0, April 2011, available at http://www.intel.com/content/www/us/en/high-performance-computing/high-performance-computingxeon-e7-analyze-business-as-it-happens-with-sap-hana-software-brief.html. SAP National Security Services TM SAP NS2 TM 3

In the SAP HANA architecture, all of the data that will be subject to logical processing is contained in main memory at all times, even very large amounts of data. External memory -- either in solid state or physical disc drives -- is no longer necessary, except for back-up and disaster recovery purposes. There is no chunking of the data or sequential delivery to the processor as in the Von Neumann architecture. Nor is there any splitting up and distribution of the logic algorithms as in the Hadoop architecture. With SAP HANA, the logic and all of the data reside and work together in main memory. Figure 3: SAP HANA Architecture The practical result of this is a remarkable increase in analytical processing speed, even on very large data sets. Therefore, SAP HANA can meet users expectations of very fast response times for transactions, analyses, and visualizations, even with Big Data sets. Freed from the limitations of the older generation of hardware (including CPUs, memory and discs), SAP HANA does not require the extensive administration and constant tuning that database administrators have had to do with traditional database management systems which were first written in the 1970s. SAP HANA can be deployed as a stand-alone in-memory database accommodating 2 to 8 terabytes of memory per server node. Alternatively, SAP HANA can be combined with Hadoop s on-disc data storage in a hybrid deployment in which the hot, frequently used data is stored in the SAP HANA in-memory columnar database, while the warm, less frequently used data is stored in a Hadoop system. In short, SAP HANA delivers speed-of-thought analysis, whether data is sourced from traditional row-based data warehouses, transactional business applications, streaming sensor data, or from the Hadoop system. SAP calls this architecture the Real-Time Data Platform. SAP National Security Services TM SAP NS2 TM 4

REAL-WORLD BENEFITS The real-world mission and business benefits that leaders can expect from employing the SAP Real-Time Data Platform include: Accelerated analyses of complex queries against granular (not summary level) Big Data. HANA s near-real-time analyses can help support your decisions, improve planning, and optimize mission execution. A single in-memory data repository can be used both for transactional applications and for analyses simultaneously. This cuts the Operation and Maintenance (O&M) costs of having to maintain a separate transactional database for mission applications; another data warehouse optimized for reporting and analysis; and multiple other replications into data marts for specific users mission needs. Integrating highly compressed in-memory data storage in SAP HANA with bulk on-disc data storage in Hadoop allows organizations to optimize the use of their infrastructure resources and save money. Integrating data from other systems inside or outside the enterprise multiplies the value of data sets by unlocking previously undiscoverable analytical insights and making them easier to access in a self-service way. The SAP Real-Time Data Platform allows for a reduction in specialized skills needed to operate and maintain the overall system, because data integration with other systems in the enterprise uses standard interfaces which are commonly available skill sets (such as SQL, MDX, XML, ODATA, JSON, and JDBC). Now engineers at SAP National Security Services (SAP NS2) are working with partners and customers to develop specific applications that exploit this game-changing technology to advance the missions of the US Intelligence Community, Department of Defense, and homeland security customers. We would welcome the opportunity to discuss how SAP NS2 can help your organization advance your mission with the SAP HANA Real-Time Data Platform. SAP National Security Services TM SAP NS2 TM 5

www.sapns2.com FOR MORE INFORMATION Contact your account manager or call us at 877-9-SAPNS2 (877-972-7672) Email: info@sapns2.com Website: www.sapns2.com AUTHOR Bob Palmer Senior Director, Solutions bob.palmer@sapns2.com 301.641.7785 About SAP National Security Services (SAP NS2 ) SAP National Security Services (SAP NS2) offers a full suite of enterprise applications, analytics, database, and mobility software solutions from SAP with specialized levels of security and support to meet the unique mission requirements of US national security and critical infrastructure customers. SAP National Security Services and SAP NS2 are trademarks owned by SAP Government Support and Services (SAP GSS). For more information, please visit www.sapns2.com. Copyright 2013 by SAP Government Support and Services. All rights reserved. May not be copied or redistributed without permission