SOME STRAIGHT TALK ABOUT THE COSTS OF DATA WAREHOUSING
|
|
|
- Peter Burke
- 10 years ago
- Views:
Transcription
1 Inmon Consulting SOME STRAIGHT TALK ABOUT THE COSTS OF DATA WAREHOUSING Inmon Consulting PO Box Wilcox Street Castle Rock, Colorado An Inmon Consulting White Paper By W H Inmon
2 By W H Inmon Everybody knows that data warehouses cost a lot of money. Everybody knows that when you have a data warehouse that the corporate budget will be affected. But what everybody doesn t know is that the costs of a data warehouse do not have to eat you out of house and home. There are some good ways to effectively mitigate the large amounts of dollars that a data warehouse is reputed to swallow up. DOES A DATA WAREHOUSE HAVE TO COST A LOT OF MONEY? The truth of the matter is that the long-term cost of the data warehouse depends more on the developers and designers and the decisions they make than on the actual cost of technology. Stated differently a data warehouse can cost a lot of money, but a data warehouse does not have to cost a lot of money. In order to understand what some of the options in the building and operation of a data warehouse are and why the design options are so important to the cost of the data warehouse, consider some time honored conventional wisdoms that have been put forth into the community of technology and business. One of those conventional wisdoms is that storage costs are getting cheaper all the time. Another of those conventional wisdoms is that processors are getting cheaper all the time, as well as storage. But what happens when cheaper storage and processors are not the problem and the costs associated with the management and expertise and additional technologies are? That s when it s time to change the rules when conventional wisdom must give way to innovation. 2
3 STORAGE IS GETTING CHEAPER ALL THE TIME The diagram in Fig 1 illustrates this well-repeated conventional wisdom. Cost per megabyte Time The classical curve of declining cost of storage Fig 1 The diagram in Fig 1 shows that the unit cost of storage is getting less expensive every year. The purveyor of this diagram quotes Moore s law and then draws the conclusion that we should not worry about the costs of storage for our future systems design and development effort. The implications drawn from this diagram are that in the future the costs of storage will tend to be free or close to it. When this costing of storage occurs, storage becomes commoditized. This conclusion that we should not worry about the costs of storage in the future is a gross distortion of the facts. Reality as we shall see is something quite different. It is no surprise that Moore s law is quoted by hardware and storage vendors who have the most to gain by the gross misinterpretation of the lowered unit costs of storage on the spending habits in the marketplace. THE CONSUMPTION CURVE There are several other pieces of information that must be taken into account when considering the long-term costs of storage. One piece of information is that while the unit cost of storage is dropping, the demand for storage far outstrips the drop in the costs of megabytes of storage. Fig 2 shows the increase in the demands for storage. Bytes Time The storage consumption curve Fig 2 3
4 The storage consumption curve is not as widely known and quoted as the declining cost of storage curve. But industry sources such as IDEMA INSIGHT (James Parker, Vol. IX, no 5) support this consumption curve of storage. The truth is that storage is being consumed at a rate faster than the rate at which storage price is dropping. And along with the consumption curve of data is a corresponding consumption curve for processors. It is not just that we have more data. We need more processors to do something with that data. And it is not just the volume of data and the cost of the infrastructure that presents a challenge. The pathways that allow the data to be accessed need to be increased as well. With a large amount of data and only a limited pathway into the data, it is inevitable that bottlenecks to performance will develop. THE INFRASTRUCTURE COST OF STORAGE But the real culprit in the storage/cost conundrum is the fact that as the volume of data increases, the infrastructure cost for the management of data increases far, far greater than the rate of the drop of the unit cost of storage. In order to see this equation at work, you only have to go to Radio Shack and buy some storage for you computer. You can buy several gigabytes for your personal computer. The cost will be minimal. And the cost of your computer which will ultimately house the storage is several hundred dollars or maybe even a few thousand dollars. So for smaller volumes of data, the cost of the management infrastructure is minimal. But for several terabytes of storage managed by IBM or Teradata in a data warehouse environment the cost of the infrastructure to manage those terabytes of data may be from $500,000 to $1,000,000 per terabyte. The actual cost of storage is a rounding error compared to the cost of the storage management infrastructure for the larger volumes of data. 4
5 Fig 3 shows this curve. Hundreds of thousands of dollars Tens of thousands of dollars Dollars Cents Kb Mb Gb Tb The storage infrastructure cost per unit of storage Fig 3 It is noted that the curve shown in Fig 3 is logarithmic, not linear. The cost of the storage management infrastructure skyrockets as the volume of data increases. Stated differently, when you buy storage for a data warehouse, you are really paying for the storage management infrastructure, not the storage itself. The same conditions hold true for processors. The most expensive processor cycles are those that are managed by the largest computers. The smaller the computer, the less expensive the processing cycle. Stated differently, the unit cost of processing cycles is more expensive, the larger the processor that you get. And when you have lots of data in a classical processing environment, the more processing cycles you need. THE FUNCTIONALITY OF THE STORAGE MANAGEMENT INFRASTRUCTURE So what kind of processing is going on that is so expensive in large processors and in large collections of data? In addition to the sheer volume of data driving the costs of storage management infrastructure higher, the actual functionality of the storage under management is a factor as well. It is not just the total volume of storage that needs to be managed that is a factor; it is what is being done to the storage by the infrastructure that costs as well. Stated differently, it is a lot more expensive to manage an OLTP infrastructure of a terabyte of data than it is to manage an archival environment of ten terabytes of data. The degree of the functionality of processing supported by the infrastructure plays a big part in the cost of the infrastructure. 5
6 In order to understand the role played by the functionality of the processing being done to the data under storage management, consider the simple diagram of Fig 4 that shows the basic components of the modern computer. The basic operations of a computer Fig 4 Fig 4 shows the basic components of the computer. There is the network manager, which sends and receives messages from a network. There is the application function where specified processing is executed. There is the operating system that manages the functions and their priorities. And there is the dbms that manages the access and interchange of data to and from the computer. These functions are pretty much generic to any computer. They are found in a $1,000 personal computer and they are found in the largest and most sophisticated parallel processor that sells for millions of dollars. It is a temptation to say that you can go to Radio Shack and buy a terabyte of storage. Indeed you can do just that. But when you come home and start to use the terabyte of data, you are going to find out that your personal computer operates quite differently from a large-scale parallel processor. You cannot run a world wide OLTP environment from your home computer. 6
7 THE FUNCTIONALITY OF A FULL-SCALE PARALLEL PROCESSOR So what infrastructure services does a full-scale parallel processor offer? What is the functionality that drives up the price of the data management infrastructure? Fig 5 shows some of the functionality found in a full-scale parallel processor. Network Application Oper sys Dbms Some of the many functions accomplished by the infrastructure for storage management Fig 5 Speed of processing - indexing - caching - hashing - direct access of data Analysis - location of data - formatting Operations - update - creation - deletion - overflow management - transaction integrity - backup/recovery - security Fig 5 shows that a full-scale parallel processor offers quite a bit of functionality. The first service offered is that of running in parallel. Running in parallel requires quite a bit of sophistication. Running in parallel requires the coordination and integration of many processors operating separately. But there are other features to the large scale management of storage. 7
8 Some of these features include speed of processing indexing caching hashing direct access of data analysis location of data formatting of data operations update of data creation of data deletion of data overflow management transaction integrity backup and recovery security and so forth. It is seen that for large amounts of data that the data management infrastructure is quite sophisticated. It is no surprise then that the costs of storage rise as the volume of data and the sophisticated use of that data escalate. Stated differently, it is no surprise that the costs of storage have much more to do with the management infrastructure than they do with the unit cost of storage. DATA WAREHOUSE COSTS Experience has shown that there are four main components to the cost of a data warehouse. These components are - Consulting and development costs - Storage and storage infrastructure costs - ETL costs - Dbms and other infrastructure costs. Consulting and development costs are those costs that relate to the building of the data warehouse. These costs include data modeling, design, mapping, ETL and data population costs. These costs can be done in-house or by contracting to a consulting firm, or both. Storage costs include the cost of storage and the infrastructure. ETL costs are software costs that include the ETL software, metadata capture, transformation logic, source and target mapping, and so forth. Dbms and other infrastructure costs include software licensing, network, processor, and other infrastructure costs. 8
9 Fig 6 shows the short-term (up to 4 year) costs associated with the building and operation of a data warehouse. $ Consulting Storage ETL Dbms Year The short-term costs of a data warehouse Fig 6 It is seen that consulting and development costs are major ticket items. A lot of money will be spent on the design, development, and deployment of the data warehouse. The good news is that these costs diminish over time. While a certain amount of ongoing work will need to be done to the data warehouse, the major part of the work will be done in the first few years of the life of the data warehouse. After the first few iterations are built, the development work degenerates to maintenance work. And as maintenance is done, the high cost of development goes away. The cost of storage in the first few years of the data warehouse is not high. In the first few years there is relatively little historical data, and there are only a few subject areas that are built into the data warehouse. It is over time that large amounts of history appear and that lots of subject areas are entered into the data warehouse. ETL (extract/transform/load) processing is the processing that occurs as data is read from the legacy application environment and is transformed into the data warehouse environment. ETL is normally done by software. However, for a really small data warehouse ETL processing can be done manually. But when ETL processing is done manually, its long-term costs exceed that of doing ETL by software. The costs of ETL are seen to diminish as the data warehouse becomes mature. 9
10 Dbms and other infrastructure costs include software and hardware. In the diagram shown in Fig 6 it is assumed that there is a new license of software that is needed and that there are more processors that are needed. If there already is a software license or where software licensing can be done on an incremental basis, then the initial costs will be less. If the initial CPU and/or CPU license cost is one-time then the maintenance costs are perpetual and accumulating unless there is a site license agreement in place. In addition, there is a cost to burn in the software. If the software is already being used, then there will be much less of a burn in cost. In addition, training is needed for the installation and usage of both ETL and dbms software. It is seen that as time passes the costs for dbms and processors diminish after the initial purchase and acquisition are made. Looking at Fig 6, it is seen that the cost of storage is not particularly significant in the life of the building and usage of the data warehouse in the short term. So why should a person worry about storage costs? In fact the costs of storage are so significant that in many organizations only summarized or aggregated data is archived. The details of data are lost because it is simply too expensive to store them over time. And unfortunately the details of data are the place where many of the most interesting and most useful data lie. THE LONG-TERM COSTS OF A DATA WAREHOUSE The long-term picture for the costs of data warehousing looks considerably different however. Fig 7 shows the long-term picture of the costs of the data warehouse. $ Year Consulting Storage The long-term costs of a data warehouse ETL Dbms Fig 7 10
11 The picture of costs of a data warehouse seen in Fig 7 is considerably different than the picture of costs seen in Fig 6. Fig 7 depicts the costs of data warehousing over a ten-year time frame. In Fig 7 it is seen that at the end of ten years far and away the costs of the data warehouse are dedicated to storage. In fact it is because of the large amounts of storage and the infrastructure needed to manage that volume of data that the cost of data warehousing is so large over a lengthy period of time. But do the costs of data warehousing have to rise at the rate seen in Fig 7? The answer is not at all. It is possible to mitigate the long-term costs and the effective ability to use the data found in a data warehouse by the introduction of what is termed a data warehouse appliance. A data warehouse appliance is a combination of hardware and software that is designed to manage very large amounts of storage but at a significantly lower rate than traditional storage. The data warehouse appliance such as that offered by Dataupia allows the data warehouse environment to remain intact. If an organization starts with an Oracle license, they make no changes to the environment as the data warehouse appliance is implemented. If an organization starts with DB2, then they make no changes to the environment as the data warehouse appliance is implemented. The deployment of the data warehouse appliance is independent of and transparent to the dbms. In the case of Dataupia, the data warehouse appliance will act as an MPP foundation for your current rdbms vendor The dbms accesses data as it has always done. Only some or all of the data being managed is being managed by the data warehouse appliance. And the cost of storage management under the data warehouse appliance is a fraction of that required by traditional storage management equipment. Fig 8 shows what happens to the costs of storage in a data warehouse environment upon the introduction of the data warehouse appliance. $ The introduction of a data warehouse appliance Year The introduction of a data warehouse appliance greatly reduces the long term costs of the data warehouse environment Fig 8 11
12 In Fig 8 it is seen that the escalating costs of storage are mitigated by the introduction of a data warehouse appliance. Once the costs of storage are mitigated, the long term costs of the data warehouse do not spiral out of control. An organization is free to build as large a data warehouse as they need with no real consideration to the costs of the data warehouse. Another way to look at the phenomenon of the introduction a data warehouse appliance into the equation is seen in Fig 9. Storage costs with no data warehouse appliance $ Storage costs with a data warehouse appliance Year The cost differential with and without a data warehouse appliance Fig 9 Fig 9 shows explicitly what the effect of the introduction of a data warehouse appliance into the data warehouse environment in the long term is. The amount of money needed to build and operate the data warehouse is greatly minimized by the usage of a data warehouse appliance. It is because of this decision to use or not to use a data warehouse appliance that it is asserted that the costs of the data warehouse depend more on the design and development decision-makers than on the actual costs of equipment for the data warehouse. If a designer or developer chooses to make the costs of a data warehouse high, then a designer or developer has the ability to do just that. But if a designer or developer chooses to make the costs of a data warehouse low, then the designer/developer has the ability to do just that as well. The money saved by designers allows them to buy other technologies and/or technologies which allow them to further increase ROI and the business value through its data warehouse user. 12
Virtual Data Warehouse Appliances
infrastructure (WX 2 and blade server Kognitio provides solutions to business problems that require acquisition, rationalization and analysis of large and/or complex data The Kognitio Technology and Data
Data Warehousing. Jens Teubner, TU Dortmund [email protected]. Winter 2015/16. Jens Teubner Data Warehousing Winter 2015/16 1
Jens Teubner Data Warehousing Winter 2015/16 1 Data Warehousing Jens Teubner, TU Dortmund [email protected] Winter 2015/16 Jens Teubner Data Warehousing Winter 2015/16 13 Part II Overview
Data Warehousing: A Technology Review and Update Vernon Hoffner, Ph.D., CCP EntreSoft Resouces, Inc.
Warehousing: A Technology Review and Update Vernon Hoffner, Ph.D., CCP EntreSoft Resouces, Inc. Introduction Abstract warehousing has been around for over a decade. Therefore, when you read the articles
Data Warehouse (DW) Maturity Assessment Questionnaire
Data Warehouse (DW) Maturity Assessment Questionnaire Catalina Sacu - [email protected] Marco Spruit [email protected] Frank Habers [email protected] September, 2010 Technical Report UU-CS-2010-021
Innovative technology for big data analytics
Technical white paper Innovative technology for big data analytics The HP Vertica Analytics Platform database provides price/performance, scalability, availability, and ease of administration Table of
Using In-Memory Data Fabric Architecture from SAP to Create Your Data Advantage
SAP HANA Using In-Memory Data Fabric Architecture from SAP to Create Your Data Advantage Deep analysis of data is making businesses like yours more competitive every day. We ve all heard the reasons: the
SQL Maestro and the ELT Paradigm Shift
SQL Maestro and the ELT Paradigm Shift Abstract ELT extract, load, and transform is replacing ETL (extract, transform, load) as the usual method of populating data warehouses. Modern data warehouse appliances
DATA MINING AND WAREHOUSING CONCEPTS
CHAPTER 1 DATA MINING AND WAREHOUSING CONCEPTS 1.1 INTRODUCTION The past couple of decades have seen a dramatic increase in the amount of information or data being stored in electronic format. This accumulation
Emerging Technologies Shaping the Future of Data Warehouses & Business Intelligence
Emerging Technologies Shaping the Future of Data Warehouses & Business Intelligence Appliances and DW Architectures John O Brien President and Executive Architect Zukeran Technologies 1 TDWI 1 Agenda What
High performance ETL Benchmark
High performance ETL Benchmark Author: Dhananjay Patil Organization: Evaltech, Inc. Evaltech Research Group, Data Warehousing Practice. Date: 07/02/04 Email: [email protected] Abstract: The IBM server iseries
Riverbed Whitewater/Amazon Glacier ROI for Backup and Archiving
Riverbed Whitewater/Amazon Glacier ROI for Backup and Archiving November, 2013 Saqib Jang Abstract This white paper demonstrates how to increase profitability by reducing the operating costs of backup
Many government agencies are requiring disclosure of security breaches. 32 states have security breach similar legislation
Is it safe? The business impact of data protection. Bruce Master IBM LTO Program Linear Tape-Open, LTO, LTO Logo, Ultrium and Ultrium Logo are trademarks of HP, IBM and Quantum in the US and other countries.
Department of Technology Services UNIX SERVICE OFFERING
Department of Technology Services UNIX SERVICE OFFERING BACKGROUND The Department of Technology Services (DTS) operates dozens of UNIX-based systems to meet the business needs of its customers. The services
James Serra Sr BI Architect [email protected] http://jamesserra.com/
James Serra Sr BI Architect [email protected] http://jamesserra.com/ Our Focus: Microsoft Pure-Play Data Warehousing & Business Intelligence Partner Our Customers: Our Reputation: "B.I. Voyage came
News and trends in Data Warehouse Automation, Big Data and BI. Johan Hendrickx & Dirk Vermeiren
News and trends in Data Warehouse Automation, Big Data and BI Johan Hendrickx & Dirk Vermeiren Extreme Agility from Source to Analysis DWH Appliances & DWH Automation Typical Architecture 3 What Business
Oracle BI EE Implementation on Netezza. Prepared by SureShot Strategies, Inc.
Oracle BI EE Implementation on Netezza Prepared by SureShot Strategies, Inc. The goal of this paper is to give an insight to Netezza architecture and implementation experience to strategize Oracle BI EE
Big Data and Its Impact on the Data Warehousing Architecture
Big Data and Its Impact on the Data Warehousing Architecture Sponsored by SAP Speaker: Wayne Eckerson, Director of Research, TechTarget Wayne Eckerson: Hi my name is Wayne Eckerson, I am Director of Research
Storage Design for High Capacity and Long Term Storage. DLF Spring Forum, Raleigh, NC May 6, 2009. Balancing Cost, Complexity, and Fault Tolerance
Storage Design for High Capacity and Long Term Storage Balancing Cost, Complexity, and Fault Tolerance DLF Spring Forum, Raleigh, NC May 6, 2009 Lecturer: Jacob Farmer, CTO Cambridge Computer Copyright
Extraction Transformation Loading ETL Get data out of sources and load into the DW
Lection 5 ETL Definition Extraction Transformation Loading ETL Get data out of sources and load into the DW Data is extracted from OLTP database, transformed to match the DW schema and loaded into the
Whitepaper. 4 Steps to Successfully Evaluating Business Analytics Software. www.sisense.com
Whitepaper 4 Steps to Successfully Evaluating Business Analytics Software Introduction The goal of Business Analytics and Intelligence software is to help businesses access, analyze and visualize data,
Well packaged sets of preinstalled, integrated, and optimized software on select hardware in the form of engineered systems and appliances
INSIGHT Oracle's All- Out Assault on the Big Data Market: Offering Hadoop, R, Cubes, and Scalable IMDB in Familiar Packages Carl W. Olofson IDC OPINION Global Headquarters: 5 Speen Street Framingham, MA
Data warehouse and Business Intelligence Collateral
Data warehouse and Business Intelligence Collateral Page 1 of 12 DATA WAREHOUSE AND BUSINESS INTELLIGENCE COLLATERAL Brains for the corporate brawn: In the current scenario of the business world, the competition
THE DATA WAREHOUSE ETL TOOLKIT CDT803 Three Days
Three Days Prerequisites Students should have at least some experience with any relational database management system. Who Should Attend This course is targeted at technical staff, team leaders and project
How To Make A Backup System More Efficient
Identifying the Hidden Risk of Data De-duplication: How the HYDRAstor Solution Proactively Solves the Problem October, 2006 Introduction Data de-duplication has recently gained significant industry attention,
CSCA0102 IT & Business Applications. Foundation in Business Information Technology School of Engineering & Computing Sciences FTMS College Global
CSCA0102 IT & Business Applications Foundation in Business Information Technology School of Engineering & Computing Sciences FTMS College Global Chapter 2 Data Storage Concepts System Unit The system unit
Archiving the insurance data warehouse
REFLECTIONS Archiving the insurance data warehouse Surajit Basu, IBEXI Solutions Surajit Basu is an IT consultant with over 20 years of experience in design, development and implementation of enterprise-wide
Scaling Your Data to the Cloud
ZBDB Scaling Your Data to the Cloud Technical Overview White Paper POWERED BY Overview ZBDB Zettabyte Database is a new, fully managed data warehouse on the cloud, from SQream Technologies. By building
Cleaning Up Your Outlook Mailbox and Keeping It That Way ;-) Mailbox Cleanup. Quicklinks >>
Cleaning Up Your Outlook Mailbox and Keeping It That Way ;-) Whether you are reaching the limit of your mailbox storage quota or simply want to get rid of some of the clutter in your mailbox, knowing where
THE DBMS SCALABILITY REPORT
THE DBMS SCALABILITY REPORT PERCEPTION VERSUS REALITY EXECUTIVE OVERVIEW David McGoveran Alternative Technologies 6221A Graham Hill Road, Suite 8001 Felton, California 95018 Telephone: 831/338-4621 Fax:
Bringing Big Data into the Enterprise
Bringing Big Data into the Enterprise Overview When evaluating Big Data applications in enterprise computing, one often-asked question is how does Big Data compare to the Enterprise Data Warehouse (EDW)?
In-Memory Analytics for Big Data
In-Memory Analytics for Big Data Game-changing technology for faster, better insights WHITE PAPER SAS White Paper Table of Contents Introduction: A New Breed of Analytics... 1 SAS In-Memory Overview...
STORAGE. Buying Guide: TARGET DATA DEDUPLICATION BACKUP SYSTEMS. inside
Managing the information that drives the enterprise STORAGE Buying Guide: DEDUPLICATION inside What you need to know about target data deduplication Special factors to consider One key difference among
Texas Digital Government Summit. Data Analysis Structured vs. Unstructured Data. Presented By: Dave Larson
Texas Digital Government Summit Data Analysis Structured vs. Unstructured Data Presented By: Dave Larson Speaker Bio Dave Larson Solu6ons Architect with Freeit Data Solu6ons In the IT industry for over
W H I T E P A P E R. Deriving Intelligence from Large Data Using Hadoop and Applying Analytics. Abstract
W H I T E P A P E R Deriving Intelligence from Large Data Using Hadoop and Applying Analytics Abstract This white paper is focused on discussing the challenges facing large scale data processing and the
The Data Warehouse ETL Toolkit
2008 AGI-Information Management Consultants May be used for personal purporses only or by libraries associated to dandelon.com network. The Data Warehouse ETL Toolkit Practical Techniques for Extracting,
Optimizing Large Arrays with StoneFly Storage Concentrators
Optimizing Large Arrays with StoneFly Storage Concentrators All trademark names are the property of their respective companies. This publication contains opinions of which are subject to change from time
Microsoft s SQL Server Parallel Data Warehouse Provides High Performance and Great Value
Microsoft s SQL Server Parallel Data Warehouse Provides High Performance and Great Value Published by: Value Prism Consulting Sponsored by: Microsoft Corporation Publish date: March 2013 Abstract: Data
Turnkey Hardware, Software and Cash Flow / Operational Analytics Framework
Turnkey Hardware, Software and Cash Flow / Operational Analytics Framework With relevant, up to date cash flow and operations optimization reporting at your fingertips, you re positioned to take advantage
Business Usage Monitoring for Teradata
Managing Big Analytic Data Business Usage Monitoring for Teradata Increasing Operational Efficiency and Reducing Data Management Costs How to Increase Operational Efficiency and Reduce Data Management
PARALLEL PROCESSING AND THE DATA WAREHOUSE
PARALLEL PROCESSING AND THE DATA WAREHOUSE BY W. H. Inmon One of the essences of the data warehouse environment is the accumulation of and the management of large amounts of data. Indeed, it is said that
Acme Corporation Enterprise Storage Assessment Prepared by:
Acme Corporation Enterprise Storage Assessment Prepared by: Rocket Software, Inc. 150 N. Mathilda Place, Suite 304 Sunnyvale, CA 94086 USA Tel# +1.650.237.6100 / Fax# +1.650.237.9183 Table of Contents
Flash Array Storage: Best Practices, Tips, and Advice from Real Users
Flash Array Storage: Best Practices, Tips, and Advice from Real Users i IT Central Station: Reviews of Enterprise Flash Storage and Flash Storage Best Practices Abstract The world of storage is being transformed
WHITE PAPER Linux Management with Red Hat Network Satellite Server: Measuring Business Impact and ROI
WHITE PAPER Linux Management with Red Hat Network Satellite Server: Measuring Business Impact and ROI Sponsored by: Red Hat Tim Grieser Randy Perry October 2009 Eric Hatcher Global Headquarters: 5 Speen
The New Economics of SAP Business Suite powered by SAP HANA. 2013 SAP AG. All rights reserved. 2
The New Economics of SAP Business Suite powered by SAP HANA 2013 SAP AG. All rights reserved. 2 COMMON MYTH Running SAP Business Suite on SAP HANA is more expensive than on a classical database 2013 2014
Hardware Configuration Guide
Hardware Configuration Guide Contents Contents... 1 Annotation... 1 Factors to consider... 2 Machine Count... 2 Data Size... 2 Data Size Total... 2 Daily Backup Data Size... 2 Unique Data Percentage...
Breaking the Storage Array Lifecycle with Cloud Storage
Breaking the Storage Array Lifecycle with Cloud Storage 2011 TwinStrata, Inc. The Storage Array Lifecycle Anyone who purchases storage arrays is familiar with the many advantages of modular storage systems
CREATING PACKAGED IP FOR BUSINESS ANALYTICS PROJECTS
CREATING PACKAGED IP FOR BUSINESS ANALYTICS PROJECTS A PERSPECTIVE FOR SYSTEMS INTEGRATORS Sponsored by Microsoft Corporation 1/ What is Packaged IP? Categorizing the Options 2/ Why Offer Packaged IP?
In-memory Tables Technology overview and solutions
In-memory Tables Technology overview and solutions My mainframe is my business. My business relies on MIPS. Verna Bartlett Head of Marketing Gary Weinhold Systems Analyst Agenda Introduction to in-memory
Beyond Conventional Data Warehousing. Florian Waas Greenplum Inc.
Beyond Conventional Data Warehousing Florian Waas Greenplum Inc. Takeaways The basics Who is Greenplum? What is Greenplum Database? The problem Data growth and other recent trends in DWH A look at different
Lection 3-4 WAREHOUSING
Lection 3-4 DATA WAREHOUSING Learning Objectives Understand d the basic definitions iti and concepts of data warehouses Understand data warehousing architectures Describe the processes used in developing
BUSINESS VALUE SPOTLIGHT
BUSINESS VALUE SPOTLIGHT Improve Data Management and Reduce Infrastructure Costs with Enterprise Application Archiving and Test Data Management: A Case Study of IKON February 2010 Sponsored by Informatica
pg. pg. pg. pg. pg. pg. Rationalizing Supplier Increases What is Predictive Analytics? Reducing Business Risk
What is Predictive Analytics? 3 Why is Predictive Analytics Important to Sourcing? 4 Rationalizing Supplier Increases 5 Better Control of Sourcing and Costs 6 Reducing Business Risk 7 How do you implement
An Accenture Point of View. Oracle Exalytics brings speed and unparalleled flexibility to business analytics
An Accenture Point of View Oracle Exalytics brings speed and unparalleled flexibility to business analytics Keep your competitive edge with analytics When it comes to working smarter, organizations that
Build an effective data integration strategy to drive innovation
IBM Software Thought Leadership White Paper September 2010 Build an effective data integration strategy to drive innovation Five questions business leaders must ask 2 Build an effective data integration
Database Schema Management
Whitemarsh Information Systems Corporation 2008 Althea Lane Bowie, Maryland 20716 Tele: 301-249-1142 Email: [email protected] Web: www.wiscorp.com Table of Contents 1. Objective...1 2. Topics Covered...2
LDA, the new family of Lortu Data Appliances
LDA, the new family of Lortu Data Appliances Based on Lortu Byte-Level Deduplication Technology February, 2011 Copyright Lortu Software, S.L. 2011 1 Index Executive Summary 3 Lortu deduplication technology
<Insert Picture Here> Oracle and/or Hadoop And what you need to know
Oracle and/or Hadoop And what you need to know Jean-Pierre Dijcks Data Warehouse Product Management Agenda Business Context An overview of Hadoop and/or MapReduce Choices, choices,
Credit Card Processing 101
Credit Card Processing 101 Customers have come to expect credit cards as a payment option. With ATM fees continuing to rise, some consumers may even exclusively choose to take their purchasing power to
Knowledge Base Data Warehouse Methodology
Knowledge Base Data Warehouse Methodology Knowledge Base's data warehousing services can help the client with all phases of understanding, designing, implementing, and maintaining a data warehouse. This
TRENDS IN THE DEVELOPMENT OF BUSINESS INTELLIGENCE SYSTEMS
9 8 TRENDS IN THE DEVELOPMENT OF BUSINESS INTELLIGENCE SYSTEMS Assist. Prof. Latinka Todoranova Econ Lit C 810 Information technology is a highly dynamic field of research. As part of it, business intelligence
HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW
HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW 757 Maleta Lane, Suite 201 Castle Rock, CO 80108 Brett Weninger, Managing Director [email protected] Dave Smelker, Managing Principal [email protected]
Big Data Open Source Stack vs. Traditional Stack for BI and Analytics
Big Data Open Source Stack vs. Traditional Stack for BI and Analytics Part I By Sam Poozhikala, Vice President Customer Solutions at StratApps Inc. 4/4/2014 You may contact Sam Poozhikala at [email protected].
POLAR IT SERVICES. Business Intelligence Project Methodology
POLAR IT SERVICES Business Intelligence Project Methodology Table of Contents 1. Overview... 2 2. Visualize... 3 3. Planning and Architecture... 4 3.1 Define Requirements... 4 3.1.1 Define Attributes...
White Paper. Recording Server Virtualization
White Paper Recording Server Virtualization Prepared by: Mike Sherwood, Senior Solutions Engineer Milestone Systems 23 March 2011 Table of Contents Introduction... 3 Target audience and white paper purpose...
Real-time Data Replication
Real-time Data Replication from Oracle to other databases using DataCurrents WHITEPAPER Contents Data Replication Concepts... 2 Real time Data Replication... 3 Heterogeneous Data Replication... 4 Different
Why Nuix doesn t believe in Magic
Why Nuix doesn t believe in Magic Nuix has participated in the Gartner Magic Quadrant for ediscovery Software for the past four years. Over that time, the ediscovery market has changed considerably. Unfortunately,
Offload Enterprise Data Warehouse (EDW) to Big Data Lake. Ample White Paper
Offload Enterprise Data Warehouse (EDW) to Big Data Lake Oracle Exadata, Teradata, Netezza and SQL Server Ample White Paper EDW (Enterprise Data Warehouse) Offloads The EDW (Enterprise Data Warehouse)
SQL Server PDW. Artur Vieira Premier Field Engineer
SQL Server PDW Artur Vieira Premier Field Engineer Agenda 1 Introduction to MPP and PDW 2 PDW Architecture and Components 3 Data Structures 4 PDW Tools Data Load / Data Output / Administrative Console
Hadoop and Relational Database The Best of Both Worlds for Analytics Greg Battas Hewlett Packard
Hadoop and Relational base The Best of Both Worlds for Analytics Greg Battas Hewlett Packard The Evolution of Analytics Mainframe EDW Proprietary MPP Unix SMP MPP Appliance Hadoop? Questions Is Hadoop
EMC BACKUP MEETS BIG DATA
EMC BACKUP MEETS BIG DATA Strategies To Protect Greenplum, Isilon And Teradata Systems 1 Agenda Big Data: Overview, Backup and Recovery EMC Big Data Backup Strategy EMC Backup and Recovery Solutions for
Big Data and Big Data Modeling
Big Data and Big Data Modeling The Age of Disruption Robin Bloor The Bloor Group March 19, 2015 TP02 Presenter Bio Robin Bloor, Ph.D. Robin Bloor is Chief Analyst at The Bloor Group. He has been an industry
Availability Digest. www.availabilitydigest.com. Data Deduplication February 2011
the Availability Digest Data Deduplication February 2011 What is Data Deduplication? Data deduplication is a technology that can reduce disk storage-capacity requirements and replication bandwidth requirements
(Refer Slide Time: 01:52)
Software Engineering Prof. N. L. Sarda Computer Science & Engineering Indian Institute of Technology, Bombay Lecture - 2 Introduction to Software Engineering Challenges, Process Models etc (Part 2) This
SAS and Oracle: Big Data and Cloud Partnering Innovation Targets the Third Platform
SAS and Oracle: Big Data and Cloud Partnering Innovation Targets the Third Platform David Lawler, Oracle Senior Vice President, Product Management and Strategy Paul Kent, SAS Vice President, Big Data What
Oracle on Oracle. Hans Peter Kipfer Vice President, Engineered Systems EMEA
Oracle on Oracle Hans Peter Kipfer Vice President, Engineered Systems EMEA MORE COMPLEXITY MEANS LESS INNOVATION IT SPENDING DISTRIBUTION WHAT IF 50% 66% RUN THE BUSINESS 20% 25% GROW THE BUSINESS 14%
Architecting for Big Data Analytics and Beyond: A New Framework for Business Intelligence and Data Warehousing
Architecting for Big Data Analytics and Beyond: A New Framework for Business Intelligence and Data Warehousing Wayne W. Eckerson Director of Research, TechTarget Founder, BI Leadership Forum Business Analytics
Safe Harbor Statement
Safe Harbor Statement "Safe Harbor" Statement: Statements in this presentation relating to Oracle's future plans, expectations, beliefs, intentions and prospects are "forward-looking statements" and are
Dell Information Management solutions
Dell Information Management solutions Uday Tekumalla Solutions Marketing, Information Management 1 10/28/2013 Information Management Solutions My introduction Uday Tekumalla, the ponytail guy Information
SAN vs. NAS: The Critical Decision
SAN vs. NAS: The Critical Decision Executive Summary The storage strategy for your organization is dictated by many factors: the nature of the documents and files you need to store, the file usage patterns
Data Deduplication in Tivoli Storage Manager. Andrzej Bugowski 19-05-2011 Spała
Data Deduplication in Tivoli Storage Manager Andrzej Bugowski 19-05-2011 Spała Agenda Tivoli Storage, IBM Software Group Deduplication concepts Data deduplication in TSM 6.1 Planning for data deduplication
