Estimating Deduplication Ratios in Large Data Sets
|
|
|
- Philip Randall
- 10 years ago
- Views:
Transcription
1 IBM Research labs - Haifa Estimating Deduplication Ratios in Large Data Sets Danny Harnik, Oded Margalit, Dalit Naor, Dmitry Sotnikov Gil Vernik
2 Estimating dedupe and compression ratios some motivation Data reduction does not come for free Incurs overheads, sometimes significant Not always worth the effort Better to know in advance Different techniques give different ratios What technique to use? Chunks? Fixed? Variable-sized? Full-file? Compression? Sometimes better to consolidate storage pools, sometimes not Different data reduction ratio: different number of disks to buy Disks = money!
3 How do you estimate efficiently? Option 1: according to data type, application type, etc Can be grossly in-accurate I ve seen the same DB application with dedupe ratio 1:2 and 1:50 Better to actually look at the data. Option 2: Sample a small portion of the data set and deduce from it Problematic when deduplication is involved
4 Data set == picture Identical block == identical picture block In real life we don t see the full data set at once Rather, we probe locations In sampling we only probe a small number of locations
5 Why sampling is problematic for dedupe estimation Dedupe ratio is 1:2 In order to see the duplicates must hit the exact same location twice Birthday paradox need ~N ½ elements to hit a collision Ω(N ½ ) to see a large number of collisions. What about triple collisions?
6 Formal limitations of sampling for Distinct Elements Lower bounds: Can show 2 data sets with far apart dedupe ratio, but same behavior under sampling of O(N 1-ε ) elements Charikar, Chaudhuri, Motwani, & Narasayya showed basic bounds and empirical failures of approximating distinct elements. Rashkodnikova, Ron, Shpilka & Smith 2009 need Ω(N 1-ε ) samples to approximate to within a multiplicative factor. Valiant & Valiant need Ω(N / logn) samples for same task. Bottom line: need to look at essentially he whole data set
7 Algorithms that see the whole data set : challenges Still a hard task High resources requirements: memory, time, disk accesses Naïve approach: simulate the actual dedupe just don t store the data Problem: simply storing the table of hash indices is too big for memory E.g. [Meyer & Bolosky, FAST 11] Example: metadata for 7 TB of data 24 GB of metadata Using an efficient dynamic data structure will require additional overhead Many solutions, none easy Goal 1: find a memory efficient estimation scheme with high accuracy Goal 2: Be as efficient as possible (time & CPU). More pronounced when need to integrate local (LZ) compression with deduplication
8 Plan for rest of talk Our Algorithm Integrating compression Analysis both formal and empirical Related work Full-file deduplication a special case
9 Our Algorithm: Sample & Scan Sample phase create a base sample and store in memory Scan phase Count appearances only of elements in the base sample Summary Average of the base sample tallies too give estimation
10 What about compression? Add compression evaluation only to elements in the base sample Only in the sample phase Crucial since compression is a time consuming operation In the summary take the compression stats into account Note: estimating dedupe and compression ratios separately is not sufficient for accurate estimation of the combination
11 Formal Analysis We prove a relation between the size of the base sample and the accuracy achieved for a given confidence parameter Accuracy by how much we may error Confidence with what probability Intuition: Each block in the data-set has a contribution to the overall reduced output Example: Suppose a block s dedupe count is 4 then each block contributes ¼ of a block to the overall output We attempt to estimate the average contribution over the whole data set Average contribution = total output size / total input size Estimating of averages is well studied Our estimation should form a normal distribution around the actual value
12 Example of Size of Sample vs Error and Confidence Dedupe Ratio Error Confidence Sample Size Memory 3: MB 5: MB 15: MB Error is a percentage of the output size Good dedupe small output 1% error is much smaller in absolute values This is why some bound on dedupe ratio is required If error percentage of original size is acceptable, then life is easier and memory requirements become even smaller
13 Empirical Analysis Evaluated our estimation on 4 real life data-sets: 1. Workstation data 2. File repository of an organization 3. Backup repository 4. Exchange DB periodical backups Largest data set was the file repository with 7TBs of data The tests back-up our formal analysis
14 Convergence of the error percentage as base sample grows
15 Error is indeed normally distributed around the actual ratio
16 Related Work Distinct elements is a heavily studied topic DB analysis Streaming algorithms [FM85],,[AMS99],,[KNW2010], many more Distinct elemnts works: Focus on one-pass algorithms Always fixed size elements not files or variable sized blocks Do not consider compression Our work: not one pass Comparable to most works (memory wise) When combined with compression we give the best performance (minimal number of chunks to compress)
17 Full-file deduplication Files have properties and metadata that can help us Different file length no dedupe Different hash on first block no dedupe Our algorithm: In sample phase keep also the file length and first block hash of base sample During scan phase: Check if file size relevant to base sample, if not, discard Check if first block hash relevant to base sample, if not, discard If still relevant, then read whole file from disk Overall: all metadata is scanned, not all files read!
18 Full-file estimation: Percentage of data read from disk vs. desired error gurantee
19 What to take home from this talk? 1. Need to be careful when doing sampling for dedupe 2. We have a good algorithm if you can run a full scan or already have metadata available. 3. Our algorithm integrates compression and dedupe naturally Practically no overhead when adding compression 4. For full-file deduplication we reduce the data reads substantially!
20 Thank You!
Data Reduction: Deduplication and Compression. Danny Harnik IBM Haifa Research Labs
Data Reduction: Deduplication and Compression Danny Harnik IBM Haifa Research Labs Motivation Reducing the amount of data is a desirable goal Data reduction: an attempt to compress the huge amounts of
STORAGE. Buying Guide: TARGET DATA DEDUPLICATION BACKUP SYSTEMS. inside
Managing the information that drives the enterprise STORAGE Buying Guide: DEDUPLICATION inside What you need to know about target data deduplication Special factors to consider One key difference among
Data Deduplication and Tivoli Storage Manager
Data Deduplication and Tivoli Storage Manager Dave Cannon Tivoli Storage Manager rchitect Oxford University TSM Symposium September 2007 Disclaimer This presentation describes potential future enhancements
SDGen: Mimicking Datasets for Content Generation in Storage Benchmarks
SDGen: Mimicking Datasets for Content Generation in Storage Benchmarks Raúl Gracia-Tinedo (Universitat Rovira i Virgili, Spain) Danny Harnik, Dalit Naor, Dmitry Sotnikov (IBM Research-Haifa, Israel) Sivan
Deduplication Demystified: How to determine the right approach for your business
Deduplication Demystified: How to determine the right approach for your business Presented by Charles Keiper Senior Product Manager, Data Protection Quest Software Session Objective: To answer burning
Data De-duplication Methodologies: Comparing ExaGrid s Byte-level Data De-duplication To Block Level Data De-duplication
Data De-duplication Methodologies: Comparing ExaGrid s Byte-level Data De-duplication To Block Level Data De-duplication Table of Contents Introduction... 3 Shortest Possible Backup Window... 3 Instant
Data Deduplication HTBackup
Data Deduplication HTBackup HTBackup and it s Deduplication technology is touted as one of the best ways to manage today's explosive data growth. If you're new to the technology, these key facts will help
HP StoreOnce D2D. Understanding the challenges associated with NetApp s deduplication. Business white paper
HP StoreOnce D2D Understanding the challenges associated with NetApp s deduplication Business white paper Table of contents Challenge #1: Primary deduplication: Understanding the tradeoffs...4 Not all
A Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique
A Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique Jyoti Malhotra 1,Priya Ghyare 2 Associate Professor, Dept. of Information Technology, MIT College of
Hardware Configuration Guide
Hardware Configuration Guide Contents Contents... 1 Annotation... 1 Factors to consider... 2 Machine Count... 2 Data Size... 2 Data Size Total... 2 Daily Backup Data Size... 2 Unique Data Percentage...
09'Linux Plumbers Conference
09'Linux Plumbers Conference Data de duplication Mingming Cao IBM Linux Technology Center [email protected] 2009 09 25 Current storage challenges Our world is facing data explosion. Data is growing in a amazing
Data Backup and Archiving with Enterprise Storage Systems
Data Backup and Archiving with Enterprise Storage Systems Slavjan Ivanov 1, Igor Mishkovski 1 1 Faculty of Computer Science and Engineering Ss. Cyril and Methodius University Skopje, Macedonia [email protected],
HP StoreOnce & Deduplication Solutions Zdenek Duchoň Pre-sales consultant
DISCOVER HP StoreOnce & Deduplication Solutions Zdenek Duchoň Pre-sales consultant HP StorageWorks Data Protection Solutions HP has it covered Near continuous data protection Disk Mirroring Advanced Backup
IMPLEMENTATION OF SOURCE DEDUPLICATION FOR CLOUD BACKUP SERVICES BY EXPLOITING APPLICATION AWARENESS
IMPLEMENTATION OF SOURCE DEDUPLICATION FOR CLOUD BACKUP SERVICES BY EXPLOITING APPLICATION AWARENESS Nehal Markandeya 1, Sandip Khillare 2, Rekha Bagate 3, Sayali Badave 4 Vaishali Barkade 5 12 3 4 5 (Department
The Curious Case of Database Deduplication. PRESENTATION TITLE GOES HERE Gurmeet Goindi Oracle
The Curious Case of Database Deduplication PRESENTATION TITLE GOES HERE Gurmeet Goindi Oracle Agenda Introduction Deduplication Databases and Deduplication All Flash Arrays and Deduplication 2 Quick Show
Understanding the Economics of Flash Storage
Understanding the Economics of Flash Storage By James Green, vexpert Virtualization Consultant and Scott D. Lowe, vexpert Co-Founder, ActualTech Media February, 2015 Table of Contents Table of Contents...
MAD2: A Scalable High-Throughput Exact Deduplication Approach for Network Backup Services
MAD2: A Scalable High-Throughput Exact Deduplication Approach for Network Backup Services Jiansheng Wei, Hong Jiang, Ke Zhou, Dan Feng School of Computer, Huazhong University of Science and Technology,
Everything you need to know about flash storage performance
Everything you need to know about flash storage performance The unique characteristics of flash make performance validation testing immensely challenging and critically important; follow these best practices
WHITE PAPER. Storage Savings Analysis: Storage Savings with Deduplication and Acronis Backup & Recovery 10
Storage Savings Analysis: Storage Savings with Deduplication and Acronis Backup & Recovery 10 Copyright Acronis, Inc., 2000 2009 Table of contents Executive Summary... 3 The Importance of Deduplication...
WAN Optimized Replication of Backup Datasets Using Stream-Informed Delta Compression
WAN Optimized Replication of Backup Datasets Using Stream-Informed Delta Compression Philip Shilane, Mark Huang, Grant Wallace, and Windsor Hsu Backup Recovery Systems Division EMC Corporation Abstract
Data Deduplication and Tivoli Storage Manager
Data Deduplication and Tivoli Storage Manager Dave annon Tivoli Storage Manager rchitect March 2009 Topics Tivoli Storage, IM Software Group Deduplication technology Data reduction and deduplication in
3Gen Data Deduplication Technical
3Gen Data Deduplication Technical Discussion NOTICE: This White Paper may contain proprietary information protected by copyright. Information in this White Paper is subject to change without notice and
Backup Software Data Deduplication: What you need to know. Presented by W. Curtis Preston Executive Editor & Independent Backup Expert
Backup Software Data Deduplication: What you need to know Presented by W. Curtis Preston Executive Editor & Independent Backup Expert When I was in the IT Department When I started as backup guy at $35B
Speeding Up Cloud/Server Applications Using Flash Memory
Speeding Up Cloud/Server Applications Using Flash Memory Sudipta Sengupta Microsoft Research, Redmond, WA, USA Contains work that is joint with B. Debnath (Univ. of Minnesota) and J. Li (Microsoft Research,
Data Deduplication in Tivoli Storage Manager. Andrzej Bugowski 19-05-2011 Spała
Data Deduplication in Tivoli Storage Manager Andrzej Bugowski 19-05-2011 Spała Agenda Tivoli Storage, IBM Software Group Deduplication concepts Data deduplication in TSM 6.1 Planning for data deduplication
Data Reduction Methodologies: Comparing ExaGrid s Byte-Level-Delta Data Reduction to Data De-duplication. February 2007
Data Reduction Methodologies: Comparing ExaGrid s Byte-Level-Delta Data Reduction to Data De-duplication February 2007 Though data reduction technologies have been around for years, there is a renewed
CA ARCserve r16.0 - Data Deduplication Frequently Asked Questions
CA ARCserve r16.0 - Data Deduplication Frequently Asked Questions Table of Contents For any backup on to Deduplication device, how many files does ARCserve generate and what do they contain?...4 Can I
Deploying De-Duplication on Ext4 File System
Deploying De-Duplication on Ext4 File System Usha A. Joglekar 1, Bhushan M. Jagtap 2, Koninika B. Patil 3, 1. Asst. Prof., 2, 3 Students Department of Computer Engineering Smt. Kashibai Navale College
Demystifying Deduplication for Backup with the Dell DR4000
Demystifying Deduplication for Backup with the Dell DR4000 This Dell Technical White Paper explains how deduplication with the DR4000 can help your organization save time, space, and money. John Bassett
Top Ten Questions. to Ask Your Primary Storage Provider About Their Data Efficiency. May 2014. Copyright 2014 Permabit Technology Corporation
Top Ten Questions to Ask Your Primary Storage Provider About Their Data Efficiency May 2014 Copyright 2014 Permabit Technology Corporation Introduction The value of data efficiency technologies, namely
Cloud De-duplication Cost Model THESIS
Cloud De-duplication Cost Model THESIS Presented in Partial Fulfillment of the Requirements for the Degree Master of Science in the Graduate School of The Ohio State University By Christopher Scott Hocker
Seriously: Tape Only Backup Systems are Dead, Dead, Dead!
Seriously: Tape Only Backup Systems are Dead, Dead, Dead! Agenda Overview Tape backup rule #1 So what s the problem? Intelligent disk targets Disk-based backup software Overview We re still talking disk
A Survey on Deduplication Strategies and Storage Systems
A Survey on Deduplication Strategies and Storage Systems Guljar Shaikh ((Information Technology,B.V.C.O.E.P/ B.V.C.O.E.P, INDIA) Abstract : Now a day there is raising demands for systems which provide
FAST 11. Yongseok Oh <[email protected]> University of Seoul. Mobile Embedded System Laboratory
CAFTL: A Content-Aware Flash Translation Layer Enhancing the Lifespan of flash Memory based Solid State Drives FAST 11 Yongseok Oh University of Seoul Mobile Embedded System Laboratory
Security of Cloud Storage: - Deduplication vs. Privacy
Security of Cloud Storage: - Deduplication vs. Privacy Benny Pinkas - Bar Ilan University Shai Halevi, Danny Harnik, Alexandra Shulman-Peleg - IBM Research Haifa 1 Remote storage and security Easy to encrypt
Reducing Replication Bandwidth for Distributed Document Databases
Reducing Replication Bandwidth for Distributed Document Databases Lianghong Xu 1, Andy Pavlo 1, Sudipta Sengupta 2 Jin Li 2, Greg Ganger 1 Carnegie Mellon University 1, Microsoft Research 2 #1 You can
Building a High Performance Deduplication System Fanglu Guo and Petros Efstathopoulos
Building a High Performance Deduplication System Fanglu Guo and Petros Efstathopoulos Symantec Research Labs Symantec FY 2013 (4/1/2012 to 3/31/2013) Revenue: $ 6.9 billion Segment Revenue Example Business
Inline Deduplication
Inline Deduplication [email protected] 1.1 Inline Vs Post-process Deduplication In target based deduplication, the deduplication engine can either process data for duplicates in real time (i.e.
Avamar. Technology Overview
1 Technology Overview 2 Growth of Remote Offices Remote offices are growing rapidly Over 4 million remote offices in the US No standard remote office size, data, staff, network, etc. The average branch
Veeam Best Practices with Exablox
Veeam Best Practices with Exablox Overview Exablox has worked closely with the team at Veeam to provide the best recommendations when using the the Veeam Backup & Replication software with OneBlox appliances.
Enterprise-class Backup Performance with Dell DR6000 Date: May 2014 Author: Kerry Dolan, Lab Analyst and Vinny Choinski, Senior Lab Analyst
ESG Lab Review Enterprise-class Backup Performance with Dell DR6000 Date: May 2014 Author: Kerry Dolan, Lab Analyst and Vinny Choinski, Senior Lab Analyst Abstract: This ESG Lab review documents hands-on
LDA, the new family of Lortu Data Appliances
LDA, the new family of Lortu Data Appliances Based on Lortu Byte-Level Deduplication Technology February, 2011 Copyright Lortu Software, S.L. 2011 1 Index Executive Summary 3 Lortu deduplication technology
NETAPP WHITE PAPER Looking Beyond the Hype: Evaluating Data Deduplication Solutions
NETAPP WHITE PAPER Looking Beyond the Hype: Evaluating Data Deduplication Solutions Larry Freeman, Network Appliance, Inc. September 2007 WP-7028-0907 Table of Contents The Deduplication Hype 3 What Is
Protect Data... in the Cloud
QUASICOM Private Cloud Backups with ExaGrid Deduplication Disk Arrays Martin Lui Senior Solution Consultant Quasicom Systems Limited Protect Data...... in the Cloud 1 Mobile Computing Users work with their
Read Performance Enhancement In Data Deduplication For Secondary Storage
Read Performance Enhancement In Data Deduplication For Secondary Storage A THESIS SUBMITTED TO THE FACULTY OF THE GRADUATE SCHOOL OF THE UNIVERSITY OF MINNESOTA BY Pradeep Ganesan IN PARTIAL FULFILLMENT
Quanqing XU [email protected]. YuruBackup: A Highly Scalable and Space-Efficient Incremental Backup System in the Cloud
Quanqing XU [email protected] YuruBackup: A Highly Scalable and Space-Efficient Incremental Backup System in the Cloud Outline Motivation YuruBackup s Architecture Backup Client File Scan, Data
DEXT3: Block Level Inline Deduplication for EXT3 File System
DEXT3: Block Level Inline Deduplication for EXT3 File System Amar More M.A.E. Alandi, Pune, India [email protected] Zishan Shaikh M.A.E. Alandi, Pune, India [email protected] Vishal Salve
Enterprise Backup and Restore technology and solutions
Enterprise Backup and Restore technology and solutions LESSON VII Veselin Petrunov Backup and Restore team / Deep Technical Support HP Bulgaria Global Delivery Hub Global Operations Center November, 2013
Data Deduplication: An Essential Component of your Data Protection Strategy
WHITE PAPER: THE EVOLUTION OF DATA DEDUPLICATION Data Deduplication: An Essential Component of your Data Protection Strategy JULY 2010 Andy Brewerton CA TECHNOLOGIES RECOVERY MANAGEMENT AND DATA MODELLING
In-Memory Databases Algorithms and Data Structures on Modern Hardware. Martin Faust David Schwalb Jens Krüger Jürgen Müller
In-Memory Databases Algorithms and Data Structures on Modern Hardware Martin Faust David Schwalb Jens Krüger Jürgen Müller The Free Lunch Is Over 2 Number of transistors per CPU increases Clock frequency
Aspirus Enterprise Backup Assessment and Implementation of Avamar and NetWorker
Aspirus Enterprise Backup Assessment and Implementation of Avamar and NetWorker Written by: Thomas Whalen Server and Storage Infrastructure Team Leader, Aspirus Information Technology Department Executive
An Authorized Duplicate Check Scheme for Removing Duplicate Copies of Repeating Data in The Cloud Environment to Reduce Amount of Storage Space
An Authorized Duplicate Check Scheme for Removing Duplicate Copies of Repeating Data in The Cloud Environment to Reduce Amount of Storage Space Jannu.Prasanna Krishna M.Tech Student, Department of CSE,
SDFS Overview. By Sam Silverberg
SDFS Overview By Sam Silverberg Why did I do this? I had an Idea that I needed to see if it worked. Design Goals Create a dedup file system capable of effective inline deduplication for Virtual Machines
Reference Guide WindSpring Data Management Technology (DMT) Solving Today s Storage Optimization Challenges
Reference Guide WindSpring Data Management Technology (DMT) Solving Today s Storage Optimization Challenges September 2011 Table of Contents The Enterprise and Mobile Storage Landscapes... 3 Increased
efficient protection, and impact-less!!
Converged, Hyper- or Flash Sample photo. Replace if desired. efficient protection, and impact-less!! Bogdan Stefanescu (..aka Bogs) EMC Data Protection Solutions [email protected] 1 ALL DATA HAS
A SCALABLE DEDUPLICATION AND GARBAGE COLLECTION ENGINE FOR INCREMENTAL BACKUP
A SCALABLE DEDUPLICATION AND GARBAGE COLLECTION ENGINE FOR INCREMENTAL BACKUP Dilip N Simha (Stony Brook University, NY & ITRI, Taiwan) Maohua Lu (IBM Almaden Research Labs, CA) Tzi-cker Chiueh (Stony
CloudBerry Dedup Server
Getting Started Guide: CloudBerry Dedup Server Getting Started Guide Getting Started with CloudBerry Dedup Server We are glad to announce our new product CloudBerry Dedup Server. CloudBerry Dedup is a
STORAGE SOURCE DATA DEDUPLICATION PRODUCTS. Buying Guide: inside
Managing the information that drives the enterprise STORAGE Buying Guide: inside 2 Key features of source data deduplication products 5 Special considerations Source dedupe products can efficiently protect
Hyper-converged IT drives: - TCO cost savings - data protection - amazing operational excellence
Hyper-converged IT drives: - TCO cost savings - data protection - amazing operational excellence Sebastian Nowicki SimpliVity is one of the biggest innovations in enterprise computing since ware. ~John
A Survey on Aware of Local-Global Cloud Backup Storage for Personal Purpose
A Survey on Aware of Local-Global Cloud Backup Storage for Personal Purpose Abhirupa Chatterjee 1, Divya. R. Krishnan 2, P. Kalamani 3 1,2 UG Scholar, Sri Sairam College Of Engineering, Bangalore. India
A block based storage model for remote online backups in a trust no one environment
A block based storage model for remote online backups in a trust no one environment http://www.duplicati.com/ Kenneth Skovhede (author, [email protected]) René Stach (editor, [email protected]) Abstract
[email protected]
www.ijecs.in International Journal Of Engineering And Computer Science ISSN:2319-7242 Volume 4 Issue 3 March 2015, Page No. 10715-10720 Data DeDuplication Using Optimized Fingerprint Lookup Method for
Data deduplication is more than just a BUZZ word
Data deduplication is more than just a BUZZ word Per Larsen Principal Systems Engineer Mr. Hansen DATA BUDGET RECOVERY & DATACENTER GROWTH PRESSURE DISCOVERY REVOLUTION More Storage Longer Backups Smaller
HP StoreOnce: reinventing data deduplication
HP : reinventing data deduplication Reduce the impact of explosive data growth with HP StorageWorks D2D Backup Systems Technical white paper Table of contents Executive summary... 2 Introduction to data
Data Deduplication in a Hybrid Architecture for Improving Write Performance
Data Deduplication in a Hybrid Architecture for Improving Write Performance Data-intensive Salable Computing Laboratory Department of Computer Science Texas Tech University Lubbock, Texas June 10th, 2013
EMC VNX2 Deduplication and Compression
White Paper VNX5200, VNX5400, VNX5600, VNX5800, VNX7600, & VNX8000 Maximizing effective capacity utilization Abstract This white paper discusses the capacity optimization technologies delivered in the
Theoretical Aspects of Storage Systems Autumn 2009
Theoretical Aspects of Storage Systems Autumn 2009 Chapter 3: Data Deduplication André Brinkmann News Outline Data Deduplication Compare-by-hash strategies Delta-encoding based strategies Measurements
bup: the git-based backup system Avery Pennarun
bup: the git-based backup system Avery Pennarun 2010 10 25 The Challenge Back up entire filesystems (> 1TB) Including huge VM disk images (files >100GB) Lots of separate files (500k or more) Calculate/store
How To Make A Backup System More Efficient
Identifying the Hidden Risk of Data De-duplication: How the HYDRAstor Solution Proactively Solves the Problem October, 2006 Introduction Data de-duplication has recently gained significant industry attention,
Tandberg Data AccuVault RDX
Tandberg Data AccuVault RDX Binary Testing conducts an independent evaluation and performance test of Tandberg Data s latest small business backup appliance. Data backup is essential to their survival
TSM Family Capacity Pricing Special Bid Offering. 2009 IBM Corporation
TSM Family Capacity Pricing Special Bid Offering 2009 IBM Corporation Agenda TSM Family Capacity Pricing Special Bid Offering overview IBM Direct Rep - how to order IBM Business Partner how to order How
A Novel Cloud Based Elastic Framework for Big Data Preprocessing
School of Systems Engineering A Novel Cloud Based Elastic Framework for Big Data Preprocessing Omer Dawelbeit and Rachel McCrindle October 21, 2014 University of Reading 2008 www.reading.ac.uk Overview
Availability Digest. www.availabilitydigest.com. Data Deduplication February 2011
the Availability Digest Data Deduplication February 2011 What is Data Deduplication? Data deduplication is a technology that can reduce disk storage-capacity requirements and replication bandwidth requirements
LEVERAGING FLASH MEMORY in ENTERPRISE STORAGE. Matt Kixmoeller, Pure Storage
LEVERAGING FLASH MEMORY in ENTERPRISE STORAGE Matt Kixmoeller, Pure Storage SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA unless otherwise noted. Member companies
Assuring Demanded Read Performance of Data Deduplication Storage with Backup Datasets
Assuring Demanded Read Performance of Data Deduplication Storage with Backup Datasets Young Jin Nam School of Computer and Information Technology Daegu University Gyeongsan, Gyeongbuk, KOREA 7-7 Email:
DEDUPLICATION NOW AND WHERE IT S HEADING. Lauren Whitehouse Senior Analyst, Enterprise Strategy Group
DEDUPLICATION NOW AND WHERE IT S HEADING Lauren Whitehouse Senior Analyst, Enterprise Strategy Group Need Dedupe? Before/After Dedupe Deduplication Production Data Deduplication In Backup Process Backup
Target Deduplication Metrics and Risk Analysis Using Post Processing Methods
Target Deduplication Metrics and Risk Analysis Using Post Processing Methods Gayathri.R 1, 1 Dr. Malathi.A 2 1 Assistant Professor, 2 Assistant Professor 1 School of IT and Science, 2 PG and Research Department
Big Data & Scripting Part II Streaming Algorithms
Big Data & Scripting Part II Streaming Algorithms 1, Counting Distinct Elements 2, 3, counting distinct elements problem formalization input: stream of elements o from some universe U e.g. ids from a set
Creating a Cloud Backup Service. Deon George
Creating a Cloud Backup Service Deon George Agenda TSM Cloud Service features Cloud Service Customer, providing a internal backup service Internal Backup Cloud Service Service Provider, providing a backup
Tiered Data Protection Strategy Data Deduplication. Thomas Störr Sales Director Central Europe November 8, 2007
Tiered Data Protection Strategy Data Deduplication Thomas Störr Sales Director Central Europe November 8, 2007 Overland Storage Tiered Data Protection = Good = Better = Best! NEO / ARCvault REO w/ expansion
PLC-Cache: Endurable SSD Cache for Deduplication-based Primary Storage
PLC-Cache: Endurable SSD Cache for Deduplication-based Primary Storage Jian Liu, Yunpeng Chai, Yuan Xiao Renmin University of China Xiao Qin Auburn University Speaker: Tao Xie MSST, June 5, 2014 Deduplication
Data Compression and Deduplication. LOC 2010 2010 Cisco Systems, Inc. All rights reserved.
Data Compression and Deduplication LOC 2010 2010 Systems, Inc. All rights reserved. 1 Data Redundancy Elimination Landscape VMWARE DeDE IBM DDE for Tank Solaris ZFS Hosts (Inline and Offline) MDS + Network
idedup: Latency-aware, inline data deduplication for primary storage
idedup: Latency-aware, inline data deduplication for primary storage Kiran Srinivasan, Tim Bisson, Garth Goodson, Kaladhar Voruganti NetApp, Inc. {skiran, tbisson, goodson, kaladhar}@netapp.com Abstract
EMC XTREMIO EXECUTIVE OVERVIEW
EMC XTREMIO EXECUTIVE OVERVIEW COMPANY BACKGROUND XtremIO develops enterprise data storage systems based completely on random access media such as flash solid-state drives (SSDs). By leveraging the underlying
WHITE PAPER. Effectiveness of Variable-block vs Fixedblock Deduplication on Data Reduction: A Technical Analysis
WHITE PAPER Effectiveness of Variable-block vs Fixedblock Deduplication on Data Reduction: A Technical Analysis CONTENTS Executive Summary... 3 Fixed vs. Variable-block Deduplication... 3 Test Configuration...
