Choosing an Enterprise-Class Deduplication Technology
|
|
- Lawrence Atkins
- 8 years ago
- Views:
Transcription
1 WHITE PAPER Choosing an Enterprise-Class Deduplication Technology 10 Key Questions to Ask Your Deduplication Vendor 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F:
2 Copyright Sepaton, Inc. All rights reserved. Sepaton, S2100, DeltaStor, DeltaRemote, and SRE are registered trademarks and ContentAware, Secure Erasure, and DeltaScale are trademarks of Sepaton, Inc. Other product and company names mentioned herein are or may be trademarks and/or registered trademarks of their respective companies v2 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F:
3 Deduplication Specifically Designed for Enterprises The volume of data generated by most companies has grown at such an explosive rate that many data centers have simply run out of space, power, cooling, and storage capacity to handle it. In large enterprise organizations, the sheer volume and variety of data to be protected requires a level of performance and scalability that few data protection technologies can deliver. Fundamental issues of insufficient capacity are being compounded by increasingly stringent regulatory requirements and business initiatives demanding higher service levels, longer online retention times, and higher levels of data protection. For small to medium-sized organizations there are several deduplication technologies available that can meet their needs. However, for large enterprise organizations with big data environments to protect, these deduplication technologies fall short. Understanding the strengths and weaknesses of deduplication technologies can help you choose a solution that best meets the needs of your large enterprise data protection environment. At a minimum, an enterprise deduplication technology should: Allow high-performance backup, replication, and recovery Reduce capacity requirements to avoid the need for data center build-out and stay within power consumption limits Meet service level agreements and regulatory requirements for fast data recovery by keeping more data online longer Minimize WAN usage and end-user disruption by completing backups within backup windows Control the cost of adding capacity and performance by avoiding technologies that require complete forklift upgrades to scale Ensure data integrity throughout backup, retention, and restore processes Minimize downtime while meeting all regulatory requirements. Provide a view into the backup, deduplication, and replication efficiency at a granular level. Support data centers that are using more than one backup application (e.g., Symantec NetBackup, IBM TSM, etc.) and backup protocol (e.g., Fibre Channel, 1 and 10 GB Ethernet). Deduplicate data from large databases (Oracle, SAP, SQL Server), and from analytical tools used to manage big data environments efficiently. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 1
4 Choosing the Right Deduplication for Your Environment There are several different types of deduplication technologies to choose from. Some are better suited to departments and small to medium-sized enterprises (SMEs); others are optimized for enterprise use and big data backup environments. Consider the strengths and drawbacks of each of these technologies to choose the solution that best meets your requirements. Ten key questions are provided on page 4 to help IT managers determine whether a deduplication technology can meet the needs of a large enterprise data center. Deduplication Technologies Understanding the distinctions between each category of deduplication technology is essential for choosing the most appropriate one for the specific environment. All deduplication technologies compare the data in each backup set to data in a baseline data set to identify duplication. New data is stored and duplicate data is replaced with a pointer to a single reference copy. The key distinctions among deduplication technologies are as follows: Source vs Target Deduplication The first distinction refers to where the deduplication process is performed. Source deduplication is typically performed either on the backup media server or on individual client systems (desktops and servers) to reduce the amount of data sent to the target, but it may also be performed before data is sent to backup media servers. The client system may leverage the file system to locate just those files that have changed since the previous backup or it may compute hashes to describe every file. Target deduplication is used to process larger data volumes. Target-based deduplication can be performed inline, post process, or concurrently. Inline, Post Process, and Concurrent Deduplication The second distinction refers to when the deduplication is performed relative to the backup data being written to the target disk. Deduplication may be performed inline before it is sent to the target, it may be performed as a post process after the backup reaches the target, or it may be performed concurrently along with data ingest to the target. Source deduplication is typically performed inline. Target deduplication may be performed inline, post process, or concurrently. In post-processing systems, the goal is to ingest and protect the backups at the maximum possible speed, minimizing the time to safety. Some systems back up data to disk, and then post process deduplication in discrete steps. While this improves the backup time, it slows completion of deduplication, replication, and restore processes. In contrast, concurrent post processing performs backup, deduplication, replication, and restore operations concurrently, load balancing these operations across multiple nodes, if necessary. This process results in an optimal balance of time to safety and capacity reduction. Hash vs. ContentAware Deduplication The third distinction refers to the methodology used to identify data as duplicate. Hash-based technologies analyze segments of data as they are backed up and assign a unique identifier 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 2
5 called a hash to each segment. Most of these technologies use an algorithm that computes a cryptographic hash value from a fixed or variable segment of data in the backup stream, independent of the data type. The hashes are stored in an index. As each backup is performed, the hashes of incoming data are compared to those in the index. If the hash already exists, the incoming data is replaced with a pointer to the hash. Some source deduplication approaches work on chunks of data rather than on entire files, where a chunk is some multiple of the block size used on the target. ContentAware deduplication reads metadata from the incoming data stream and uses a unique process to identify duplicate data at the byte level. Table 1 compares deduplication technologies. Backward vs Forward Referencing The fourth distinction refers to the way data is identified and tracked as duplicate in the backup. Deduplication technologies all compare incoming data to a baseline or reference data set to identify duplicates. Hash-based technologies, whether they are inline or postprocessed, use the first (oldest) backup as the baseline. Every new backup is compared against it and duplicate data in the latest backup is replaced with pointers back to the older baseline data. In contrast, concurrent, ContentAware deduplication technologies use the most recent backup as the baseline and replace pointers in older data with pointers forward to it. Each new backup replaces the previous one as the baseline and pointers are refreshed to point forward to it in older backups. With this methodology, the system maintains a full, intact copy of the most recent backup for immediate restores. The questions on the following pages can help IT managers ensure that they choose a deduplication technology that can meet the specific needs of large enterprises. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 3
6 10 Key Questions to Ask Your Deduplication Vendor 1. What impact will deduplication have on backup performance both now and over time? High performance is essential to large enterprises that need to move massive data volumes to the safety of a backup environment within a finite backup window. Inline deduplication, whether it is performed at the source or at the target, is not recommended for large data volumes because it can cause bottlenecks that degrade ingest performance for a number of reasons. First, inline deduplication systems cannot be scaled across multiple processors. All backup and deduplication processing has to be performed on a single node 1. If a restore request is processed during deduplication, the restore also has to be performed on the single node. Performance is not only slowed, it is unpredictable. If a company has backup volume and data protection requirements that need the performance of multiple nodes, the IT staff has to divide the backup onto multiple separate systems. Data is not deduplicated among the individual systems, creating an inherently less efficient capacity optimization environment. Second, because hash calculations and lookups take time, data ingestion in inline/hashbased system can be slower, therefore reducing the backup throughput of the device. This bottleneck worsens over time, as the index table grows, and more data in the index needs to be located and compared. The benefit of the hash-based approach is that in a steady-state condition where there is a small rate of change between successive backups (typically < 5 percent); only the new data is written to disk. The performance of these systems is constrained by both the index lookup speed and the storage subsystem, which is optimized for a relatively low write throughput. As the index size grows larger than available memory, the index must be paged to/from disk, resulting in lower performance. A new full backup (due to a new server being added to the backup mix) will result in rapid index updates combined with maximum write requests, resulting in extremely low performance. Third, hash-based inline technologies perform a very CPU-intensive process called housecleaning. During this process, which can take as long as 12 hours nightly, the system cleans and organizes the hash index and actually reclaims the capacity gained through deduplication in a batch process. As a result, administrators need to carefully plan and schedule cleaning windows or delay backups until capacity becomes available. ContentAware deduplication systems can scale both capacity and performance. They backup data at wire speed and perform the analysis, comparison, deduplication, and capacity reclamation processes concurrently. Although this method needs slightly more capacity, its performance and capacity can be scaled easily to backup and deduplicate petabytes of data on a single appliance without degrading performance. Capacity reclamation is continuous and automatic without CPU- housekeeping processes or the need for complex scheduling. 1 Some vendors sell two-node clustered systems, but they do not support the same connectivity and features of their single-node products. In the case of one vendor, a failure of either node renders the entire system unavailable. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 4
7 2. Will deduplication degrade my restore performance? Understand the time required to restore files that were backed within the previous 30 days (the most common category of restore request). As described in Backward vs Forward Referencing above, hash-based deduplication typically uses backward referencing. Forward referencing delivers faster time to recovery for the most frequently needed data. (According to analysts, 90 percent of recovery requests are for data that is less than 30 days old.) Forward referencing technologies keep the last backup available for instant restore and tape vaulting with no need to find and replace pointers with baseline data. 3. How will you scale capacity and performance as your environment grows? Calculate how much data you will be able to store on a single backup system with deduplication given your specific deduplication ratios, policies, and data types. Understand the implications of exceeding that capacity in terms of administrative complexity, capital expense, and disruption to your environment. Some deduplication technologies, particularly those that are performed in-line, cannot easily scale. To perform backups fast enough to stay within your backup window, you need to add multiple independently managed appliances. Overall efficiency of deduplication is reduced because the data comparisons that identify duplicate data are only performed within individual devices. Many deduplication solutions top out at backup rates of 800 Gb/hr. per appliance. At this rate, to backup 10 TB of data in an eight-hour backup window, you would need numerous appliances. That would add significant complexity and require you to modify backup infrastructure/policies. As your data grows, more appliances need to be deployed and managed. This creates silos of deduplication and a management challenge. Overall efficiency of deduplication is also dramatically reduced because the data comparisons that identify duplicate data are only performed within individual devices. Truly enterprise-class ContentAware deduplication solutions can backup data as fast as 43.2 TB/hr. and handle petabytes of data in a single appliance. 4. Can you efficiently deduplicate large multiplexed, multi-streamed databases (e.g., Oracle, SAP, SQL Server)? Deduplication technologies face the challenge of delivering more granular comparisons of data without slowing the backup process. In large enterprises with massive data volumes, the more closely that data is examined, the more CPU resources are required to deduplicate it. Inline deduplication technologies cannot afford to examine data in sections smaller than 8 KB because doing so would severely slow backup performance and increase the size of the hash table index. This granularity of comparison is particularly important in deduplicating large databases, such as Oracle, SAP, and Exchange, as well as analytical tools used to manage big data environments that store data in segments of 8 KB or smaller. For these critical applications, a large volume of duplicate data is completely unidentified by inline solutions. ContentAware technology, however, performs byte-level comparison, enabling it to deliver significantly greater deduplication efficiency in large data base environments. Inline/hash-based systems are also challenged by multiplexed, multi-streamed databases. They force enterprise data managers to choose between fast multi-streaming/multiplexing backups and deduplication. Although some hash-based systems, such as EMC Data Domain Boost integration with Oracle RMAN have attempted to address this challenge for Oracle databases, but they require the installation of third party software on the database servers, which adds risk and complexity to the environment and requires weeks of best practices scripting. These systems also have no deduplication option for non-oracle databases such as SQL or SAP. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 5
8 5. How efficient is your deduplication technology in progressive incremental backup environments such as Tivoli Storage Manager (TSM) and in NetBackup OST? Inline solutions are also are inefficient in deduplicating non-file backups, TSM progressive incremental backups, and backups from applications that fragment their data, such as NetWorker, HP Data Protector, and Symantec OST. ContentAware deduplication technology uses the metadata from these backup applications to identify the areas of data that are likely to contain duplicate data. As described in question #4 above, they perform a byte-level comparison of data for optimal capacity reduction. 6. What are realistic expectations for capacity reduction? The efficiency of a deduplication solution to remove duplicate data determines how much space can be saved. However, if there are no duplicate copies of data in the backup, there can be no space savings. Some data types and application environments typically contain more duplicate copies than others. Data deduplication ratios can vary widely depending on a number of factors, from type of data to frequency of data change and retention period. Although many deduplication vendors claim very high deduplication ratios, these ratios assume a 1 percent data change rate and that datasets are the typical mix of data type found in SMB companies. Capacity reduction savings are summarized in Table 1. Note that a 5:1 ratio reduces storage needs by 80 percent and yields tremendous cost savings. However, there is only a 10 percent difference between a 5:1 and a 10:1 reduction ratio. The bulk of actual capacity savings is between a 2:1 and 5:1 deduplication ratio. The amount of duplicate data in a specific environment is also determined by a number of factors, such as: Data type. Unstructured, semi-structured, and structured data each result in different deduplication results as follows. Unstructured data files. File data such as Microsoft Word, PowerPoint, or Excel files created by office workers often contain redundant data and are frequently distributed or copied. Semi-structured and structured data types. Data created by specific business applications that ensure that operations can be run on a day-to-day basis. For example, Microsoft Exchange for and Oracle for a transactional database clearly fall into the must protect category and require frequent backup. Table 1: Reduction Ratio Comparison Space Reduction Ratio Space Reduction Percentage = 1 (1/Space Reduction Ratio) 2:1 1/2 = 50% 5:1 4/5 = 80% 10:1 9/10 = 90% 20:1 19/20 = 95% 100:1 99/100 = 99% 500:1 499/500 = 99.8% Frequency of data change. The less frequently that data is modified the greater the chance that copies of that data will contain duplicate data. The data deduplication ratio will be higher when the change rate is lower. This also implies that a higher data deduplication ratio should be expected as the percentage of reference data to total active data increases, because reference data is not changed. Retention period. A longer retention period increases the likelihood that duplicate data will be found. Assume a retention window of 90 days and an internal process by which full backups are performed weekly. In this example, the initial full backup will only be deduplicated against itself and will result in a small reduction in storage footprint. When the subsequent full weekly backup for week 2 is performed, only the unique data that has been changed will be stored. When the full backup for week 8 is performed, it will be compared against all the unique data for weeks 1, 2, 3, 4, 5, 6, and 7, which increases the chances that duplicate data will be found. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 6
9 Single vs. multiple pools. Some data protection platforms allow you to separate data into discrete storage pools for increased effectiveness for specific data types. For example, a storage pool can be created for all Exchange data. Deduplication will compare a new backup to the previous Exchange data backup, resulting in a much higher reduction ratio than without storage pools, where the same policy and rules are applied across all data, regardless of type. Rather than pushing for higher generic ratios, a more effective strategy is to choose a solution that guarantees the ability to move data to safety within backup windows while also providing efficient deduplication. Concurrent processing and deterministic ingest rate, deduplication, and replication are key enablers to an enterprise environment. 7. Can administrators monitor backup, deduplication, replication, and restore enterprisewide? Deduplication technologies that do not scale, force IT managers to separate backup operations onto multiple independent systems, each of which has to be managed and optimized separately. Scalable, ContentAware deduplication systems enable administrators to monitor and manage the precise status of data as it passes through backup, deduplication, replication and restore and archive operations. This holistic view of the data protection environment enables them to manage more data per administrator, fine tune the backup environment for optimal efficiency, and to plan accurately for future performance and capacity requirements. 8. Can your deduplication help reduce replication bandwidth requirements for large enterprise data volumes? Some inline deduplication technologies enable companies to replicate data across a WAN efficiently. However, these inline solutions cannot handle large data volumes efficiently and do not enable efficient restore times. ContentAware deduplication works in conjunction with replication software to reduce bandwidth requirements by as much as 97 percent. 9. Can I tune the deduplication to meet my needs? Enterprise data protection environments may have data types that have special deduplication requirements. For example, IT manager may not want to deduplicate some volumes of backup data for regulatory compliance purposes. Or, they may not want to waste processing cycles on image data or other data that is unlikely to contain duplicates. Hash-based deduplication technologies are all or nothing and are capable of performing only one level of deduplication comparison. ContentAware solutions enable IT managers to choose the volumes of data that they want to deduplicate by server, backup application, and by data volume. They also automatically detect the type of data being backed-up automatically perform the method of deduplication that is most efficient for that data type. 10. How much experience does your vendor have with your enterprise backup environment? Enterprise data centers with massive data volumes and complex policies need a data protection vendor with experience with enterprise-class backup applications, such as NetBackup, NetBackup OST, and Tivoli Storage Manager. They should be prepared to provide technical advice and guidance on how to optimize the data deduplication technology for optimal results in these environments. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 7
10 Conclusion For large enterprises with massive data volumes, inline deduplication (at the source or target) cannot provide sufficient time to safety, capacity reduction efficiency, or replication bandwidth optimization to be practical or cost-efficient. The Sepaton DeltaStor data deduplication software was designed with a variety of advanced features to specifically address the needs of large enterprises. DeltaStor delivers the fastest, most efficient, and most cost-effective way for enterprises to back up, deduplicate, replicate, and restore large data volumes. DeltaStor s innovative, enterprise-class features include concurrent processing, which enables Sepaton products to back up petabytes of data within stringent backup windows while providing fast, efficient deduplication processing. Concurrent processing also enables DeltaStor to take advantage of Sepaton s unique grid architecture to leverage the processing power of up to eight nodes to perform backup, deduplication, replication, and restore simultaneously at the industry s highest rates. DeltaStor was also designed with unique forward referencing pointers to enable instant restores of the most recently backed up and most recently replicated data for optimal business continuity and RTOs. This feature also enables the industry s most efficient tape vaulting by eliminating the need to reconstitute deduplicated data before it is written to tape. Automated transparent space reclamation saves administration time and frees capacity faster by reclaiming space continuously. Administrators no longer need to plan and schedule cleaning windows or delay backups until capacity becomes available. Because enterprises rely on large, complex databases such as Oracle and SAP to run critical business operations, Sepaton designed DeltaStor to examine data in sub-8 KB increments used in these systems. Patented technologies enable DeltaStor to examine data at a more granular level and to find duplicate data more efficiently than does any other deduplication technology. As a result, Sepaton delivers the most efficient capacity reduction in the industry. As demonstrated in several use case scenarios, many of the world s largest enterprises have standardized on Sepaton S2100 data protection products with DeltaStor. Sepaton delivers the only products with performance, scalability, single-system capacity, deduplication efficiency, and advanced automation needed in today s large enterprise data protection environments. The enterprise-class features of Sepaton products are summarized in Table 2 on the next page. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 8
11 Table 2: DeltaStor Features Specifically Designed for Large Enterprises Enterprise Deduplication Requirement Allow highperformance backup, replication, and recovery Immediate restores Management tools for optimization Reduce space, power Pay as you grow grid scalability Massive single-system performance and capacity Fast, cost-effective replication Lower TCO and admin cost Tape vaulting Enterprise-grade, highreliability design DeltaStor Features Industry-leading deterministic ingest rates and concurrent, multinode processing to meet backup/replication windows and restore SLAs. Forward referencing enables immediate restore of the most commonly needed data. Management reporting and dashboards to enable a single administrator to ensure that petabytes of data are protected, deduplicated, and replicated efficiently. Detailed status and trend reporting pinpoints sources of inefficiency, enabling unparalleled control and cost savings. Single-system global deduplication of all data enables superior deduplication efficiency and reduces data center space, power, and cooling requirements. Enterprise architecture. Add nodes from one to eight as needed to increase performance. Add disk shelves as needed to scale capacity in cost-saving increments. Enterprise-grade scale and performance. Deploy as many as eight processing nodes to achieve industry-leading performance of up to 43.2 TB/hour. Bandwidth-optimized replication of large data volumes by transmitting only new data. Up to 97% bandwidth reduction. Advanced automation simplifies IT management. Bring added performance or storage online, configure devices to the optimal settings, and load balance without operator intervention. Forward referencing keeps the most recent backup in an un-deduplicated form for fast, efficient tape vaulting % data availability; RAID-6 based Hitachi Data Systems enterpriseclass storage; redundant control and data paths; internally mirrored drives; and real-time 24x7 platform monitoring, diagnostics, and automated alerts. About Sepaton Sepaton is the only company in the world that designs and implements data protection solutions specifically for large enterprises. As a result, many of the world s largest enterprises have deployed Sepaton s purpose-built, integrated scale-out data-protection platform. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 9
Improving Efficiency in Tivoli Storage Manager Environments Sepaton S2100 Data Protection Platform
Improving Efficiency in Tivoli Storage Manager Environments Sepaton S2100 Data Protection Platform White Paper 2013 TABLE OF CONTENTS Introduction... 1 Overview of the TSM Backup and Housekeeping Processes...
More informationDeltaStor Data Deduplication: A Technical Review
White Paper DeltaStor Data Deduplication: A Technical Review DeltaStor software is a next-generation data deduplication application for the SEPATON S2100 -ES2 virtual tape library that enables enterprises
More informationHitachi Protection Platform
Hitachi Protection Platform Powerful, Scalable Backup Appliance for Enterprise Data Protection By Victor Nemechek April 2015 Contents Introduction... 3 Cost Saving Solution... 3 Grid Scalability and Massive
More informationMaximizing Deduplication ROI in a NetBackup Environment
Use Case Maximizing Deduplication ROI in a NetBackup Environment For many companies, backing up and restoring data is becoming a slow, complicated, expensive process. The volume of data to be protected
More informationProtect Microsoft Exchange databases, achieve long-term data retention
Technical white paper Protect Microsoft Exchange databases, achieve long-term data retention HP StoreOnce Backup systems, HP StoreOnce Catalyst, and Symantec NetBackup OpenStorage Table of contents Introduction...
More informationEMC DATA DOMAIN OPERATING SYSTEM
EMC DATA DOMAIN OPERATING SYSTEM Powering EMC Protection Storage ESSENTIALS High-Speed, Scalable Deduplication Up to 58.7 TB/hr performance Reduces requirements for backup storage by 10 to 30x and archive
More informationEMC DATA DOMAIN OPERATING SYSTEM
ESSENTIALS HIGH-SPEED, SCALABLE DEDUPLICATION Up to 58.7 TB/hr performance Reduces protection storage requirements by 10 to 30x CPU-centric scalability DATA INVULNERABILITY ARCHITECTURE Inline write/read
More informationSTORAGE. Buying Guide: TARGET DATA DEDUPLICATION BACKUP SYSTEMS. inside
Managing the information that drives the enterprise STORAGE Buying Guide: DEDUPLICATION inside What you need to know about target data deduplication Special factors to consider One key difference among
More informationExaGrid - A Backup and Data Deduplication appliance
Detailed Product Description ExaGrid Backup Storage Appliances with Deduplication 2014 ExaGrid Systems, Inc. All rights reserved. Table of Contents Executive Summary...2 ExaGrid Basic Concept...2 ExaGrid
More informationProtect Data... in the Cloud
QUASICOM Private Cloud Backups with ExaGrid Deduplication Disk Arrays Martin Lui Senior Solution Consultant Quasicom Systems Limited Protect Data...... in the Cloud 1 Mobile Computing Users work with their
More informationEight Considerations for Evaluating Disk-Based Backup Solutions
Eight Considerations for Evaluating Disk-Based Backup Solutions 1 Introduction The movement from tape-based to disk-based backup is well underway. Disk eliminates all the problems of tape backup. Backing
More informationData De-duplication Methodologies: Comparing ExaGrid s Byte-level Data De-duplication To Block Level Data De-duplication
Data De-duplication Methodologies: Comparing ExaGrid s Byte-level Data De-duplication To Block Level Data De-duplication Table of Contents Introduction... 3 Shortest Possible Backup Window... 3 Instant
More informationDetailed Product Description
Detailed Product Description ExaGrid Disk Backup with Deduplication 2014 ExaGrid Systems, Inc. All rights reserved. Table of Contents Executive Summary...2 ExaGrid Basic Concept...2 Product Benefits...
More informationEnterprise-class Backup Performance with Dell DR6000 Date: May 2014 Author: Kerry Dolan, Lab Analyst and Vinny Choinski, Senior Lab Analyst
ESG Lab Review Enterprise-class Backup Performance with Dell DR6000 Date: May 2014 Author: Kerry Dolan, Lab Analyst and Vinny Choinski, Senior Lab Analyst Abstract: This ESG Lab review documents hands-on
More informationExaGrid Product Description. Cost-Effective Disk-Based Backup with Data Deduplication
ExaGrid Product Description Cost-Effective Disk-Based Backup with Data Deduplication 1 Contents Introduction... 3 Considerations When Examining Disk-Based Backup Approaches... 3 ExaGrid A Disk-Based Backup
More informationTurnkey Deduplication Solution for the Enterprise
Symantec NetBackup 5000 Appliance Turnkey Deduplication Solution for the Enterprise Mayur Dewaikar Sr. Product Manager, Information Management Group White Paper: A Deduplication Appliance Solution for
More informationWHITE PAPER Improving Storage Efficiencies with Data Deduplication and Compression
WHITE PAPER Improving Storage Efficiencies with Data Deduplication and Compression Sponsored by: Oracle Steven Scully May 2010 Benjamin Woo IDC OPINION Global Headquarters: 5 Speen Street Framingham, MA
More informationTechnology Fueling the Next Phase of Storage Optimization
White Paper HP StoreOnce Deduplication Software Technology Fueling the Next Phase of Storage Optimization By Lauren Whitehouse June, 2010 This ESG White Paper was commissioned by Hewlett-Packard and is
More informationUsing HP StoreOnce Backup Systems for NDMP backups with Symantec NetBackup
Technical white paper Using HP StoreOnce Backup Systems for NDMP backups with Symantec NetBackup Table of contents Executive summary... 2 Introduction... 2 What is NDMP?... 2 Technology overview... 3 HP
More informationLDA, the new family of Lortu Data Appliances
LDA, the new family of Lortu Data Appliances Based on Lortu Byte-Level Deduplication Technology February, 2011 Copyright Lortu Software, S.L. 2011 1 Index Executive Summary 3 Lortu deduplication technology
More informationEfficient Backup with Data Deduplication Which Strategy is Right for You?
Efficient Backup with Data Deduplication Which Strategy is Right for You? Rob Emsley Senior Director, Product Marketing CPU Utilization CPU Utilization Exabytes Why So Much Interest in Data Deduplication?
More informationIBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE
White Paper IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE Abstract This white paper focuses on recovery of an IBM Tivoli Storage Manager (TSM) server and explores
More informationSymantec NetBackup 5220
A single-vendor enterprise backup appliance that installs in minutes Data Sheet: Data Protection Overview is a single-vendor enterprise backup appliance that installs in minutes, with expandable storage
More informationSymantec NetBackup 5000 Appliance Series
A turnkey, end-to-end, global deduplication solution for the enterprise. Data Sheet: Data Protection Overview Symantec NetBackup 5000 series offers your organization a content aware, end-to-end, and global
More informationBest Practices Guide. Symantec NetBackup with ExaGrid Disk Backup with Deduplication. 2012 ExaGrid Systems, Inc. All rights reserved.
Best Practices Guide Symantec NetBackup with ExaGrid Disk Backup with Deduplication 2012 ExaGrid Systems, Inc. All rights reserved. Table of Contents EXECUTIVE SUMMARY... 2 INTRODUCTION... 2 AUDIENCE...
More informationEMC Data Domain Boost for Oracle Recovery Manager (RMAN)
White Paper EMC Data Domain Boost for Oracle Recovery Manager (RMAN) Abstract EMC delivers Database Administrators (DBAs) complete control of Oracle backup, recovery, and offsite disaster recovery with
More informationDemystifying Deduplication for Backup with the Dell DR4000
Demystifying Deduplication for Backup with the Dell DR4000 This Dell Technical White Paper explains how deduplication with the DR4000 can help your organization save time, space, and money. John Bassett
More informationProtecting enterprise servers with StoreOnce and CommVault Simpana
Technical white paper Protecting enterprise servers with StoreOnce and CommVault Simpana HP StoreOnce Backup systems Table of contents Introduction 2 Technology overview 2 HP StoreOnce Backup systems key
More informationWHY DO I NEED FALCONSTOR OPTIMIZED BACKUP & DEDUPLICATION?
WHAT IS FALCONSTOR? FalconStor Optimized Backup and Deduplication is the industry s market-leading virtual tape and LAN-based deduplication solution, unmatched in performance and scalability. With virtual
More informationHow To Make A Backup System More Efficient
Identifying the Hidden Risk of Data De-duplication: How the HYDRAstor Solution Proactively Solves the Problem October, 2006 Introduction Data de-duplication has recently gained significant industry attention,
More informationHP StoreOnce: reinventing data deduplication
HP : reinventing data deduplication Reduce the impact of explosive data growth with HP StorageWorks D2D Backup Systems Technical white paper Table of contents Executive summary... 2 Introduction to data
More informationIdentifying the Hidden Risk of Data Deduplication: How the HYDRAstor TM Solution Proactively Solves the Problem
Identifying the Hidden Risk of Data Deduplication: How the HYDRAstor TM Solution Proactively Solves the Problem Advanced Storage Products Group Table of Contents 1 - Introduction 2 Data Deduplication 3
More informationData Deduplication: An Essential Component of your Data Protection Strategy
WHITE PAPER: THE EVOLUTION OF DATA DEDUPLICATION Data Deduplication: An Essential Component of your Data Protection Strategy JULY 2010 Andy Brewerton CA TECHNOLOGIES RECOVERY MANAGEMENT AND DATA MODELLING
More informationProtecting Information in a Smarter Data Center with the Performance of Flash
89 Fifth Avenue, 7th Floor New York, NY 10003 www.theedison.com 212.367.7400 Protecting Information in a Smarter Data Center with the Performance of Flash IBM FlashSystem and IBM ProtecTIER Printed in
More informationEMC BACKUP MEETS BIG DATA
EMC BACKUP MEETS BIG DATA Strategies To Protect Greenplum, Isilon And Teradata Systems 1 Agenda Big Data: Overview, Backup and Recovery EMC Big Data Backup Strategy EMC Backup and Recovery Solutions for
More informationReal-time Compression: Achieving storage efficiency throughout the data lifecycle
Real-time Compression: Achieving storage efficiency throughout the data lifecycle By Deni Connor, founding analyst Patrick Corrigan, senior analyst July 2011 F or many companies the growth in the volume
More informationReducing Backups with Data Deduplication
The Essentials Series: New Techniques for Creating Better Backups Reducing Backups with Data Deduplication sponsored by by Eric Beehler Reducing Backups with Data Deduplication... 1 Explaining Data Deduplication...
More informationOracle Data Protection Concepts
Oracle Data Protection Concepts Matthew Ellis Advisory Systems Engineer BRS Database Technologist, EMC Corporation Accelerating Transformation EMC Backup Recovery Systems Division 1 Agenda Market Conditions.
More informationUsing HP StoreOnce Backup systems for Oracle database backups
Technical white paper Using HP StoreOnce Backup systems for Oracle database backups Table of contents Introduction 2 Technology overview 2 HP StoreOnce Backup systems key features and benefits 2 HP StoreOnce
More informationEMC DATA DOMAIN PRODUCT OvERvIEW
EMC DATA DOMAIN PRODUCT OvERvIEW Deduplication storage for next-generation backup and archive Essentials Scalable Deduplication Fast, inline deduplication Provides up to 65 PBs of logical storage for long-term
More informationExaGrid s EX32000E is its newest and largest appliance, taking in a 32TB full backup with an ingest rate of 7.5TB/hour.
Nominee: ExaGrid Nomination title: ExaGrid s next-generation appliances continue market disruption with bigger/faster appliances and larger scalable GRID systems in 2015 ExaGrid s EX32000E is its newest
More informationData Deduplication in Tivoli Storage Manager. Andrzej Bugowski 19-05-2011 Spała
Data Deduplication in Tivoli Storage Manager Andrzej Bugowski 19-05-2011 Spała Agenda Tivoli Storage, IBM Software Group Deduplication concepts Data deduplication in TSM 6.1 Planning for data deduplication
More informationEMC Data Domain Boost for Oracle Recovery Manager (RMAN)
White Paper EMC Data Domain Boost for Oracle Recovery Manager (RMAN) Abstract EMC delivers Database Administrators (DBAs) complete control of Oracle backup, recovery, and offsite disaster recovery with
More informationGIVE YOUR ORACLE DBAs THE BACKUPS THEY REALLY WANT
Why Data Domain Series GIVE YOUR ORACLE DBAs THE BACKUPS THEY REALLY WANT Why you should take the time to read this paper Speed up backups (Up to 58.7 TB/hr, Data Domain systems are about 1.5 times faster
More informationESG REPORT. Data Deduplication Diversity: Evaluating Software- vs. Hardware-Based Approaches. By Lauren Whitehouse. April, 2009
ESG REPORT : Evaluating Software- vs. Hardware-Based Approaches By Lauren Whitehouse April, 2009 Table of Contents ESG REPORT Table of Contents... i Introduction... 1 External Forces Contribute to IT Challenges...
More informationHP Data Protector software and HP StoreOnce backup systems for federated deduplication and flexible deployment
Brochure Maximize storage efficiency across the enterprise HP Data Protector software and HP StoreOnce backup systems for federated deduplication and flexible deployment Maximize storage efficiency across
More informationBackup and Recovery for SAP Environments using EMC Avamar 7
White Paper Backup and Recovery for SAP Environments using EMC Avamar 7 Abstract This white paper highlights how IT environments deploying SAP can benefit from efficient backup with an EMC Avamar solution.
More informationEvaluation Guide. Software vs. Appliance Deduplication
Evaluation Guide Software vs. Appliance Deduplication Table of Contents Introduction... 2 Data Deduplication Overview... 3 Backup Requirements... 6 Backup Application Client Side Deduplication... 7 Backup
More informationSYMANTEC NETBACKUP APPLIANCE FAMILY OVERVIEW BROCHURE. When you can do it simply, you can do it all.
SYMANTEC NETBACKUP APPLIANCE FAMILY OVERVIEW BROCHURE When you can do it simply, you can do it all. SYMANTEC NETBACKUP APPLIANCES Symantec understands the shifting needs of the data center and offers NetBackup
More informationLong term retention and archiving the challenges and the solution
Long term retention and archiving the challenges and the solution NAME: Yoel Ben-Ari TITLE: VP Business Development, GH Israel 1 Archive Before Backup EMC recommended practice 2 1 Backup/recovery process
More informationDXi Accent Technical Background
TECHNOLOGY BRIEF NOTICE This Technology Brief contains information protected by copyright. Information in this Technology Brief is subject to change without notice and does not represent a commitment on
More informationData Reduction Methodologies: Comparing ExaGrid s Byte-Level-Delta Data Reduction to Data De-duplication. February 2007
Data Reduction Methodologies: Comparing ExaGrid s Byte-Level-Delta Data Reduction to Data De-duplication February 2007 Though data reduction technologies have been around for years, there is a renewed
More information3Gen Data Deduplication Technical
3Gen Data Deduplication Technical Discussion NOTICE: This White Paper may contain proprietary information protected by copyright. Information in this White Paper is subject to change without notice and
More informationEMC DATA DOMAIN OVERVIEW. Copyright 2011 EMC Corporation. All rights reserved.
EMC DATA DOMAIN OVERVIEW 1 2 With Data Domain Deduplication Storage Systems, You Can WAN Retain longer Keep backups onsite longer with less disk for fast, reliable restores, and eliminate the use of tape
More informationNetApp Syncsort Integrated Backup
WHITE PAPER NetApp Syncsort Integrated Backup Protect your Microsoft and VMware Environment with NetApp Syncsort Integrated Backup Protecting Microsoft and VMware Executive Summary 3 Microsoft and VMware
More informationEvery organization has critical data that it can t live without. When a disaster strikes, how long can your business survive without access to its
DISASTER RECOVERY STRATEGIES: BUSINESS CONTINUITY THROUGH REMOTE BACKUP REPLICATION Every organization has critical data that it can t live without. When a disaster strikes, how long can your business
More informationHow To Protect Data On Network Attached Storage (Nas) From Disaster
White Paper EMC FOR NETWORK ATTACHED STORAGE (NAS) BACKUP AND RECOVERY Abstract This white paper provides an overview of EMC s industry leading backup and recovery solutions for NAS systems. It also explains
More informationDetailed Product Description
Detailed Product Description ExaGrid Backup Storage Appliances with Data Deduplication 2015 ExaGrid Systems, Inc. All rights reserved. Table of Contents Executive Summary...2 ExaGrid Basic Concept...2
More informationDeduplication and Beyond: Optimizing Performance for Backup and Recovery
Beyond: Optimizing Gartner clients using deduplication for backups typically report seven times to 25 times the reductions (7:1 to 25:1) in the size of their data, and sometimes higher than 100:1 for file
More informationDEDUPLICATION NOW AND WHERE IT S HEADING. Lauren Whitehouse Senior Analyst, Enterprise Strategy Group
DEDUPLICATION NOW AND WHERE IT S HEADING Lauren Whitehouse Senior Analyst, Enterprise Strategy Group Need Dedupe? Before/After Dedupe Deduplication Production Data Deduplication In Backup Process Backup
More informationQuantum DXi6500 Family of Network-Attached Disk Backup Appliances with Deduplication
PRODUCT BRIEF Quantum DXi6500 Family of Network-Attached Disk Backup Appliances with Deduplication NOTICE This Product Brief contains proprietary information protected by copyright. Information in this
More informationBusiness Benefits of Data Footprint Reduction
Business Benefits of Data Footprint Reduction Why and how reducing your data footprint provides a positive benefit to your business and application service objectives By Greg Schulz Founder and Senior
More informationActifio Big Data Director. Virtual Data Pipeline for Unstructured Data
Actifio Big Data Director Virtual Data Pipeline for Unstructured Data Contact Actifio Support As an Actifio customer, you can get support for all Actifio products through the Support Portal at http://support.actifio.com/.
More informationBackup Software Data Deduplication: What you need to know. Presented by W. Curtis Preston Executive Editor & Independent Backup Expert
Backup Software Data Deduplication: What you need to know Presented by W. Curtis Preston Executive Editor & Independent Backup Expert When I was in the IT Department When I started as backup guy at $35B
More informationEMC Disk Library with EMC Data Domain Deployment Scenario
EMC Disk Library with EMC Data Domain Deployment Scenario Best Practices Planning Abstract This white paper is an overview of the EMC Disk Library with EMC Data Domain deduplication storage system deployment
More informationGet Success in Passing Your Certification Exam at first attempt!
Get Success in Passing Your Certification Exam at first attempt! Exam : E22-290 Title : EMC Data Domain Deduplication, Backup and Recovery Exam Version : DEMO 1.A customer has a Data Domain system with
More informationABOUT DISK BACKUP WITH DEDUPLICATION
Disk Backup with Data Deduplication ABOUT DISK BACKUP WITH DEDUPLICATION www.exagrid.com What appears to be simple & straightforward Built for Backup is often more complex & risky than you think. 2 Agenda
More informationDeduplication s Role in Disaster Recovery. Thomas Rivera, SEPATON
Thomas Rivera, SEPATON SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA. Member companies and individual members may use this material in presentations and literature
More informationData deduplication is more than just a BUZZ word
Data deduplication is more than just a BUZZ word Per Larsen Principal Systems Engineer Mr. Hansen DATA BUDGET RECOVERY & DATACENTER GROWTH PRESSURE DISCOVERY REVOLUTION More Storage Longer Backups Smaller
More informationSolving Data Growth Issues using Deduplication
Solving Data Growth Issues using Deduplication Reducing Storage Costs and Speeding Backups with Dell Deduplication Solutions Abstract Data growth is increasing at record rates, and ensuring the integrity
More informationDeduplication s Role in Disaster Recovery. Thomas Rivera, SEPATON
Thomas Rivera, SEPATON SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA. Member companies and individual members may use this material in presentations and literature
More informationWHITE PAPER BRENT WELCH NOVEMBER
BACKUP WHITE PAPER BRENT WELCH NOVEMBER 2006 WHITE PAPER: BACKUP TABLE OF CONTENTS Backup Overview 3 Background on Backup Applications 3 Backup Illustration 4 Media Agents & Keeping Tape Drives Busy 5
More information<Insert Picture Here> Refreshing Your Data Protection Environment with Next-Generation Architectures
1 Refreshing Your Data Protection Environment with Next-Generation Architectures Dale Rhine, Principal Sales Consultant Kelly Boeckman, Product Marketing Analyst Program Agenda Storage
More informationVeritas Backup Exec 15: Deduplication Option
Veritas Backup Exec 15: Deduplication Option Who should read this paper Technical White Papers are designed to introduce IT professionals to key technologies and technical concepts that are associated
More informationTop Ten Questions. to Ask Your Primary Storage Provider About Their Data Efficiency. May 2014. Copyright 2014 Permabit Technology Corporation
Top Ten Questions to Ask Your Primary Storage Provider About Their Data Efficiency May 2014 Copyright 2014 Permabit Technology Corporation Introduction The value of data efficiency technologies, namely
More informationBarracuda Backup Deduplication. White Paper
Barracuda Backup Deduplication White Paper Abstract Data protection technologies play a critical role in organizations of all sizes, but they present a number of challenges in optimizing their operation.
More informationFuture-Proofed Backup For A Virtualized World!
! Future-Proofed Backup For A Virtualized World! Prepared by: Colm Keegan, Senior Analyst! Prepared: January 2014 Future-Proofed Backup For A Virtualized World Like death and taxes, growing backup windows
More informationUsing HP StoreOnce D2D systems for Microsoft SQL Server backups
Technical white paper Using HP StoreOnce D2D systems for Microsoft SQL Server backups Table of contents Executive summary 2 Introduction 2 Technology overview 2 HP StoreOnce D2D systems key features and
More informationDEDUPLICATION SOLUTIONS ARE NOT ALL CREATED EQUAL, WHY DATA DOMAIN?
Why Data Domain Series DEDUPLICATION SOLUTIONS ARE NOT ALL CREATED EQUAL, WHY DATA DOMAIN? Why you should take the time to read this paper Speed up your backups (Achieve up to 31 TB/hr - 3 times faster
More informationWHITE PAPER. Effectiveness of Variable-block vs Fixedblock Deduplication on Data Reduction: A Technical Analysis
WHITE PAPER Effectiveness of Variable-block vs Fixedblock Deduplication on Data Reduction: A Technical Analysis CONTENTS Executive Summary... 3 Fixed vs. Variable-block Deduplication... 3 Test Configuration...
More informationRiverbed Whitewater/Amazon Glacier ROI for Backup and Archiving
Riverbed Whitewater/Amazon Glacier ROI for Backup and Archiving November, 2013 Saqib Jang Abstract This white paper demonstrates how to increase profitability by reducing the operating costs of backup
More informationAn Oracle White Paper November 2010. Backup and Recovery with Oracle s Sun ZFS Storage Appliances and Oracle Recovery Manager
An Oracle White Paper November 2010 Backup and Recovery with Oracle s Sun ZFS Storage Appliances and Oracle Recovery Manager Introduction...2 Oracle Backup and Recovery Solution Overview...3 Oracle Recovery
More informationSales Tool. Summary DXi Sales Messages November 2009 6 NOVEMBER 2009. ST00431-v06
Summary DXi Sales Messages November 2009 6 NOVEMBER 2009 ST00431-v06 Notice This Sales Tool contains proprietary information protected by copyright. Information in this Sales Tool is subject to change
More informationDeduplication has been around for several
Demystifying Deduplication By Joe Colucci Kay Benaroch Deduplication holds the promise of efficient storage and bandwidth utilization, accelerated backup and recovery, reduced costs, and more. Understanding
More informationSLOW BACKUPS GOT YOU DOWN?
Why Data Domain Series SLOW BACKUPS GOT YOU DOWN? Why you should take the time to read this paper Speed up backups by 50% (Finish backups within backup windows with breathing room for data growth. With
More informationDisaster Recovery Strategies: Business Continuity through Remote Backup Replication
W H I T E P A P E R S O L U T I O N : D I S A S T E R R E C O V E R Y T E C H N O L O G Y : R E M O T E R E P L I C A T I O N Disaster Recovery Strategies: Business Continuity through Remote Backup Replication
More informationReduce your data storage footprint and tame the information explosion
IBM Software White paper December 2010 Reduce your data storage footprint and tame the information explosion 2 Reduce your data storage footprint and tame the information explosion Contents 2 Executive
More informationNetApp Data Fabric: Secured Backup to Public Cloud. Sonny Afen Senior Technical Consultant NetApp Indonesia
NetApp Data Fabric: Secured Backup to Public Cloud Sonny Afen Senior Technical Consultant NetApp Indonesia Agenda Introduction Solution Overview Solution Technical Overview 2 Introduction 3 Hybrid cloud:
More informationEMC DATA DOMAIN EXTENDED RETENTION SOFTWARE: MEETING NEEDS FOR LONG-TERM RETENTION OF BACKUP DATA ON EMC DATA DOMAIN SYSTEMS
SOLUTION PROFILE EMC DATA DOMAIN EXTENDED RETENTION SOFTWARE: MEETING NEEDS FOR LONG-TERM RETENTION OF BACKUP DATA ON EMC DATA DOMAIN SYSTEMS MAY 2012 Backups are essential for short-term data recovery
More informationTechnical White Paper for the Oceanspace VTL6000
Document No. Technical White Paper for the Oceanspace VTL6000 Issue V2.1 Date 2010-05-18 Huawei Symantec Technologies Co., Ltd. Copyright Huawei Symantec Technologies Co., Ltd. 2010. All rights reserved.
More informationBackup Exec 15: Deduplication Option
TECHNICAL BRIEF: BACKUP EXEC 15: DEDUPLICATION OPTION........................................ Backup Exec 15: Deduplication Option Who should read this paper Technical White Papers are designed to introduce
More informationBackup Exec 2014: Deduplication Option
TECHNICAL BRIEF: BACKUP EXEC 2014: DEDUPLICATION OPTION........................................ Backup Exec 2014: Deduplication Option Who should read this paper Technical White Papers are designed to
More informationVMware vsphere Data Protection
FREQUENTLY ASKED QUESTIONS VMware vsphere Data Protection vsphere Data Protection Advanced Overview Q. What is VMware vsphere Data Protection Advanced? A. VMware vsphere Data Protection Advanced is a backup
More informationUnderstanding EMC Avamar with EMC Data Protection Advisor
Understanding EMC Avamar with EMC Data Protection Advisor Applied Technology Abstract EMC Data Protection Advisor provides a comprehensive set of features that reduce the complexity of managing data protection
More informationHow To Store Data On Disk On Data Domain
WHITE PAPER Deduplication Storage Deduplication Storage for Nearline Applications September 2007 w w w. d a t a d o m a i n. c o m - 2 0 0 7 DATA DOMAIN I Contents Consolidated Support for Backup, Archiving,
More informationIBM PROTECTIER: FROM BACKUP TO RECOVERY
SOLUTION PROFILE IBM PROTECTIER: FROM BACKUP TO RECOVERY NOVEMBER 2011 When it comes to backup and recovery, backup performance numbers rule the roost. It s understandable really: far more data gets backed
More informationArchiving, Backup, and Recovery for Complete the Promise of Virtualization
Archiving, Backup, and Recovery for Complete the Promise of Virtualization Unified information management for enterprise Windows environments The explosion of unstructured information It is estimated that
More informationHP StorageWorks D2D Backup Systems and StoreOnce
AUtOMATEyour data protection. HP StorageWorks D2D Backup Systems and StoreOnce The combination that right-sizes your storage capacity. Solution brief Regardless of size and industry, many of today s organizations
More informationData Deduplication and Tivoli Storage Manager
Data Deduplication and Tivoli Storage Manager Dave Cannon Tivoli Storage Manager rchitect Oxford University TSM Symposium September 2007 Disclaimer This presentation describes potential future enhancements
More informationIBM Storwize V7000 Unified and Storwize V7000 storage systems
IBM Storwize V7000 Unified and Storwize V7000 storage systems Transforming the economics of data storage Highlights Meet changing business needs with virtualized, enterprise-class, flashoptimized modular
More information