Choosing an Enterprise-Class Deduplication Technology

Size: px
Start display at page:

Download "Choosing an Enterprise-Class Deduplication Technology"

Transcription

1 WHITE PAPER Choosing an Enterprise-Class Deduplication Technology 10 Key Questions to Ask Your Deduplication Vendor 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F:

2 Copyright Sepaton, Inc. All rights reserved. Sepaton, S2100, DeltaStor, DeltaRemote, and SRE are registered trademarks and ContentAware, Secure Erasure, and DeltaScale are trademarks of Sepaton, Inc. Other product and company names mentioned herein are or may be trademarks and/or registered trademarks of their respective companies v2 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F:

3 Deduplication Specifically Designed for Enterprises The volume of data generated by most companies has grown at such an explosive rate that many data centers have simply run out of space, power, cooling, and storage capacity to handle it. In large enterprise organizations, the sheer volume and variety of data to be protected requires a level of performance and scalability that few data protection technologies can deliver. Fundamental issues of insufficient capacity are being compounded by increasingly stringent regulatory requirements and business initiatives demanding higher service levels, longer online retention times, and higher levels of data protection. For small to medium-sized organizations there are several deduplication technologies available that can meet their needs. However, for large enterprise organizations with big data environments to protect, these deduplication technologies fall short. Understanding the strengths and weaknesses of deduplication technologies can help you choose a solution that best meets the needs of your large enterprise data protection environment. At a minimum, an enterprise deduplication technology should: Allow high-performance backup, replication, and recovery Reduce capacity requirements to avoid the need for data center build-out and stay within power consumption limits Meet service level agreements and regulatory requirements for fast data recovery by keeping more data online longer Minimize WAN usage and end-user disruption by completing backups within backup windows Control the cost of adding capacity and performance by avoiding technologies that require complete forklift upgrades to scale Ensure data integrity throughout backup, retention, and restore processes Minimize downtime while meeting all regulatory requirements. Provide a view into the backup, deduplication, and replication efficiency at a granular level. Support data centers that are using more than one backup application (e.g., Symantec NetBackup, IBM TSM, etc.) and backup protocol (e.g., Fibre Channel, 1 and 10 GB Ethernet). Deduplicate data from large databases (Oracle, SAP, SQL Server), and from analytical tools used to manage big data environments efficiently. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 1

4 Choosing the Right Deduplication for Your Environment There are several different types of deduplication technologies to choose from. Some are better suited to departments and small to medium-sized enterprises (SMEs); others are optimized for enterprise use and big data backup environments. Consider the strengths and drawbacks of each of these technologies to choose the solution that best meets your requirements. Ten key questions are provided on page 4 to help IT managers determine whether a deduplication technology can meet the needs of a large enterprise data center. Deduplication Technologies Understanding the distinctions between each category of deduplication technology is essential for choosing the most appropriate one for the specific environment. All deduplication technologies compare the data in each backup set to data in a baseline data set to identify duplication. New data is stored and duplicate data is replaced with a pointer to a single reference copy. The key distinctions among deduplication technologies are as follows: Source vs Target Deduplication The first distinction refers to where the deduplication process is performed. Source deduplication is typically performed either on the backup media server or on individual client systems (desktops and servers) to reduce the amount of data sent to the target, but it may also be performed before data is sent to backup media servers. The client system may leverage the file system to locate just those files that have changed since the previous backup or it may compute hashes to describe every file. Target deduplication is used to process larger data volumes. Target-based deduplication can be performed inline, post process, or concurrently. Inline, Post Process, and Concurrent Deduplication The second distinction refers to when the deduplication is performed relative to the backup data being written to the target disk. Deduplication may be performed inline before it is sent to the target, it may be performed as a post process after the backup reaches the target, or it may be performed concurrently along with data ingest to the target. Source deduplication is typically performed inline. Target deduplication may be performed inline, post process, or concurrently. In post-processing systems, the goal is to ingest and protect the backups at the maximum possible speed, minimizing the time to safety. Some systems back up data to disk, and then post process deduplication in discrete steps. While this improves the backup time, it slows completion of deduplication, replication, and restore processes. In contrast, concurrent post processing performs backup, deduplication, replication, and restore operations concurrently, load balancing these operations across multiple nodes, if necessary. This process results in an optimal balance of time to safety and capacity reduction. Hash vs. ContentAware Deduplication The third distinction refers to the methodology used to identify data as duplicate. Hash-based technologies analyze segments of data as they are backed up and assign a unique identifier 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 2

5 called a hash to each segment. Most of these technologies use an algorithm that computes a cryptographic hash value from a fixed or variable segment of data in the backup stream, independent of the data type. The hashes are stored in an index. As each backup is performed, the hashes of incoming data are compared to those in the index. If the hash already exists, the incoming data is replaced with a pointer to the hash. Some source deduplication approaches work on chunks of data rather than on entire files, where a chunk is some multiple of the block size used on the target. ContentAware deduplication reads metadata from the incoming data stream and uses a unique process to identify duplicate data at the byte level. Table 1 compares deduplication technologies. Backward vs Forward Referencing The fourth distinction refers to the way data is identified and tracked as duplicate in the backup. Deduplication technologies all compare incoming data to a baseline or reference data set to identify duplicates. Hash-based technologies, whether they are inline or postprocessed, use the first (oldest) backup as the baseline. Every new backup is compared against it and duplicate data in the latest backup is replaced with pointers back to the older baseline data. In contrast, concurrent, ContentAware deduplication technologies use the most recent backup as the baseline and replace pointers in older data with pointers forward to it. Each new backup replaces the previous one as the baseline and pointers are refreshed to point forward to it in older backups. With this methodology, the system maintains a full, intact copy of the most recent backup for immediate restores. The questions on the following pages can help IT managers ensure that they choose a deduplication technology that can meet the specific needs of large enterprises. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 3

6 10 Key Questions to Ask Your Deduplication Vendor 1. What impact will deduplication have on backup performance both now and over time? High performance is essential to large enterprises that need to move massive data volumes to the safety of a backup environment within a finite backup window. Inline deduplication, whether it is performed at the source or at the target, is not recommended for large data volumes because it can cause bottlenecks that degrade ingest performance for a number of reasons. First, inline deduplication systems cannot be scaled across multiple processors. All backup and deduplication processing has to be performed on a single node 1. If a restore request is processed during deduplication, the restore also has to be performed on the single node. Performance is not only slowed, it is unpredictable. If a company has backup volume and data protection requirements that need the performance of multiple nodes, the IT staff has to divide the backup onto multiple separate systems. Data is not deduplicated among the individual systems, creating an inherently less efficient capacity optimization environment. Second, because hash calculations and lookups take time, data ingestion in inline/hashbased system can be slower, therefore reducing the backup throughput of the device. This bottleneck worsens over time, as the index table grows, and more data in the index needs to be located and compared. The benefit of the hash-based approach is that in a steady-state condition where there is a small rate of change between successive backups (typically < 5 percent); only the new data is written to disk. The performance of these systems is constrained by both the index lookup speed and the storage subsystem, which is optimized for a relatively low write throughput. As the index size grows larger than available memory, the index must be paged to/from disk, resulting in lower performance. A new full backup (due to a new server being added to the backup mix) will result in rapid index updates combined with maximum write requests, resulting in extremely low performance. Third, hash-based inline technologies perform a very CPU-intensive process called housecleaning. During this process, which can take as long as 12 hours nightly, the system cleans and organizes the hash index and actually reclaims the capacity gained through deduplication in a batch process. As a result, administrators need to carefully plan and schedule cleaning windows or delay backups until capacity becomes available. ContentAware deduplication systems can scale both capacity and performance. They backup data at wire speed and perform the analysis, comparison, deduplication, and capacity reclamation processes concurrently. Although this method needs slightly more capacity, its performance and capacity can be scaled easily to backup and deduplicate petabytes of data on a single appliance without degrading performance. Capacity reclamation is continuous and automatic without CPU- housekeeping processes or the need for complex scheduling. 1 Some vendors sell two-node clustered systems, but they do not support the same connectivity and features of their single-node products. In the case of one vendor, a failure of either node renders the entire system unavailable. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 4

7 2. Will deduplication degrade my restore performance? Understand the time required to restore files that were backed within the previous 30 days (the most common category of restore request). As described in Backward vs Forward Referencing above, hash-based deduplication typically uses backward referencing. Forward referencing delivers faster time to recovery for the most frequently needed data. (According to analysts, 90 percent of recovery requests are for data that is less than 30 days old.) Forward referencing technologies keep the last backup available for instant restore and tape vaulting with no need to find and replace pointers with baseline data. 3. How will you scale capacity and performance as your environment grows? Calculate how much data you will be able to store on a single backup system with deduplication given your specific deduplication ratios, policies, and data types. Understand the implications of exceeding that capacity in terms of administrative complexity, capital expense, and disruption to your environment. Some deduplication technologies, particularly those that are performed in-line, cannot easily scale. To perform backups fast enough to stay within your backup window, you need to add multiple independently managed appliances. Overall efficiency of deduplication is reduced because the data comparisons that identify duplicate data are only performed within individual devices. Many deduplication solutions top out at backup rates of 800 Gb/hr. per appliance. At this rate, to backup 10 TB of data in an eight-hour backup window, you would need numerous appliances. That would add significant complexity and require you to modify backup infrastructure/policies. As your data grows, more appliances need to be deployed and managed. This creates silos of deduplication and a management challenge. Overall efficiency of deduplication is also dramatically reduced because the data comparisons that identify duplicate data are only performed within individual devices. Truly enterprise-class ContentAware deduplication solutions can backup data as fast as 43.2 TB/hr. and handle petabytes of data in a single appliance. 4. Can you efficiently deduplicate large multiplexed, multi-streamed databases (e.g., Oracle, SAP, SQL Server)? Deduplication technologies face the challenge of delivering more granular comparisons of data without slowing the backup process. In large enterprises with massive data volumes, the more closely that data is examined, the more CPU resources are required to deduplicate it. Inline deduplication technologies cannot afford to examine data in sections smaller than 8 KB because doing so would severely slow backup performance and increase the size of the hash table index. This granularity of comparison is particularly important in deduplicating large databases, such as Oracle, SAP, and Exchange, as well as analytical tools used to manage big data environments that store data in segments of 8 KB or smaller. For these critical applications, a large volume of duplicate data is completely unidentified by inline solutions. ContentAware technology, however, performs byte-level comparison, enabling it to deliver significantly greater deduplication efficiency in large data base environments. Inline/hash-based systems are also challenged by multiplexed, multi-streamed databases. They force enterprise data managers to choose between fast multi-streaming/multiplexing backups and deduplication. Although some hash-based systems, such as EMC Data Domain Boost integration with Oracle RMAN have attempted to address this challenge for Oracle databases, but they require the installation of third party software on the database servers, which adds risk and complexity to the environment and requires weeks of best practices scripting. These systems also have no deduplication option for non-oracle databases such as SQL or SAP. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 5

8 5. How efficient is your deduplication technology in progressive incremental backup environments such as Tivoli Storage Manager (TSM) and in NetBackup OST? Inline solutions are also are inefficient in deduplicating non-file backups, TSM progressive incremental backups, and backups from applications that fragment their data, such as NetWorker, HP Data Protector, and Symantec OST. ContentAware deduplication technology uses the metadata from these backup applications to identify the areas of data that are likely to contain duplicate data. As described in question #4 above, they perform a byte-level comparison of data for optimal capacity reduction. 6. What are realistic expectations for capacity reduction? The efficiency of a deduplication solution to remove duplicate data determines how much space can be saved. However, if there are no duplicate copies of data in the backup, there can be no space savings. Some data types and application environments typically contain more duplicate copies than others. Data deduplication ratios can vary widely depending on a number of factors, from type of data to frequency of data change and retention period. Although many deduplication vendors claim very high deduplication ratios, these ratios assume a 1 percent data change rate and that datasets are the typical mix of data type found in SMB companies. Capacity reduction savings are summarized in Table 1. Note that a 5:1 ratio reduces storage needs by 80 percent and yields tremendous cost savings. However, there is only a 10 percent difference between a 5:1 and a 10:1 reduction ratio. The bulk of actual capacity savings is between a 2:1 and 5:1 deduplication ratio. The amount of duplicate data in a specific environment is also determined by a number of factors, such as: Data type. Unstructured, semi-structured, and structured data each result in different deduplication results as follows. Unstructured data files. File data such as Microsoft Word, PowerPoint, or Excel files created by office workers often contain redundant data and are frequently distributed or copied. Semi-structured and structured data types. Data created by specific business applications that ensure that operations can be run on a day-to-day basis. For example, Microsoft Exchange for and Oracle for a transactional database clearly fall into the must protect category and require frequent backup. Table 1: Reduction Ratio Comparison Space Reduction Ratio Space Reduction Percentage = 1 (1/Space Reduction Ratio) 2:1 1/2 = 50% 5:1 4/5 = 80% 10:1 9/10 = 90% 20:1 19/20 = 95% 100:1 99/100 = 99% 500:1 499/500 = 99.8% Frequency of data change. The less frequently that data is modified the greater the chance that copies of that data will contain duplicate data. The data deduplication ratio will be higher when the change rate is lower. This also implies that a higher data deduplication ratio should be expected as the percentage of reference data to total active data increases, because reference data is not changed. Retention period. A longer retention period increases the likelihood that duplicate data will be found. Assume a retention window of 90 days and an internal process by which full backups are performed weekly. In this example, the initial full backup will only be deduplicated against itself and will result in a small reduction in storage footprint. When the subsequent full weekly backup for week 2 is performed, only the unique data that has been changed will be stored. When the full backup for week 8 is performed, it will be compared against all the unique data for weeks 1, 2, 3, 4, 5, 6, and 7, which increases the chances that duplicate data will be found. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 6

9 Single vs. multiple pools. Some data protection platforms allow you to separate data into discrete storage pools for increased effectiveness for specific data types. For example, a storage pool can be created for all Exchange data. Deduplication will compare a new backup to the previous Exchange data backup, resulting in a much higher reduction ratio than without storage pools, where the same policy and rules are applied across all data, regardless of type. Rather than pushing for higher generic ratios, a more effective strategy is to choose a solution that guarantees the ability to move data to safety within backup windows while also providing efficient deduplication. Concurrent processing and deterministic ingest rate, deduplication, and replication are key enablers to an enterprise environment. 7. Can administrators monitor backup, deduplication, replication, and restore enterprisewide? Deduplication technologies that do not scale, force IT managers to separate backup operations onto multiple independent systems, each of which has to be managed and optimized separately. Scalable, ContentAware deduplication systems enable administrators to monitor and manage the precise status of data as it passes through backup, deduplication, replication and restore and archive operations. This holistic view of the data protection environment enables them to manage more data per administrator, fine tune the backup environment for optimal efficiency, and to plan accurately for future performance and capacity requirements. 8. Can your deduplication help reduce replication bandwidth requirements for large enterprise data volumes? Some inline deduplication technologies enable companies to replicate data across a WAN efficiently. However, these inline solutions cannot handle large data volumes efficiently and do not enable efficient restore times. ContentAware deduplication works in conjunction with replication software to reduce bandwidth requirements by as much as 97 percent. 9. Can I tune the deduplication to meet my needs? Enterprise data protection environments may have data types that have special deduplication requirements. For example, IT manager may not want to deduplicate some volumes of backup data for regulatory compliance purposes. Or, they may not want to waste processing cycles on image data or other data that is unlikely to contain duplicates. Hash-based deduplication technologies are all or nothing and are capable of performing only one level of deduplication comparison. ContentAware solutions enable IT managers to choose the volumes of data that they want to deduplicate by server, backup application, and by data volume. They also automatically detect the type of data being backed-up automatically perform the method of deduplication that is most efficient for that data type. 10. How much experience does your vendor have with your enterprise backup environment? Enterprise data centers with massive data volumes and complex policies need a data protection vendor with experience with enterprise-class backup applications, such as NetBackup, NetBackup OST, and Tivoli Storage Manager. They should be prepared to provide technical advice and guidance on how to optimize the data deduplication technology for optimal results in these environments. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 7

10 Conclusion For large enterprises with massive data volumes, inline deduplication (at the source or target) cannot provide sufficient time to safety, capacity reduction efficiency, or replication bandwidth optimization to be practical or cost-efficient. The Sepaton DeltaStor data deduplication software was designed with a variety of advanced features to specifically address the needs of large enterprises. DeltaStor delivers the fastest, most efficient, and most cost-effective way for enterprises to back up, deduplicate, replicate, and restore large data volumes. DeltaStor s innovative, enterprise-class features include concurrent processing, which enables Sepaton products to back up petabytes of data within stringent backup windows while providing fast, efficient deduplication processing. Concurrent processing also enables DeltaStor to take advantage of Sepaton s unique grid architecture to leverage the processing power of up to eight nodes to perform backup, deduplication, replication, and restore simultaneously at the industry s highest rates. DeltaStor was also designed with unique forward referencing pointers to enable instant restores of the most recently backed up and most recently replicated data for optimal business continuity and RTOs. This feature also enables the industry s most efficient tape vaulting by eliminating the need to reconstitute deduplicated data before it is written to tape. Automated transparent space reclamation saves administration time and frees capacity faster by reclaiming space continuously. Administrators no longer need to plan and schedule cleaning windows or delay backups until capacity becomes available. Because enterprises rely on large, complex databases such as Oracle and SAP to run critical business operations, Sepaton designed DeltaStor to examine data in sub-8 KB increments used in these systems. Patented technologies enable DeltaStor to examine data at a more granular level and to find duplicate data more efficiently than does any other deduplication technology. As a result, Sepaton delivers the most efficient capacity reduction in the industry. As demonstrated in several use case scenarios, many of the world s largest enterprises have standardized on Sepaton S2100 data protection products with DeltaStor. Sepaton delivers the only products with performance, scalability, single-system capacity, deduplication efficiency, and advanced automation needed in today s large enterprise data protection environments. The enterprise-class features of Sepaton products are summarized in Table 2 on the next page. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 8

11 Table 2: DeltaStor Features Specifically Designed for Large Enterprises Enterprise Deduplication Requirement Allow highperformance backup, replication, and recovery Immediate restores Management tools for optimization Reduce space, power Pay as you grow grid scalability Massive single-system performance and capacity Fast, cost-effective replication Lower TCO and admin cost Tape vaulting Enterprise-grade, highreliability design DeltaStor Features Industry-leading deterministic ingest rates and concurrent, multinode processing to meet backup/replication windows and restore SLAs. Forward referencing enables immediate restore of the most commonly needed data. Management reporting and dashboards to enable a single administrator to ensure that petabytes of data are protected, deduplicated, and replicated efficiently. Detailed status and trend reporting pinpoints sources of inefficiency, enabling unparalleled control and cost savings. Single-system global deduplication of all data enables superior deduplication efficiency and reduces data center space, power, and cooling requirements. Enterprise architecture. Add nodes from one to eight as needed to increase performance. Add disk shelves as needed to scale capacity in cost-saving increments. Enterprise-grade scale and performance. Deploy as many as eight processing nodes to achieve industry-leading performance of up to 43.2 TB/hour. Bandwidth-optimized replication of large data volumes by transmitting only new data. Up to 97% bandwidth reduction. Advanced automation simplifies IT management. Bring added performance or storage online, configure devices to the optimal settings, and load balance without operator intervention. Forward referencing keeps the most recent backup in an un-deduplicated form for fast, efficient tape vaulting % data availability; RAID-6 based Hitachi Data Systems enterpriseclass storage; redundant control and data paths; internally mirrored drives; and real-time 24x7 platform monitoring, diagnostics, and automated alerts. About Sepaton Sepaton is the only company in the world that designs and implements data protection solutions specifically for large enterprises. As a result, many of the world s largest enterprises have deployed Sepaton s purpose-built, integrated scale-out data-protection platform. 400 Nickerson Road, Marlborough, MA P: 866.Sepaton or F: Page 9

Improving Efficiency in Tivoli Storage Manager Environments Sepaton S2100 Data Protection Platform

Improving Efficiency in Tivoli Storage Manager Environments Sepaton S2100 Data Protection Platform Improving Efficiency in Tivoli Storage Manager Environments Sepaton S2100 Data Protection Platform White Paper 2013 TABLE OF CONTENTS Introduction... 1 Overview of the TSM Backup and Housekeeping Processes...

More information

DeltaStor Data Deduplication: A Technical Review

DeltaStor Data Deduplication: A Technical Review White Paper DeltaStor Data Deduplication: A Technical Review DeltaStor software is a next-generation data deduplication application for the SEPATON S2100 -ES2 virtual tape library that enables enterprises

More information

Hitachi Protection Platform

Hitachi Protection Platform Hitachi Protection Platform Powerful, Scalable Backup Appliance for Enterprise Data Protection By Victor Nemechek April 2015 Contents Introduction... 3 Cost Saving Solution... 3 Grid Scalability and Massive

More information

Maximizing Deduplication ROI in a NetBackup Environment

Maximizing Deduplication ROI in a NetBackup Environment Use Case Maximizing Deduplication ROI in a NetBackup Environment For many companies, backing up and restoring data is becoming a slow, complicated, expensive process. The volume of data to be protected

More information

Protect Microsoft Exchange databases, achieve long-term data retention

Protect Microsoft Exchange databases, achieve long-term data retention Technical white paper Protect Microsoft Exchange databases, achieve long-term data retention HP StoreOnce Backup systems, HP StoreOnce Catalyst, and Symantec NetBackup OpenStorage Table of contents Introduction...

More information

EMC DATA DOMAIN OPERATING SYSTEM

EMC DATA DOMAIN OPERATING SYSTEM EMC DATA DOMAIN OPERATING SYSTEM Powering EMC Protection Storage ESSENTIALS High-Speed, Scalable Deduplication Up to 58.7 TB/hr performance Reduces requirements for backup storage by 10 to 30x and archive

More information

EMC DATA DOMAIN OPERATING SYSTEM

EMC DATA DOMAIN OPERATING SYSTEM ESSENTIALS HIGH-SPEED, SCALABLE DEDUPLICATION Up to 58.7 TB/hr performance Reduces protection storage requirements by 10 to 30x CPU-centric scalability DATA INVULNERABILITY ARCHITECTURE Inline write/read

More information

STORAGE. Buying Guide: TARGET DATA DEDUPLICATION BACKUP SYSTEMS. inside

STORAGE. Buying Guide: TARGET DATA DEDUPLICATION BACKUP SYSTEMS. inside Managing the information that drives the enterprise STORAGE Buying Guide: DEDUPLICATION inside What you need to know about target data deduplication Special factors to consider One key difference among

More information

ExaGrid - A Backup and Data Deduplication appliance

ExaGrid - A Backup and Data Deduplication appliance Detailed Product Description ExaGrid Backup Storage Appliances with Deduplication 2014 ExaGrid Systems, Inc. All rights reserved. Table of Contents Executive Summary...2 ExaGrid Basic Concept...2 ExaGrid

More information

Protect Data... in the Cloud

Protect Data... in the Cloud QUASICOM Private Cloud Backups with ExaGrid Deduplication Disk Arrays Martin Lui Senior Solution Consultant Quasicom Systems Limited Protect Data...... in the Cloud 1 Mobile Computing Users work with their

More information

Eight Considerations for Evaluating Disk-Based Backup Solutions

Eight Considerations for Evaluating Disk-Based Backup Solutions Eight Considerations for Evaluating Disk-Based Backup Solutions 1 Introduction The movement from tape-based to disk-based backup is well underway. Disk eliminates all the problems of tape backup. Backing

More information

Data De-duplication Methodologies: Comparing ExaGrid s Byte-level Data De-duplication To Block Level Data De-duplication

Data De-duplication Methodologies: Comparing ExaGrid s Byte-level Data De-duplication To Block Level Data De-duplication Data De-duplication Methodologies: Comparing ExaGrid s Byte-level Data De-duplication To Block Level Data De-duplication Table of Contents Introduction... 3 Shortest Possible Backup Window... 3 Instant

More information

Detailed Product Description

Detailed Product Description Detailed Product Description ExaGrid Disk Backup with Deduplication 2014 ExaGrid Systems, Inc. All rights reserved. Table of Contents Executive Summary...2 ExaGrid Basic Concept...2 Product Benefits...

More information

Enterprise-class Backup Performance with Dell DR6000 Date: May 2014 Author: Kerry Dolan, Lab Analyst and Vinny Choinski, Senior Lab Analyst

Enterprise-class Backup Performance with Dell DR6000 Date: May 2014 Author: Kerry Dolan, Lab Analyst and Vinny Choinski, Senior Lab Analyst ESG Lab Review Enterprise-class Backup Performance with Dell DR6000 Date: May 2014 Author: Kerry Dolan, Lab Analyst and Vinny Choinski, Senior Lab Analyst Abstract: This ESG Lab review documents hands-on

More information

ExaGrid Product Description. Cost-Effective Disk-Based Backup with Data Deduplication

ExaGrid Product Description. Cost-Effective Disk-Based Backup with Data Deduplication ExaGrid Product Description Cost-Effective Disk-Based Backup with Data Deduplication 1 Contents Introduction... 3 Considerations When Examining Disk-Based Backup Approaches... 3 ExaGrid A Disk-Based Backup

More information

Turnkey Deduplication Solution for the Enterprise

Turnkey Deduplication Solution for the Enterprise Symantec NetBackup 5000 Appliance Turnkey Deduplication Solution for the Enterprise Mayur Dewaikar Sr. Product Manager, Information Management Group White Paper: A Deduplication Appliance Solution for

More information

WHITE PAPER Improving Storage Efficiencies with Data Deduplication and Compression

WHITE PAPER Improving Storage Efficiencies with Data Deduplication and Compression WHITE PAPER Improving Storage Efficiencies with Data Deduplication and Compression Sponsored by: Oracle Steven Scully May 2010 Benjamin Woo IDC OPINION Global Headquarters: 5 Speen Street Framingham, MA

More information

Technology Fueling the Next Phase of Storage Optimization

Technology Fueling the Next Phase of Storage Optimization White Paper HP StoreOnce Deduplication Software Technology Fueling the Next Phase of Storage Optimization By Lauren Whitehouse June, 2010 This ESG White Paper was commissioned by Hewlett-Packard and is

More information

Using HP StoreOnce Backup Systems for NDMP backups with Symantec NetBackup

Using HP StoreOnce Backup Systems for NDMP backups with Symantec NetBackup Technical white paper Using HP StoreOnce Backup Systems for NDMP backups with Symantec NetBackup Table of contents Executive summary... 2 Introduction... 2 What is NDMP?... 2 Technology overview... 3 HP

More information

LDA, the new family of Lortu Data Appliances

LDA, the new family of Lortu Data Appliances LDA, the new family of Lortu Data Appliances Based on Lortu Byte-Level Deduplication Technology February, 2011 Copyright Lortu Software, S.L. 2011 1 Index Executive Summary 3 Lortu deduplication technology

More information

Efficient Backup with Data Deduplication Which Strategy is Right for You?

Efficient Backup with Data Deduplication Which Strategy is Right for You? Efficient Backup with Data Deduplication Which Strategy is Right for You? Rob Emsley Senior Director, Product Marketing CPU Utilization CPU Utilization Exabytes Why So Much Interest in Data Deduplication?

More information

IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE

IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE White Paper IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE Abstract This white paper focuses on recovery of an IBM Tivoli Storage Manager (TSM) server and explores

More information

Symantec NetBackup 5220

Symantec NetBackup 5220 A single-vendor enterprise backup appliance that installs in minutes Data Sheet: Data Protection Overview is a single-vendor enterprise backup appliance that installs in minutes, with expandable storage

More information

Symantec NetBackup 5000 Appliance Series

Symantec NetBackup 5000 Appliance Series A turnkey, end-to-end, global deduplication solution for the enterprise. Data Sheet: Data Protection Overview Symantec NetBackup 5000 series offers your organization a content aware, end-to-end, and global

More information

Best Practices Guide. Symantec NetBackup with ExaGrid Disk Backup with Deduplication. 2012 ExaGrid Systems, Inc. All rights reserved.

Best Practices Guide. Symantec NetBackup with ExaGrid Disk Backup with Deduplication. 2012 ExaGrid Systems, Inc. All rights reserved. Best Practices Guide Symantec NetBackup with ExaGrid Disk Backup with Deduplication 2012 ExaGrid Systems, Inc. All rights reserved. Table of Contents EXECUTIVE SUMMARY... 2 INTRODUCTION... 2 AUDIENCE...

More information

EMC Data Domain Boost for Oracle Recovery Manager (RMAN)

EMC Data Domain Boost for Oracle Recovery Manager (RMAN) White Paper EMC Data Domain Boost for Oracle Recovery Manager (RMAN) Abstract EMC delivers Database Administrators (DBAs) complete control of Oracle backup, recovery, and offsite disaster recovery with

More information

Demystifying Deduplication for Backup with the Dell DR4000

Demystifying Deduplication for Backup with the Dell DR4000 Demystifying Deduplication for Backup with the Dell DR4000 This Dell Technical White Paper explains how deduplication with the DR4000 can help your organization save time, space, and money. John Bassett

More information

Protecting enterprise servers with StoreOnce and CommVault Simpana

Protecting enterprise servers with StoreOnce and CommVault Simpana Technical white paper Protecting enterprise servers with StoreOnce and CommVault Simpana HP StoreOnce Backup systems Table of contents Introduction 2 Technology overview 2 HP StoreOnce Backup systems key

More information

WHY DO I NEED FALCONSTOR OPTIMIZED BACKUP & DEDUPLICATION?

WHY DO I NEED FALCONSTOR OPTIMIZED BACKUP & DEDUPLICATION? WHAT IS FALCONSTOR? FalconStor Optimized Backup and Deduplication is the industry s market-leading virtual tape and LAN-based deduplication solution, unmatched in performance and scalability. With virtual

More information

How To Make A Backup System More Efficient

How To Make A Backup System More Efficient Identifying the Hidden Risk of Data De-duplication: How the HYDRAstor Solution Proactively Solves the Problem October, 2006 Introduction Data de-duplication has recently gained significant industry attention,

More information

HP StoreOnce: reinventing data deduplication

HP StoreOnce: reinventing data deduplication HP : reinventing data deduplication Reduce the impact of explosive data growth with HP StorageWorks D2D Backup Systems Technical white paper Table of contents Executive summary... 2 Introduction to data

More information

Identifying the Hidden Risk of Data Deduplication: How the HYDRAstor TM Solution Proactively Solves the Problem

Identifying the Hidden Risk of Data Deduplication: How the HYDRAstor TM Solution Proactively Solves the Problem Identifying the Hidden Risk of Data Deduplication: How the HYDRAstor TM Solution Proactively Solves the Problem Advanced Storage Products Group Table of Contents 1 - Introduction 2 Data Deduplication 3

More information

Data Deduplication: An Essential Component of your Data Protection Strategy

Data Deduplication: An Essential Component of your Data Protection Strategy WHITE PAPER: THE EVOLUTION OF DATA DEDUPLICATION Data Deduplication: An Essential Component of your Data Protection Strategy JULY 2010 Andy Brewerton CA TECHNOLOGIES RECOVERY MANAGEMENT AND DATA MODELLING

More information

Protecting Information in a Smarter Data Center with the Performance of Flash

Protecting Information in a Smarter Data Center with the Performance of Flash 89 Fifth Avenue, 7th Floor New York, NY 10003 www.theedison.com 212.367.7400 Protecting Information in a Smarter Data Center with the Performance of Flash IBM FlashSystem and IBM ProtecTIER Printed in

More information

EMC BACKUP MEETS BIG DATA

EMC BACKUP MEETS BIG DATA EMC BACKUP MEETS BIG DATA Strategies To Protect Greenplum, Isilon And Teradata Systems 1 Agenda Big Data: Overview, Backup and Recovery EMC Big Data Backup Strategy EMC Backup and Recovery Solutions for

More information

Real-time Compression: Achieving storage efficiency throughout the data lifecycle

Real-time Compression: Achieving storage efficiency throughout the data lifecycle Real-time Compression: Achieving storage efficiency throughout the data lifecycle By Deni Connor, founding analyst Patrick Corrigan, senior analyst July 2011 F or many companies the growth in the volume

More information

Reducing Backups with Data Deduplication

Reducing Backups with Data Deduplication The Essentials Series: New Techniques for Creating Better Backups Reducing Backups with Data Deduplication sponsored by by Eric Beehler Reducing Backups with Data Deduplication... 1 Explaining Data Deduplication...

More information

Oracle Data Protection Concepts

Oracle Data Protection Concepts Oracle Data Protection Concepts Matthew Ellis Advisory Systems Engineer BRS Database Technologist, EMC Corporation Accelerating Transformation EMC Backup Recovery Systems Division 1 Agenda Market Conditions.

More information

Using HP StoreOnce Backup systems for Oracle database backups

Using HP StoreOnce Backup systems for Oracle database backups Technical white paper Using HP StoreOnce Backup systems for Oracle database backups Table of contents Introduction 2 Technology overview 2 HP StoreOnce Backup systems key features and benefits 2 HP StoreOnce

More information

EMC DATA DOMAIN PRODUCT OvERvIEW

EMC DATA DOMAIN PRODUCT OvERvIEW EMC DATA DOMAIN PRODUCT OvERvIEW Deduplication storage for next-generation backup and archive Essentials Scalable Deduplication Fast, inline deduplication Provides up to 65 PBs of logical storage for long-term

More information

ExaGrid s EX32000E is its newest and largest appliance, taking in a 32TB full backup with an ingest rate of 7.5TB/hour.

ExaGrid s EX32000E is its newest and largest appliance, taking in a 32TB full backup with an ingest rate of 7.5TB/hour. Nominee: ExaGrid Nomination title: ExaGrid s next-generation appliances continue market disruption with bigger/faster appliances and larger scalable GRID systems in 2015 ExaGrid s EX32000E is its newest

More information

Data Deduplication in Tivoli Storage Manager. Andrzej Bugowski 19-05-2011 Spała

Data Deduplication in Tivoli Storage Manager. Andrzej Bugowski 19-05-2011 Spała Data Deduplication in Tivoli Storage Manager Andrzej Bugowski 19-05-2011 Spała Agenda Tivoli Storage, IBM Software Group Deduplication concepts Data deduplication in TSM 6.1 Planning for data deduplication

More information

EMC Data Domain Boost for Oracle Recovery Manager (RMAN)

EMC Data Domain Boost for Oracle Recovery Manager (RMAN) White Paper EMC Data Domain Boost for Oracle Recovery Manager (RMAN) Abstract EMC delivers Database Administrators (DBAs) complete control of Oracle backup, recovery, and offsite disaster recovery with

More information

GIVE YOUR ORACLE DBAs THE BACKUPS THEY REALLY WANT

GIVE YOUR ORACLE DBAs THE BACKUPS THEY REALLY WANT Why Data Domain Series GIVE YOUR ORACLE DBAs THE BACKUPS THEY REALLY WANT Why you should take the time to read this paper Speed up backups (Up to 58.7 TB/hr, Data Domain systems are about 1.5 times faster

More information

ESG REPORT. Data Deduplication Diversity: Evaluating Software- vs. Hardware-Based Approaches. By Lauren Whitehouse. April, 2009

ESG REPORT. Data Deduplication Diversity: Evaluating Software- vs. Hardware-Based Approaches. By Lauren Whitehouse. April, 2009 ESG REPORT : Evaluating Software- vs. Hardware-Based Approaches By Lauren Whitehouse April, 2009 Table of Contents ESG REPORT Table of Contents... i Introduction... 1 External Forces Contribute to IT Challenges...

More information

HP Data Protector software and HP StoreOnce backup systems for federated deduplication and flexible deployment

HP Data Protector software and HP StoreOnce backup systems for federated deduplication and flexible deployment Brochure Maximize storage efficiency across the enterprise HP Data Protector software and HP StoreOnce backup systems for federated deduplication and flexible deployment Maximize storage efficiency across

More information

Backup and Recovery for SAP Environments using EMC Avamar 7

Backup and Recovery for SAP Environments using EMC Avamar 7 White Paper Backup and Recovery for SAP Environments using EMC Avamar 7 Abstract This white paper highlights how IT environments deploying SAP can benefit from efficient backup with an EMC Avamar solution.

More information

Evaluation Guide. Software vs. Appliance Deduplication

Evaluation Guide. Software vs. Appliance Deduplication Evaluation Guide Software vs. Appliance Deduplication Table of Contents Introduction... 2 Data Deduplication Overview... 3 Backup Requirements... 6 Backup Application Client Side Deduplication... 7 Backup

More information

SYMANTEC NETBACKUP APPLIANCE FAMILY OVERVIEW BROCHURE. When you can do it simply, you can do it all.

SYMANTEC NETBACKUP APPLIANCE FAMILY OVERVIEW BROCHURE. When you can do it simply, you can do it all. SYMANTEC NETBACKUP APPLIANCE FAMILY OVERVIEW BROCHURE When you can do it simply, you can do it all. SYMANTEC NETBACKUP APPLIANCES Symantec understands the shifting needs of the data center and offers NetBackup

More information

Long term retention and archiving the challenges and the solution

Long term retention and archiving the challenges and the solution Long term retention and archiving the challenges and the solution NAME: Yoel Ben-Ari TITLE: VP Business Development, GH Israel 1 Archive Before Backup EMC recommended practice 2 1 Backup/recovery process

More information

DXi Accent Technical Background

DXi Accent Technical Background TECHNOLOGY BRIEF NOTICE This Technology Brief contains information protected by copyright. Information in this Technology Brief is subject to change without notice and does not represent a commitment on

More information

Data Reduction Methodologies: Comparing ExaGrid s Byte-Level-Delta Data Reduction to Data De-duplication. February 2007

Data Reduction Methodologies: Comparing ExaGrid s Byte-Level-Delta Data Reduction to Data De-duplication. February 2007 Data Reduction Methodologies: Comparing ExaGrid s Byte-Level-Delta Data Reduction to Data De-duplication February 2007 Though data reduction technologies have been around for years, there is a renewed

More information

3Gen Data Deduplication Technical

3Gen Data Deduplication Technical 3Gen Data Deduplication Technical Discussion NOTICE: This White Paper may contain proprietary information protected by copyright. Information in this White Paper is subject to change without notice and

More information

EMC DATA DOMAIN OVERVIEW. Copyright 2011 EMC Corporation. All rights reserved.

EMC DATA DOMAIN OVERVIEW. Copyright 2011 EMC Corporation. All rights reserved. EMC DATA DOMAIN OVERVIEW 1 2 With Data Domain Deduplication Storage Systems, You Can WAN Retain longer Keep backups onsite longer with less disk for fast, reliable restores, and eliminate the use of tape

More information

NetApp Syncsort Integrated Backup

NetApp Syncsort Integrated Backup WHITE PAPER NetApp Syncsort Integrated Backup Protect your Microsoft and VMware Environment with NetApp Syncsort Integrated Backup Protecting Microsoft and VMware Executive Summary 3 Microsoft and VMware

More information

Every organization has critical data that it can t live without. When a disaster strikes, how long can your business survive without access to its

Every organization has critical data that it can t live without. When a disaster strikes, how long can your business survive without access to its DISASTER RECOVERY STRATEGIES: BUSINESS CONTINUITY THROUGH REMOTE BACKUP REPLICATION Every organization has critical data that it can t live without. When a disaster strikes, how long can your business

More information

How To Protect Data On Network Attached Storage (Nas) From Disaster

How To Protect Data On Network Attached Storage (Nas) From Disaster White Paper EMC FOR NETWORK ATTACHED STORAGE (NAS) BACKUP AND RECOVERY Abstract This white paper provides an overview of EMC s industry leading backup and recovery solutions for NAS systems. It also explains

More information

Detailed Product Description

Detailed Product Description Detailed Product Description ExaGrid Backup Storage Appliances with Data Deduplication 2015 ExaGrid Systems, Inc. All rights reserved. Table of Contents Executive Summary...2 ExaGrid Basic Concept...2

More information

Deduplication and Beyond: Optimizing Performance for Backup and Recovery

Deduplication and Beyond: Optimizing Performance for Backup and Recovery Beyond: Optimizing Gartner clients using deduplication for backups typically report seven times to 25 times the reductions (7:1 to 25:1) in the size of their data, and sometimes higher than 100:1 for file

More information

DEDUPLICATION NOW AND WHERE IT S HEADING. Lauren Whitehouse Senior Analyst, Enterprise Strategy Group

DEDUPLICATION NOW AND WHERE IT S HEADING. Lauren Whitehouse Senior Analyst, Enterprise Strategy Group DEDUPLICATION NOW AND WHERE IT S HEADING Lauren Whitehouse Senior Analyst, Enterprise Strategy Group Need Dedupe? Before/After Dedupe Deduplication Production Data Deduplication In Backup Process Backup

More information

Quantum DXi6500 Family of Network-Attached Disk Backup Appliances with Deduplication

Quantum DXi6500 Family of Network-Attached Disk Backup Appliances with Deduplication PRODUCT BRIEF Quantum DXi6500 Family of Network-Attached Disk Backup Appliances with Deduplication NOTICE This Product Brief contains proprietary information protected by copyright. Information in this

More information

Business Benefits of Data Footprint Reduction

Business Benefits of Data Footprint Reduction Business Benefits of Data Footprint Reduction Why and how reducing your data footprint provides a positive benefit to your business and application service objectives By Greg Schulz Founder and Senior

More information

Actifio Big Data Director. Virtual Data Pipeline for Unstructured Data

Actifio Big Data Director. Virtual Data Pipeline for Unstructured Data Actifio Big Data Director Virtual Data Pipeline for Unstructured Data Contact Actifio Support As an Actifio customer, you can get support for all Actifio products through the Support Portal at http://support.actifio.com/.

More information

Backup Software Data Deduplication: What you need to know. Presented by W. Curtis Preston Executive Editor & Independent Backup Expert

Backup Software Data Deduplication: What you need to know. Presented by W. Curtis Preston Executive Editor & Independent Backup Expert Backup Software Data Deduplication: What you need to know Presented by W. Curtis Preston Executive Editor & Independent Backup Expert When I was in the IT Department When I started as backup guy at $35B

More information

EMC Disk Library with EMC Data Domain Deployment Scenario

EMC Disk Library with EMC Data Domain Deployment Scenario EMC Disk Library with EMC Data Domain Deployment Scenario Best Practices Planning Abstract This white paper is an overview of the EMC Disk Library with EMC Data Domain deduplication storage system deployment

More information

Get Success in Passing Your Certification Exam at first attempt!

Get Success in Passing Your Certification Exam at first attempt! Get Success in Passing Your Certification Exam at first attempt! Exam : E22-290 Title : EMC Data Domain Deduplication, Backup and Recovery Exam Version : DEMO 1.A customer has a Data Domain system with

More information

ABOUT DISK BACKUP WITH DEDUPLICATION

ABOUT DISK BACKUP WITH DEDUPLICATION Disk Backup with Data Deduplication ABOUT DISK BACKUP WITH DEDUPLICATION www.exagrid.com What appears to be simple & straightforward Built for Backup is often more complex & risky than you think. 2 Agenda

More information

Deduplication s Role in Disaster Recovery. Thomas Rivera, SEPATON

Deduplication s Role in Disaster Recovery. Thomas Rivera, SEPATON Thomas Rivera, SEPATON SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA. Member companies and individual members may use this material in presentations and literature

More information

Data deduplication is more than just a BUZZ word

Data deduplication is more than just a BUZZ word Data deduplication is more than just a BUZZ word Per Larsen Principal Systems Engineer Mr. Hansen DATA BUDGET RECOVERY & DATACENTER GROWTH PRESSURE DISCOVERY REVOLUTION More Storage Longer Backups Smaller

More information

Solving Data Growth Issues using Deduplication

Solving Data Growth Issues using Deduplication Solving Data Growth Issues using Deduplication Reducing Storage Costs and Speeding Backups with Dell Deduplication Solutions Abstract Data growth is increasing at record rates, and ensuring the integrity

More information

Deduplication s Role in Disaster Recovery. Thomas Rivera, SEPATON

Deduplication s Role in Disaster Recovery. Thomas Rivera, SEPATON Thomas Rivera, SEPATON SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA. Member companies and individual members may use this material in presentations and literature

More information

WHITE PAPER BRENT WELCH NOVEMBER

WHITE PAPER BRENT WELCH NOVEMBER BACKUP WHITE PAPER BRENT WELCH NOVEMBER 2006 WHITE PAPER: BACKUP TABLE OF CONTENTS Backup Overview 3 Background on Backup Applications 3 Backup Illustration 4 Media Agents & Keeping Tape Drives Busy 5

More information

<Insert Picture Here> Refreshing Your Data Protection Environment with Next-Generation Architectures

<Insert Picture Here> Refreshing Your Data Protection Environment with Next-Generation Architectures 1 Refreshing Your Data Protection Environment with Next-Generation Architectures Dale Rhine, Principal Sales Consultant Kelly Boeckman, Product Marketing Analyst Program Agenda Storage

More information

Veritas Backup Exec 15: Deduplication Option

Veritas Backup Exec 15: Deduplication Option Veritas Backup Exec 15: Deduplication Option Who should read this paper Technical White Papers are designed to introduce IT professionals to key technologies and technical concepts that are associated

More information

Top Ten Questions. to Ask Your Primary Storage Provider About Their Data Efficiency. May 2014. Copyright 2014 Permabit Technology Corporation

Top Ten Questions. to Ask Your Primary Storage Provider About Their Data Efficiency. May 2014. Copyright 2014 Permabit Technology Corporation Top Ten Questions to Ask Your Primary Storage Provider About Their Data Efficiency May 2014 Copyright 2014 Permabit Technology Corporation Introduction The value of data efficiency technologies, namely

More information

Barracuda Backup Deduplication. White Paper

Barracuda Backup Deduplication. White Paper Barracuda Backup Deduplication White Paper Abstract Data protection technologies play a critical role in organizations of all sizes, but they present a number of challenges in optimizing their operation.

More information

Future-Proofed Backup For A Virtualized World!

Future-Proofed Backup For A Virtualized World! ! Future-Proofed Backup For A Virtualized World! Prepared by: Colm Keegan, Senior Analyst! Prepared: January 2014 Future-Proofed Backup For A Virtualized World Like death and taxes, growing backup windows

More information

Using HP StoreOnce D2D systems for Microsoft SQL Server backups

Using HP StoreOnce D2D systems for Microsoft SQL Server backups Technical white paper Using HP StoreOnce D2D systems for Microsoft SQL Server backups Table of contents Executive summary 2 Introduction 2 Technology overview 2 HP StoreOnce D2D systems key features and

More information

DEDUPLICATION SOLUTIONS ARE NOT ALL CREATED EQUAL, WHY DATA DOMAIN?

DEDUPLICATION SOLUTIONS ARE NOT ALL CREATED EQUAL, WHY DATA DOMAIN? Why Data Domain Series DEDUPLICATION SOLUTIONS ARE NOT ALL CREATED EQUAL, WHY DATA DOMAIN? Why you should take the time to read this paper Speed up your backups (Achieve up to 31 TB/hr - 3 times faster

More information

WHITE PAPER. Effectiveness of Variable-block vs Fixedblock Deduplication on Data Reduction: A Technical Analysis

WHITE PAPER. Effectiveness of Variable-block vs Fixedblock Deduplication on Data Reduction: A Technical Analysis WHITE PAPER Effectiveness of Variable-block vs Fixedblock Deduplication on Data Reduction: A Technical Analysis CONTENTS Executive Summary... 3 Fixed vs. Variable-block Deduplication... 3 Test Configuration...

More information

Riverbed Whitewater/Amazon Glacier ROI for Backup and Archiving

Riverbed Whitewater/Amazon Glacier ROI for Backup and Archiving Riverbed Whitewater/Amazon Glacier ROI for Backup and Archiving November, 2013 Saqib Jang Abstract This white paper demonstrates how to increase profitability by reducing the operating costs of backup

More information

An Oracle White Paper November 2010. Backup and Recovery with Oracle s Sun ZFS Storage Appliances and Oracle Recovery Manager

An Oracle White Paper November 2010. Backup and Recovery with Oracle s Sun ZFS Storage Appliances and Oracle Recovery Manager An Oracle White Paper November 2010 Backup and Recovery with Oracle s Sun ZFS Storage Appliances and Oracle Recovery Manager Introduction...2 Oracle Backup and Recovery Solution Overview...3 Oracle Recovery

More information

Sales Tool. Summary DXi Sales Messages November 2009 6 NOVEMBER 2009. ST00431-v06

Sales Tool. Summary DXi Sales Messages November 2009 6 NOVEMBER 2009. ST00431-v06 Summary DXi Sales Messages November 2009 6 NOVEMBER 2009 ST00431-v06 Notice This Sales Tool contains proprietary information protected by copyright. Information in this Sales Tool is subject to change

More information

Deduplication has been around for several

Deduplication has been around for several Demystifying Deduplication By Joe Colucci Kay Benaroch Deduplication holds the promise of efficient storage and bandwidth utilization, accelerated backup and recovery, reduced costs, and more. Understanding

More information

SLOW BACKUPS GOT YOU DOWN?

SLOW BACKUPS GOT YOU DOWN? Why Data Domain Series SLOW BACKUPS GOT YOU DOWN? Why you should take the time to read this paper Speed up backups by 50% (Finish backups within backup windows with breathing room for data growth. With

More information

Disaster Recovery Strategies: Business Continuity through Remote Backup Replication

Disaster Recovery Strategies: Business Continuity through Remote Backup Replication W H I T E P A P E R S O L U T I O N : D I S A S T E R R E C O V E R Y T E C H N O L O G Y : R E M O T E R E P L I C A T I O N Disaster Recovery Strategies: Business Continuity through Remote Backup Replication

More information

Reduce your data storage footprint and tame the information explosion

Reduce your data storage footprint and tame the information explosion IBM Software White paper December 2010 Reduce your data storage footprint and tame the information explosion 2 Reduce your data storage footprint and tame the information explosion Contents 2 Executive

More information

NetApp Data Fabric: Secured Backup to Public Cloud. Sonny Afen Senior Technical Consultant NetApp Indonesia

NetApp Data Fabric: Secured Backup to Public Cloud. Sonny Afen Senior Technical Consultant NetApp Indonesia NetApp Data Fabric: Secured Backup to Public Cloud Sonny Afen Senior Technical Consultant NetApp Indonesia Agenda Introduction Solution Overview Solution Technical Overview 2 Introduction 3 Hybrid cloud:

More information

EMC DATA DOMAIN EXTENDED RETENTION SOFTWARE: MEETING NEEDS FOR LONG-TERM RETENTION OF BACKUP DATA ON EMC DATA DOMAIN SYSTEMS

EMC DATA DOMAIN EXTENDED RETENTION SOFTWARE: MEETING NEEDS FOR LONG-TERM RETENTION OF BACKUP DATA ON EMC DATA DOMAIN SYSTEMS SOLUTION PROFILE EMC DATA DOMAIN EXTENDED RETENTION SOFTWARE: MEETING NEEDS FOR LONG-TERM RETENTION OF BACKUP DATA ON EMC DATA DOMAIN SYSTEMS MAY 2012 Backups are essential for short-term data recovery

More information

Technical White Paper for the Oceanspace VTL6000

Technical White Paper for the Oceanspace VTL6000 Document No. Technical White Paper for the Oceanspace VTL6000 Issue V2.1 Date 2010-05-18 Huawei Symantec Technologies Co., Ltd. Copyright Huawei Symantec Technologies Co., Ltd. 2010. All rights reserved.

More information

Backup Exec 15: Deduplication Option

Backup Exec 15: Deduplication Option TECHNICAL BRIEF: BACKUP EXEC 15: DEDUPLICATION OPTION........................................ Backup Exec 15: Deduplication Option Who should read this paper Technical White Papers are designed to introduce

More information

Backup Exec 2014: Deduplication Option

Backup Exec 2014: Deduplication Option TECHNICAL BRIEF: BACKUP EXEC 2014: DEDUPLICATION OPTION........................................ Backup Exec 2014: Deduplication Option Who should read this paper Technical White Papers are designed to

More information

VMware vsphere Data Protection

VMware vsphere Data Protection FREQUENTLY ASKED QUESTIONS VMware vsphere Data Protection vsphere Data Protection Advanced Overview Q. What is VMware vsphere Data Protection Advanced? A. VMware vsphere Data Protection Advanced is a backup

More information

Understanding EMC Avamar with EMC Data Protection Advisor

Understanding EMC Avamar with EMC Data Protection Advisor Understanding EMC Avamar with EMC Data Protection Advisor Applied Technology Abstract EMC Data Protection Advisor provides a comprehensive set of features that reduce the complexity of managing data protection

More information

How To Store Data On Disk On Data Domain

How To Store Data On Disk On Data Domain WHITE PAPER Deduplication Storage Deduplication Storage for Nearline Applications September 2007 w w w. d a t a d o m a i n. c o m - 2 0 0 7 DATA DOMAIN I Contents Consolidated Support for Backup, Archiving,

More information

IBM PROTECTIER: FROM BACKUP TO RECOVERY

IBM PROTECTIER: FROM BACKUP TO RECOVERY SOLUTION PROFILE IBM PROTECTIER: FROM BACKUP TO RECOVERY NOVEMBER 2011 When it comes to backup and recovery, backup performance numbers rule the roost. It s understandable really: far more data gets backed

More information

Archiving, Backup, and Recovery for Complete the Promise of Virtualization

Archiving, Backup, and Recovery for Complete the Promise of Virtualization Archiving, Backup, and Recovery for Complete the Promise of Virtualization Unified information management for enterprise Windows environments The explosion of unstructured information It is estimated that

More information

HP StorageWorks D2D Backup Systems and StoreOnce

HP StorageWorks D2D Backup Systems and StoreOnce AUtOMATEyour data protection. HP StorageWorks D2D Backup Systems and StoreOnce The combination that right-sizes your storage capacity. Solution brief Regardless of size and industry, many of today s organizations

More information

Data Deduplication and Tivoli Storage Manager

Data Deduplication and Tivoli Storage Manager Data Deduplication and Tivoli Storage Manager Dave Cannon Tivoli Storage Manager rchitect Oxford University TSM Symposium September 2007 Disclaimer This presentation describes potential future enhancements

More information

IBM Storwize V7000 Unified and Storwize V7000 storage systems

IBM Storwize V7000 Unified and Storwize V7000 storage systems IBM Storwize V7000 Unified and Storwize V7000 storage systems Transforming the economics of data storage Highlights Meet changing business needs with virtualized, enterprise-class, flashoptimized modular

More information