Hardware Configuration Guide

Size: px
Start display at page:

Download "Hardware Configuration Guide"

Transcription

1 Hardware Configuration Guide Contents Contents... 1 Annotation... 1 Factors to consider... 2 Machine Count... 2 Data Size... 2 Data Size Total... 2 Daily Backup Data Size... 2 Unique Data Percentage... 2 Backup Type... 3 Backup Scheme... 3 Backup Window... 3 Retention Period... 3 Hardware... 4 Storage... 4 Backup data... 4 Deduplication database... 7 Catalog database... 8 Other Storage Node databases... 9 CPU... 9 RAM... 9 Network Data Transfer Rate Number of Storage Nodes Time to process backups Time to reindex backups Annotation The following document describes what are the main factors affecting the backup and deduplication speed. This document can be used to estimate the amount of backup data occurred after some period of time and hardware configuration (primarily of the Storage Node) needed to manage this data.

2 Factors to consider The affordable Storage Node configuration depends on a lot of factors. There is a bunch of questions to be answered to define the explicit hardware component parameters to be used for storing and effectively processing backups. The most important factors to be taken into account are represented as parameters and appropriate formulas are provided. Machine Count The Machine Count parameter roughly shows the amount of work for Storage Node to be done. 1-4 machines can be handled by ASN with almost any configuration. Having 500 machines even with 5-10 GB of data will most probably require a high-performance server for the Storage Node. The Machine Count parameter helps to decide how many Storage Nodes you need, what network bandwidth is most appropriate. In conjunction with average data size information this number can be used to estimate the capacity of backup storage. Note that the Machine Count parameter should consider the future growth. Data Size Data Size parameter, which means the average amount of data on the machine, is used for estimation of the backup storage size. Additionally it is a basis for calculating the Daily Backup Data Size. Data Size Total Sometimes it is more convenient to use the Data Size Total parameter which means a total amount of data on all machines in the environment to be backed up. This parameter value is calculated by the formula: Data Size Total = Machine Count * Data Size Daily Backup Data Size Daily Backup Data Size parameter shows how much original backup data to be processed appears every day. This data needs to be backed up so it is taken into account in capacity calculations. Additionally this amount is used to calculate the needed backup window. It is convenient to use the Daily Backup Data Percentage parameter which is specified in percents. Backup Data Size can be calculated by the formula: Backup Data Size = Data Size * Daily Backup Data Percentage / 100 Unique Data Percentage Unique Data Percentage parameter means how much unique data are there on the machine in general. User data are usually unique. Operating system or program files are usually duplicated. This parameter value depends on the backed up machine purpose it can be office machine with low percentage of unique data or file server with high percentage of unique data or something else. Usually this parameter value it is taken as 10-20%. If you already have a Storage Node, one of the ways how to calculate this amount is to perform full back up of several machines and use the resultant Deduplication Ratio in the following formula:

3 Unique Data Percentage (%) = (Deduplication Ratio * Machine Count / 100 1) / (Number of Backups - 1) Unique Data Size can be calculated by the formula: Unique Data Size = Data Size * Unique Data Percentage / 100 This parameter affects the deduplication ratio and as a result backup storage savings. Additionally the less unique data is on a machine, the less backup traffic to the deduplicating vault is. Backup Type Disk-level backups are used several times more often than file-level backups especially for servers as this type of backups can be used for system recovery. File-level backups are convenient for storing user data when the data is important and system itself is not needed to be recovered. Disk-level backups are performed faster than file-level ones. Backup Scheme The backup scheme defines the method and frequency of backups. Here is a list of available backup schemes: - The Simple scheme is designed for quick setting up daily backup. Backups generally depend on the previous ones up to the very first one. - The Grandfather-Father-Son scheme allows you to set the days of week when the daily backup will be performed and select from these days the day of weekly/monthly backup. - The Tower of Hanoi backup scheme allows you to schedule when and how often to back up and select the number of backup levels. By setting up the backup schedule and selecting backup levels, you automatically obtain the rollback period the guaranteed number of sessions that is possible to go back at any time. TOH provides the most effective distribution of backups on the timeline. - The Custom scheme provides the most flexibility in defining backup schedules and retention rules. Backup Window Backup window defines the time when backups are allowed. Backups are usually configured for night time to avoid affecting performance of working machines in business hours. Backup window affects the number of Storage Nodes to be used. If one Storage Node cannot process backups of all machines, an additional one should be added. That is defined by comparing the length of the backup window and a time needed for backups. Retention Period Retention period defines how long the backups should be stored. Backup schemes provide the ability to adjust the retention period. Retention period affects the capacity of a backup storage.

4 Hardware The backup requirements are the basis for finding the affordable configuration for a Storage Node. Each configuration parameter depends on one or more requirements. This section describes how to define the configuration parameters of a Storage Node: storage size and type, RAM size, CPU speed. Storage Amount of space occupied by the backups is one of the most important parameters of a Storage Node configuration. The backups and their metadata are stored in several places: deduplication data store, deduplication database, catalog database. The following sections show how to estimate the size of each of them. Backup data Capacity The capacity which is taken by the backups mainly depends on backup data size and used backup schemes/schedules. Here are the details specific for the backup schemes: - Simple scheme daily incremental backups are performed. First full backup can be performed in an implementation phase which should not be taken into account. The size of daily backup is a size of daily incremental data. - Grandfather-Father-Son scheme supposes daily backups. Full backups are made on a repetitive basis. The backup window should fit the time of full backup creation (or several backup windows can be defined for each type of backup). The size of largest differential backup is 15 sizes of daily incremental data but as it is made at the same day as the full backup, the backup window for this day must fit full backup time. - Tower of Hanoi supposes the creation of incremental backups every 2 nd day. Differential backups are created every other second day. Once on a period a full backup is created. The frequency of making full backup depends on a level of the scheme. With the 6 th level (which is by default) a full backup is created every 16 th day. If the creation of full backups is rare, a special backup window can be scheduled for each of them. Otherwise the full backup should be fit into standard backup window. The size of the largest differential backup is the size of daily incremental backups multiplied by the period length. - Custom scheme is the most flexible one so it can be configured with relevance of specified backup windows. The following table shows the numbers of backups for different backup schedules/schemes for the specified periods of time and retention periods ( f is for full, i is for incremental, d is for differential, w is for weeks): GFS (keep monthly backups indefinitely, no backups on weekends) Retention\Due date 1 month 3 months 6 months 1 year 2 years 5 years Indefinitely 2f,3d, 15i 4f, 5d, 4i 7f, 5d, 4i 14f, 5d, 4i 26f, 5d, 4i 65f, 5d, 4i GFS (keep monthly backups for 1 year, no backups on weekends) Retention\Due date 1 month 3 months 6 months 1 year 2 years 5 years

5 1 year (104w) 2f,3d, 15i 4f, 5d, 4i 7f, 5d, 4i 14f, 5d, 4i 14f, 5d, 4i 14f, 5d, 4i Daily full (make full backup every day, no backups on weekends) Retention\Due date 1 month 3 months 6 months 1 year 2 years 5 years 1 week (1w) 5f 5f 5f 5f 5f 5f 1 month (4w) 20f 20f 20f 20f 20f 20f 3 months (13w) 20f 65f 65f 65f 65f 65f 6 months (26w) 20f 65f 130f 130f 130f 130f 1 year (52w) 20f 65f 130f 260f 260f 260f 2 years (104w) 20f 65f 130f 260f 520f 520f 5 years (260w) 20f 65f 130f 260f 520f 1300f Daily incremental (make incremental backups every day, no backups on weekends) Retention\Due date 1 month 3 months 6 months 1 year 2 years 5 years 1 week (1w) 1f, 4i 1f, 4i 1f, 4i 1f, 4i 1f, 4i 1f, 4i 1 month (4w) 1f, 19i 1f, 19i 1f, 19i 1f, 19i 1f, 19i 1f, 19i 3 months (13w) 1f, 19i 1f, 64i 1f, 64i 1f, 64i 1f, 64i 1f, 64i 6 months (26w) 1f, 19i 1f, 64i 1f, 129i 1f, 129i 1f, 129i 1f, 129i 1 year (52w) 1f, 19i 1f, 64i 1f, 129i 1f, 259f 1f, 259f 1f, 259f 2 years (104w) 1f, 19i 1f, 64i 1f, 129i 1f, 259f 1f, 519i 1f, 519i 5 years (260w) 1f, 19i 1f, 64i 1f, 129i 1f, 259f 1f, 519i 1f, 1299i Weekly full, daily incremental (make full backup once a week and make incremental all other days, no backups on weekends) Retention\Due date 1 month 3 months 6 months 1 year 2 years 5 years 1 week (1w) 1f, 4i 1f, 4i 1f, 4i 1f, 4i 1f, 4i 1f, 4i 1 month (4w) 4f, 16i 4f, 16i 4f, 16i 4f, 16i 4f, 16i 4f, 16i 3 months (13w) 4f, 16i 13f, 52i 13f, 52i 13f, 52i 13f, 52i 13f, 52i 6 months (26w) 4f, 16i 13f, 52i 26f, 104i 26f, 104i 26f, 104i 26f, 104i 1 year (52w) 4f, 16i 13f, 52i 26f, 104i 52f, 208i 52f, 208i 52f, 208i 2 years (104w) 4f, 16i 13f, 52i 26f, 104i 52f, 208i 104f, 416i 104f, 416i 5 years (260w) 4f, 16i 13f, 52i 26f, 104i 52f, 208i 104f, 416i 260f, 1040i The size of full backup changes because of changed data. For example if daily change percentage is 1% and backups are performed on workdays only, full backup size after 52 weeks will be 3.6 times higher than the original data backup. For rough capacity estimations it is recommended to use the backup size at the end of the retention period. The more accurate results you can get taking the average between the initial backup and the last backup sizes. Incremental backup size depends on the frequency of backups. Daily changed backup size is one of the parameters defined initially. It is possible to configure backups to be performed only weekly, monthly or at any day so the real incremental backup size must be calculated. Differential backup size is based on changed data percentage and the amount of days from the last full backup. To estimate highest differential backup size take the amount of days between first full and last of its differential backups. For example in case of GFS the longest distance between full backup and differential backup is 15 days so the largest differential backup size will be the size of daily incremental data multiplied

6 by 15. More accurate estimation is using average between the first and the last differential backups of the same full backup. The data in all types of backups can be compressed. Normal level of compression, which is used by default, usually makes the backup data size about 1.5 times smaller. Deduplication significantly affects the amount of data occupied by backups. Here is the formula for calculating the initial storage space: Storage Space (GB) = (Data Size Total * Unique Data Percentage / Data Size Total * (100 - Unique Data Percentage) / 100 / Machine Count) / Compression Ratio To calculate the storage space after some period of time, backup scheme and retention period should be considered (above table can be used, see example at the end of this section for more details). Storage type The main backup storage requirement is having enough capacity to store all backup data. Here are the recommendations regarding the storage type: 1. Deduplicated vault data can be organized: a. on the Storage Node s local hard drives recommended for higher performance b. on a network share c. on a Storage Area Network (SAN) d. on a Network Attached Storage (NAS) 2. The storage device can be relatively slow comparing to the vault database disk. 3. Use RAID for redundancy. 4. Place vault data and vault database on drives controlled by different controllers. 5. Vault data should not be placed on the same drive with operating system. 6. There should be plenty of free space to store all the backups and perform service operations. Storage size depends on the amount and type of data you are going to back up and appropriate retention rules. To estimate the storage size, take the full amount of backups of all your machines, divided this by the deduplication factor and multiply this by the number of full backups from each machine you are going to retain. Add daily incremental data size multiplied by the number of working days the data is retained. Flat recommendation: use 7200RPM disks in a RAID. 30 workstations are backed up with GFS scheme (full backups are stored infinitely) about 1500 GB of data in total. Daily Backup Data Percentage is 1% a day. Unique data size is 20%. Need to calculate the capacity of the storage needed for backups after one year (52 weeks). With GFS scheme in a year there will be 14 full backups, 5 differential and 4 incremental backups for each machine. To calculate full backup size we need to know the initial and the final ASN backup data sizes. During calculations we should take into account deduplication and compression.

7 Initial ASN backup data size is: Initial Backup Data Size = (Data Size Total * Unique Data Percentage / Data Size Total * (100 - Unique Data Percentage) / 100 / Machine Count) / Compression Ratio = (1500 GB * 20 / GB * (100-20) / 100 / 30) / 1.5 = (300 GB + 40 GB) / 1.5 = GB. Now we have to count final backup data size. We count it as initial backup size plus a half of the daily change for each day (this is specific for GFS scheme). Final Backup Data Size = Initial Backup Data Size + (Initial Backup Data Size * Daily Backup Data Percentage / 100) * (Due Date * 5 1) / 2 = GB + ( GB * 1 / 100) * (52 * 5 1) / 2 = GB GB = GB. We will take average full backup size: Average Full Backup Size = (Initial Backup Data Size + Final Backup Data Size) / 2 = (227 GB GB) / 2 = 374 GB Average Incremental Backup Size is taken as a daily change of average full backup: Average Incremental Backup Size = Average Full Backup Size * Daily Backup Data Percentage / 100 = 374 GB * 1 / 100 = 3.74 GB For differential backup we will take the following estimation: the first differential backup in GFS scheme contains the data for 5 days so it is 5 * 1% = 5%. The last differential in the chain is 15 * 1% = 15%. Average is 10% of average backup data. In bytes that will be: Average Differential Backup Size = Average Full Backup Size * 10 / 100 = 374 GB * 10 / 100 = 37.4 GB Total data size for all types of backups for the specified period will be (we take upper storage limit estimation so Final Backup Data Size is used for full backups): Data Size Total = Final Backup Data Size * 14 + Average Differential Backup * 5 + Average Incremental Backup Size * 4 = 521 GB * GB * GB * 4 = 7294 GB GB + 15 GB = 7496 GB Additional space is occupied with deduplication database which stores information about where to find deduplicated data blocks and Catalog database which allows fast browsing and search inside archive content. Deduplication database Deduplication database stores hash values and offsets for each file 256 kb block (file-level backups) or disk 4 kb block (disk-level backups) stored in the vault. Deduplication database size can be roughly estimated based on the fact that in case of disk-level backups Deduplication database for 500 GB of unique data takes 8 GB. For the case of file-level backups the Deduplication database for 500 GB will take 64 times less space (0.13 GB) because of difference in deduplicated block size. Meanwhile the size of Deduplication database is much lower, it is recommended to store it on the reliable drives with minimal access time. Such drives are more expensive that s why the size of deduplication database should be considered. Deduplication database can be placed on a separate drive for higher performance reasons.

8 Here is the formula for deduplication database for disk-level backups: Deduplication Database size (GB) = * Data size 2.12 GB And here is a variant for file-level backups: Deduplication Database size (GB) = * Data size GB Storage type For higher Storage Node performance follow the provided recommendations: 1. The folder must reside on a fixed drive. 2. The folder size may become large estimation is 40 GB per 2.5 TB of used space, or about 2.4 percent. 3. The folder should not be placed on the same drive with operating system. 4. Minimal access time is extremely important. If you are backing up more than GB a day an enterprise-grade SSD device is highly recommended. If SSD s are not available, you can use locally attached 10000RPM or 7200RPM in a RAID10. Processing speeds will, however, be slower than enterprise-grade SSD devices but about 5-7%. Flat recommendation: use 7200RPM disks in a RAID10. There are 30 servers with 50 GB of data on each of them. Unique data size is 20%. Need to calculate the size of deduplication database. As the first step the data size is calculated. Data Size = (Data Size Total * Unique Data Percentage / Data Size Total * (100 - Unique Data Percentage) / 100 / Machine Count) / Compression Ratio = (1500 GB * 20 / GB * (100-20) / 100 / 30) / 1.5 = (300 GB + 40 GB) / 1.5 = 227 GB And now calculate deduplication database size: Deduplication Database Size = * Data Size GB = * 227 GB GB = 3.86 GB GB = 1.74 GB Catalog database Catalog database contains index with information about all the files in the vault. This database size can be roughly estimated based on the fact that catalog database with 1 million items (file information blocks) take about 250 Mb. This is correct for both disk-level and file-level backups. So here is the formula: Catalog Database Size (GB) = GB * Number of Files It is more convenient to operate with the same parameters in all the formulas. For such purposes the Data Size Total can be converted to Number of files based on average file size. Average file size varies depending on the machine type. Average file size for office workstations is about 0.5 Mb. If the machine stores big files such as music or video, the average size is higher.

9 Storage type Catalog database processing speed does not affect backup performance. This database is usually placed on the same disk where operating system resides. There are 5 servers with 100 GB of data on each of them. Need to calculate the size of catalog database. Suppose the average file size is about 0.5 Mb = GB (actually for such a small amount of machines it is possible to count the number of files on each of them). Number of Files = 100 * 5 / = files Catalog Database Size = GB * = 0.25 GB Other Storage Node databases Storage Node maintains several additional databases where it stores information about logs, tasks and other things. The size of these databases is small and does not depend on backup parameters and can be skipped from calculations. CPU CPU speed is generally not a bottleneck so it is recommended that the Storage Node has a CPU with 2 cores x 3 GHz or 4 cores x 2.5 GHz. This is true regardless of the number of client machines that use the Storage Node. RAM RAM becomes a vital configuration parameter of a Storage Node in Acronis Backups & Recovery 11. Because of widely using caching algorithms and because of 64bit architecture Storage Node became much more scalable. Addition of more RAM for the Storage Node server in most cases helps to make the amount of data effectively handled by a Storage Node higher. The minimal recommended amount of RAM is based on the amount of unique data to be processed by ASN. In any case there should be not less than 8 GB of RAM. Having 8 GB of RAM (which is minimal recommended RAM size) allows effective processing of 800 GB of unique data which confirms to 12 GB of Deduplication database size. Having 32 GB of RAM allows to process 3700 GB of unique data (61 GB of Deduplication DB). The formula is the following: Minimal RAM Size (GB) = (4000 GB + 24 * Data Size) / 2900 Note that incremental data are mostly unique so the amount of unique data on the machines will grow. There are 10 servers with 2000 GB of total compressed deduplicated backup data. Need to calculate the minimal RAM size. The calculations are simple:

10 Minimal RAM Size = (4000 GB + 24 * Data Size) / 2900 = (4000 GB + 24 * 2000 GB) / 2900 = 18 GB Network Data Transfer Rate Client-side deduplication is always turned on Storage node 11 so it does not send duplicate data to deduplicated vaults. That minimizes network traffic during backing up making it lower up to times, depending on the amount of unique data and availability of duplicated data on the backup storage. First backup to empty backup storage has high traffic as all the backup data is transferred to the server. When indexing of these data completes, the deduplication ratio of consequent backup data raises. That s why it is recommended to perform initial phase of backing up first one or several machines and wait until indexing completes. The formula for disk-level backups is the following: Average Network Traffic (Mbit/sec) = Unique Data Percentage / 100 * 28 Mbit/sec + 2 Mbit/sec The formula for file-level backups is the following: Average Network Traffic (Mbit/sec) = Unique Data Percentage / 100 * 17.5 Mbit/sec Mbit/sec For the case of simultaneous backing up of several clients the traffic is multiplied by the number of these clients. There are 10 clients with unique data percentage 10%. Need to define average network traffic for the case of simultaneous disk-level backing up of these 10 clients. First, calculate the traffic for the initial phase as there are no data on the storage yet, all the backup data will be transferred so the traffic will be maximal. Average Network Traffic = 100 /100 * 28 Mbit/sec + 2 Mbit/sec = = 30 Mbit/sec Second, define the network traffic for one machine: Average Network Traffic = 10 / 100 * 28 Mbit/sec + 2 Mbit/sec = 2.8 Mbit/sec + 2 Mbit/sec = 4.8 Mbit/sec Then, multiply this value by the number of simultaneous backups: Average Total Network Traffic = 4.8 Mbit/sec * 10 = 48 Mbit/sec Number of Storage Nodes The number of Storage Nodes to be used can be based on the general backup processing speeds on the server with recommended configuration described above. Here is a list: Disk-level backup speed: 135 GB/hour Disk-level indexing speed: 85 GB/hour File-level backup speed: 45 GB/hour

11 File-level indexing speed: 204 GB/hour To estimate how much Storage Nodes is needed, calculate the time needed to process all the backups and then compare it with backup window. Backup window is specified for the client software so only backup time is compared with it. Indexing speed should be taken into account to calculate if Storage Node can reindex all the backups (additionally to the time of backups) for the time provided for its work (for dedicated Storage Node servers 24 hours a day). So in short Storage Node should accept all the backups in a specified backup window and accept and reindex all the backups in a time provided for its work. If the time is not enough, additional Storage Node should be added or more time provided. One more possible solution is to use Custom backup scheme and configure longer-time backups to be performed on special long backup windows, for example on holidays. Time to process backups The amount of time depends on the amount of data to be processed and the selected backup scheme/schedule. Full backups performed much longer than incremental and differential ones. If all full backups do not fit into backup window, protected machines can be grouped to perform full backups of each group in at a separate day. For example if full backups for all 10 machines cannot be performed in one backup window, split machines to several groups and configure appropriate amount of backup plans for full backups to be performed in different days. As backups can be accepted by Storage Node in parallel, the time for backups can be shorter. So the time needed to accept all the backups is usually divided by 2 because of parallelism. Here is a common formula: Backup Time (hours) = Data Size / Backup Speed / 2 Based on the data above take the backup sizes, calculate the amount of time for it and divide the appropriate backup window time by this calculated value. Do it for all backup types. The biggest calculated value will show how much Storage Nodes are needed. Number of Storage Nodes = max (Backup Time / Backup Window) (all backup types and appropriate backup windows are taken) 5 servers (500 GB of data in total) are being backed up with GFS scheme. Disk-level backups are performed. Daily incremental data size is 2%. Backup window is 8 hours a day for any type of backups. Need to calculate how much Storage Nodes needed to accept all the backups. Let s calculate how much time needed to process full backups. Full Backup Time = 500 GB / 135 GB/hour / 2 = 1.85 hours Now count how much Storage Nodes are needed. Number of Storage Nodes = 1.85 hours/ 8 hours = 0.23, less than 1 so one Storage Node is enough to accept all the backups.

12 We do not calculate here the number for Incremental backups as they are performed at the similar backup window but have much smaller size. Time to reindex backups The backups are reindexed one by one so the average speed of backups indexing is lower than the backup speed. From the other side the time for reindexing is usually 24 hours a day (for dedicated Storage Node servers). For calculating how much Storage Nodes needed to reindex backups from all the machines, the amount of backup data should be estimated based on the backup scheme/schedule. Reindex time is based on the amount of backup data and reindex speed: Reindex Time (hours) = Data Size / Reindex Speed Based on the data above take the backup sizes, calculate the amount of time for it and divide the appropriate Storage Node work hours by this calculated value: Number of Storage Nodes = Reindex Time / Work Hours 10 servers and 100 workstations contain 6000 GB of data to back up. Disk-level backups are performed. Storage Nodes are dedicated servers (work 24 hours a day). Need to calculate how much Storage Nodes needed to reindex all the backups. Reindex time is calculated: Reindex Time = 6000 GB / 85 GB/hour = 71 hours Storage Node works in full time so the amount of time for reindex is 24 hours and the number of Storage Nodes will be: Number of Storage Nodes = 71 hours / 24 hours = 2.96 which shows that 3 Storage Nodes will be needed to reindex all full backups in one backup window. Otherwise the machines can be divided to three groups with a separate backup plan for each of them. In this case one Storage Node will be able to reindex all backup data.

Acronis Backup Deduplication. Technical Whitepaper

Acronis Backup Deduplication. Technical Whitepaper Acronis Backup Deduplication Technical Whitepaper Table of Contents Table of Contents Table of Contents... 1 Introduction... 3 Storage Challenges... 4 How Deduplication Helps... 5 How It Works... 6 Deduplication

More information

Acronis Backup & Recovery for Mac. Acronis Backup & Recovery & Acronis ExtremeZ-IP REFERENCE ARCHITECTURE

Acronis Backup & Recovery for Mac. Acronis Backup & Recovery & Acronis ExtremeZ-IP REFERENCE ARCHITECTURE Acronis Backup & Recovery for Mac Acronis Backup & Recovery & Acronis ExtremeZ-IP This document describes the technical requirements and best practices for implementation of a disaster recovery solution

More information

Acronis Backup & Recovery 10 Advanced Server Virtual Edition. Quick Start Guide

Acronis Backup & Recovery 10 Advanced Server Virtual Edition. Quick Start Guide Acronis Backup & Recovery 10 Advanced Server Virtual Edition Quick Start Guide Table of contents 1 Main components...3 2 License server...3 3 Supported operating systems...3 3.1 Agents... 3 3.2 License

More information

NETGEAR ReadyNAS and Acronis Backup & Recovery 10 Configuring ReadyNAS as an Acronis Backup & Recovery 10 Vault

NETGEAR ReadyNAS and Acronis Backup & Recovery 10 Configuring ReadyNAS as an Acronis Backup & Recovery 10 Vault NETGEAR ReadyNAS and Acronis Backup & Recovery 10 Configuring ReadyNAS as an Acronis Backup & Recovery 10 Vault Table of Contents Contents...2 Concepts...3 Data Deduplication...3 Acronis Vaults...4 Components...4

More information

WHITE PAPER. How Deduplication Benefits Companies of All Sizes An Acronis White Paper

WHITE PAPER. How Deduplication Benefits Companies of All Sizes An Acronis White Paper How Deduplication Benefits Companies of All Sizes An Acronis White Paper Copyright Acronis, Inc., 2000 2009 Table of contents Executive Summary... 3 What is deduplication?... 4 File-level deduplication

More information

Backup and Recovery FAQs

Backup and Recovery FAQs May 2013 Page 1 This document answers frequently asked questions regarding the Emerson system Backup and Recovery application. www.deltav.com May 2013 Page 2 Table of Contents Introduction... 6 General

More information

Backup and Recovery. Introduction. Benefits. Best-in-class offering. Easy-to-use Backup and Recovery solution.

Backup and Recovery. Introduction. Benefits. Best-in-class offering. Easy-to-use Backup and Recovery solution. DeltaV Distributed Control System Product Data Sheet Backup and Recovery Best-in-class offering. Easy-to-use Backup and Recovery solution. Data protection and disaster recovery in a single solution. Scalable

More information

LDA, the new family of Lortu Data Appliances

LDA, the new family of Lortu Data Appliances LDA, the new family of Lortu Data Appliances Based on Lortu Byte-Level Deduplication Technology February, 2011 Copyright Lortu Software, S.L. 2011 1 Index Executive Summary 3 Lortu deduplication technology

More information

Backup and Recovery. Backup and Recovery. Introduction. DeltaV Product Data Sheet. Best-in-class offering. Easy-to-use Backup and Recovery solution

Backup and Recovery. Backup and Recovery. Introduction. DeltaV Product Data Sheet. Best-in-class offering. Easy-to-use Backup and Recovery solution April 2013 Page 1 Protect your plant data with the solution. Best-in-class offering Easy-to-use solution Data protection and disaster recovery in a single solution Scalable architecture and functionality

More information

Get Success in Passing Your Certification Exam at first attempt!

Get Success in Passing Your Certification Exam at first attempt! Get Success in Passing Your Certification Exam at first attempt! Exam : E22-290 Title : EMC Data Domain Deduplication, Backup and Recovery Exam Version : DEMO 1.A customer has a Data Domain system with

More information

Data Deduplication in Tivoli Storage Manager. Andrzej Bugowski 19-05-2011 Spała

Data Deduplication in Tivoli Storage Manager. Andrzej Bugowski 19-05-2011 Spała Data Deduplication in Tivoli Storage Manager Andrzej Bugowski 19-05-2011 Spała Agenda Tivoli Storage, IBM Software Group Deduplication concepts Data deduplication in TSM 6.1 Planning for data deduplication

More information

Advanced Knowledge and Understanding of Industrial Data Storage

Advanced Knowledge and Understanding of Industrial Data Storage Dec. 3 rd 2013 Advanced Knowledge and Understanding of Industrial Data Storage By Jesse Chuang, Senior Software Manager, Advantech With the popularity of computers and networks, most enterprises and organizations

More information

Building an efficient and inexpensive PACS system. OsiriX - dcm4chee - JPEG2000

Building an efficient and inexpensive PACS system. OsiriX - dcm4chee - JPEG2000 Building an efficient and inexpensive PACS system OsiriX - dcm4chee - JPEG2000 The latest version of OsiriX greatly improves compressed DICOM support, specifically JPEG2000 1 support. In this paper, we

More information

Acronis Backup & Recovery 11

Acronis Backup & Recovery 11 Acronis Backup & Recovery 11 Quick Start Guide Applies to the following editions: Advanced Server Virtual Edition Advanced Server SBS Edition Advanced Workstation Server for Linux Server for Windows Workstation

More information

WHITE PAPER. Storage Savings Analysis: Storage Savings with Deduplication and Acronis Backup & Recovery 10

WHITE PAPER. Storage Savings Analysis: Storage Savings with Deduplication and Acronis Backup & Recovery 10 Storage Savings Analysis: Storage Savings with Deduplication and Acronis Backup & Recovery 10 Copyright Acronis, Inc., 2000 2009 Table of contents Executive Summary... 3 The Importance of Deduplication...

More information

EMC BACKUP-AS-A-SERVICE

EMC BACKUP-AS-A-SERVICE Reference Architecture EMC BACKUP-AS-A-SERVICE EMC AVAMAR, EMC DATA PROTECTION ADVISOR, AND EMC HOMEBASE Deliver backup services for cloud and traditional hosted environments Reduce storage space and increase

More information

CONFIGURATION GUIDELINES: EMC STORAGE FOR PHYSICAL SECURITY

CONFIGURATION GUIDELINES: EMC STORAGE FOR PHYSICAL SECURITY White Paper CONFIGURATION GUIDELINES: EMC STORAGE FOR PHYSICAL SECURITY DVTel Latitude NVMS performance using EMC Isilon storage arrays Correct sizing for storage in a DVTel Latitude physical security

More information

GET. tech brief FASTER BACKUPS

GET. tech brief FASTER BACKUPS GET tech brief FASTER BACKUPS Faster Backups Local. Offsite. Remote Office. Why Should You Care? According to a recent survey from the IDG Research Group, the biggest challenge facing IT managers responsible

More information

WHITE PAPER. Acronis Backup & Recovery 10 Overview

WHITE PAPER. Acronis Backup & Recovery 10 Overview Acronis Backup & Recovery 10 Overview Copyright Acronis, Inc., 2000 2009 Table of contents Executive Summary... 3 The Next Generation Backup and Recovery Solution from Acronis... 4 New! Scaleable Architecture

More information

EaseTag Cloud Storage Solution

EaseTag Cloud Storage Solution EaseTag Cloud Storage Solution The Challenge For most companies, data especially unstructured data, continues to grow by 50 percent annually. The impact of spending more every year on storage, and on protecting

More information

Dell Migration Manager for Archives 7.3. SQL Best Practices

Dell Migration Manager for  Archives 7.3. SQL Best Practices Dell Migration Manager for Email Archives 7.3 SQL Best Practices 2016 Dell Inc. All rights reserved. This product is protected by U.S. and international copyright and intellectual property laws. Dell and

More information

Distributed File System. MCSN N. Tonellotto Complements of Distributed Enabling Platforms

Distributed File System. MCSN N. Tonellotto Complements of Distributed Enabling Platforms Distributed File System 1 How do we get data to the workers? NAS Compute Nodes SAN 2 Distributed File System Don t move data to workers move workers to the data! Store data on the local disks of nodes

More information

Acronis Backup & Recovery 10 Server for Windows. Workstation. Quick Start Guide

Acronis Backup & Recovery 10 Server for Windows. Workstation. Quick Start Guide Acronis Backup & Recovery 10 Server for Windows Acronis Backup & Recovery 10 Workstation Quick Start Guide 1. About this document This document describes how to install and start using any of the following

More information

Terminal Server Software and Hardware Requirements. Terminal Server. Software and Hardware Requirements. Datacolor Match Pigment Datacolor Tools

Terminal Server Software and Hardware Requirements. Terminal Server. Software and Hardware Requirements. Datacolor Match Pigment Datacolor Tools Terminal Server Software and Hardware Requirements Datacolor Match Pigment Datacolor Tools January 21, 2011 Page 1 of 8 Introduction This document will provide preliminary information about the both the

More information

EMC DATA DOMAIN SISL SCALING ARCHITECTURE

EMC DATA DOMAIN SISL SCALING ARCHITECTURE EMC DATA DOMAIN SISL SCALING ARCHITECTURE A Detailed Review ABSTRACT While tape has been the dominant storage medium for data protection for decades because of its low cost, it is steadily losing ground

More information

Backup and Recovery. Backup and Recovery. Introduction. DeltaV Product Data Sheet. Best-in-class offering. Easy-to-use Backup and Recovery solution

Backup and Recovery. Backup and Recovery. Introduction. DeltaV Product Data Sheet. Best-in-class offering. Easy-to-use Backup and Recovery solution September 2013 Page 1 Protect your plant data with the solution. Best-in-class offering Easy-to-use solution Data protection and disaster recovery in a single solution Scalable architecture and functionality

More information

Secret Server Architecture and Sizing Guide

Secret Server Architecture and Sizing Guide This document contains information for planning Secret Server architecture and resource allocation within your environment. Read through or use one of the following links to skip ahead to the relevant

More information

Backup and Recovery 1

Backup and Recovery 1 Backup and Recovery What is a Backup? Backup is an additional copy of data that can be used for restore and recovery purposes. The Backup copy is used when the primary copy is lost or corrupted. This Backup

More information

Optimize VMware and Hyper-V Protection with HP and Veeam

Optimize VMware and Hyper-V Protection with HP and Veeam Optimize VMware and Hyper-V Protection with HP and Veeam John DeFrees, Global Alliance Solution Architect, Veeam Markus Berber, HP LeftHand P4000 Product Marketing Manager, HP Key takeaways from today

More information

Reference Guide WindSpring Data Management Technology (DMT) Solving Today s Storage Optimization Challenges

Reference Guide WindSpring Data Management Technology (DMT) Solving Today s Storage Optimization Challenges Reference Guide WindSpring Data Management Technology (DMT) Solving Today s Storage Optimization Challenges September 2011 Table of Contents The Enterprise and Mobile Storage Landscapes... 3 Increased

More information

PARALLELS CLOUD STORAGE

PARALLELS CLOUD STORAGE PARALLELS CLOUD STORAGE Performance Benchmark Results 1 Table of Contents Executive Summary... Error! Bookmark not defined. Architecture Overview... 3 Key Features... 5 No Special Hardware Requirements...

More information

Solution Brief: Creating Avid Project Archives

Solution Brief: Creating Avid Project Archives Solution Brief: Creating Avid Project Archives Marquis Project Parking running on a XenData Archive Server provides Fast and Reliable Archiving to LTO or Sony Optical Disc Archive Cartridges Summary Avid

More information

Redefining Backup for VMware Environment. Copyright 2009 EMC Corporation. All rights reserved.

Redefining Backup for VMware Environment. Copyright 2009 EMC Corporation. All rights reserved. Redefining Backup for VMware Environment 1 Agenda VMware infrastructure backup and recovery challenges Introduction to EMC Avamar Avamar solutions for VMware infrastructure Key takeaways Copyright 2009

More information

Paragon Protect & Restore

Paragon Protect & Restore Paragon Protect & Restore ver. 3 Centralized and Disaster Recovery for virtual and physical environments Tight Integration with hypervisors for agentless backups, VM replication and seamless restores Paragon

More information

Double-Take Pagefile Configuration

Double-Take Pagefile Configuration Double-Take Pagefile Configuration Double-Take Pagefile Configuration published August 2002 NSI and Double-Take are registered trademarks of Network Specialists, Inc. All other products are trademarks

More information

Tandberg Data AccuVault RDX

Tandberg Data AccuVault RDX Tandberg Data AccuVault RDX Binary Testing conducts an independent evaluation and performance test of Tandberg Data s latest small business backup appliance. Data backup is essential to their survival

More information

Windows Server Performance Monitoring

Windows Server Performance Monitoring Spot server problems before they are noticed The system s really slow today! How often have you heard that? Finding the solution isn t so easy. The obvious questions to ask are why is it running slowly

More information

Online Backup Plus Frequently Asked Questions

Online Backup Plus Frequently Asked Questions Online Backup Plus Frequently Asked Questions 1 INSTALLATION 1.1 Who installs the Redstor Online Backup Plus service? 1.2 How does the installed client connect to Redstor s Cloud Platform? 1.3 On which

More information

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...

More information

WHITE PAPER BRENT WELCH NOVEMBER

WHITE PAPER BRENT WELCH NOVEMBER BACKUP WHITE PAPER BRENT WELCH NOVEMBER 2006 WHITE PAPER: BACKUP TABLE OF CONTENTS Backup Overview 3 Background on Backup Applications 3 Backup Illustration 4 Media Agents & Keeping Tape Drives Busy 5

More information

OPTIMIZING EXCHANGE SERVER IN A TIERED STORAGE ENVIRONMENT WHITE PAPER NOVEMBER 2006

OPTIMIZING EXCHANGE SERVER IN A TIERED STORAGE ENVIRONMENT WHITE PAPER NOVEMBER 2006 OPTIMIZING EXCHANGE SERVER IN A TIERED STORAGE ENVIRONMENT WHITE PAPER NOVEMBER 2006 EXECUTIVE SUMMARY Microsoft Exchange Server is a disk-intensive application that requires high speed storage to deliver

More information

Identifying the Hidden Risk of Data De-duplication: How the HYDRAstor Solution Proactively Solves the Problem

Identifying the Hidden Risk of Data De-duplication: How the HYDRAstor Solution Proactively Solves the Problem Identifying the Hidden Risk of Data De-duplication: How the HYDRAstor Solution Proactively Solves the Problem October, 2006 Introduction Data de-duplication has recently gained significant industry attention,

More information

A SCALABLE DEDUPLICATION AND GARBAGE COLLECTION ENGINE FOR INCREMENTAL BACKUP

A SCALABLE DEDUPLICATION AND GARBAGE COLLECTION ENGINE FOR INCREMENTAL BACKUP A SCALABLE DEDUPLICATION AND GARBAGE COLLECTION ENGINE FOR INCREMENTAL BACKUP Dilip N Simha (Stony Brook University, NY & ITRI, Taiwan) Maohua Lu (IBM Almaden Research Labs, CA) Tzi-cker Chiueh (Stony

More information

DEDUPLICATION BASICS

DEDUPLICATION BASICS DEDUPLICATION BASICS 4 DEDUPE BASICS 12 HOW DO DISASTER RECOVERY & ARCHIVING FIT IN? 6 WHAT IS DEDUPLICATION 14 DEDUPLICATION FOR EVERY BUDGET QUANTUM DXi4000 and vmpro 4000 8 METHODS OF DEDUPLICATION

More information

A Deduplication File System & Course Review

A Deduplication File System & Course Review A Deduplication File System & Course Review Kai Li 12/13/12 Topics A Deduplication File System Review 12/13/12 2 Traditional Data Center Storage Hierarchy Clients Network Server SAN Storage Remote mirror

More information

Disk-to-Disk-to-Offsite Backups for SMBs with Retrospect

Disk-to-Disk-to-Offsite Backups for SMBs with Retrospect Disk-to-Disk-to-Offsite Backups for SMBs with Retrospect Abstract Retrospect backup and recovery software provides a quick, reliable, easy-to-manage disk-to-disk-to-offsite backup solution for SMBs. Use

More information

Effective Planning and Use of TSM V6 Deduplication

Effective Planning and Use of TSM V6 Deduplication Effective Planning and Use of IBM Tivoli Storage Manager V6 Deduplication 08/17/12 1.0 Authors: Jason Basler Dan Wolfe Page 1 of 42 Document Location This is a snapshot of an on-line document. Paper copies

More information

Acronis True Image 2015 REVIEWERS GUIDE

Acronis True Image 2015 REVIEWERS GUIDE Acronis True Image 2015 REVIEWERS GUIDE Table of Contents INTRODUCTION... 3 What is Acronis True Image 2015?... 3 System Requirements... 4 INSTALLATION... 5 Downloading and Installing Acronis True Image

More information

Using Synology SSD Technology to Enhance System Performance Synology Inc.

Using Synology SSD Technology to Enhance System Performance Synology Inc. Using Synology SSD Technology to Enhance System Performance Synology Inc. Synology_SSD_Cache_WP_ 20140512 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges...

More information

Acronis Backup & Recovery 11. Backing Up Microsoft Exchange Server Data

Acronis Backup & Recovery 11. Backing Up Microsoft Exchange Server Data Acronis Backup & Recovery 11 Backing Up Microsoft Exchange Server Data Copyright Acronis, Inc., 2000-2012. All rights reserved. Acronis and Acronis Secure Zone are registered trademarks of Acronis, Inc.

More information

Online Backup Frequently Asked Questions

Online Backup Frequently Asked Questions Online Backup Frequently Asked Questions 1 INSTALLATION 1.1 Who installs the Redstor Online Backup service? 1.2 How does the installed client connect to Redstor s Cloud Platform? 1.3 On which machines

More information

Acronis Backup & Recovery 10 Management Server reports. Technical white paper

Acronis Backup & Recovery 10 Management Server reports. Technical white paper Acronis Backup & Recovery 10 Management Server reports Technical white paper Table of Contents 1 Report data... 3 2 Time format... 4 3 Relationship between views... 4 4 Relationship diagram... 6 5 Current

More information

NEXTGEN v5.8 HARDWARE VERIFICATION GUIDE CLIENT HOSTED OR THIRD PARTY SERVERS

NEXTGEN v5.8 HARDWARE VERIFICATION GUIDE CLIENT HOSTED OR THIRD PARTY SERVERS This portion of the survey is for clients who are NOT on TSI Healthcare s ASP and are hosting NG software on their own server. This information must be collected by an IT staff member at your practice.

More information

The Microsoft Large Mailbox Vision

The Microsoft Large Mailbox Vision WHITE PAPER The Microsoft Large Mailbox Vision Giving users large mailboxes without breaking your budget Introduction Giving your users the ability to store more e mail has many advantages. Large mailboxes

More information

Symantec Endpoint Protection 11.0 Architecture, Sizing, and Performance Recommendations

Symantec Endpoint Protection 11.0 Architecture, Sizing, and Performance Recommendations Symantec Endpoint Protection 11.0 Architecture, Sizing, and Performance Recommendations Technical Product Management Team Endpoint Security Copyright 2007 All Rights Reserved Revision 6 Introduction This

More information

Cost Effective Backup with Deduplication. Copyright 2009 EMC Corporation. All rights reserved.

Cost Effective Backup with Deduplication. Copyright 2009 EMC Corporation. All rights reserved. Cost Effective Backup with Deduplication Agenda Today s Backup Challenges Benefits of Deduplication Source and Target Deduplication Introduction to EMC Backup Solutions Avamar, Disk Library, and NetWorker

More information

Data De-duplication Methodologies: Comparing ExaGrid s Byte-level Data De-duplication To Block Level Data De-duplication

Data De-duplication Methodologies: Comparing ExaGrid s Byte-level Data De-duplication To Block Level Data De-duplication Data De-duplication Methodologies: Comparing ExaGrid s Byte-level Data De-duplication To Block Level Data De-duplication Table of Contents Introduction... 3 Shortest Possible Backup Window... 3 Instant

More information

Enterprise Backup and Restore technology and solutions

Enterprise Backup and Restore technology and solutions Enterprise Backup and Restore technology and solutions LESSON VII Veselin Petrunov Backup and Restore team / Deep Technical Support HP Bulgaria Global Delivery Hub Global Operations Center November, 2013

More information

SECURE Web Gateway Sizing Guide

SECURE Web Gateway Sizing Guide Technical Guide Version 02 26/02/2015 Contents Introduction... 3 Overview... 3 Example one... 4 Example two... 4 Maximum throughput... 4 Gateway Reporter... 4 Gateway Reporter server specification... 5

More information

Data Backup and Archiving with Enterprise Storage Systems

Data Backup and Archiving with Enterprise Storage Systems Data Backup and Archiving with Enterprise Storage Systems Slavjan Ivanov 1, Igor Mishkovski 1 1 Faculty of Computer Science and Engineering Ss. Cyril and Methodius University Skopje, Macedonia slavjan_ivanov@yahoo.com,

More information

Drobo How-To Guide. Use a Drobo iscsi Array as a Target for Veeam Backups

Drobo How-To Guide. Use a Drobo iscsi Array as a Target for Veeam Backups This document shows you how to use a Drobo iscsi array with Veeam Backup & Replication version 6.5 in a VMware environment. Veeam provides fast disk-based backup and recovery of virtual machines (VMs),

More information

TARRANT COUNTY PURCHASING DEPARTMENT

TARRANT COUNTY PURCHASING DEPARTMENT JACK BEACHAM, C.P.M., A.P.P. PURCHASING AGENT TARRANT COUNTY PURCHASING DEPARTMENT AUGUST 4, 2010 RFP NO. 2010-103 ROB COX, C.P.M., A.P.P. ASSISTANT PURCHASING AGENT RFP FOR DIGITAL ASSET MANAGEMENT SYSTEM

More information

Deduplication Demystified: How to determine the right approach for your business

Deduplication Demystified: How to determine the right approach for your business Deduplication Demystified: How to determine the right approach for your business Presented by Charles Keiper Senior Product Manager, Data Protection Quest Software Session Objective: To answer burning

More information

WHITE PAPER Improving Storage Efficiencies with Data Deduplication and Compression

WHITE PAPER Improving Storage Efficiencies with Data Deduplication and Compression WHITE PAPER Improving Storage Efficiencies with Data Deduplication and Compression Sponsored by: Oracle Steven Scully May 2010 Benjamin Woo IDC OPINION Global Headquarters: 5 Speen Street Framingham, MA

More information

Hardware/Software Guidelines

Hardware/Software Guidelines There are many things to consider when preparing for a TRAVERSE v11 installation. The number of users, application modules and transactional volume are only a few. Reliable performance of the system is

More information

Data Protection with IBM TotalStorage NAS and NSI Double- Take Data Replication Software

Data Protection with IBM TotalStorage NAS and NSI Double- Take Data Replication Software Data Protection with IBM TotalStorage NAS and NSI Double- Take Data Replication September 2002 IBM Storage Products Division Raleigh, NC http://www.storage.ibm.com Table of contents Introduction... 3 Key

More information

Acronis Backup & Recovery 11.5. Update 2. Backing Up Microsoft Exchange Server Data

Acronis Backup & Recovery 11.5. Update 2. Backing Up Microsoft Exchange Server Data Acronis Backup & Recovery 11.5 Update 2 Backing Up Microsoft Exchange Server Data Copyright Statement Copyright Acronis International GmbH, 2002-2013. All rights reserved. Acronis and Acronis Secure Zone

More information

Paragon Backup Retention Wizard

Paragon Backup Retention Wizard Paragon Backup Retention Wizard User Guide Getting Started with the Paragon Backup Retention Wizard In this guide you will find all the information necessary to get the product ready to use. System Requirements

More information

Acronis Backup & Recovery 11.5 Quick Start Guide

Acronis Backup & Recovery 11.5 Quick Start Guide Acronis Backup & Recovery 11.5 Quick Start Guide Applies to the following editions: Advanced Server for Windows Virtual Edition Advanced Server SBS Edition Advanced Workstation Server for Linux Server

More information

Solid State Storage in Massive Data Environments Erik Eyberg

Solid State Storage in Massive Data Environments Erik Eyberg Solid State Storage in Massive Data Environments Erik Eyberg Senior Analyst Texas Memory Systems, Inc. Agenda Taxonomy Performance Considerations Reliability Considerations Q&A Solid State Storage Taxonomy

More information

Turnkey Deduplication Solution for the Enterprise

Turnkey Deduplication Solution for the Enterprise Symantec NetBackup 5000 Appliance Turnkey Deduplication Solution for the Enterprise Mayur Dewaikar Sr. Product Manager, Information Management Group White Paper: A Deduplication Appliance Solution for

More information

Using Data Domain Storage with Symantec Enterprise Vault 8. White Paper. Michael McLaughlin Data Domain Technical Marketing

Using Data Domain Storage with Symantec Enterprise Vault 8. White Paper. Michael McLaughlin Data Domain Technical Marketing Using Data Domain Storage with Symantec Enterprise Vault 8 White Paper Michael McLaughlin Data Domain Technical Marketing Charles Arconi Cornerstone Technologies - Principal Consultant Data Domain, Inc.

More information

Identifying the Hidden Risk of Data Deduplication: How the HYDRAstor TM Solution Proactively Solves the Problem

Identifying the Hidden Risk of Data Deduplication: How the HYDRAstor TM Solution Proactively Solves the Problem Identifying the Hidden Risk of Data Deduplication: How the HYDRAstor TM Solution Proactively Solves the Problem Advanced Storage Products Group Table of Contents 1 - Introduction 2 Data Deduplication 3

More information

Disaster Recovery Strategies: Business Continuity through Remote Backup Replication

Disaster Recovery Strategies: Business Continuity through Remote Backup Replication W H I T E P A P E R S O L U T I O N : D I S A S T E R R E C O V E R Y T E C H N O L O G Y : R E M O T E R E P L I C A T I O N Disaster Recovery Strategies: Business Continuity through Remote Backup Replication

More information

EMC Backup and Recovery for Microsoft Exchange 2007 SP2

EMC Backup and Recovery for Microsoft Exchange 2007 SP2 EMC Backup and Recovery for Microsoft Exchange 2007 SP2 Enabled by EMC Celerra and Microsoft Windows 2008 Copyright 2010 EMC Corporation. All rights reserved. Published February, 2010 EMC believes the

More information

Effective Planning and Use of IBM Tivoli Storage Manager V6 and V7 Deduplication

Effective Planning and Use of IBM Tivoli Storage Manager V6 and V7 Deduplication Effective Planning and Use of IBM Tivoli Storage Manager V6 and V7 Deduplication 02/17/2015 2.1 Authors: Jason Basler Dan Wolfe Page 1 of 52 Document Location This is a snapshot of an on-line document.

More information

(Formerly Double-Take Backup)

(Formerly Double-Take Backup) (Formerly Double-Take Backup) An up-to-the-minute copy of branch office data and applications can keep a bad day from getting worse. Double-Take RecoverNow for Windows (formerly known as Double-Take Backup)

More information

Acronis Backup & Recovery Online Backup Evaluation Guide

Acronis Backup & Recovery Online Backup Evaluation Guide Acronis Backup & Recovery Online Backup Evaluation Guide Introduction Acronis Backup & Recovery Online is a service that enables you to back up data to Acronis Online Storage. The service is available

More information

SAP HANA - Main Memory Technology: A Challenge for Development of Business Applications. Jürgen Primsch, SAP AG July 2011

SAP HANA - Main Memory Technology: A Challenge for Development of Business Applications. Jürgen Primsch, SAP AG July 2011 SAP HANA - Main Memory Technology: A Challenge for Development of Business Applications Jürgen Primsch, SAP AG July 2011 Why In-Memory? Information at the Speed of Thought Imagine access to business data,

More information

Corporate PC Backup - Best Practices

Corporate PC Backup - Best Practices A Druva Whitepaper Corporate PC Backup - Best Practices This whitepaper explains best practices for successfully implementing laptop backup for corporate workforce. White Paper WP /100 /009 Oct 10 Table

More information

Backup and Archiving Explained. White Paper

Backup and Archiving Explained. White Paper Backup and Archiving Explained White Paper Backup vs. Archiving The terms backup and archiving are often referenced together and sometimes incorrectly used interchangeably. While both technologies are

More information

Why Computers Are Getting Slower (and what we can do about it) Rik van Riel Sr. Software Engineer, Red Hat

Why Computers Are Getting Slower (and what we can do about it) Rik van Riel Sr. Software Engineer, Red Hat Why Computers Are Getting Slower (and what we can do about it) Rik van Riel Sr. Software Engineer, Red Hat Why Computers Are Getting Slower The traditional approach better performance Why computers are

More information

Every organization has critical data that it can t live without. When a disaster strikes, how long can your business survive without access to its

Every organization has critical data that it can t live without. When a disaster strikes, how long can your business survive without access to its DISASTER RECOVERY STRATEGIES: BUSINESS CONTINUITY THROUGH REMOTE BACKUP REPLICATION Every organization has critical data that it can t live without. When a disaster strikes, how long can your business

More information

Trends in Enterprise Backup Deduplication

Trends in Enterprise Backup Deduplication Trends in Enterprise Backup Deduplication Shankar Balasubramanian Architect, EMC 1 Outline Protection Storage Deduplication Basics CPU-centric Deduplication: SISL (Stream-Informed Segment Layout) Data

More information

Eight Considerations for Evaluating Disk-Based Backup Solutions

Eight Considerations for Evaluating Disk-Based Backup Solutions Eight Considerations for Evaluating Disk-Based Backup Solutions 1 Introduction The movement from tape-based to disk-based backup is well underway. Disk eliminates all the problems of tape backup. Backing

More information

White Paper for Data Protection with Synology Snapshot Technology. Based on Btrfs File System

White Paper for Data Protection with Synology Snapshot Technology. Based on Btrfs File System White Paper for Data Protection with Synology Snapshot Technology Based on Btrfs File System 1 Table of Contents Introduction 3 Data Protection Technologies 4 Btrfs File System Snapshot Technology How

More information

Acronis Backup & Recovery 11.5

Acronis Backup & Recovery 11.5 Acronis Backup & Recovery 11.5 Update 2 Backing Up Microsoft Exchange Server Data Copyright Statement Copyright Acronis International GmbH, 2002-2013. All rights reserved. Acronis and Acronis Secure Zone

More information

Deduplication on SNC NAS: UI Configurations and Impact on Capacity Utilization

Deduplication on SNC NAS: UI Configurations and Impact on Capacity Utilization Deduplication on SNC NAS: UI Configurations and Impact on Capacity Utilization Application Note Abstract This application note describes how to configure the deduplication function on SNC NAS systems,

More information

Quantifying Hardware Selection in an EnCase v7 Environment

Quantifying Hardware Selection in an EnCase v7 Environment Quantifying Hardware Selection in an EnCase v7 Environment Introduction and Background The purpose of this analysis is to evaluate the relative effectiveness of individual hardware component selection

More information

Acronis Backup & Recovery Online Stand-alone. User Guide

Acronis Backup & Recovery Online Stand-alone. User Guide Acronis Backup & Recovery Online Stand-alone User Guide Table of contents 1 Introduction to Acronis Backup & Recovery Online...4 1.1 What is Acronis Backup & Recovery Online?... 4 1.2 What data can I back

More information

Deep Dive on SimpliVity s OmniStack A Technical Whitepaper

Deep Dive on SimpliVity s OmniStack A Technical Whitepaper Deep Dive on SimpliVity s OmniStack A Technical Whitepaper By Hans De Leenheer and Stephen Foskett August 2013 1 Introduction This paper is an in-depth look at OmniStack, the technology that powers SimpliVity

More information

STORAGE. Buying Guide: TARGET DATA DEDUPLICATION BACKUP SYSTEMS. inside

STORAGE. Buying Guide: TARGET DATA DEDUPLICATION BACKUP SYSTEMS. inside Managing the information that drives the enterprise STORAGE Buying Guide: DEDUPLICATION inside What you need to know about target data deduplication Special factors to consider One key difference among

More information

Veeam Cloud Connect. Version 8.0. Administrator Guide

Veeam Cloud Connect. Version 8.0. Administrator Guide Veeam Cloud Connect Version 8.0 Administrator Guide April, 2015 2015 Veeam Software. All rights reserved. All trademarks are the property of their respective owners. No part of this publication may be

More information

W H I T E P A P E R R e a l i z i n g t h e B e n e f i t s o f Deduplication in a Backup and Restore System

W H I T E P A P E R R e a l i z i n g t h e B e n e f i t s o f Deduplication in a Backup and Restore System W H I T E P A P E R R e a l i z i n g t h e B e n e f i t s o f Deduplication in a Backup and Restore System Sponsored by: HP Noemi Greyzdorf November 2008 Robert Amatruda INTRODUCTION Global Headquarters:

More information

Acronis Backup & Recovery 11 Virtual Edition

Acronis Backup & Recovery 11 Virtual Edition Acronis Backup & Recovery 11 Virtual Edition Backing Up Virtual Machines Copyright Acronis, Inc., 2000-2011. All rights reserved. Acronis and Acronis Secure Zone are registered trademarks of Acronis, Inc.

More information

Using HP StoreOnce Backup Systems for NDMP backups with Symantec NetBackup

Using HP StoreOnce Backup Systems for NDMP backups with Symantec NetBackup Technical white paper Using HP StoreOnce Backup Systems for NDMP backups with Symantec NetBackup Table of contents Executive summary... 2 Introduction... 2 What is NDMP?... 2 Technology overview... 3 HP

More information

es T tpassport Q&A * K I J G T 3 W C N K V [ $ G V V G T 5 G T X K E G =K ULLKX LXKK [VJGZK YKX\OIK LUX UTK _KGX *VVR YYY VGUVRCUURQTV EQO

es T tpassport Q&A * K I J G T 3 W C N K V [ $ G V V G T 5 G T X K E G =K ULLKX LXKK [VJGZK YKX\OIK LUX UTK _KGX *VVR YYY VGUVRCUURQTV EQO Testpassport Q&A Exam : E22-280 Title : Avamar Backup and Data Deduplication Exam Version : Demo 1 / 9 1. What are key features of EMC Avamar? A. Disk-based archive RAID, RAIN, clustering and replication

More information

A Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique

A Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique A Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique Jyoti Malhotra 1,Priya Ghyare 2 Associate Professor, Dept. of Information Technology, MIT College of

More information

Acronis Backup & Recovery 10 Advanced Workstation. Update 5. User Guide

Acronis Backup & Recovery 10 Advanced Workstation. Update 5. User Guide Acronis Backup & Recovery 10 Advanced Workstation Update 5 User Guide Copyright Acronis, Inc., 2000-2011. All rights reserved. Acronis and Acronis Secure Zone are registered trademarks of Acronis, Inc.

More information

Protect Microsoft Exchange databases, achieve long-term data retention

Protect Microsoft Exchange databases, achieve long-term data retention Technical white paper Protect Microsoft Exchange databases, achieve long-term data retention HP StoreOnce Backup systems, HP StoreOnce Catalyst, and Symantec NetBackup OpenStorage Table of contents Introduction...

More information