Backup architectures in the modern data center. Author: Edmond van As edmond@competa.com Competa IT b.v.



Similar documents
Massive SAN Backups. HP World 2002 Los Angeles, CA September, 2002 Denys Beauchemin. The Innovators. We watch your data

WHITE PAPER BRENT WELCH NOVEMBER

Using HP StoreOnce Backup Systems for NDMP backups with Symantec NetBackup

Protect Microsoft Exchange databases, achieve long-term data retention

Sawmill Log Analyzer Best Practices!! Page 1 of 6. Sawmill Log Analyzer Best Practices

Using HP StoreOnce D2D systems for Microsoft SQL Server backups

Storage Backup and Disaster Recovery: Using New Technology to Develop Best Practices

WHITE PAPER: customize. Best Practice for NDMP Backup Veritas NetBackup. Paul Cummings. January Confidence in a connected world.

Protecting enterprise servers with StoreOnce and CommVault Simpana

Design and Implementation of a Storage Repository Using Commonality Factoring. IEEE/NASA MSST2003 April 7-10, 2003 Eric W. Olsen

Aspirus Enterprise Backup Assessment and Implementation of Avamar and NetWorker

Implementing an Enterprise Class Database Backup and Recovery Plan

Keys to Successfully Architecting your DSI9000 Virtual Tape Library. By Chris Johnson Dynamic Solutions International

Eliminating Backup System Bottlenecks: Taking Your Existing Backup System to the Next Level. Jacob Farmer, CTO, Cambridge Computer

Rumors, Rumors, Rumors

Technical White Paper. Symantec Backup Exec 10d System Sizing. Best Practices For Optimizing Performance of the Continuous Protection Server

EMC CLARiiON Backup Storage Solutions

CA ARCserve Family r15

Considerations when Choosing a Backup System for AFS

VERITAS NetBackup 6.0 Enterprise Server INNOVATIVE DATA PROTECTION DATASHEET. Product Highlights

Data Backup Options for SME s

Liteon True Speed White Paper SPEED UP NETWORK STORAGE PERFORMANCE WITH LITEON TRUE SPEED TECHNOLOGY

ZFS Backup Platform. ZFS Backup Platform. Senior Systems Analyst TalkTalk Group. Robert Milkowski.

BMC Recovery Manager for Databases: Benchmark Study Performed at Sun Laboratories

Veritas NetBackup 6.0 Server Now from Symantec

Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle

STORAGE SOURCE DATA DEDUPLICATION PRODUCTS. Buying Guide: inside

SQL-BackTrack the Smart DBA s Power Tool for Backup and Recovery

VERITAS NetBackup BusinesServer

CYBERNETICS. Virtualization of Tape Storage

Distribution One Server Requirements

IBM ^ xseries ServeRAID Technology

XenData Archive Series Software Technical Overview

Symantec NetBackup Backup Planning and Performance Tuning Guide

Configuring Apache Derby for Performance and Durability Olav Sandstå

Quantum StorNext. Product Brief: Distributed LAN Client

Google File System. Web and scalability

WHITE PAPER Improving Storage Efficiencies with Data Deduplication and Compression

MySQL Enterprise Backup

Using HP StoreOnce Backup systems for Oracle database backups

Enterprise Backup and Restore technology and solutions

AFS Usage and Backups using TiBS at Fermilab. Presented by Kevin Hill

BrightStor ARCserve Backup for Windows

Keys to optimizing your backup environment: Legato NetWorker

Business-Centric Storage FUJITSU Storage ETERNUS CS800 Data Protection Appliance

Business-centric Storage FUJITSU Storage ETERNUS CS800 Data Protection Appliance

How To Store Data On A Computer (For A Computer)

Continuous Data Protection. PowerVault DL Backup to Disk Appliance

Implementing an Automated Digital Video Archive Based on the Video Edition of XenData Software

Performance Report Modular RAID for PRIMERGY

Reducing Backups with Data Deduplication

Whitepaper: performance of SqlBulkCopy

How To Use External Usb Hard Drives With Backupassist With Backuphelp With Backupa2.Com (A Usb Hard Drive) With Backupahd (A Thumbdrive) With A Backupassistant (A Flash) With An External Hard Drive

PrimeArray Data Storage Solutions Network Attached Storage (NAS) iscsi Storage Area Networks (SAN) Optical Storage Systems (CD/DVD)

1 Storage Devices Summary

Keys to Optimizing Your Backup Environment: Tivoli Storage Manager

RFP - Equipment for the Replication of Critical Systems at Bank of Mauritius Tower and at Disaster Recovery Site. 06 March 2014

Amazon Cloud Storage Options

How To Limit Volume In Bacula

WHITE PAPER PPAPER. Symantec Backup Exec Quick Recovery & Off-Host Backup Solutions. for Microsoft Exchange Server 2003 & Microsoft SQL Server

NetBackup Performance Tuning on Windows

How To Backup Your Computer With A File Copy Engine

Symantec NetBackup Backup Planning and Performance Tuning Guide

Optimizing LTO Backup Performance

Tape or Disk Backup Which is Best?

Storage Architecture in XProtect Corporate

Introduction. Setup of Exchange in a VM. VMware Infrastructure

REMOTE BACKUP-WHY SO VITAL?

Implementing a Digital Video Archive Using XenData Software and a Spectra Logic Archive

Implementing Offline Digital Video Storage using XenData Software

Why Backup? What is YOUR Cost of Downtime?

Cost Effective Backup with Deduplication. Copyright 2009 EMC Corporation. All rights reserved.

Disk-to-Disk-to-Offsite Backups for SMBs with Retrospect

Maximizing Backup and Restore Performance of Large Databases

EMC Backup Storage Solutions: The Value of EMC Disk Library with TSM

An Oracle White Paper September Oracle Exadata Database Machine - Backup & Recovery Sizing: Tape Backups

Yiwo Tech Development Co., Ltd. EaseUS Todo Backup. Reliable Backup & Recovery Solution. EaseUS Todo Backup Solution Guide. All Rights Reserved Page 1

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

Summer Student Project Report

Barracuda Backup Deduplication. White Paper

SAN TECHNICAL - DETAILS/ SPECIFICATIONS

EMC CLARiiON Backup Storage Solutions

Enterprise Data Backup and Recovery

Considerations when Choosing a Backup System for AFS

Cloud Based Application Architectures using Smart Computing

WHITE PAPER. Storage Savings Analysis: Storage Savings with Deduplication and Acronis Backup & Recovery 10

White Paper. Overland REO SERIES. Implementation Best Practices

Parallels Cloud Server 6.0

Implementing a Digital Video Archive Based on XenData Software

Virtual Infrastructure Security

Everything you wanted to know about data tape bar codes but were afraid to ask

Archive Data Retention & Compliance. Solutions Integrated Storage Appliances. Management Optimized Storage & Migration

Virtual Tape Systems for IBM Mainframes A comparative analysis

June Blade.org 2009 ALL RIGHTS RESERVED

May 27, Version: Draft 1.0 MARTIN COUNTY ADMINISTRATION INFORMATION TECHNOLOGY SERVICES RECOMMENDATION ON PRODUCTS & SERVICES FOR

How To Back Up Your Computer With A Hard Drive On A Usb Or Usb 2 (For Small Businesses)

Muse Server Sizing. 18 June Document Version Muse

WHITE PAPER FUJITSU PRIMERGY SERVER BASICS OF DISK I/O PERFORMANCE

Backup Technologies. An Overview of the Different Technologies Available for Backup. What do I need to think about when choosing a backup solution

<Insert Picture Here> RMAN Configuration and Performance Tuning Best Practices

Transcription:

Backup architectures in the modern data center. Author: Edmond van As edmond@competa.com Competa IT b.v.

Existing backup methods Most companies see an explosive growth in the amount of data that they have to process and store. This fact given, it s usually necessary to take a closer look to how backups are done and how the methods of backing up data can/must be expanded or changed. To come to the right decision on what backup method(s) to use, there first must be a good understanding of what methods there are and what benefits and disadvantages these methods have. An overview. 1. Backups to disk 2. Removable media. like CDR s (CD recordable), MO (Magneto Optical)drives, ZIP drives, etc. 3. Local (per server) tape drive(s). 4. Shared tape drive(s). (1 or more backup servers) 5. Tape stackers (shared or not). 6. Robotic tape libraries (shared or not). There are more methods you can think of. I chose these because they are the most common in my opinion. 1. The first option, is probably the most expensive. You are in for a shock when you compare the cost per Gigabyte of diskdrives to cost per Gigabyte of tapes. However, it seems that more and more companies are using or going to use this option to increase the speed of important backups and dramatically decrease restore times. A common application of backing up to disk is snapshotting databases to third mirror disks or volumes. Using this mechanism you have an instant copy of your data without downtime. In case of data loss, you can synchronize the production disk with the backup disk you made earlier. It will still be necessary to stream this data to tapes for long time storage. Another advantage of using disks as backup medium is that you can actually use the data on those disks for testing of new applications for instance before putting them on your production environment. 2. It seems obvious that the second option in the list is merely useful for relatively small amounts of data. It is way too time consuming to start backing up big chunks of data to a backup medium with a storage capacity of less than a Gb. However, there are situations where these kinds of media could come in handy. You can think of data that has to be kept for an extensive period of time, like archives of tax data for instance. Or, backups of important correspondence that will be deleted from main storage but has to be kept for possible future use. Technologies keep on being improved every few months. Two years ago you could write 650 Mb to an MO disk. Today you can write 9 Gb to that MO disk. This makes it a perfect solution for backing up your desktop computers. 3. A widely spread method of backing up data is that every server has it s own tape drive(s). The advantage is that you do not have to worry about network bandwidth or the availability of a backup server. You might get in trouble though when the amount of data that you have to back up, outgrows the capacity of the locally attached hardware and used media. In this case, you will have to invest in a mechanism that can replace full media with empty media in order to continue the backup from were it left of. This mechanism can be a person changing the tape by hand what makes it very labour intensive for large sites, and is prone to human error. Or you can increase the capacity by using extra hardware, like an extra drive or a tape stacker.

4. Another common method is using a backup server. This machine has the task of gathering data of other machines spread over a network. This can be a dedicated machine, but you can also use a machine that is idle at the time you want to perform your backups. Of course, this machine has to be well equipped for this task. I will come to system requirements later. In some situations it is sensible to have more than one backup server. For instance, you have 30 machines to back up. You do not have the money to invest in a big dedicated backup server. In this case, you could point out 3 machines that are idle over night which will backup the data of 10 machines each. Like in the former example, you must carefully choose the type of hardware. Backing up several machines without making a careful inventory can easily cause your machine or chosen media to run out of capacity. Other considerations are: Are there any future plans for these machines, what kind of support contract do you have on them and are they fast enough to write all the data to tape in the designated backup window? 5. A tape stacker is an obvious choice when you run out of your single drive s capacity. A stacker needs a special kind of attention. Especially when you use your own scripts or other solutions to get your backups to tape. A stacker works according to the sequential system. The first tape is written until it is completely full. After that, the next tape has its turn. This routine repeats itself until the last tape is used up. Using a simple schedule, this will force you to change tapes as soon as the stacker is full. When you use your own smart schedule with some kind of retention period or a third party solution, it might be possible round robbin to the first tape in the sequence if the capacity of the stacker allows you to. 6. The Robotic tape library uses a totally different mechanism than the tape stacker. Unless you are a superb programmer, you need vendor or third party software solutions to make use of your purchase. The tape library is inventory aware. This means that the library knows about how many tapes there are in the library. It often uses barcodes attached to the tapes as a reference. These barcodes are read with laser or camera techniques. The software that controls the library keeps track of the content of the tapes. This offers you the possibility to mix several sets of different clients, schedules and retention periods. Do you need changes? It is of course possible that you run your backups very smoothly with your existing hardware and software. Only make changes if you feel or know that something is wrong. However, changes may be necessary if the policies of your users change. For instance, when you normally have 6 hours of downtime for a database backup, it could happen that your users want a higher availability of that database. This could force you in having to do online backups of that database. Another example is that hardware is added to the environment. When a department is in the process of implementing a 1 Terabyte fileserver, your are likely to get in to trouble when you do not investigate what the impact will be on your existing backup architecture. It is a commonly made error that project teams start looking at how they are going to get the enormous amounts of data to tape after the machine is installed. This is too late. Another reason for change is that data has to be more and more available. If availability increases, backup window size decreases. No more database downtime.

If changing, what do you need? This is a tough question to answer. First you must know what kind of data you have to back up (e.g. Databases, fileservers, webservers, etc.) and the different combinations that (have to) coexist in the environment. Use the following list of basics to begin planning your backup environment: Dataset size and average number of files. Network bandwidth. Memory. Tape drive Technologies. Software packages. Off site storage. Backup frequency. Dataset size and average number of files. To determine the total tape capacity you need, there are a number of things to consider. First of all; What is the total amount of data that you have to back up when doing a full backup (at this point it is not yet important to know if you are going to back up locally or to a backup server). If you are going to use a third party or OEM ed (Original Equipment Manufacturer) product, you must know the number of files that you have to backup too. Products like NetBackup from Veritas or NetWorker from Legato use about 150 bytes to store the location, retention period and backup time of the backed up file in its index database. So, backing up one million files leaves you with a theoretical index database size of around 150 Mb. To be able to tune your backup system in a later stage, the total number of files that reside on a file system could come in handy too. With these two values (total amount of data and number of files), you can calculate the average file size. On large file servers, you can improve backup and restore performance by tuning the block size written to tape to the average file size. On relatively small machines you will hardly notice any differences in performance. If you really want to be sure that you are not going to buy too many tapes, you can add compression too as a parameter. Also very important, the on drive hardware compression helps to improve performance because the amount of data written to the drives cache is in fact more than actually is written to tape. In the following table, the typical compression ratios and gained performance when using compression is shown. Data Type Gained Performance Compression Text 1.46:1 1.44:1 File server 1.60:1 1.63:1 Web server 1.57:1 1.82:1 Database 1.60:1 1.57:1 Ref. 1

The native write capacity of a DLT drive is 5Mb/s. This means that you can write 5Mb/s of uncompressed data to the drive. According to the table, writing speeds for a fileserver improves with 1.60:1 which produces a theoretical data stream of 8Mb/s. Network Bandwidth. In the case of a network related or centralized backup architecture, knowing about the way your network will be utilized during backups in the peak hours is essential. The following table shows a number of commonly used network technologies with their theoretical speeds and their more realistic speeds based on vendor specifications and personal experience. This table is only a guideline and transfer rates depend on actual network load at any one time. Network adapter Speed in theory Speed in real life 10BaseT Ethernet 10Mbps 0.65 0.85 MB/sec 100BaseT Ethernet 100Mbps 6.5 8.5 MB/sec Gigabit Ethernet 1000Mbps 25 53 MB/sec Ref. 1 Please note that normally the transfer rate of a network interface card is noted in bits, where 1 Mbps is actually 1000x1000 bits. We are far more interested in how many bytes, or Mega bytes for that matter, we can pump through the network. A good way of testing your network transfer rate is to create a reasonably large file on a client and setup an ftp connection to the backup server (if you have one already). If you copy the file to /dev/null on the receiving machine, you can get a pretty good picture of your network transfer rate in combination with the read capabilities of the hard drive(s) on the client. When ftp finishes the file transfer, it gives you the number of Kbytes/sec. Dividing this number by 1024 gives you Mbytes/sec. If you divide the total amount of data that has to be backup up from the client by the result of the ftp transfer, you get the estimated time the backup would take. The actual results of a live backup depends on how many clients you back up at the same time, hardware capabilities of the backup server and the type of backup hardware/media you use. If you merely want to test actual TCP traffic from network card to network card, you can use tools like ttcp for example. Using the /dev/null method can produce unreliable results on some Unixes though. Let s assume your backup server has one 100Mbps link to the network. This should be enough to get two DLT (Digital Linear Tape) drives to stream. If you would experience slow backups, adding tape drives would probably slow down your backups because the drives must share the available bandwidth causing them to slow down because they can not get into streaming mode. So, you should keep the minimum drive throughput in combination with your network speed on the receiving side in mind. The client machines are less important in this case. If you have slow clients, you can start more than one backup at the same time to increase the data stream to your backup server. For this reason it is important to know what link speeds there are on your client machines. Create an inventory list of client hardware to plan the schedules for your different server groups.

Memory. On the client side, available memory is usually enough because the system load during a backup can be compared with normal operation (more or less depending on server task). If your system behaves acceptably during operating hours, it should not degrade during backup hours. However, there are some memory considerations on the backup server, like the following: Some third party backup solutions use chunks of shared memory to buffer data before it is transferred to the tape drives used during a backup or restore. The calculation of the size of shared memory depends on 4 parameters: 1. The size of the buffers, 2. the number of buffers, 3. the number of drives and 4. the number of multiplexes (Veritas) or the number of parallelisms (Legato). Multiplexes or parallelisms are the number of backups that you stream to a single drive simultaneously. This can be multiple file systems from one client or more clients at the same time or a mix of those two. Note that you can multiplex on a per client basis or multiplex on a per drive basis. So, if you have 2 drives and 4 multiplexes per drive, you can backup 4 clients with 2 multiplexes each at the same time. Veritas suggests the following calculation to determine the size of shared memory: Shared_memory=(buffer_size*nBuffers)*nDrives*nMPX Ref. 2 So, if you have 4 drives you can use simultaneously, 4 multiplexes per drive, a buffer size of 64 Kb and 16 buffers, you would get the following: (65536*16)*4*4=16777216 which is 16 Megabytes of shared memory. Note that the calculation is done using the full notation of the number of bytes. Shared memory is a limited resource. Setting the buffer size and number of buffers too high for available shared memory can cause you problems because the software expects there to be enough shared memory to store its buffers in. A buffer size of 65536 bytes seems to give best performance on DLT drives. Tape drive technologies. The following table shows a number of different tape drive technologies: Technology Native capacity in Gb Speed in Mb/second DDS 3 12 1 DDS 4 20 3 Mammoth (Exabyte) 20 3 DLT 7000 35 5 DLT 8000 40 6 Ref. 1

Choose the technology that best fits your needs and budget. Even if you choose to use a robotic tape library, you can still decide what kind of tape technology you use to a certain level because some libraries support different kinds of tape drives. Let s say you have to back up four machines with a total of 60 Gb of data. It will probably be cheaper to buy one DLT drive that you connect to one machine and perform remote backups with ufsdump. It s not only cheaper than buying four drives for every machine (extra SCSI cards perhaps?), it is better to manage because you only have to change tapes in one drive instead of four. Next to that, administration concentrates on one location instead of four. Software packages. There are a wide variety of backup/restore software packages available on the market today. A lot of these packages support a broad range of soft and hardware platforms. There are already a number of packages designed for the Linux community too. A small selection of backup/restore software packages: Armanda Veritas NetBackup Legato Networker BRU from EST (Linux) Arkeia from Knox (Linux) HP Omniback The decision on what software to choose should be based on the needs of your environment, personal preferences (evaluate that stuff!) and, as always, budgetary limitations (or absolute freedom). Off site storage. Many companies require to have business critical data stored outside the company walls. You can imagine that in case of a fire, you do not want all your data (disks, tapes) at one location. The problem here is that you still need your data within reach to ensure quick restores if needed. To solve this problem you can clone or duplicate the data that has been written to tape during backup hours, usually at night. The most ideal way is to have a dedicated backup server that can perform these duplications when it is idle during non backup hours. This mechanism can/will influence the way you set up multiplexes or parallelisms. For instance, when you want to duplicate a tape with 8 multiplexes, the duplication process will run over the tape 8 times to construct 8 files on the duplication tape. The reason for this is to decrease restore times dramatically when you are forced to use the duplicate to restore data. Duplication can be a motivation to buy extra tape drives even if the network bandwidth of the backup server would withhold you to use them.

Backup frequency. Basically you can designate three kinds of data: 1. Static data, 2. dynamic data, 3. Machine generated data. 1. Making regular backups of static data can be a waste of time, tape space and therefore money. With static data, you can think of files the OS consists of or application files. It should be sufficient to make full backups, let s say, once a month, and an incremental backup in between full backups. It all depends on how dynamic your configuration files are. 2. Dynamic data is data that applications use or data that resides in databases such as Oracle or Sybase. This data can change very quickly. It can change so quickly that you even might consider not to make incremental backups at all because in the end they will just be full backups. 3. In some cases, machine produce data from, for instance, mathematical calculations or end of month financial calculations. The decision whether you back this data up should be based on how long it will take for the machine to regenerate or the time you need to get it back from tape. How to motivate management. Implementing a brand new backup system can be very expensive. This investment is usually a painful one for management. Most IT systems are used to run the business more efficiently and they pay themselves back in the long run. Backup systems cost money and they just sit there doing backups. Of course this is a wrong assumption. The reason you make backups is to protect your precious data from hardware failures, software failures, human error and foul play. Loosing data is way more expensive that maintaining a backup system to prevent data loss in the first place. The question you must ask yourself is: What is the company going to loose when data is completely lost or it will take an extensive period of time to get it back, one day for example. A good backup mechanism should be considered as an insurance policy. Compared with the value of the data they protect, it is a relatively cheap insurance policy too. Conclusion. Implementing a data center wide backup solution is not an easy task. Talk to the different departments and with the end users about there requirements and data types. Compare the wide range of products there are on the market and evaluate software packages. And, when it all runs? Test your restore procedures on a regular basis.

Resources Ref. 1. Planning you backup architecture. White paper. Sun Microsystems. Ref. 2. Size/number of data buffers. How it works and how to determine what to use. TechNote. Veritas software.