An On-line Backup Function for a Clustered NAS System (X-NAS)



Similar documents
THE EXPAND PARALLEL FILE SYSTEM A FILE SYSTEM FOR CLUSTER AND GRID COMPUTING. José Daniel García Sánchez ARCOS Group University Carlos III of Madrid

Shared Parallel File System

Application Brief: Using Titan for MS SQL

Performance and scalability of a large OLTP workload

Eloquence Training What s new in Eloquence B.08.00

Distributed File System Choices: Red Hat Storage, GFS2 & pnfs

Network Attached Storage. Jinfeng Yang Oct/19/2015

New!! - Higher performance for Windows and UNIX environments

Performance Comparison of Fujitsu PRIMERGY and PRIMEPOWER Servers

Integrated Grid Solutions. and Greenplum

TotalStorage Network Attached Storage 300G Cost effective integration of NAS and LAN solutions

Microsoft Windows Server 2003 with Internet Information Services (IIS) 6.0 vs. Linux Competitive Web Server Performance Comparison

BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB

Egnyte Local Cloud Architecture. White Paper

EMC Backup and Recovery for Microsoft Exchange 2007 SP2

AIX NFS Client Performance Improvements for Databases on NAS

SCALABILITY AND AVAILABILITY

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

High Availability Solutions with MySQL

SERVER CLUSTERING TECHNOLOGY & CONCEPT

File System Suite of Benchmarks

Oracle TimesTen In-Memory Database on Oracle Exalogic Elastic Cloud

Fusionstor NAS Enterprise Server and Microsoft Windows Storage Server 2003 competitive performance comparison

Distributed RAID Architectures for Cluster I/O Computing. Kai Hwang

SCI Briefing: A Review of the New Hitachi Unified Storage and Hitachi NAS Platform 4000 Series. Silverton Consulting, Inc.

IBM TotalStorage Network Attached Storage 100

EMC Unified Storage for Microsoft SQL Server 2008

EMC Business Continuity for Microsoft SQL Server 2008

Top 10 Reasons why MySQL Experts Switch to SchoonerSQL - Solving the common problems users face with MySQL

Implementing Offline Digital Video Storage using XenData Software

Technical White Paper. Symantec Backup Exec 10d System Sizing. Best Practices For Optimizing Performance of the Continuous Protection Server

Running a Workflow on a PowerCenter Grid

WHITE PAPER. Permabit Albireo Data Optimization Software. Benefits of Albireo for Virtual Servers. January Permabit Technology Corporation

Remote Copy Technology of ETERNUS6000 and ETERNUS3000 Disk Arrays

Hardware/Software Guidelines

IBM TotalStorage Network Attached Storage 300G

ADVANCED NETWORK CONFIGURATION GUIDE

The Benefits of Virtualizing

A Survey of Shared File Systems

Accelerate SQL Server 2014 AlwaysOn Availability Groups with Seagate. Nytro Flash Accelerator Cards

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage

High-Performance Nested Virtualization With Hitachi Logical Partitioning Feature

Enabling Technologies for Distributed Computing

Selling Compellent NAS: File & Block Level in the Same System Chad Thibodeau

IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE

Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1

Enabling Technologies for Distributed and Cloud Computing

Planning and Administering Windows Server 2008 Servers

The virtualization of SAP environments to accommodate standardization and easier management is gaining momentum in data centers.

Double-Take Replication in the VMware Environment: Building DR solutions using Double-Take and VMware Infrastructure and VMware Server

Protect SQL Server 2012 AlwaysOn Availability Group with Hitachi Application Protector

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Flexible Storage Allocation

On- Prem MongoDB- as- a- Service Powered by the CumuLogic DBaaS Platform

Parallels Cloud Server 6.0

Red Hat Enterprise Linux as a

PATROL Console Server and RTserver Getting Started

How To Build A Clustered Storage Area Network (Csan) From Power All Networks

Introduction to NetApp Infinite Volume

Exploiting Remote Memory Operations to Design Efficient Reconfiguration for Shared Data-Centers over InfiniBand

Microsoft Windows Server 2003 vs. Linux Competitive File Server Performance Comparison

Oracle Database Scalability in VMware ESX VMware ESX 3.5

Client/Server Computing Distributed Processing, Client/Server, and Clusters

Evaluation of Enterprise Data Protection using SEP Software

HP Intelligent Management Center User Access Management Software

NetApp Storage System Plug-In for Oracle Enterprise Manager 12c Installation and Administration Guide

Introduction to Windows Storage Server 2003 Architecture and Deployment

Red Hat Network Satellite Management and automation of your Red Hat Enterprise Linux environment

System Requirements Version 8.0 July 25, 2013

EMC Virtual Infrastructure for Microsoft SQL Server

Red Hat Satellite Management and automation of your Red Hat Enterprise Linux environment

Performance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage

New Features in SANsymphony -V10 Storage Virtualization Software

Planning Domain Controller Capacity

Simplify Data Management and Reduce Storage Costs with File Virtualization

Real-time Protection for Hyper-V

FUJITSU Software Interstage Big Data Parallel Processing Server V User's Guide. Linux(64) J2UL ENZ0(00) October 2013 PRIMERGY

HP reference configuration for entry-level SAS Grid Manager solutions

Emulex OneConnect 10GbE NICs The Right Solution for NAS Deployments

Configuration Maximums VMware Infrastructure 3

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014

NEC Corporation of America Intro to High Availability / Fault Tolerant Solutions

High Availability Solutions for the MariaDB and MySQL Database

System Requirements. Version 8.2 November 23, For the most recent version of this document, visit our documentation website.

W H I T E P A P E R. VMware Infrastructure Architecture Overview

Centralized Orchestration and Performance Monitoring

How To Evaluate Netapp Ethernet Storage System For A Test Drive

Key Messages of Enterprise Cluster NAS Huawei OceanStor N8500

EMC Backup and Recovery for Oracle Database 11g Without Hot Backup Mode using DNFS and Automatic Storage Management on Fibre Channel

Implementing an Automated Digital Video Archive Based on the Video Edition of XenData Software

Transcription:

_ An On-line Backup Function for a Clustered NAS System (X-NAS) Yoshiko Yasuda, Shinichi Kawamoto, Atsushi Ebata, Jun Okitsu, and Tatsuo Higuchi Hitachi, Ltd., Central Research Laboratory 1-28 Higashi-koigakubo, Kokubunji-shi, Tokyo 185-861, Japan Tel: +81-423-23-1111, Fax: +81-423-27-7743 e-mail: {yoshikoy, skawamo, ebata, j-okitsu, higuchi}@crl.hitachi.co.ip Abstract An on-line backup function for X-NAS, a clustered NAS system designed for entry-level NAS, has been developed. The on-line backup function can replicate file objects on X- NAS to a remote NAS in real-. It makes use of the virtualized global file system of X-NAS, and sends NFS write operations to both X-NAS and the remote backup NAS at the same. The performance of the on-line backup function was evaluated and the evaluation results show that the on-line backup function of X-NAS improves the system reliability while maintaining 8% of the throughput of the X-NAS without this function. 1. Introduction An entry-level NAS system is convenient in terms of the cost and the ease of management for offices with no IT experts. However, it is not scalable. To solve this problem, X-NAS, which is a simple, scalable clustered NAS architecture designed for entry-level NAS, has been proposed [6]. Like conventional NAS systems, it can be used for various clients, such as those using UNIX and Windows 1. X-NAS aims at the following four goals. Cost reduction by using entry-level NAS as an element Ease of use by providing a single-file-system view for various kinds of clients Ease of management by providing a centralized management function Ease of scaling-up by providing several system-reconfiguration functions To achieve these goals, X-NAS virtualizes multiple entry-level NAS systems as a unified system without changing clients' environments. In addition, X-NAS maintains the manageability and the performance of the entry-level NAS. It also can easily be reconfigured without stopping file services or changing setting information. However, when one of the X-NAS elements suffers a fault, file objects on the faulty NAS system may be lost if there are no backups. To improve the X-NAS reliability, a file-replication function must therefore be developed. The goal of the present work is to introduce an on-line backup function of X-NAS that replicates original file objects on X-NAS to a remote NAS for each file access request in real- without changing the clients' environments. The performance of the on-line backup function was evaluated and the evaluation results indicate that X-NAS with the on-line backup function improves the system reliability while maintaining 8% of the throughput of standard X-NAS. 1 Windows and DFS are trademarks of Microsoft Corporation. Double Take is a trademark of Network Specialists, Inc. All other products are trademarks of their respective corporations. 11

2. On-line backup function for X-NAS To improve the reliability of X-NAS, an on-line backup function for X-NAS has been developed. (Since the details of the X-NAS structure are discussed in another paper [6], they are not described here.) The on-line backup function consists of many sub-functions. Among these sub-functions, we focus on on-line replication, the heart of the on-line backup function, in this paper. The on-line replication replicates files of X-NAS to a remote NAS, which is called a backup NAS, in real- for each file access request. 2.1. Requirements The on-line backup function of X-NAS must meet the following requirements: Generate replicas of file objects in real- in order to eliminate the lag between the original data and the replicas. Use a standard file-access protocol such as NFS to communicate between X-NAS and the backup NAS in order to apply as many kinds of NAS as clients need. Do not change clients' environments in order to curb their management cost. 2.2. On-line replication There are several methods for replicating file objects to remote systems via an IP network. One method is to use a block I/O [5]. Since using a block I/O is a fine-grain process, all file objects are completely consistent with copied objects. However, the system structure is limited because the logical disk blocks of the objects must be allocated to the same address between the original data and its replica. Another method is to change the client's system. DFS [1] is a simple method for replicating file objects to many NASs. It replicates file objects in constant intervals but not in real-. X and the management partition in X-NAS enable the centralized management of many NAS elements and provide a unified file system view for clients (Fig. 1). X is a wrapper daemon and receives an NFS operation in place of the NFS server and sends the operation to others. On-line replication of X-NAS makes use of X in order to copy file objects to the backup NAS. By extending this function, X sends the NFS operation not only to the NFS servers on the X-NAS but also to the NFS servers on the backup NAS. All file objects can thus be replicated in real- for each NFS operation. 2.2.1. Operations NFS operations handled in X-NAS can be divided into four categories. Category 1 is reading files; category 2 is writing files; category 3 is reading directories; and category 4 is writing directories. X sends NFS operations belonging to categories 2 and 4 to both X-NAS and the backup NAS at the same. On the other hand, NFS operations belonging to categories 1 and 3 are not sent to the backup NAS. When a UNIX client sends a WRITE operation for file f to X-NAS, X on P-NAS (parent NAS) receives the operation in place of the NFS daemon. Figure 1 shows the flow of this operation, and Figure 2 shows the timing chart with or without the on-line backup function. Firstly, X specifies a that stores the file entity by using the inode number of the dummy file f on the management partition. Secondly, X invokes a sub thread and then sends the WRITE operation to the backup NAS by 12

using the thread. Thirdly, X sends the WRITE operation to the NFS daemon on the specified C-NAS (child NAS), and then the C-NAS processes the operation. Finally, X waits for the s of the operations from the NFS server on the C- NAS and from the backup NAS, and then it makes one from all the s and sends it back to the client. We call this procedure a synchronized backup. LAN UNIX (NFS) client WRITE file f Windows (CIFS) client Internal LAN #3 #2 main sub #1 X Samba #4 filec file f filea filea filef filec dummy files filea filef filec management partition C-NAS C-NAS P-NAS X-NAS Backup NAS Figure 1: Flow of WRITE operation with online backup function. client X elapsed client X-main X-sub send to backup NAS client Xmain X-sub send to backup NAS access to management partition invoke sub-thread access to wait sub-thread completion (a) X-NAS without on-line backup (b) X-NAS (c) X-NAS with partial asynchronized backup Figure 2. Timing charts of WRITE operation with or without on-line backup function. 2.2.2. Key features An on-line backup function must guarantee the consistency of data between X-NAS and the backup NAS. To achieve this, X waits for all s from both one of the NFS servers on the X-NAS and the backup NAS for each NFS operation. However, waiting for the s degrades total performance. To solve this problem, the performance of the on-line replication function must be improved through three key features as follows. (1) Multi-threaded wrapper daemon X waits for all s from both NAS systems. This incurs an overhead because of frequent accesses to the network and the disk drives. To reduce this cost, the main 13

thread of X invokes a sub thread to send the file I/Os to the backup NAS. This feature enables X-NAS to process the disk accesses of both X-NAS and the backup NAS in parallel. (2) File-handle cache The cost of specifying the full path name and the file handle on the backup NAS is high because of frequent accesses to the network and the disk drives. To reduce this cost, X- NAS makes use of a file-handle cache, which records the correspondence between the file handle of the dummy file, i.e., the global file handle, and the file handle of the backup NAS. (3) Partial asynchronized backup Although the synchronized backup is a simple method, the execution cost is high because this method waits for all the s from the NFS servers on the X-NAS and the backup NAS. A method that does not wait for the from the backup NAS achieves the same performance as X-NAS without the on-line backup function. However, when X-NAS or the backup NAS becomes faulty, it is difficult to guarantee the consistency of data between X-NAS and the backup NAS. Using a log is one solution to guarantee the consistency. However, since the log size is limited, it is not a perfect solution for entry-level NAS, which usually has a small-sized memory. Furthermore, according to the X-NAS concept, the architecture must be simplified as much as possible. X thus supports a partial asynchronized backup method in addition to the synchronized backup. Figure 2(c) shows the timing chart of the WRITE operation with partial asynchronized backup. In the method, after processing disk accesses to the data partition on the X-NAS element, X sends back a to a client without waiting for the from the backup NAS. As a result, the client can send the next operation. The main thread of X can perform the disk accesses to the management partition for the next operation during the waiting for the from the backup NAS. 3. Performance evaluation To evaluate the on-line backup function of X-NAS, an X-NAS prototype based on the NFSv3 implementation was developed. We ran NetBench [3] and SPECsfs97 [4] on the X-NAS prototype with or without on-line backup function. In this evaluation, by taking account of permissible range for the entry-level NAS's users, we set the performance objective for X-NAS with the on-line backup function at 8% of the performance of X- NAS without the function. Throughput and average are used as the performance metrics. In this evaluation, we implemented the partial asynchronized backup function in the WRITE operation. This is because the ratio of the WRITE operations to all operations is higher than other operations in the workload mix of the benchmarks. Furthermore, since the file sizes used by the benchmark programs are from 1 to 3 KB, many WRITE operations are issued continuously and then each process in a WRITE operation could be overlapped. 3.1. Experimental environment In the experimental environment, the maximum number of X-NAS elements is fixed to four. Each X-NAS element and the backup NAS configured with one 1-GHz Pentium III 14

processor, 1 GB of RAM and a 35-GB Ultra 16 SCSI disk drive running Red Hat Linux 7.2. For the NetBench test, one to eight clients running Windows 2 Professional were used. The clients, P-NAS, C-NASs, and the backup NAS were connected by 1-Megabit Ethernet because most offices still use this type of LAN. 3.2. Results Figures 3 and 4 show the results of our performance evaluation in terms of throughput and average. The throughputs of X-NAS with the synchronized backup function are about 8% of those without the function. Under an experimental environment with NetBench, the average for X-NAS with the function is about 1.2 s higher than that for X-NAS without the function. On the other hand, under an experimental environment with SPECsfs, the average for X-NAS with the function is about 1.4 s higher than the for X-NAS without it. Although the partial asynchronized backup can improve both throughput and average by several percentage, the performance objective for the in the case of SPECsfs cannot be achieved yet. Throughput (Mbit/sec) 5 4 3 2 without on-line backup with partial-asynchronized backup 1 1 2 3 4 5 6 7 8 Number of clients (a) Throughput Response (msec) 4 without on-line backup with partial-asynchronized backup 3 1 2 3 4 5 6 7 8 Number of clients (b) Average Figure 3. Throughput and average of X-NAS with or without on-line backup function in the case of NetBench. 2 1 Delivered load (NFSOPS) 8 3 Response (msec) Response (msec) 1 6 2 8 6 4 4 1 2 2 2 4 6 8 2 4 6 8 2 4 6 8 Offered load (NFSOPS) without on-line backup Offered load (NFSOPS) with partial-asynchronized backup Offered load (NFSOPS) (a) Total throughput (b) Total average (c) Average s for WRITE operations Figure 4. Throughput and average of X-NAS with or without on-line backup function in the case of SPECsfs. 3.3. Discussion To specify the reason for the longer in the case of SPECsfs, the average 15

of each NFS operation in the case of the synchronized backup was analyzed. The average s for some write requests such as WRITE, SETATTR and CREATE are longer than those for X-NAS without that function. In particular, the average for WRITE operations is 2.5 s higher than that for the other operations. Profiling results of the WRITE operations shows that waiting for the sub-thread completion is about 24% of the total processing and access to the via an IP network is about 48% of that. By applying the partial asynchronized backup to X-NAS, this waiting can be reduced to almost zero. Figure 4(c) shows the effects of the partial asynchronized backup in the case of WRITE operations. The average for WRITE operations with the partial asynchronized backup can be reduced to from 2.5 s to 1.8 s the for X-NAS without the function. As a result, the total average for SPECsfs with the function can be reduced to 1.3 s that without it. However since the ratio of data transmission to the total processing is still higher in the case of the 1- Megabit Ethernet, using a Gigabit network is effective because it can reduce the data transmission for 1-Megabit Ethernet to at least one-fifth. Furthermore, by optimizing other operations such as CREATE and COMMIT, the performance objective of 1.2 s can be achieved. 4. Related work There are several methods for replicating file objects between several NAS systems via the network. DFS [1] is a simple and easy file-replication function on Windows systems. DRBD [5] is a kernel module for building a two-node HA cluster under Linux. Double Take [2] is a third-vendor software to replicate file objects on the master NAS to the slave NAS. 5. Conclusions An on-line backup function for X-NAS, a clustered NAS system, has been developed. On-line replication, the core of the on-line backup function, replicates file objects on X- NAS to a remote backup NAS in real- for each NFS operation. A multi-threaded wrapper daemon with a low overhead, the developed file-handle cache and the partial asynchronized backup method can reduce the overhead for accessing the backup NAS. An X-NAS prototype with the on-line backup function, based on NFSv3 running the NetBench and SPECsfs97 programs attains 8% of the performance of X-NAS without the function. This function improves the dependability of entry-level NAS while maintaining its manageability. References [1] Deploying Windows Powered NAS Using Dfs with or without Active Directory. http://www.microsoft.com, 21. [2] Double-Take Theory of Operations. http://www.nsisoftware.com, 21. [3] NetBench 7..3. http://www.etestomglabs.com/benchmarks/netbench, 22. [4] SFS3. Documentation Version 1.. http://www.spec.org, 22. [5] P. Reisner. DRBD. In Proceedings of the 7 th International Linux Kongress, 2. [6] Y. Yasuda et al. Concept and Evaluation of X-NAS: a highly scalable NAS system. In Proceedings of the 2 th IEEE/11 th NASA MSST23. 16