1 An Implementation of Storage-Based Synchronous Remote Mirroring for SANs SHU Ji-wu, YAN Rui, WEN Dongchan, ZHENG Weimin Department of Computer Science and Technology, Tsinghua University, Beijing , China Abstract. Remote mirroring ensures that all data written to a primary storage device are also written to a remote secondary storage device to support disaster recoverability. In this study, we designed and implemented a storage-based synchronous remote mirroring for SAN-attached storage nodes. Taking advantage of the high bandwidth and long-distance linking ability of dedicated fiber connections, this approach provides a consistent and up-to-date copy in a remote location to meet the demand for disaster recovery. This system has no host or application overhead, and it is also independent of the actual storage unit. In addition, we present a disk failover solution. The performance results indicate that the bandwidth of the storage node with mirroring under a heavy load was 98.67% of the bandwidth without mirroring, which was only a slight performance loss. This means that our synchronous remote mirroring has little impact on the host s average response time and the actual bandwidth of the storage node. 1 Introduction Remote mirroring ensures that all data written to a primary storage are also written to a remote secondary storage to support disaster recoverability. It can be implemented at various levels, including the file system, the volume manager, the driver, the host bus adapter (HBA) and the storage control unit . In general, there are two locations at which mirroring are implemented: the storage control unit and the host. Each location has its own advantages and disadvantages. IBM s Peer-to-Peer Remote Copy (PPRC) and EMC s Symmetrix Remote Data FacilitySRDF use a synchronous protocol at the level of the storage control unit. Today s storage control units contain general-purpose processors, in addition to special-purpose elements for moving and computing blocks of data. Therefore, remote mirroring is provided by storage subsystems as advanced copy functions. But it costs a lot and depends on the actual disk s storage subsystem. Veritas s Volume Replicator  is a remote mirroring solution at the level of the host s device driver. It intercepts write operations at the host device driver level and sends the changes to a remote device. So it is kept independent of the actual storage unit. However, it takes a toll on the host CPU cycles and communication bandwidth, and it is difficult to manage because it needs to interact with all the
2 hosts where replication software is installed. NetApp s SnapMirror is a remote mirroring solution using an asynchronous protocol at the level of the host s file system, but it is not suitable for block-level I/O access in the SAN environment. In this study, we designed and implemented synchronous remote mirroring at the level of the storage control unit for Tsinghua Mass Storage Network System (TH-MSNS) , an implementation of the FC-SAN. Based on the high bandwidth and long-distance linking ability of dedicated fiber connections, this approach provides a consistent and up-to-date data copy in a remote location to meet the demand for disaster recovery. This implementation of remote mirroring has no host or application overhead, and it is also independent of the actual storage unit. In addition, we present a failover solution for disk failure. The performance results indicate that our synchronous remote mirroring does not have a significant effect on the average command response time or on the actual bandwidth of the storage node. 2 Introduction of the TH-MSNS The TH-MSNS is an implementation of an FC-SAN. In the TH-MSNS, the storage nodes provide storage services. A storage node is composed of a general-purpose server, SCSI disk arrays, and fibre channel adapters, and it has a software module named the SCSI target simulator  running on it. By using the SCSI target simulator to control the I/O process to access disk arrays, the TH-MSNS can implement basic functions of FC disk arrays while only using general SCSI disk arrays. Because of this, it is low cost, highly flexible and can achieve considerable performance . Figure 1 shows the I/O path of the TH- MSNS. The file system and SCSI driver in the host converts the application s I/O Fig. 1. The I/O path of the TH-MSNS requests to SCSI commands and data, and then the FC HBA driver encapsulates them into FC frames using Fibre Channel Protocol (FCP) and sends them to the SAN. When the FC HBA on the storage node receives the frames, the FC
3 target driver transforms them back into SCSI commands and data. Then the SCSI target simulator, a kernel module running on the storage node, queues and fills up the SCSI requests structures, and finally prompts the API of the SCSI driver layer to commit the SCSI requests to the SCSI subsystem. After the SCSI subsystem has completed the SCSI commands, the SCSI target simulator returns the command status or data to the host. Therefore, by the coordination of the SCSI target simulator and the FC target driver, the SCSI disk arrays of the storage node can be directly mapped to the host as its own local disks. So the storage node is equal to the storage control unit in the SAN environment at the basic function of storage service. 3 Design and implementation of remote mirroring for the TH-MSNS 3.1 The Architecture of Remote Mirroring Figure 2 shows the architecture of remote mirroring for the TH-MSNS. We Fig. 2. The architecture of remote mirroring for TH-MSNS added a remote storage node with the same structure and configuration as the local storage node. By adding an FC HBA in the local storage node to connect point-to-point with the remote storage node, the remote storage node s disks are regarded as the local storage node s own disks. Therefore, the storage node and remote storage node can constitute a mirrored pair. The write commands from the host can be mirrored to the remote storage node by the SCSI target simulator on the local storage node. So the data on the storage node can be mirrored to the remote storage node. The remote storage node can be located up to 10km away from the local storage node using fibre channel technology, and this distance can also been extended by using extenders. The advantages of this approach of remote mirroring are as follows. It has low cost and high flexibility because the remote mirroring is implemented by software modules, not by special hardware.
4 The actual storage unit is independent because the host s SCSI commands are redirected to the storage node s SCSI driver layer, which provides a common API for I/O access and hides the detail of the low-level storage subsystem. The mirroring process is transparent to the host, and does not consume any of the host s resource. Moreover, the remote storage node is also transparent to the host, because the SCSI target simulator can prevent the remote disks being mapped by the host. 3.2 Synchronous Mirroring Synchronous remote mirroring writes data not only to a local disk but also to a remote mirror disk at the same time. The acknowledgement is not be sent back until all the data is written to both disks. In many cases, both copies of the data are also locally protected by RAID. This approach provides a consistent and up-to-date copy in a remote location to meet the demand for disaster recovery. In the mirroring architecture we presented, the SCSI target simulator on the local storage node receives SCSI commands and data from the FC target driver. Then it converts each write command into a pair of write commands to mirrored disks, queues them into different request queues, and finally prompts the SCSI driver to process them. Actually, the local write command is sent to the local disk, and the remote write command is sent to the network disk mapped by the remote storage node. The remote write command is received by the SCSI target simulator on the remote storage node through the pointto-point fiber connection. The acknowledgement is not sent back to the host until both the local and remote write commands have been completed. Figure 3 shows the local and remote I/O path of synchronous mirroring. The dashed line represents the I/O path of the remote write commands, and the solid line represents the I/O path of the local read/write commands. In order to reduce Fig. 3. The I/O path of remote mirroring for the TH-MSNS
5 the command processing time, local and remote write commands can share the same data buffer. It is not necessary to apply the data buffer for the remote write command; the data pointer only has to be pointed to the data buffer of the local write command. Local write and remote commands can be committed to the SCSI driver at almost the same time, and can be executed by different HBAs concurrently. In this way, commands can be executed more efficiently. There are two device chain structures in the SCSI target simulator: the local disk chain and the remote disk chain. Figure 4 shows the local and remote disk chains. The structure of the local disk in the local disk chain contains a pointer Fig. 4. The local and remote disk chains which points to the structure of its mirrored disk. The relationship of mirrored disks can also be created between two local disks, just like the software RAID 1. When a SCSI command arrives, the SCSI target simulator analyzes the SCSI command s target disk and finds the remote mirrored disk through the pointer mentioned above. Then the SCSI target simulator fills the mirrored write request using the information of the mirrored disk such as the number of the SCSI host bus, target, or LUN and so on. Furthermore, the SCSI target simulator only maps the local disk to the host, so the remote disk is invisible to host. 3.3 Disk fail over Although both the local and remote disks are also locally protected by RAID, many circumstances could cause a SCSI command s failure. Some examples include the power failure of the disk array, an unplugged SCSI cable, a break in the fiber connection or the failure of the RAID. Because of this, it is necessary to monitor all write commands locally and remotely. When a command is returned with an abnormal status, some actions must be taken immediately to ensure the services continuity. For the accidents mentioned above, SCSI commands will timeout. The SCSI driver layer will try to recover it or retry command. If these actions fail, the SCSI driver will return the SCSI command with the timeout status to the SCSI target simulator. The SCSI target simulator analyzes the status of the local and remote command, and adopts different measures to meet different instances. The remote write command fails, but the local command is ok. In this case, the remote disk is assumed to be defunct. The mirror relationship needs to be broken; the corresponding information should be recorded for later resynchronization.
6 The local write command fails, but the remote command is ok. In this case, the local disk is assumed to be defunct and the mirror relationship must be severed. The corresponding information must been recorded for later resynchronization. The most important action is to redirect the read and write commands on the local disk to the remote mirrored disk. The local read command fails. In this case, the local disk is assumed to be defunct. The mirror relationship is broken, and the corresponding information is recorded for later resynchronize. The most important action is to redirect the read and write commands on local disk to the remote mirrored disk. In order to perform the disk failover, the local disk s structure contains a status identifier to record the status of the local disk. Table 1 shows the possible status of the local disks. In addition, some key remote mirroring implementation tech- Device s state Description DEVICE OK Disk is ok. DEVICE DEFUNCT Disk is defunct. If disk has mirrored, it means that both local and remote disks are defunct. DEVICE MIRRORED Disk has mirrored, both local and remote disks are ok. DEVICE LOCAL DEFUNCT Disk has mirrored. Remote disk is ok, but local disk is defunct. DEVICE MIRROR DEFUNCT Disk has mirrored. Local disk is ok, but remote disk is defunct. DEVICE SYNCING Disk is synchronizing. Table 1. The status of the local disk niques, such as software LUN masking, online resynchronization and disaster tolerance, have been introduced in another paper . 4 Performance evaluation The synchronous remote mirroring system was tested, and its performance was evaluated and analyzed. Because the read command is executed locally in the process of mirroring, our testing only used the write command. Table 2 shows the test configuration of the host and storage nodes. Host A CPU Intel Xeon 700MHz 4 Memory 1G OS Linux (kernel: ) FC HBA Emulex LP9822Gb/s Host B CPU Intel Itanium 2 1GHz 2 Memory 2G OS Linux (kernel: e12) FC HBA Emulex LP982(2Gb/s) Storage node and its storage subsystem
7 CPU Intel Xeon 2.4GHz 1 Memory 1G OS Linux (kernel: ) FC HBA Emulex LP982 (Initiator mode2gb/s) Qlogic ISP 2300 (Target mode2gb/s) RAID Controller Adaptec Ultra160 RAID Controller 2110S SCSI Disks Seagate Cheetah (73GB 10KRPM) 7 configured as JBOD Table 2. Test configuration of hosts and storage node 4.1 Comparison of average command response times The average response time of the command is a very important factor to evaluate the performance and quality of services. In this test, a host issues commands with different data block sizes to its one network disk, which is provided by the local storage node. The goal is to compare the average response time of each command both with mirroring and without. The Iometer benchmarking kit was used. The host issues sequential write commands with different block size ranging from 64KB to 2048 KB. This test adopts a host configured as host A (given above) and two storage nodes: a local storage node and a remote storage node. Table 3 shows the test results. The results show that synchronous mirroring has little Block Size Command Average Response Time No mirror(ms) Mirror(ms) 64KB KB KB KB KB KB KB KB KB Table 3. Comparison of average command response time on different block sizes impact on the average command response time for different block sizes. 4.2 The total execution time for replication in a large amount of data In this test, we used command dd on the host to replicate a large amount of data (10GB-50GB) to the network disk. By comparing the total execution time with and without mirroring, we evaluated the performance of synchronous mirroring. Command dd is able to directly generate read/write block-level requests. This test adopts a host configured as host A and two storage nodes. Figure 5 shows the test results. The results show that synchronous mirroring has little impact on the total execution time for replicating a large amount of data.
8 Fig. 5. The total execution time for replication of a large amount of data 4.3 The storage node bandwidth with heavy loads In this test, we adopted seven hosts, local and remote storage nodes which both have 7 SCSI disks and are mapped to the hosts. Each host accesses its one network disk respectively. We ran IOzone benchmarking kit on each host to offer the local storage node a heavy load. The test used sequential 100% write requests with different record sizes ranging from 4 KB to 4096 KB. The hosts file system were ext2, and the test files size was 15 GB. We compared the bandwidth of local storage node with mirroring and without mirroring. The seven hosts were configured as host B. Figure 6 shows the results. The results show the bandwidth of storage node with different record sizes. With the host cache, the results appear higher than actual bandwidth of the storage node. In the figure, the black bars represent the results without mirroring, with an average bandwidth of MB/s, while the white bars represent the results with mirroring, with an average bandwidth of MB/s. The bandwidth of storage node with mirroring is 98.67% of that without mirroring. The performance loss is very little. In addition, the SCSI HBA we adopted in the test was an Ultra 160 card, so the max throughput in theory was only 160Mbyte/s. Furthermore, the storage node is a server which is only responsible for I/O operations without caring other services, so its CPU has not heavily load. Actually, its CPU utilization is less than 20% at most time.
9 Fig. 6. A comparison of storage node bandwidth with heavy loads 5 Conclusion In this study, we designed and implemented a storage-based synchronous remote mirroring system for the TH-MSNS. Based on the high bandwidth and longdistance linking ability of dedicated fiber connections, this approach provides a consistent and up-to-date data copy in a remote location to meet the demands of disaster recovery. This system is independent of the actual storage unit, and the process of mirroring is transparent to the hosts. In addition, we present a disk failover solution. In the performance evaluation, we compared the average command response time with different block sizes, the total execute time in replicating a large amount of data, and the bandwidth of storage node under heavy loads, both with and without synchronous mirroring. The performance results indicate that the bandwidth of storage node with mirroring under heavy loads is 98.67% of the bandwidth of storage node without performing mirroring, a result showing only slight performance loss. This means that our synchronous remote mirroring has little impact on the host s average response time and the actual bandwidth of the storage node. Acknowledgements The work described in this paper was supported by the National High-Tech Research and Development Plan of China under Grant No. 2001AA References 1. A.C.Azagury,M.E.Factor,W.F.Micka. Advanced Functions for Storage Subsystems: Supporting Continuous Availability, IBM SYSTEM Journal VOL 42,NO 2,2003.
10 2. Using EMC SnapView and MirrorView for Remote Backup, Engineering White Paper,EMC Corporation(April 2002). 3. VERITAS Volume Replicator Successful Replication and Disaster Recovery,Veritas Software. Corporation.See replicator whitepaper.pdf 4. Patterson, R.H., et al. SnapMirror: File-System-Based Asynchronous Mirroring for Disaster Recovery. In First USENIX conference on File and Storage Technologies Monterey, CA, USA 5. Shu Ji-wu, et al. A Highly Efficient FC-SAN Based On Load Stream. Xingming Zhou, Stefan Jahnichen, Ming Xu, and Jiannong Cao eds., The Fifth International Workshop on Advanced Parallel Processing Technologies, LECTURE NOTES IN COMPUTER SCIENCE 2834, pp.31-40,2003, 6. Technical Report: Design and Implementation of the TH-MSNS (In Chinese), Computer Science Department, Tsinghua University, P.R. China, 2003, 7. Ashish Palekar, Narendran Ganapathy. Design and Implement of A LINUX SCSI Target for Storage Area Networks. Proceedings of the 5th Annual Linux Showcase & Conference, 2001.Oakland,USA. 8. Jae-Chang Namgoong,Chan-Ik Park. Design and Implement of a Fibre Channel Network Driver for SAN-Attached RAID Controllers. IEEE Parallel and Distributed Systems, Fibre Channel Protocol for SCSI (FCP), ANSI X.272, rev4.5, D.Patterson, G.Gibson, and R.Katz, A Case for Redundant Arrays of Inexpensive Disks (RAID), ACM SIGMOD, June YAO Jun, SHU Ji-wu, ZHENG Wei-minA Distributed Storage Cluster Design for Remote Mirroring based on Storage Area NetworkIEEE Transactions on computers(submitted), also in Jerry Sievert. Iometer: The I/O Performance Analysis Tool for Servers IOzone Filesystem Benchmark Richard Barker, Paul Massiglia,Storage Area Network Essentials: a complete guide to understanding and implementing SANS. Publish by John Wiley&Sons, Inc.,New York, Draft, T10 Project 1561-D, SCSI Architecture Model - 3 (SAM-3), Revision 3 16 September 2002,
Yao J, Shu JW, Zheng WM. Distributed storage cluster design for remote mirroring based on storage area network. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY 22(4): 521ο526 July 2007 Distributed Storage Cluster
TH-VSS: An Asymmetric Storage Virtualization System for the SAN Environment Da Xiao, Jiwu Shu, Wei Xue, and Weimin Zheng Department of Computer Science and Technology, Tsinghua University, 100084 Beijing,
Oracle Database Deployments with EMC CLARiiON AX4 Storage Systems Applied Technology Abstract This white paper investigates configuration and replication choices for Oracle Database deployment with EMC
Remote Copy Technology of ETERNUS6000 and ETERNUS3000 Disk Arrays V Tsutomu Akasaka (Manuscript received July 5, 2005) This paper gives an overview of a storage-system remote copy function and the implementation
Storage I/O Performance on VMware vsphere 5.1 over 16 Gigabit Fibre Channel Performance Study TECHNICAL WHITE PAPER Table of Contents Introduction... 3 Executive Summary... 3 Setup... 3 Workload... 4 Results...
Performance Study PVSCSI Storage Performance VMware ESX 4.0 VMware vsphere 4 offers Paravirtualized SCSI (PVSCSI), a new virtual storage adapter. This document provides a performance comparison of PVSCSI
A Method of Deduplication for Data Remote Backup Jingyu Liu 1,2, Yu-an Tan 1, Yuanzhang Li 1, Xuelan Zhang 1, Zexiang Zhou 3 1 School of Computer Science and Technology, Beijing Institute of Technology,
TECHNICAL NOTE VMware Infrastructure 3 SAN Conceptual and Design Basics VMware ESX Server can be used in conjunction with a SAN (storage area network), a specialized high speed network that connects computer
Redundant Optical Storage System Using DVD-RAM Library Takaya Tanabe, Makoto Takayanagi, Hidetoshi Tatemiti, Tetsuya Ura, and Manabu Yamamoto firstname.lastname@example.org NTT Integrated Information & Energy Systems
Fourth International Workshop on Storage Network Architecture and Parallel I/Os A Packet Forwarding Method for the ISCSI Virtualization Switch Yi-Cheng Chung a, Stanley Lee b Network & Communications Technology,
A Method of Deduplication for Data Remote Backup Jingyu Liu 1,2, Yu-an Tan 1, Yuanzhang Li 1, Xuelan Zhang 1, and Zexiang Zhou 3 1 School of Computer Science and Technology, Beijing Institute of Technology,
The functionality and advantages of a high-availability file server system This paper discusses the benefits of deploying a JMR SHARE High-Availability File Server System. Hardware and performance considerations
VMware SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Table of Contents Introduction.... 3 vsphere Architectural Overview... 4 SAN Backup
Using Multipathing Technology to Achieve a High Availability Solution Table of Contents Introduction...3 Multipathing Technology...3 Multipathing I/O Implementations...5 Storage Redundancy...5 Infortrend
Achieving a Million I/O Operations per Second from a Single VMware vsphere 5.0 Host Performance Study TECHNICAL WHITE PAPER Table of Contents Introduction... 3 Executive Summary... 3 Software and Hardware...
A Performance Analysis of the iscsi Protocol Stephen Aiken email@example.com Dirk Grunwald firstname.lastname@example.org Jesse Willeke email@example.com Andrew R. Pleszkun firstname.lastname@example.org
Performance and scalability of a large OLTP workload ii Performance and scalability of a large OLTP workload Contents Performance and scalability of a large OLTP workload with DB2 9 for System z on Linux..............
VMware vsphere 6 and Oracle Database Scalability Study Scaling Monster Virtual Machines TECHNICAL WHITE PAPER Table of Contents Executive Summary... 3 Introduction... 3 Test Environment... 3 Virtual Machine
At one point in time only a single processor was needed to power a server and all its applications. Then came multiprocessing, in which two or more processors shared a pool of memory and could handle more
Enterprise Linux Business Continuity Solutions for Critical Applications Introducing SteelEye Protection Suite for Linux Multi-site Cluster May 15 2009 EMEA DataKeeper Our Expanding Family of Solutions
Agenda Enterprise Performance Factors Overall Enterprise Performance Factors Best Practice for generic Enterprise Best Practice for 3-tiers Enterprise Hardware Load Balancer Basic Unix Tuning Performance
Best practices for fully automated disaster recovery of Microsoft SQL Server 2008 using HP Continuous Access EVA with Cluster Extension EVA Subtitle Table of contents Overview... 2 Key findings... 3 Solution
Overview of I/O Performance and RAID in an RDBMS Environment By: Edward Whalen Performance Tuning Corporation Abstract This paper covers the fundamentals of I/O topics and an overview of RAID levels commonly
White Paper EMC MID-RANGE STORAGE AND THE MICROSOFT SQL SERVER I/O RELIABILITY PROGRAM Abstract This white paper explains the integration of EMC Mid-range Storage arrays with the Microsoft SQL Server I/O
EMC Business Continuity for Microsoft SQL Server 2008 Enabled by EMC Celerra Fibre Channel, EMC MirrorView, VMware Site Recovery Manager, and VMware vsphere 4 Reference Architecture Copyright 2009, 2010
Performance Report Modular RAID for PRIMERGY Version 1.1 March 2008 Pages 15 Abstract This technical documentation is designed for persons, who deal with the selection of RAID technologies and RAID controllers
Network Attached Storage Jinfeng Yang Oct/19/2015 Outline Part A 1. What is the Network Attached Storage (NAS)? 2. What are the applications of NAS? 3. The benefits of NAS. 4. NAS s performance (Reliability
A Fast Disaster Recovery Mechanism for Volume Replication Systems Yanlong Wang, Zhanhuai Li, and Wei Lin School of Computer Science, Northwestern Polytechnical University, No.127 West Youyi Road, Xi'an,
E-Series NetApp E-Series Storage Systems Mirroring Feature Guide NetApp, Inc. 495 East Java Drive Sunnyvale, CA 94089 U.S. Telephone: +1 (408) 822-6000 Fax: +1 (408) 822-4501 Support telephone: +1 (888)
Post-production Video Editing Solution Guide with Quantum StorNext File System AssuredSAN 4000 Dot Hill Systems introduction 1 INTRODUCTION Dot Hill Systems offers high performance network storage products
CS341: Operating System Lect 36: 1 st Nov 2014 Dr. A. Sahu Dept of Comp. Sc. & Engg. Indian Institute of Technology Guwahati File System & Device Drive Mass Storage Disk Structure Disk Arm Scheduling RAID
EMC Backup and Recovery for SAP Microsoft SQL Server 2005 on Windows Server 2003 Enabled by EMC CLARiiON CX3, EMC Disk Library, EMC Replication Manager, EMC NetWorker, and Symantec Veritas NetBackup Reference
Page 1 15-Drive Supports over 5 TB of reliable, low-cost, high performance storage 15200 Product Highlights First to deliver a full HW iscsi solution with SATA drives - Lower CPU utilization - Higher data
from to HP 2013 A.W. Krings 1 Background NonStop Systems - Cyclone Commercial Database Systems with very long MTTF Modularity» units of service, failure, diagnosis, repair, growth» fault containment regions»
HP recommended configuration for Microsoft Exchange Server 2010: HP LeftHand P4000 SAN Table of contents Executive summary... 2 Introduction... 2 Solution criteria... 3 Hyper-V guest machine configurations...
Violin: A Framework for Extensible Block-level Storage Michail Flouris Dept. of Computer Science, University of Toronto, Canada email@example.com Angelos Bilas ICS-FORTH & University of Crete, Greece
Infortrend ESVA Family Enterprise Scalable Virtualized Architecture R Optimized ROI Ensures the most efficient allocation of consolidated capacity and computing power, and meets wide array of service level
BY Shashwath Veerappa Devaru CS615 Aspects of System Administration Using Multiple Hard Drives for Performance and Reliability RAID is the term used to describe a storage systems' resilience to disk failure
Exploring RAID Configurations J. Ryan Fishel Florida State University August 6, 2008 Abstract To address the limits of today s slow mechanical disks, we explored a number of data layouts to improve RAID
AX4 5 Series Software Overview March 6, 2008 This document presents an overview of all software you need to configure and monitor any AX4 5 series storage system running the Navisphere Express management
Eloquence Training What s new in Eloquence B.08.00 2010 Marxmeier Software AG Rev:100727 Overview Released December 2008 Supported until November 2013 Supports 32-bit and 64-bit platforms HP-UX Itanium
658 IEEE TRANSACTIONS ON NUCLEAR SCIENCE, VOL. 57, NO. 2, APRIL 2010 A High-Performance Storage System for the LHCb Experiment Juan Manuel Caicedo Carvajal, Jean-Christophe Garnier, Niko Neufeld, and Rainer
SMB Direct for SQL Server and Private Cloud Increased Performance, Higher Scalability and Extreme Resiliency June, 2014 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server
White Paper Deploying Oracle Data Guard on Oracle Database 10g R2 Enterprise Edition Real Application Clusters with Red Hat Enterprise Linux 4 on Dell PowerEdge Servers and Dell EMC Storage Abstract Oracle
IBM ^ xseries ServeRAID Technology Reliability through RAID technology Executive Summary: t long ago, business-critical computing on industry-standard platforms was unheard of. Proprietary systems were
SANtricity Storage Manager Remote Volume Mirroring Feature Guide for Version 8.3x TI12932-E1, First Edition Proprietary Rights Notice This document contains proprietary information of LSI Logic and Storage
WHITE PAPER How To Build a SAN The Essential Guide for Turning Your Windows Server Into Shared Storage on Your IP Network TABLE OF CONTENTS Introduction... 3 What is a SAN?... 4 Why iscsi Storage?... 4
Virtualized SQL Server Database Performance Study TECHNICAL WHITE PAPER Table of Contents Executive Summary... 3 Introduction... 3 Factors Affecting Storage vmotion... 3 Workload... 4 Experiments and Results...
EMC Solutions for Microsoft SQL Server 2005 on Windows 2003 in VMware ESX Server EMC CLARiiON CX3 Series FCP EMC Global Solutions 42 South Street Hopkinton, MA 01748-9103 1-508-435-1000 www.emc.com www.emc.com
Performance Analysis and Testing of Storage Area Network Yao-Long Zhu, Shu-Yu Zhu and Hui Xiong Data Storage Institute, Singapore Email: firstname.lastname@example.org http://www.dsi.nubs.edu.sg Motivations What
Chapter 13 Selected Storage Systems and Interface Chapter 13 Objectives Appreciate the role of enterprise storage as a distinct architectural entity. Expand upon basic I/O concepts to include storage protocols.
8Gb Fibre Channel Adapter of Choice in Microsoft Hyper-V Environments QLogic 8Gb Adapter Outperforms Emulex QLogic Offers Best Performance and Scalability in Hyper-V Environments Key Findings The QLogic
Performance Advantages for Oracle Database At a Glance This Technical Brief illustrates that even for smaller online transaction processing (OLTP) databases, the Sun 8Gb/s Fibre Channel Host Bus Adapter
, pp.325-336 http://dx.doi.org/10.14257/ijca.2015.8.11.31 Research and Implementation of CDP Disaster Recovery System based on Oracle Yongjian Ren ab, Jiaolong Ye ab, Jilin Zhang abc, Li Zhou ab, Jue Wang
WHITE PAPER FUJITSU PRIMERGY AND PRIMEPOWER SERVERS Performance Comparison of Fujitsu PRIMERGY and PRIMEPOWER Servers CHALLENGE Replace a Fujitsu PRIMEPOWER 2500 partition with a lower cost solution that
What s the best disk storage for my i5/os workload? Sue Baker IBM System i Advanced Technical Support Agenda Storage management styles Storage technologies for i5/os Considerations for implementing and
HP Data Protector Software Performance White Paper Executive summary... 3 Overview... 3 Objectives and target audience... 4 Introduction and review of test configuration... 4 Storage array... 5 Storage
OSDsim - a Simulation and Design Platform of an Object-based Storage Device WeiYa Xi Wei-Khing For DongHong Wang Renuga Kanagavelu Wai-Kit Goh Data Storage Institute, Singapore xi email@example.com
World Telecom Congress 2012 Workshop on Cloud Computing in the Telecom Environment, Bridging the Gap A Filesystem Layer Data Replication Method for Cloud Computing Masanori Itoh, Kei-ichi Yuyama, Kenjirou
Implementing Enterprise Disk Arrays Using Open Source Software Marc Smith Mott Community College - Flint, MI Merit Member Conference 2012 Mott Community College (MCC) Mott Community College is a mid-sized
DATASHEET VERITAS Storage Foundation 4.3 for Windows Advanced Volume Management Technology for Windows In distributed client/server environments, users demand that databases, mission-critical applications
Large Scale Storage Orlando Richards, Information Services firstname.lastname@example.org LCFG Users Day, University of Edinburgh 18 th January 2013 Overview My history of storage services What is (and is not)
Using Synology SSD Technology to Enhance System Performance Synology Inc. Synology_SSD_Cache_WP_ 20140512 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges...
Protect SQL Server 2012 AlwaysOn Availability Group with Hitachi Application Protector Tech Note Nathan Tran The purpose of this tech note is to show how organizations can use Hitachi Applications Protector
Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...
Redundant Array of Independent Disks (RAID) Technology Overview What is RAID? The basic idea behind RAID is to combine multiple small, inexpensive disk drives into an array to accomplish performance or
Technical white paper HP ProLiant BL660c Gen9 and Microsoft SQL Server 2014 technical brief Scale-up your Microsoft SQL Server environment to new heights Table of contents Executive summary... 2 Introduction...
White Paper Technology Review iscsi- Internet Small Computer System Interface Author: TULSI GANGA COMPLEX, 19-C, VIDHAN SABHA MARG, LUCKNOW 226001 Uttar Pradesh, India March 2004 Copyright 2004 Tata Consultancy
Best practices for Implementing Lotus Domino in a Storage Area Network (SAN) Environment With the implementation of storage area networks (SAN) becoming more of a standard configuration, this paper describes
, pp.265-274 http://dx.doi.org/10.14257/ijdta.2016.9.6.27 Research on Database Remote Disaster Recovery and Backup Technology Based on Multi Point and Multi Hop Guoyong Lin and Fan Huang, Department of
NETAPP WHITE PAPER USING A NETWORK APPLIANCE SAN WITH VMWARE INFRASTRUCTURE 3 TO FACILITATE SERVER AND STORAGE CONSOLIDATION Network Appliance, Inc. March 2007 TABLE OF CONTENTS 1 INTRODUCTION... 3 2 BACKGROUND...
Sistemas Operativos: Input/Output Disks Pedro F. Souto (email@example.com) April 28, 2012 Topics Magnetic Disks RAID Solid State Disks Topics Magnetic Disks RAID Solid State Disks Magnetic Disk Construction
Choosing and Architecting Storage for Your Environment Lucas Nguyen Technical Alliance Manager Mike DiPetrillo Specialist Systems Engineer Agenda VMware Storage Options Fibre Channel NAS iscsi DAS Architecture
Technical White Paper Comparison of Novell, Polyserve, and Microsoft s Clustering Performance J Wolfgang Goerlich Written December 2006 Business Abstract Performance is a priority in highly utilized, asynchronous
VMware ESX Server Multiple Workload Performance on Dell PowerEdge 2850 and PowerEdge 6850 Servers Enterprise Product Group (EPG) Dell White Paper By Todd Muirhead, Dave Jaffe and Scott Stanford July 2005
Enabling Technologies for Distributed Computing Dr. Sanjay P. Ahuja, Ph.D. Fidelity National Financial Distinguished Professor of CIS School of Computing, UNF Multi-core CPUs and Multithreading Technologies
Whitepaper Storwize V7000 replication over TCP/IP with IBM i Version 1 Jana Jamsek Storage Advanced technical skills, Europe Jens Liebetrau IT Specialist, European Storage Competence Center, Mainz Table
Array-based Data Protection: Making Data Recovery Fast, Easy, and Cost-Effective Disk-based data protection is increasingly essential as a supplement to tape-based protection to meet rapid recovery requirements