White Paper FUJITSU Storage ETERNUS DX Storage Cluster



Similar documents
Enhancements of ETERNUS DX / SF

Technical White Paper Integration of ETERNUS DX Storage Systems in VMware Environments

Microsoft SharePoint 2010 on VMware Availability and Recovery Options. Microsoft SharePoint 2010 on VMware Availability and Recovery Options

Veritas Storage Foundation High Availability for Windows by Symantec

High availability and disaster recovery with Microsoft, Citrix and HP

BUSINESS CONTINUITY AND DISASTER RECOVERY FOR ORACLE 11g

Affordable Remote Data Replication

Mastering Disaster Recovery: Business Continuity and Virtualization Best Practices W H I T E P A P E R

High Availability and Disaster Recovery for Exchange Servers Through a Mailbox Replication Approach

EMC Business Continuity for Microsoft SQL Server Enabled by SQL DB Mirroring Celerra Unified Storage Platforms Using iscsi

Datasheet. FUJITSU Storage ETERNUS SF Storage Cruiser V16.0 ETERNUS SF AdvancedCopy Manager V16.0 ETERNUS SF Express V16.0

Best practices for fully automated disaster recovery of Microsoft SQL Server 2008 using HP Continuous Access EVA with Cluster Extension EVA

Oracle Database Disaster Recovery Using Dell Storage Replication Solutions

EMC VPLEX FAMILY. Continuous Availability and Data Mobility Within and Across Data Centers

E-Series. NetApp E-Series Storage Systems Mirroring Feature Guide. NetApp, Inc. 495 East Java Drive Sunnyvale, CA U.S.

Implementing Storage Concentrator FailOver Clusters

Microsoft Cross-Site Disaster Recovery Solutions

Symantec Storage Foundation High Availability for Windows

EMC VPLEX FAMILY. Continuous Availability and data Mobility Within and Across Data Centers

Disaster Recovery with EonStor DS Series &VMware Site Recovery Manager

EMC VPLEX FAMILY. Transparent information mobility within, across, and between data centers ESSENTIALS A STORAGE PLATFORM FOR THE PRIVATE CLOUD

SQL Server Storage Best Practice Discussion Dell EqualLogic

Cloud Computing Disaster Recovery (DR)

STORAGE CENTER. The Industry s Only SAN with Automated Tiered Storage STORAGE CENTER

Achieving High Availability & Rapid Disaster Recovery in a Microsoft Exchange IP SAN April 2006

Deploying Global Clusters for Site Disaster Recovery via Symantec Storage Foundation on Infortrend Systems

Disaster Recovery Solution Achieved by EXPRESSCLUSTER

Symantec and VMware: Virtualizing Business Critical Applications with Confidence WHITE PAPER

Informix Dynamic Server May Availability Solutions with Informix Dynamic Server 11

Implementing and Managing Windows Server 2008 Clustering

HIGHLY AVAILABLE MULTI-DATA CENTER WINDOWS SERVER SOLUTIONS USING EMC VPLEX METRO AND SANBOLIC MELIO 2010

IBM Virtualization Engine TS7700 GRID Solutions for Business Continuity

Business-centric Storage for small and medium-sized enterprises. How ETERNUS DX powered by Intel Xeon processors improves data management

VMware vcloud Air - Disaster Recovery User's Guide

Module: Business Continuity

Business-centric Storage for small and medium-sized enterprises. How ETERNUS DX powered by Intel Xeon processors improves data management

Remote Copy Technology of ETERNUS6000 and ETERNUS3000 Disk Arrays

HP StorageWorks Data Protection Strategy brief

Disaster Recovery for Oracle Database

The case for cloud-based disaster recovery

The functionality and advantages of a high-availability file server system

Westek Technology Snapshot and HA iscsi Replication Suite

StoneFly SCVM TM for ESXi

EonStor DS remote replication feature guide

Leveraging Virtualization for Disaster Recovery in Your Growing Business

W H I T E P A P E R. Disaster Recovery Virtualization Protecting Production Systems Using VMware Virtual Infrastructure and Double-Take

Using Live Sync to Support Disaster Recovery

EMC CLARiiON Guidelines for VMware Site Recovery Manager with EMC MirrorView and Microsoft Exchange

Availability Guide for Deploying SQL Server on VMware vsphere. August 2009

VMware Business Continuity & Disaster Recovery Solution VMware Inc. All rights reserved

Optimized Storage Solution for Enterprise Scale Hyper-V Deployments

AUTOMATED DISASTER RECOVERY SOLUTION USING AZURE SITE RECOVERY FOR FILE SHARES HOSTED ON STORSIMPLE

Disaster Recovery Feature of Symfoware DBMS

SAN Conceptual and Design Basics

SAP Solutions on VMware Business Continuance Protecting Against Unplanned Downtime

SanDisk ION Accelerator High Availability

Server and Storage Virtualization with IP Storage. David Dale, NetApp

Backup and Redundancy

High Availability with Postgres Plus Advanced Server. An EnterpriseDB White Paper

VMware Site Recovery Manager with EMC RecoverPoint

Abhi Rathinavelu Foster School of Business

Abstract Introduction Overview of Insight Dynamics VSE and Logical Servers... 2

Planning for the Worst SAS Grid Manager and Disaster Recovery

IBM System Storage DS5020 Express

DeltaV Virtualization High Availability and Disaster Recovery

Panorama High Availability

Preface Introduction... 1 High Availability... 2 Users... 4 Other Resources... 5 Conventions... 5

EMC PowerPath Family

Symantec Storage Foundation for Windows

Contents. SnapComms Data Protection Recommendations

Implementing disaster recovery solutions with IBM Storwize V7000 and VMware Site Recovery Manager

A SWOT ANALYSIS ON CISCO HIGH AVAILABILITY VIRTUALIZATION CLUSTERS DISASTER RECOVERY PLAN

Synology High Availability (SHA)

HP Data Protector software Zero Downtime Backup and Instant Recovery

Windows Server 2008 Hyper-V Backup and Replication on EMC CLARiiON Storage. Applied Technology

Constant Replicator: An Introduction

Skelta BPM and High Availability

VERITAS Business Solutions. for DB2

High Availability Cluster for RC18015xs+

Continuous Data Protection for any Point-in-Time Recovery: Product Options for Protecting Virtual Machines or Storage Array LUNs

Server and Storage Virtualization: A Complete Solution A SANRAD White Paper

Implementing the Right High Availability and Disaster Recovery Plan for Your Business Date: August 2010 Author: Mark Peters, Senior Analyst

High Availability and MetroCluster Configuration Guide For 7-Mode

Sanovi DRM for Oracle Database

Vicom Storage Virtualization Engine. Simple, scalable, cost-effective storage virtualization for the enterprise

Virtualization, Business Continuation Plan & Disaster Recovery for EMS -By Ramanj Pamidi San Diego Gas & Electric

High Availability and Disaster Recovery Solutions for Perforce

Zerto Virtual Manager Administration Guide

Total Business Continuity with Cyberoam High Availability

HP Data Protector software Zero Downtime Backup and Instant Recovery. Data sheet

Synology High Availability (SHA)

Synchronous Data Replication

Integrated Application and Data Protection. NEC ExpressCluster White Paper

Stretched Clusters and VMware

Solution Brief Availability and Recovery Options: Microsoft Exchange Solutions on VMware

Transcription:

White Paper FUJITSU Storage ETERNUS DX Storage Cluster Introduction to the concept, technology and setup of the ETERNUS DX transparent failover solution for optimal business continuity Content Management Summary 2 Business Continuity, Sever Virtualization and Storage 3 Basic Architecture 4 Storage Cluster Features and Benefits 5 Storage Cluster in Detail 6 Primary and Secondary Storage 6 Infrastructure 6 7 ETERNUS SF Management Server 7 Failover Mechanism 8 Failover and Failback 9 Storage Cluster in Virtualized Server Environments 10 Conclusion 11 Related Documents 12 Page 1 of 12

Management Summary As data is one of the most important assets in any company, storage systems play a vital role in the IT infrastructures of all enterprises. IT administrators need to ensure that mission-critical data is continuously accessible, ensuring round-the-clock business continuity despite any system or site failures, as well as unplanned or planned outages. Fujitsuʼs ETERNUS DX family, with its Business-centric Storage approach, addresses this issue with a unique transparent application and server failover solution Storage Cluster. Storage Cluster is the high-availability feature of the ETERNUS DX S3 family of storage arrays. Data is synchronously mirrored between two interlinked storage systems. If the primary (active) system should fail, all primary host connections are switched instantly to the secondary (standby) system. This failover is transparent for both servers and applications, and it ensures uninterrupted operations. This advanced feature helps fulfill service levels and delivers predictable operation for businesscritical applications particularly in virtualized server environments. Storage Cluster is integrated and controlled by the ETERNUS SF Management Suite and features simple setup and administration functionalities. Page 2 of 12

Business Continuity, Sever Virtualization and Storage A business continuity plan is built on the 3R fundamental objectives: RLO, RTO and RPO. The Recovery Level Objective (RLO) is defined according to the priority of the business application it defines the granularity with which the recovered data must be enabled i.e., whether it is necessary to recover a whole database, a web application, a site collection, a site, a list or library, or just an item. Based on the RLO, acceptable values are defined for both Recovery Point Objective (RPO) and Recovery Time Objective (RTO All objectives should be set for each business application based on the expected loss for the business with the objective the cost of achieving the objective Recovery flow from disaster, tentative recovery to complete recovery Normal operation Operation down Operation recovered (tentatively) Recovered to normal operation OPERATION LEVEL RPO RTO System reconfiguration Suspended data recovery Latest Backup Disaster Partial Recovery Recovery completed RLO TIME RTO (Recovery Time Objective) Metric for the amount of time it takes to recover from a data loss event, and how long it takes to return to service. RPO (Recovery Point Objective) Metric for the amount of data at risk of being lost during a recovery. RLO (Recovery Level Objective) Granularity level of the recovered data. Page 3 of 12

Basic Architecture The concept of Storage Cluster is based on the deployment of a secondary storage system and a supervising instance the Storage Cluster Controller. mirror mirror synchronous replication Outage replication suspended ETERNUS DX monitoring monitoring ETERNUS DX ETERNUS DX monitoring monitoring ETERNUS DX As long as the primary storage system is running, data is transferred from it to the secondary system via a synchronous replication function. The continuously checks the status of the primary storage. If a failure is detected, it runs the failover logic, and the primary storage information (e.g., LUN ID/WWN) is shifted over to the secondary storage in order to transparently identify the volume via the server I/O. Hence, operations run smoothly and ensure business continuity Due to the restrictions on synchronous mirroring between the storage systems, Storage Cluster can be deployed in building, campus and metro environments. For regional or global business continuity and disaster recovery scenarios, ETERNUS DX can be deployed with asynchronous remote mirroring functionality resulting in increased RPO and RTO, and without automated and transparent failover. Page 4 of 12

Storage Cluster Features and Benefits Storage Cluster technology is the foundation for a non-disruptive business continuity solution. By transferring data in real time and switching storage resources automatically and transparently for servers and applications, data is protected 100 percent and the business process is safe from any interruption. on Regional-level disaster Planned power outage, building or floor-level disaster, fire Solution Storage Cluster off Flood, earthquake, terrorism Wide-level disaster Solution Disaster Recovery Solution Data consistency is managed by the storage systems without the need for specific volume management software. This frees up server resources while the secondary storage is securely synced with consistent data sets. Data is replicated along with the automated storage tiering layout and auto QoS settings. Failover to the standby resources is triggered automatically there is no need for any human intervention. For manual failover, failover test or restoration after failover (failback), the failover operation can also be triggered from the user interface it is simple and secure. Seamless server affinity no change in access path, drive letter or mount point information after the failover. Volumes continue to be accessed transparently from the business application. No workload for volume reallocation or remount is needed or generated. After the recovery of the failed system, Storage Cluster automatically executes the failback to the initial state. System operation is always stable and secure. Limitations and hints Maximum distance between the sites is limited by the roundtrip time for the synchronous replication which must not exceed 10 milliseconds. To avoid misbehavior in case of complete site outages it is recommended to deploy the and the ETERNUS SF Management Server on an independent third site. A redundant multipath configuration with at least one path to each storage controller is recommended. can be run as virtual machine without needing dedicated hardware. Page 5 of 12

Storage Cluster in Detail Primary and Secondary Storage The primary and secondary ETERNUS DX S3 storage array should preferably be located in different fire compartments even better in different buildings or in metropolitan dispersion. Storage Cluster is set up using Transparent Failover Volumes (TFOV) which are part of a special copy group the TFO group. Layout of the TFOVs and TFO groups are identical in the primary and secondary storage, even including the configuration settings for automated storage tiering, snapshots, etc. TFOVs are synchronously replicated from the primary to the secondary array. I/O Mirror data Primary Storage Remote equivalent copy (synchronous) Secondary Storage Management LAN ETERNUS SF Management Server Infrastructure Primary and secondary storage have paired channel adapter (CA) ports. Paired means that the Fibre Channel ports on both sites have the same identity regarding WWN/WWPN. Under normal conditions the CA port on the primary site is in link up status, and the port on the secondary site is in link down status, so all server I/O is processed to the primary storage. The CA port states, as well as the REC state, are controlled by the Storage Cluster feature. I/O CA port link up CA port link down Mirror data Primary Storage Remote equivalent copy (synchronous) Secondary Storage Management LAN ETERNUS SF Management Server Page 6 of 12

The is a server or virtual machine including an agent to connect with the ETERNUS SF management software. It monitors the health of primary and secondary storage in order to detect outages of the active system. I/O The triggers the automatic failover in this scenario only and is not involved in cases of administrator-triggered manual failover or in cases of automated failover caused by RAID failures. Primary Storage Remote equivalent copy (synchronous) Mirror data Secondary Storage Management LAN ETERNUS SF Management Server ETERNUS SF Management Server ETERNUS SF management is the prerequisite for setting up the Storage Cluster configuration with regard to TFO groups, TFOVs, copy groups and REC pairs. It also executes the failover and failback operations, either triggered by the in cases of automatic failover, or manually by an operator. It also executes the automated failover in cases of RAID failures on the primary array. I/O Mirror data ETERNUS SF and the can be installed on the same physical or virtual server. Primary Storage Remote equivalent copy (synchronous) Secondary Storage Management LAN ETERNUS SF Management Server Page 7 of 12

Failover Mechanism Storage Cluster reroutes I/O access from one array to the other as seen in this figure. If an outage occurs the failover sequence is executed as follows: 1 The server sends I/O requests to the primary storage. 2 The primary CA port does not respond, the detects the primary ETERNUS is unreachable and reports it to ETERNUS SF. 3 The server retries the I/O after a preset time-out 4 ETERNUS SF suspends the remote mirroring (REC) session, the replicated data becomes the actual business data. 5 The CA port on the secondary array is activated (link up) with the same identity (WWN/WWPN) as the primary CA port. 6 The server I/O is processed by the secondary storage before the retry time-out is exceeded. The application continues running without any restrictions. I/O Retry /error Primary Storage CA port link up Mirror data Suspended Secondary Storage Management LAN ETERNUS SF Management Server Such an automatic failover is typically completed within less than three seconds, which is sufficient for most applications to keep on running smoothly. Conditions for Automatic Failover Link status Array Status CA Outage Failover I/O CA (a) (b) (c) Primary Secondary 1 OK OK OK Normal Normal 2 X OK OK 3 OK X OK Normal Normal 4 OK OK X Primary Storage Secondary Storage 5 X X OK Down Normal 6 X OK X Normal Down ESC Agent LAN 7 OK X X Normal Normal 8 X X X Down Down Automatic failover is only triggered if both the primary storage health supervision from (Link b) and the remote replication (Link a) are down (5). In all other cases a failover is either not necessary or not possible (8) Page 8 of 12

Failover and Failback Storage Cluster can handle different failover types and scenarios: Auto failover: Failover is triggered automatically when the primary storage becomes unreachable or a RAID group becomes unavailable. This mode ensures business continuity in cases of unpredictable failures or a disaster at the primary site. Manual failover: Failover is triggered manually from the ETERNUS SF user interface by stopping access on the primary storage and activating the secondary storage. This mode ensures business continuity when planned downtime is required on the primary site, e.g., for maintenance, disruptive upgrades or planned power shutdowns. It can also be used for general testing of the failover mechanism. Force failover: Failover is triggered manually from the ETERNUS SF user interface by activating the secondary storage regardless of the status of the primary storage. This mode ensures business continuity in cases of emergency when the primary storage is unreachable, and for any reason the auto failover cannot be executed. Auto failback: The failover back from the secondary site to the primary site is triggered automatically under these conditions: The confirms that all systems are operative, the REC session is established, and the business data and mirror data are consistent. Manual failback: The failover back from the secondary site to the primary site is triggered manually via the ETERNUS SF user interface. This mode resets normal operation manually. The conditions of auto failback apply as well. Page 9 of 12

Storage Cluster in Virtualized Server Environments In combination with ware s HA/FT functionalities, Storage Cluster helps overcome even complete site outages in virtualized multisite server configurations. It provides instantaneous, non-disruptive failover in the event of server or site failures, delivering protection from even the slightest lapse, disruption or data loss. Storage Cluster configuration is based on remotely replicated TFOVs which can be freely configured and paired. Thus on one site both primary and secondary TFO groups and TFOVs can be configured connected via linked-up (active) or passive (linked-down) CA ports, respectively. Site 1 Site 2 Operation Server Operation Server 1 2 3 4 1 2 3 4 A Fibre Channel C B D 1 2 3 4 Mirroring 3 4 1 2 ETERNUS A ETERNUS B Site 3 LAN ETERNUS SF Storage Cruiser Management Server (Windows Server) Under normal conditions, 1 and 2 at site 1 and 1 and 2 at site 2 are connected via Fibre Channel to the respective ETERNUS DX arrays located at the same site this is called the active (primary) site. All data is synchronously replicated to the standby (secondary) site, while in this example ETERNUS DX 2 is secondary for 1 and 2 and ETERNUS DX 1 is secondary for 3 and 4, respectively. Both Fibre Channel ports on both sites have the same identity regarding WWPN. As the Fibre Channel port on the primary site is in link-up status and the port on the secondary site is in link-down status, all server inputs/outputs are processed to the primary storage. Page 10 of 12

Conclusion ETERNUS DX Storage Cluster provides zero-downtime, zero-data-loss continuous availability for any application, without the cost or complexity of traditional solutions like additional storage virtualization appliances or instances. Eliminates expensive downtime or data loss due to storage or site failures. Copes with both unplanned and planned downtime. Provides continuous service to any application, regardless of operating system. Provides easy setup through an intuitive administrative interface. Advantages at a glance Feature Transparent Failover Advantages Ensures business continuity for any kind of outages Automated control Manual control Symmetrical configuration Ease of use Flexible infrastructure options Protection against hardware, network or environmental failures to prevent unplanned downtime Administrator-initiated failover supports business continuity in cases of planned downtime like disruptive upgrades End-to-end continuous data availability for virtualized server environments Simple system administration with wizard-based setup of transparent failover volumes and relationships Can be deployed for in-house fire compartments, campus-wide or metro-wide infrastructures Page 11 of 12

Related documents: Global ETERNUS DX Storage Cluster Website Flyer: ETERNUS Storage Cluster Whitepaper: ETERNUS DX Feature Set Whitepaper: ETERNUS SF Published by Fujitsu Limited Copyright 2015 Fujitsu Limited All rights reserved, including intellectual property rights. Technical data subject to modifications and delivery subject to availability. Any liability that the data and illustrations are complete, actual or correct is excluded. Designations may be trademarks and/or copyrights of the respective manufacturer, the use of which by third parties for their own purposes may infringe the rights of such owner. For further information see Page 12 of 12