Extended Distance SAN with MC/ServiceGuard Opens New Disaster Recovery Opportunities



Similar documents
How To Build A Network For Storage Area Network (San)

Designing Disaster Tolerant High Availability Clusters

Cisco Active Network Abstraction Gateway High Availability Solution

SAN Conceptual and Design Basics

IBM TotalStorage SAN Switch F16

Best practices for fully automated disaster recovery of Microsoft SQL Server 2008 using HP Continuous Access EVA with Cluster Extension EVA

Hewlett Packard - NBU partnership : SAN (Storage Area Network) или какво стои зад облаците

An Oracle White Paper October Oracle RAC and Oracle RAC One Node on Extended Distance (Stretched) Clusters

HP Serviceguard Cluster Configuration for HP-UX 11i or Linux Partitioned Systems April 2009

Understanding and Designing Serviceguard Disaster Tolerant Architectures Fifth Edition

Stretching A Wolfpack Cluster Of Servers For Disaster Tolerance. Dick Wilkins Program Manager Hewlett-Packard Co. Redmond, WA dick_wilkins@hp.

Storage Area Network

VERITAS Backup Exec 9.0 for Windows Servers

HP Cisco data replication solutions. Partnering to deliver business continuity and availability

Designing HP SAN Networking Solutions

IP Storage On-The-Road Seminar Series

Optical Networks for Next Generation Disaster Recovery Networking Solutions with WDM Systems Cloud Computing and Security

High Availability Server Clustering Solutions

Continuous Availability for Business-Critical Data

E-Series. NetApp E-Series Storage Systems Mirroring Feature Guide. NetApp, Inc. 495 East Java Drive Sunnyvale, CA U.S.

HP StorageWorks Data Protection Strategy brief

Leveraging Embedded Fiber Optic Infrastructures for Scalable Broadband Services

Virtualized Converged Data Centers & Cloud how these trends are effecting Optical Networks

Storage Area Network Configurations for RA8000/ESA12000 on Windows NT Intel

EXTENDED ORACLE RAC with EMC VPLEX Metro

Disaster Recovery Design Ehab Ashary University of Colorado at Colorado Springs

SAP Solutions on VMware Infrastructure 3: Customer Implementation - Technical Case Study

Storage Networking Management & Administration Workshop

Data Center Fabric Convergence for Cloud Computing (the Debate of Ethernet vs. Fibre Channel is Over)

HP Serviceguard Cluster Configuration for Partitioned Systems

Server Clusters : Geographically Dispersed Clusters For Windows 2000 and Windows Server 2003

HP Serviceguard for Linux Order and Configuration Guide

High Availability and MetroCluster Configuration Guide For 7-Mode

COMPARING STORAGE AREA NETWORKS AND NETWORK ATTACHED STORAGE

Continuous Availability for Business-Critical Data. MetroCluster Technical Overview

Traditionally, a typical SAN topology uses fibre channel switch wiring while a typical NAS topology uses TCP/IP protocol over common networking

SCSI vs. Fibre Channel White Paper

Using High Availability Technologies Lesson 12

Storage Networking Foundations Certification Workshop

Solution Brief Availability and Recovery Options: Microsoft Exchange Solutions on VMware

Network Design. Yiannos Mylonas

Value Proposition for Data Centers

High Availability Databases based on Oracle 10g RAC on Linux

Top of Rack: An Analysis of a Cabling Architecture in the Data Center

Extending SANs Over TCP/IP by Richard Froom & Erum Frahim

Synchronous Data Replication

A SURVEY OF POPULAR CLUSTERING TECHNOLOGIES

Backup Exec 9.1 for Windows Servers. SAN Shared Storage Option

Optimizing Infrastructure Support For Storage Area Networks

SanDisk ION Accelerator High Availability

How To Design A Data Centre

Implementing Storage Concentrator FailOver Clusters

RFP - Equipment for the Replication of Critical Systems at Bank of Mauritius Tower and at Disaster Recovery Site. 06 March 2014

Appendix A Core Concepts in SQL Server High Availability and Replication

Redundancy in enterprise storage networks using dual-domain SAS configurations

A Continuous-Availability Solution for VMware vsphere and NetApp

10G CWDM Conversion Technology

OVERVIEW. CEP Cluster Server is Ideal For: First-time users who want to make applications highly available

Data Compression and Deduplication. LOC Cisco Systems, Inc. All rights reserved.

Remote Copy Technology of ETERNUS6000 and ETERNUS3000 Disk Arrays

Deploying Global Clusters for Site Disaster Recovery via Symantec Storage Foundation on Infortrend Systems

Virtualization, Business Continuation Plan & Disaster Recovery for EMS -By Ramanj Pamidi San Diego Gas & Electric

RFP-MM Enterprise Storage Addendum 1

High Availability and Clustering

Red Hat Cluster Suite

Server Consolidation and Remote Disaster Recovery: The Path to Lower TCO and Higher Reliability

Availability Digest. Stratus Avance Brings Availability to the Edge February 2009

DAS, NAS or SAN: Choosing the Right Storage Technology for Your Organization

Redundancy & the Netnod Internet Exchange Points

Fibre Channel Overview of the Technology. Early History and Fibre Channel Standards Development

Customer Education Services Course Overview

FICON Extended Distance Solution (FEDS)

SummitStack in the Data Center

Westek Technology Snapshot and HA iscsi Replication Suite

Contingency Planning and Disaster Recovery

HP Multimode OM3 LC/LC Optical Cables Overview. Models HP 50 m Multimode OM3 LC/LC Optical Cable

DATA CENTER INTERCONNECT SICHERER UND FLEXIBLER DATENAUSTAUSCH ZWISCHEN RECHENZENTREN COPYRIGHT 2014 ALCATEL-LUCENT. ALL RIGHTS RESERVED.

QuickStart Guide vcenter Server Heartbeat 5.5 Update 2

HIGH AVAILABILITY STRATEGIES

CISCO WDM SERIES OF CWDM PASSIVE DEVICES

Affordable Remote Data Replication

Building Reliable, Scalable AR System Solutions. High-Availability. White Paper

Enterprise Linux Business Continuity Solutions for Critical Applications

High Performance Cloud Connect and DCI Solution at Optimum Cost

Redundant Gigabit Backbone Adds Speed and Reliability to Industrial Networks

SmartOptics. EMC Distance Extension Support. July 31, 2015

IBM Virtualization Engine TS7700 GRID Solutions for Business Continuity

Virtual PortChannels: Building Networks without Spanning Tree Protocol

Windows Server Failover Clustering April 2010

Tomorrow s IP-Based Access Control System on Today s Network Infrastructure using PCN IP-485 Technology

Primary Data Center. Remote Data Center Plans (COOP), Business Continuity (BC), Disaster Recovery (DR), and data

Transcription:

Extended Distance SAN with MC/ServiceGuard Opens New Disaster Recovery Opportunities Joseph Algieri Senior Consultant Network Storage Solutions Hewlett-Packard Company

Overview What is a SAN Extended distance SAN Deploying MC/ServiceGuard clusters on extended distance SAN

What is a SAN? SAN Storage Area Network A specialized network used to connect servers and storage devices together Built using Fibre Channel switches

SAN Example Servers connect to edge switches Storage connects to core switches ISLs (Inter Switch Links) are used to connect switches together Generally, only ISLs between core switches will be > 500 meters

SAN Example Edge switches Server Inter switch links Storage Device *Note: no hubs shown. Hubs would connect to edge switches Core switches

Extended Distance SAN Generally, long distance ISLs are used between core switches and short distance are used between edge switches ISL distances of up to 100km between switches

Extended Distance SAN Distance provided in two manners Long distance GBICs (GigaBit Interface Controller) DWDM (Dense Wavelength Division Multiplexing)

ISL Guidelines and Issues In general,the maximum number of ISL s between any pair of switches should equal one-half the number of ports on the switch (e.g. 1-8 for Brocade 2800; 1-4 for Brocade 2400) It is strongly recommended that a minimum of 2 ISL s exist between any pair of switches for both performance, redundancy, and to limit SAN fabric reconfiguration Not all switches support dynamic load balancing between ISL s. Port-port connections are assigned on a round-robin basis at switch startup and remain fixed regardless of loading. ISL reassignment after a link failure is done automatically by the switch

SAN ISL Considerations Distance between sites SAN Bandwidth requirements Ensure a sufficient number of ISLs to support IO workload

SAN ISL Considerations Number of optical fibers available between sites Fiber Cost Consider DWDM if a large number of fibers are necessary (cost analysis) DWDM can support multiple connections over a single pair of fibers

ISL Distances Short-wave GBIC Distances up to 500 meters Long-wave GBICs Distances up to 10km Long-haul GBICs Distances up to 80km DWDM (Dense Wavelength Division Multiplexing) Distances up to 100km

Distance, Wavelength, and Optical Fiber Specification Optical Fiber Specification 62.5/125 50/125 9/125 Short Wave GBIC 175 meters 500 meters NA Long-wave GBIC NA NA 10 Km Long-haul GBIC NA NA 80 Km DWDM* NA NA 100KM *can accept either long-wave or short-wave input. Consult your DWDM vendor for details

Short-Wave GBIC Up to 500 meters 62.5 or 50 micron optical fiber SAN Cloud

Long-Wave GBIC Up to 10 kilometers (6.6 miles) 9 micron optical fiber SAN Cloud

Long-Haul GBIC Up to 80 kilometers (50 miles) 9 micron optical fiber SAN Cloud

DWDM Up to 100 kilometers (66 miles) 9 micron optical fiber SAN Cloud The FC switch to DWDM connection can be a short-wave or long-wave connection

DWDM Illustrated Fibre Channel 850 nm Wavelength Converter Ch 1 Passive Optical Multiplexer Passive Optical Multiplexer Ch 1 Wavelength Converter 850 nm Fibre Channel Fibre Channel 850 nm Wavelength Converter Ch 2 Ch 2 Wavelength Converter 850 nm Fibre Channel Gigabit Ether 850 nm Wavelength Converter Ch 3 Ch 1 Ch 2 Ch 3 Ch 3 Wavelength Converter Gigabit Ether 850 nm The number of channels available depends on the vendor and model of the DWDM equipment used

Example storage and server attach over extended distance SAN* Up to 100 kilometers (66 miles) SAN Cloud 1 SAN Cloud 2 * This is an HA SAN example

Deploying ServiceGuard Clusters on Extended Distance SANs Why? SAN Design Considerations Solution Design Considerations Other issues

Why Extended ServiceGuard Clusters? Clustering w/ MC/ServiceGuard + Data Replication w/ MirrorDisk/UX + Storage Infrastructure w/ Extended Distance SAN = Low cost entry level DR solution

Why Extended ServiceGuard Clusters? Any storage supported by MC/ServiceGuard can be used in an extended distance cluster solution Use low cost modular storage in DR solutions Leverage existing storage

Extended Distance SAN Design Considerations Redundant SAN connectivity between servers and storage is highly recommended. SAN ISL cables between data centers must follow separate physical paths All legal and supported SAN configurations from HP are supported for extended distance clusters

Extended Distance SAN Examples Dual SANs without PV-Link support Dual SANs with PV-Links support Dual SANs and networks sharing a DWDM site interconnect

Dual SAN clouds without PV-Link support DATA CENTER 1 Data LAN + Heartbeat DATA CENTER 2 Data LAN + Heartbeat Heartbeat Heartbeat SAN CLOUD 2 SAN CLOUD 1 Primary lock disk Secondary lock disk

Failure without PV-Links - interrupted data replication DATA CENTER 1 Data LAN + Heartbeat DATA CENTER 2 Data LAN + Heartbeat Heartbeat Heartbeat SAN CLOUD 2 SAN CLOUD 1 Primary lock disk Secondary lock disk

Dual SAN clouds with PV-links preferred design DATA CENTER 1 Data LAN + Heartbeat DATA CENTER 2 Data LAN + Heartbeat Heartbeat Heartbeat SAN CLOUD 2 Primary lock disk SAN CLOUD 1 Secondary lock disk

Uninterrupted data replication with PV-Links DATA CENTER 1 Data LAN + Heartbeat DATA CENTER 2 Data LAN + Heartbeat Heartbeat Heartbeat SAN CLOUD 2 Primary lock disk SAN CLOUD 1 Secondary lock disk

Extended Cluster with DWDM Site Interconnect DATA CENTER 1 Data LAN + Heartbeat Heartbeat Physical Conduit 1 DATA CENTER 2 Data LAN + Heartbeat Heartbeat DWDM Physical Conduit 2 DWDM Primary lock disk Secondary lock disk

MC/ServiceGuard Cluster in a Single Data Center MC/ServiceGuard s intended use A properly designed solution protects from all single points of failure in the infrastructure. Protection limited to failures within the data center

MC/ServiceGuard Cluster - Single Data Center DATA CENTER Data LAN + Heartbeat Heartbeat SAN

Extended Campus Cluster Solution Extended Campus clusters moves MC/ServiceGuard into the DR space MC/ServiceGuard is the basis for all of HP s cluster DR solution product ( MetroClusters & ContinentalClusters)

Extended Cluster Solution Takes an MC/ServiceGuard cluster and turns it into a low cost DR solution Uses an extended distance SAN to provide a medium for host based replication of the cluster data

Extended Cluster Solution Considerations Many special considerations There are no special MC/ServiceGuard configuration requirements All MC/ServiceGuard configuration rules and limitation must be adhered to for the solution to be supportable

Extended Cluster Solution Considerations There are special cluster architecture requirements There are special solution infrastructure requirements

Extended Cluster Solution Considerations To ensure a complete DR solution, solution monitoring must be addressed Protection from a rolling-disaster is not automatic and can be very tricky

Extended Cluster Solution Considerations A rolling disaster occurs when a data center failure occurs while recovery from an initial failure requiring data resynchronization is in progress Results in a total loss of data requiring a restore from backup

Extended Cluster Solution Considerations Protecting from a rolling disaster is difficult because MirrorDisk/UX is the data replication tool being used MirrorDisk/UX does not inherently contain functionality required to protect from a rolling disaster

Extended Cluster Solution Considerations Other tools must be integrated into the solution to help protect from a rolling disaster This portion of the solution will be custom and will be driven by the customer s availability requirements

Extended Cluster Architecture Requirements Primary data centers must always contain the same number of nodes No lock disk support for clusters with > 4 nodes (MC/ServiceGuard limitation) Single cluster lock disk solutions ARE NOT supported

Extended Cluster Architecture Requirements Solutions requiring more than 4 nodes must either use a three data center solution or a two data center solution with quorum server (lock disks cannot be used) Maximum number of nodes in the cluster is 16 (MC/ServiceGuard limitation)

Extended Cluster Architecture Requirements Data replication is host based via MirrorDisk/UX over extended distance SAN Can put performance pressure on the servers Cluster must be deployed so that the failure of an entire data center can be recovered from by the MC/ServiceGuard cluster quorum protocols

Extended Cluster Infrastructure Requirements Two separate SANs clouds required between data centers Two separate IP networks required between data centers to carry cluster heartbeat Redundant cables between data centers must follow different physical routes to avoid the back-hoe problem

Combining Extended Distance SAN and MC/ServiceGuard No special or custom configurations or licenses are required Solution utilizes standard SAN support Solution utilizes standard MC/ServiceGuard support

Extended Cluster Monitoring Requirements Extended clusters require extensive solution monitoring Event Management Service OpenView Clusterview BMC CA Operator vigilance??

General Considerations Extended distance cluster solutions are very complex to design, monitor, and operate properly Properly designing, deploying, and managing an extended distance cluster is much more difficult to do than it looks Engage HP Consulting to ensure proper solution design, deployment, and management

Cluster Quorum Following the failure of a node or nodes in a cluster, a new cluster consisting of the surviving nodes must be able to form Care must be taken to ensure the cluster protocols can achieve quorum and a new cluster can form following a complete data center failure

Cluster Quorum Three ways to achieve quorum after a failure New cluster contains > ½ of the nodes from the prior cluster New cluster contains ½ of the nodes from the prior cluster and the quorum server votes for it New cluster contains ½ of the nodes from the prior cluster and a cluster lock disk

Cluster Quorum Both three-site clusters and two-site clusters with cluster quorum server will ensure quorum can be achieved after a site failure without the use of cluster lock disks A two site cluster with dual cluster lock disks will be able to achieve quorum after a complete data center failure

Extended Cluster Supported Topologies Two Data Center Design Cluster Quorum Server Dual Cluster Lock Disks Three Data Center Design

Two Data Center Design with Lock Disks DATA CENTER 1 Data LAN + Heartbeat DATA CENTER 2 Data LAN + Heartbeat Heartbeat Heartbeat UP TO 100 KM with DWDM SAN SAN Primary lock disk Secondary lock disk SOFTWARE MIRRORING MirrorDisk/UX

Dual Data Center Design with Quorum Server IP Cloud Quorum Server DATA CENTER 1(major) Data LAN + Heartbeat DATA CENTER 2 (major) Data LAN + Heartbeat Heartbeat Heartbeat UP TO 100 KM with DWDM SAN SAN SOFTWARE MIRRORING MirrorDisk/UX

Three Data Center Design DATA CENTER 3 Heartbeat Heartbeat DATA CENTER 1(major) Data LAN + Heartbeat DATA CENTER 2 (major) Data LAN + Heartbeat Heartbeat Heartbeat UP TO 100 KM with DWDM SAN SAN SOFTWARE MIRRORING MirrorDisk/UX

Dual Cluster Lock Disk Algorithm Flowchart Form a New cluster > 1/2 # nodes < 1/2 = 1/2 No Primary Lock disk Accessible? Yes Wait timeout period Claim primary Lock disk? Primary already claimed TOC Don t form cluster Yes TOC Don t form cluster No Claim Secondary lock? Claim secondary Lock disk* Yes Form a New cluster *The secondary cluster lock will not be claimed here if it is not accessible

Dual Data Center Extended Cluster Issues Dual data center solutions require the use of dual cluster lock disks or a cluster quorum server Use of dual cluster lock disks opens the door to Split-Brain syndrome Single cluster lock disk configurations are NOT supported as a single cluster lock is a SPOF

Split-Brain Syndrome Split-brain syndrome is when two separate viable clusters form from a single cluster Split-Brain can occur after a failure that breaks all network and SAN links between the data centers The back-hoe problem Failure of non redundant DWDM equipment

Split-Brain Syndrome Only a cluster configured with dual cluster lock disks can suffer split-brain Requires a failure of all inter-data center network and SAN links

Split-Brain Syndrome Three-data center clusters and two-data center clusters using cluster quorum server cannot suffer split-brain syndrome Only a slight chance of split-brain occurring in a properly designed and deployed solution

Split-Brain example: Single cluster before failure DATA CENTER 1 DATA CENTER 2 Data LAN + Heartbeat Data LAN + Heartbeat Heartbeat Heartbeat Single 4 node Custer Primary lock disk Secondary lock disk

Split-Brain example: Multiple connection failure DATA CENTER 1 DATA CENTER 2 Data LAN + Heartbeat Data LAN + Heartbeat Heartbeat Heartbeat Primary lock disk Secondary lock disk

Split-Brain example: Two clusters form DATA CENTER 1 DATA CENTER 2 Data LAN + Heartbeat Data LAN + Heartbeat CLUSTER 1 2 node cluster Heartbeat Heartbeat CLUSTER 2 2 node cluster Primary lock disk Secondary lock disk

Other Issues No Oracle Ops support Not supported with Veritas Volume Manager mirroring