HCS Redundancy and High Availability
|
|
- Mervin Doyle
- 8 years ago
- Views:
Transcription
1 Data Center Infrastructure, page 1 Data Center Interconnect, page 2 Geo-Redundancy, page 2 Geo-Redundancy for Cisco Unified Communications Domain Manager, page 7 Data Center Redundancy, page 15 Application Redundancy, page 16 Unified Communications, page 16 Cisco HCS Management, page 18 Data Center Infrastructure Cisco HCS leverages the Cisco Virtualized MultiServices Data Center (VMDC) for the data center architecture design. The VMDC system is the Cisco reference architecture for Infrastructure as a Service (IaaS) cloud deployments and utilizes a hierarchical network design for high availability and scalability. The hierarchical or layered DC design uses redundant switches at each layer of the network topology for device-level failover that creates a highly available transport between end nodes using the network. DC networks often require additional services beyond basic packet forwarding, such as Server Load Balancing (SLB), firewall, and intrusion prevention. These services might be introduced as modules populating a slot of one of the switching nodes in the network or as standalone appliance devices. Each service approach also supports the deployment of redundant hardware to preserve high availability standards set by the network topology. This layered approach is the basic foundation of the VMDC design to provide scalability, performance, flexibility, resiliency, and service assurance. VLANs and Virtual Routing and Forwarding (VRF) instances are used to provide customer isolation within the data center architecture, and routing protocols within the VRFs are utilized to interconnect the different networking and service devices. Leveraging VMDC, the Cisco HCS system is built around the Cisco Unified Computing System (Cisco UCS), Nexus 1000V, Nexus 5000 and Nexus 7000 switches, multilayer Director Switch (MDS), Aggregation Services Router (ASR) 9000, ASR 1000, Adaptive Security Appliance (ASA) 5585-X or Adaptive Security Appliance Services Module (ASASM), Nexus 1000V Virtual Security Gateway (VSG), VMware vsphere, and supports several shared storage options such as EMC and NetApp. 1
2 Data Center Interconnect For more information, see Cisco Virtualized MultiTenant Data Center, Version 2.2 Design Guide available at and UC Virtualization Storage System Design Requirements available at Data Center Interconnect The term Data Center Interconnect (DCI) is relevant in all scenarios where you require different levels of connectivity between two or more data center locations to provide flexibility for deploying applications and resiliency schemes. Geo redundancy or high availability of a given data center requires infrastructure and applications running in multiple data centers to be available without any major down time. Several options are available to interconnect data centers in both active-active and active-standby deployment scenarios. There is a need to extend applications (IP/LAN extension) and datastores (SAN extension) across the data centers. In typical active-standby data center deployments, where the applications are brought up on the standby data center only after the primary data center has an outage, there is no need for LAN extension between data centers. You only need to extend the SAN for data store replication. In active-active data center deployments, where the applications are extended between data centers, there is the need for LAN and SAN extension between data centers. Stretching the network space across two or more data centers may be needed by clustering applications, which require Layer 2 connectivity between cluster nodes. Putting these nodes in separate data centers can help build resilience into the cloud. This also provides the flexibility of being able to shift compute resources around geographically as needs dictate. Depending on application types, there are several possible data center interconnect methods: Layer 2 Extension: Provide a single Layer 2 domain across data centers. The data center applications are often legacy or use embedded IP addressing that drives Layer 2 expansion across data centers. Workload mobility and VMotion of virtual machines across data centers also requires Layer 2 extension. Layer 3 Extension (Recommended method): Provide routed connectivity between data centers used for segmentation and virtualization and redundant applications. This may be Layer 3 VPN-based connectivity or plain IP connectivity. SAN Extension: Storing and replicating data for business continuity, disaster recovery, and regulatory compliance. However, this presents different types of challenges and considerations because of the requirements in terms of distance and latency and the fact that Fiber Channel cannot natively be transported over an IP network. For more information, see Cisco Validated Designs using Data Center Interconnect: en/us/netsol/ns749/networking_solutions_sub_program_home.html. Geo-Redundancy Level 3 geo-redundancy (with no SAN mirroring) is offered in Cisco HCS deployments. As shown in the following figure, in normal operation Level 3 geo-redundancy (with no SAN mirroring) includes the following: Two data centers with no real-time SAN mirroring Data Center A has its management layer and publishers and subscribers backed up to a third-party site through SFTP daily Load is shared between subscriber 1 and subscriber 2 2
3 Geo-Redundancy Lower cost capital expenditure than with SAN mirroring Figure 1: Normal Layer 3 Geo-Redundant Operation - No SAN Mirroring When an outage occurs for a customer using Level 3 geo-redundancy, the following occurs: There is no call processing outage if the surviving data center can take the load. VM rebuild takes at least 4 hours. Restoring Unified Communications applications and resync takes at least another 4 hours. Total recovery time per Unified Communications cluster is 8+ hours. Outage also depends on time to schedule restore in addition to the restoration time. Not all VM restoration tasks can run in parallel due to input/output operations per second (IOPS). Total recovery time for HCS Management is approximately 8 hours. Outage also depends on time to schedule restore in addition to the restoration time. Not all VM restoration tasks can run in parallel due to input/output operations per second (IOPS). MACD outage until recovered is 8+ to 16+ hours (approximately 16+ hours for the first customer and approximately 8 hours for each subsequent customer based on all manual efforts). MACD outage is repeated after the affected data center is back in service. Cisco Unified Communications components and HCS management components must maintain same IP address during disaster recovery. Change of IP address during recovery is not supported, nor are there plans to support this in the future. For more information, see Recover from Level 3 Geo-Redundant Design Outage, on page 4 3
4 Recover from Level 3 Geo-Redundant Design Outage Operational expenditures for Level 3 geo-redundancy with no SAN mirroring is as follows: Only periodic data backup is created. All disaster recovery is performed manually, including routing change. Procedure must be repeated after the affected data center is back in service. Recover from Level 3 Geo-Redundant Design Outage This procedure outlines the high-level steps that are performed to recover from an outage when using the Level 3 geo-redundant design (with no SAN mirroring). Procedure Step 1 Data Center A down If Data Center A goes down, all HCS Management, publishers, and subscribers are down in Data Center A. There is no loss of call processing because subscriber 2 in Data Center B takes over. A MACD outage occurs. Figure 2: Layer 3 Geo-redundant Design - Data Center A Down Step 2 Implement manual network changes to allow subnet A in data center B. Step 3 Build virtual machines in Data Center B and restore from SFTP Build the virtual machines in the following order: HCS Management first, then publishers, then subscribers. 4
5 Recover from Level 3 Geo-Redundant Design Outage Note The HCS Management, publishers and subscribers on Subnet B must use the same IP address as Subnet A. Figure 3: Layer 3 Geo-Redundant Design - Build VMs in DC B and Restore from SFTP Estimated time to complete this step: For HCS Management and other applications Time to install VMs: 1 hour each Full restore from SFTP: 2 to 4 hours for each application For Unified Communications Manager only Time to install VMs: 1 hour each Basic restore from SFTP backup: 10 minutes each Full restore from subscriber 2: 2 to 4 hours for each cluster Step 4 Data Center A comes back online When Data Center A comes back online, the VMs in Data Center B must be stopped. Start to recopy VMs from Data Center B to Data Center A (30 to 60 minutes for each VM). Data Center B VMs must be stopped before restarting Data Center A. Start VMs in Data Center A (15 minutes for each VM). Note Manual network changes should be reverted. 5
6 Recover from Level 3 Geo-Redundant Design Outage Figure 4: Layer 3 Geo-Redundant Design - Data Center A Comes Back Online Figure 5: Layer 3 Geo-Redundant Design: Cisco HCS Management, Pub, Sub 1 Stopped in Data Center B 6
7 Geo-Redundancy for Cisco Unified Communications Domain Manager Note It may take many hours for Data Center B to get to MACD restored state. Therefore an alternative recovery option, if Data Center A will be offline for only a few hours (less than a total catastrophic destruction of Data Center A,) is to do nothing and simply restore Data Center A when it comes back online. Geo-Redundancy for Cisco Unified Communications Domain Manager Note This process applies to all versions of Cisco Unified Communications Domain Manager. Multinode Cluster with Unified Nodes The recommended multinode deployment using Unified nodes has the following characteristics: Four Unified nodes - each node combining Application and Database roles - are clustered and split over two geographically disparate locations. 7
8 Multinode Cluster with Unified Nodes Two Web Proxy nodes to provide High Availability that ensure an Application role failure is gracefully handled. It may be omitted if an external load balancer is available. Web Proxy and Unified nodes can be contained in separate firewalled networks. Database synchronization takes places between all Database roles, thus offering Disaster Recovery, and High Availability. All nodes in the cluster are active. Primary and fall-back Secondary Database servers can be configured manually. Note For a geo-redundant multinode deployment, please contact your Cisco support representative. Note To install a geo-redundant multinode solution, refer to thecisco Hosted Collaboration Solution, Release 10.6(1) Install Guide. Figure 6: A Graphical Representation of a Recommended Multinode Cluster 8
9 Multinode Cluster Hardware Specifications Note WebProxies can be configured to load balance across two or four Unified nodes. Multinode Cluster Hardware Specifications Table 1: Virtual Machine Requirements Migration from the Stand-Alone Cisco Unified Communications Domain Manager 10.6(x) system to a 4 Unified Node/2 WebProxy Node cluster is supported. The resource requirements for each configuration are provided in Table 1. Node Type Quantity VM Level Memory 1 CPU Disk Network Unified >= 4 VMware 5.1 or later 8 GB 2 GHz 370 GB partitioned as follows: 20 GB for OS 1 Gbit/s minimum 50 GB for application 50 GB for compressed backups 250 GB for database WebProxy >=2 VMware 5.1 or later 4 GB 2 GHz 120 GB partitioned as follows: 20 GB for OS 1 Gbit/s minimum 50 GB for application 1 For the 4 Unified Node/2 WebProxy node deployments where the total number of end users is expected to exceed 100,000, it is recommended that 16 GB of RAM be provisioned for each Unified node. The recommended memory allocation for the WebProxy nodes does not change. The size of the database storage partition supports the maximum deployment size for the release. Further increase in the size of the partition is not required when new customers are on-boarded. To set up the disk requirements, the disk should be set up on the VMware GUI Resources tab where a disk can be created. This task should be done after the OVA import but prior to the boot of the system. Alternative Geo-redundant Cluster Hardware Specification Virtual machine requirements are specified in the table below. Purpose Quantity VM Memory CPU Disk Network Application nodes >= 2 >= VMWare 4.1 8GB 2GHz 150GB 1Gbit/s recommended 9
10 Clustering Considerations Purpose Quantity VM Memory CPU Disk Network WebProxy nodes >= 2 >= VMWare 4.1 4GB 2GHz 150GB 1Gbit/s recommended Database nodes >= 3 >= VMWare 4.1 4GB 2GHz 160GB + 100GB dedicated backup 1Gbit/s recommended Clustering Considerations The cluster contains multiple nodes that can be contained in separate firewalled networks. Open network ports on firewalls to allow internode communication. Port requirements are described in the Platform Guide. All communication between nodes is encrypted. Node type WebProxy Unified Ports 22 (SSH and SFTP), 80 (HTTP), 161 and 162 (SNMP), 443 and 8443 (HTTPS) 22 (SSH and SFTP), 80 (HTTP), 161 and 162 (SNMP), 443 and 8443 (HTTPS), 27019, 27020, and (database) 22/SSH is used for remote administration. 80 and 443 are used for the web server. 161 and 162 are used for sending and receiving SNMP is used for intercluster communication , 27020, and are used for database queries and replication. Cisco Unified Communications Domain Manager 10.6(x) Redundancy and Disaster Recovery High Availability (HA) is an approach to IT system design and configuration that ensures Cisco Unified Communications Domain Manager 10.6(x) is operational and accessible during a specified timeframe. High Availability is achieved using redundant hardware and resources. Cisco recommends the use of two physical data centers, where the primary site contains three VMs and the secondary site contains three VMs. If there is a failure, an automatic failover to the secondary DR (Disaster Recovery) site takes place. Web server proxy nodes perform load-balancing between application roles, so that load is distributed. During provisioning, the web server proxy is provided with all the IP addresses of the application nodes. The web server software then does load balancing among these nodes, according to its configuration. If a node fails to 10
11 Cluster ure Scenarios respond in a set time, the proxy sends the transaction to another node. If an Application role is lost, the WebProxy transparently bypasses the faulty Application role. The proxy web server that is configured to be located in the primary site normally load balances to the two unified nodes in the primary site. The proxy web server falls back to the two nodes in the Disaster Recovery site if the nodes in the primary site are down. The web proxy nodes in the secondary site defaults load balancing to the two unified nodes configured for the secondary site. Data is replicated between unified node roles, and role failure is recoverable. Data replication is done using the database replication facility. Automatic failover between unified node roles occurs while there is greater than 50% unified node role availability. Once there is insufficient role availability, the system must be manually reprovisioned. HA can be increased by adding nodes to the cluster. Application performance and availability can be increased by adding more application role servers. Backups can be scheduled to run automatically across the cluster. Backups include application data, configuration, and software. Backups can take place to both local disk and remote network location. Every node upgrade includes a snapshot backup which allows any upgrade to be rolled back. Cluster ure Scenarios The status of the cluster can be displayed from the command line on any node using the command: cluster status The system can automatically signal and/or SNMP events in the event that a node is found to be down. Refer to the diagrams in the section on deployments. Loss of an Application role The Web Proxy will keep directing traffic to alternate application role servers. There is no downtime. Loss of a Web Proxy Communication via the lost Web Proxy will fail, unless some another load balancing infrastructure is in place (DNS, external load balancer, VIP technology). The node can be installed as a HA pair so that the VMware infrastructure will restore the node if it fails. Downtime takes place while updating the DNS entry or returning the Web Proxy to service. For continued service, traffic can be directed to an alternate Web Proxy or directly to an Application node if available. Traffic can be directed manually (i.e., network elements must be configured to forward traffic to the alternate Web Proxy). In the sections that follow, it is assumed that the partner has assigned the recommended database (DB) weights as follows: Primary Unified Node (UN1) database weight = 40 Secondary Unified Node (UN2) database weight = 30 Secondary Unified Node (UN3) database weight = 20 Secondary Unified Node (UN4 ) database weight = 10 The Unified CDM system automatically assigns Voting Members to the Unified Nodes based on their DB weights noted previously: Primary Unified Node (UN1) Voting Members = 2 11
12 Cluster ure Scenarios Table 2: Single Unified Node ure Scenarios Secondary Unified Node (UN2) Voting Members = 2 Secondary Unified Node (UN3) Voting Members = 2 Secondary Unified Node (UN4 ) Voting Members = 1 The relative size of the database weights determines the hierarchy of the nodes in a failure scenario. The Unified Node (UN) with the highest database weight is the Primary. The number of Voting Members is used to establish a quorum, and is the basis for the behavior of the system in a multi-node failure scenario. There are a total of 7 Voting Members ( ) in the cluster by default. The subsequent tables illustrate the behavior of the system in various failure scenarios. In all cases where a UN has failed, it is assumed that the entire node (and therefore the database function referenced earlier in this section) is out of service. Any in-flight transactions on nodes that failed will be lost. UN1 State [Voting Members = 2] UN2 State [Voting Members = 2] UN3 State [Voting Members = 2] UN4 State [Voting Members = 1] System Status under Scenario System is functioning normally. System continues functioning normally. System continues functioning normally. System continues functioning normally. Some downtime occurs. System automatically fails over to UN2. Table 3: Multiple Unified Node ure Scenarios UN1 State [Voting Members = 2] UN2 State [Voting Members = 2] UN3 State [Voting Members = 2] UN4 State [Voting Members = 1] Voting Members in Service System Status under Scenario 7 System is functioning normally. 3 Manual Intervention is required to restore service. 12
13 Cluster ure Scenarios UN1 State [Voting Members = 2] UN2 State [Voting Members = 2] UN3 State [Voting Members = 2] UN4 State [Voting Members = 1] Voting Members in Service System Status under Scenario 4 System continues functioning normally. 3 Manual Intervention is required to restore service. 4 System continues functioning normally. In all of the scenarios, whenever 50% or more Voting Members are in service, the system will remain in service. Manual steps will be required to restore the cluster. Loss of a Database role If the primary Database service is lost, the system will automatically revert to the secondary Database. The primary and secondary database nodes can be configured via the command-line interface using database weight <ip> <weight>. For example, the primary can be configured with a weight of 50, and the secondary with a weight of 20. If both the primary and the secondary Database servers are lost, the remaining Database servers will vote to elect a new primary Database server. There is downtime (usually no more than a few seconds) during election and failover, with a possible loss of data in transit (a single transaction). The GUI web-frontend transaction status can be queried to determine if any transactions failed. The downtime for a Primary to Secondary failover is significantly less and the risk of data loss likewise reduced. A full election (with higher downtime and risk) is therefore limited only to cases of severe outages where it is unavoidable. Although any values can be used, for 4 database nodes the weights: 4, 3, 2, 1 is recommended. Loss of a site Unified and Database nodes have database roles. The status of the roles can be displayed using cluster status. If 50% or more of the database roles are down, then there is insufficient availability for the cluster to function as is. Either additional role servers must be added, or the nodes with down roles must be removed from the cluster and the cluster needs to be reprovisioned. If there is insufficient (less than 50% means the system is down) Database role availability, manual intervention is required to reprovision the system downtime is dependent on the size of the cluster. Refer to the Operations Guide for details on DR over. Database role availability can be increased by adding Database roles, providing greater probability of automatic failover. To delete a failed node and replace it with a new one if database primary is for example lost: The node can be deleted using cluster del <ip>. Additional nodes can be deployed and added to the cluster with cluster add <ip>. The database weights can be adjusted using database weight <ip> <weight>. Finally, the cluster can be reprovisioned with cluster provision. The console output below shows examples of these commands. 13
14 Cluster ure Scenarios The cluster status: cluster status Data Centre: jhb application : cpt-bld2-cluster-04[ ] cpt-bld2-cluster-03[ ] webproxy : cpt-bld2-cluster-06[ ] cpt-bld2-cluster-04[ ] cpt-bld2-cluster-03[ ] database : cpt-bld2-cluster-04[ ] cpt-bld2-cluster-03[ ] Data Centre: cpt application : cpt-bld2-cluster-02[ ] cpt-bld2-cluster-01[ ] (services down) webproxy : cpt-bld2-cluster-05[ ] cpt-bld2-cluster-02[ ] cpt-bld2-cluster-01[ ] (services down) database : cpt-bld2-cluster-02[ ] cpt-bld2-cluster-01[ ] (services down) Deleting a node: platform@cpt-bld2-cluster-01:~$ cluster del You are about to delete a host from the cluster. Do you wish to continue? y Cluster successfully deleted node Please run 'cluster provision' to reprovision the services in the cluster Please note that the remote host may still be part of the database clustering and should either be shut down or reprovisioned as a single node BEFORE this cluster is reprovisioned You have new mail in /var/mail/platform Adding a node: platform@cpt-bld2-cluster-01:~$ cluster add Cluster successfully invited node Please run 'cluster provision' to provision the services in the cluster Database weights: listing and adding platform@cpt-bld2-cluster-01:~$ database weight list : weight: : weight: : weight: : weight: 1 platform@cpt-bld2-cluster-01:~$ database weight : weight: : weight: : weight: : weight: 1 14
15 High Availability ure Protection High Availability ure Protection The table below provides information on Cisco Unified Communications Domain Manager 10.6(1) HA failure protection. Deployment Model Downtime Manual Recovery Steps Comments Single Unified Node (Non-Primary) No See the Cisco Unified Communications Domain Manager, Release 10.6(2) Maintain and Operate Guide Application failure vs. VM failure (Detection Mechanism varies) Unified Node Service ure : Mongo DB Yes See the Cisco Unified Communications Domain Manager, Release 10.6(2) Maintain and Operate Guide Automatic failover and automatic service restart Unified Node Service ure : VOSS Yes See the Cisco Unified Communications Domain Manager, Release 10.6(2) Maintain and Operate Guide Automatic service restart Unified Node Service ure : Nginx No N/A Automatic failover by proxies and automatic service restart Single Unified Node (Primary) Yes See the Cisco Unified Communications Domain Manager, Release 10.6(2) Maintain and Operate Guide Possible loss of transactions in flight. See transaction logs for lost transaction information. Single Proxy Node Yes See the Cisco Unified Communications Domain Manager, Release 10.6(2) Maintain and Operate Guide Update DNS to use alternate proxy. CUCDM system remains operational. Data Center Redundancy When application data is duplicated at multiple data centers, clients go to the available data center in the event of catastrophic failure at one site. You can deploy redundant data centers in several ways. Active-Standby Data Centers Deploying more than one data center provides redundancy through site-to-site recovery mechanisms. Site-to-site recovery is the ability to recover from a site failure by ensuring failover to a secondary or backup site. Active-standby data center deployments can include one active and one standby data center (1:1); or one standby data center for N active data centers (1:N). 15
16 Active-Active Data Centers You can deploy the standby data center as a hot or warm standby. In a hot standby deployment, the standby data center has some applications running actively and has some traffic processing responsibilities. Resources are not kept idle in the standby data center, and this improves overall application scalability and equipment utilization. The standby data center is active for some applications or services hosted within this data center but it has spare capacity to be in standby mode for applications and services hosted in another data center. In a warm standby deployment, applications at the standby data center are activated, and the traffic is processed by the standby data center only when the primary data center goes out of service. Active-Active Data Centers You can also use multiple data centers concurrently to improve performance and scalability. You can deploy two or more data centers in active-active mode to provide scalable and highly available cloud services. The active data centers are in hot standby mode for each other. Deploying applications at multiple active data centers is analogous to building a global server farm, which increases the number of requests and number of clients that can be handled. After content is distributed to multiple data centers, requests for distributed content must be managed. You can manage the load by routing user requests for content to the appropriate data center. The selection of the appropriate data center can be based on server and application availability, content availability, network distance (proximity) from the client to the data center, and other parameters. Application Redundancy Within Cisco HCS, we recommend that all Unified Communications (UC) applications be deployed with application level redundancy to protect against any complete failure of a Unified Communications application. In addition, we recommend that you deploy Unified Communications applications in a geo-redundant fashion if at all possible, using well known techniques called clustering over the WAN. This connectivity requires Layer 3 connectivity between the Unified Communications applications with specific round trip time delay, which are different for each Unified Communications application. We recommend that you have a 40 millisecond maximum delay to support all the Unified Communications application redundancy over WAN. If applications are only deployed within a single data center, we recommend that you deploy the virtualized applications on a different Unified Communications chassis if possible or at least upon different VMware ESXi hosts (server blades). Unified Communications High Availability for Deployment Models Unified Communications services offer many capabilities aimed at achieving high availability. They may be implemented in various ways, such as the following. over Redundancy For services that are considered essential, you should deploy redundant elements so that no single point of failure is present in the design. The redundancy between the two (or more) elements is automated. For example, the clustering technology used in Cisco Unified Communications Manager allows for up to three servers to provide backup for each other. This type of redundancy may cross technological boundaries. For example, a phone may have as its first three preferred call control agents, three separate Unified Communications Manager 16
17 Cisco Unified Communications Manager Cluster Overview servers belonging to the same call processing cluster. As a fourth choice, you can configure the phone to rely on a Cisco IOS router for call processing services. Redundant Links In some instances, it is advantageous to deploy redundant IP links, such as IP WAN links, to guard against the failure of a single WAN link. Geographical Diversity Some products support the distribution of redundant service nodes across WAN links so that, if an entire site is off-line (such as would be the case during an extended power outage exceeding the capabilities of provisioned UPS and generator backup systems), another site in a different location can ensure business continuance. Cisco Unified Communications Manager Cluster Overview You may deploy a single Unified Communications Manager cluster across multiple sites that are connected by an IP WAN with QoS features enabled. Clustering over the WAN can support two types of deployments: Local over Deployment Model: Local failover requires that you place the Unified Communications Manager subscriber and backup servers at the same site, with no WAN between them. This type of deployment is ideal for two to four sites with Unified Communications Manager. Remote over Deployment Model: Remote failover allows you to deploy primary and backup call processing servers split across the WAN. Using this type of deployment, you may have multiple sites with Unified Communications Manager subscribers being backed up by Unified Communications Manager subscribers at another site. You can also use a combination of the two deployment models to satisfy specific site requirements. For example, two main sites may each have primary and backup subscribers, with another two sites containing only a primary server each and utilizing either shared backups or dedicated backups at the two main sites. Some of the key advantages of clustering over the WAN are: Single point of administration for users for all sites within the cluster Feature transparency Shared line appearances Extension mobility within the cluster Unified dial plan These features make this solution ideal as a disaster recovery plan for business continuance sites or as a single solution for multiple small or medium sites. For more information, see Cisco Unified Communications System SRNDs, available at go/ucsrnd. Cisco Unity Connection Cluster Overview Cisco Unity Connection supports a Connection cluster configuration of two Connection servers to provide high availability and redundancy. The Connection servers handle calls, HTTP, and IMAP requests. If only 17
18 Cisco HCS for Contact Center High Availability one server in the Connection cluster is functioning, the remaining server preserves the system functionality by handling all calls, HTTP requests, and IMAP requests for the Connection cluster. Each server in the Connection cluster must have enough voice messaging ports to handle all calls for the Connection cluster. The first server installed is the publisher server for the Connection cluster; the second server installed is the subscriber server. These terms are used to define the database relationship during installation. The separation of roles is consistent with the Cisco Unified Communications Manager cluster schema in which there is always one publisher server and multiple subscriber servers. (Connection runs on the Unified Communications Manager platform). Unlike a Unified Communications Manager cluster, however, Connection supports only two Connection servers in the Connection cluster. For more information, see Cisco Unity Connection Design Guides, available at: products-implementation-design-guides-list.html Cisco Untity Connection Design Guides Cisco HCS for Contact Center High Availability All Cisco Unified Contact Center products provide high availability. For example, when you integrate Unified IP IVR with Unified Communications Manager, you could add a second Unified IP IVR server to provide high availability. One of the servers is the active server and handles all the call processing. The other server is in standby mode and becomes active only if the primary server fails. Unified CVP also supports high available deployments with multiple Unified CVP servers, voice gateways, VXML gateways, SIP proxies, and so forth. In addition to the redundancy of the Unified Contact Center components themselves, their integration with Unified Communications Manager can also be redundant. For example, each Unified IP IVR server can connect to a primary CTI Manager and also to a backup CTI Manager in case the primary CTI Manager fails. With Unified CCE, a PG side A connects to a primary CTI Manager, while the redundant PG side B connects to the secondary CTI Manager, thus providing high availability if one CTI Manager fails. For more information, see Cisco Unified Contact Center SRNDs, available at: Cisco Community for Hosted Collaboration Solutions for Contact Center Cisco Hosted Collaboration Solution Maintain and Operate Guides Cisco HCS Management Cisco UCS and UCSM Cisco Unified Computing System (UCS) is deployed as a high-availability cluster for management-plane redundancy and increased data-plane bandwidth. The system is an integrated, scalable, multichassis platform in which all resources participate in a unified management domain. Managed as a single system whether it has one server or 320 servers with thousands of virtual machines, the Cisco UCS decouples scale from 18
19 Vmware vsphere and vcenter Server complexity. Cisco UCS provides a highly available computing environment and can eliminate single points of failure. Cisco Unified Computing System Manager (UCSM) provides unified, embedded management of all software and hardware components in the Cisco UCS. It controls multiple chassis and manages resources for thousands of virtual machines. For more information to install or configure the Cisco UCS, see the appropriate document for your installation at: Vmware vsphere and vcenter Server VMware vsphere makes it possible for organizations to dramatically reduce planned downtime. Because workloads in a vsphere environment can be dynamically moved to different physical servers without downtime or service interruption, you can perform server maintenance without requiring application and service downtime. VMware vsphere also provides features to support availability, security, and scalability and builds on the vcenter Server to support cloud computing across the Internet and to secure the VM access layer. We highly recommend that you protect your vcenter Server, Cisco Unified Communications applications, and Cisco HCS management applications with the use of vsphere high availability (HA). VMware HA protects application availability in the following ways: It protects against hardware failure by restarting the virtual machines on other hosts within the cluster. It protects against operating system failure by continuously monitoring a virtual machine and resetting it in the event that an operating system (OS) failure is detected. Additionally the vsphere vmotion and Storage vmotion functionality in vsphere makes it possible for organizations to reduce planned downtime because workloads in a VMware environment can be dynamically moved to different physical servers or to different underlying storage without service interruption. Administrators can perform faster and completely transparent maintenance operations, without being forced to schedule inconvenient maintenance windows. For more information, see: Best practices and advanced features for VMware High Availability ( Installing or upgrading to ESXi 5.1 best practices ( ) Upgrading to ESXi 5.0 best practices ( Upgrading to vcenter Server 5.1 best practices ( Upgrading to vcenter Server 5.0 best practices ( Installing vcenter Server 5.1 best practices ( Installing vcenter Server 5.0 best practices ( Cisco HCS Service Fulfillment and Service Assurance Domain Managers Cisco Unified Communications Domain Manager, Cisco HCM-F, and Cisco Prime Collaboration rely upon embedded high availability capabilities provided by the Cisco UCS and VMware HA. However, Cisco Unified Communications Domain Manager does provide additional redundancy and disaster recovery capabilities. These disaster recovery capabilities differ from Automatic over and provide the capability for a business 19
20 Cisco HCS Service Fulfillment and Service Assurance Domain Managers to recover and continue operating in the event of a major failure or event. For more information about this functionality see the Cisco Unified Communications Domain Manager Documentation for Cisco HCS 9.0(1) - zip archive. 20
Deployment Topologies
, page 1 Multinode Cluster with Unified Nodes, page 2 Clustering Considerations, page 3 Cisco Unified Communications Domain Manager 10.6(x) Redundancy and Disaster Recovery, page 4 Capacity Considerations,
More informationSystem Requirements and Server Configuration
Hardware Requirements for C Series, page 1 Open Virtualization Files, page 2 VMware Hosting, page 2 Software, page 2 Side A Server Component Configurations, page 5 Side B Server Component Configurations,
More informationAvaya Aura Virtualized Environment
Avaya Aura Virtualized Environment Table of Contents Redundancy Options for Avaya Aura Virtualized Environment (AAVE)... 1 Avaya Aura Communication Manager (CM) Software Duplication and Statefull Failover...
More informationCisco Application Networking Manager Version 2.0
Cisco Application Networking Manager Version 2.0 Cisco Application Networking Manager (ANM) software enables centralized configuration, operations, and monitoring of Cisco data center networking equipment
More informationCloud Optimize Your IT
Cloud Optimize Your IT Windows Server 2012 The information contained in this presentation relates to a pre-release product which may be substantially modified before it is commercially released. This pre-release
More informationVMware vsphere Data Protection
VMware vsphere Data Protection Replication Target TECHNICAL WHITEPAPER 1 Table of Contents Executive Summary... 3 VDP Identities... 3 vsphere Data Protection Replication Target Identity (VDP-RT)... 3 Replication
More informationHRG Assessment: Stratus everrun Enterprise
HRG Assessment: Stratus everrun Enterprise Today IT executive decision makers and their technology recommenders are faced with escalating demands for more effective technology based solutions while at
More informationRemote PC Guide Series - Volume 1
Introduction and Planning for Remote PC Implementation with NETLAB+ Document Version: 2016-02-01 What is a remote PC and how does it work with NETLAB+? This educational guide will introduce the concepts
More informationCisco Active Network Abstraction Gateway High Availability Solution
. Cisco Active Network Abstraction Gateway High Availability Solution White Paper This white paper describes the Cisco Active Network Abstraction (ANA) Gateway High Availability solution developed and
More informationBackup and Redundancy
Backup and Redundancy White Paper NEC s UC for Business Backup and Redundancy allow businesses to operate with confidence, providing security for themselves and their customers. When a server goes down
More informationDR-to-the- Cloud Best Practices
DR-to-the- Cloud Best Practices HOW TO EFFECTIVELY CONFIGURE YOUR OWN SELF-MANAGED RECOVERY PLANS AND THE REPLICATION OF CRITICAL VMWARE VIRTUAL MACHINES FROM ON-PREMISES TO A CLOUD SERVICE PROVIDER CONTENTS
More informationVMware vrealize Automation
VMware vrealize Automation Reference Architecture Version 6.0 and Higher T E C H N I C A L W H I T E P A P E R Table of Contents Overview... 4 What s New... 4 Initial Deployment Recommendations... 4 General
More informationVMware vsphere-6.0 Administration Training
VMware vsphere-6.0 Administration Training Course Course Duration : 20 Days Class Duration : 3 hours per day (Including LAB Practical) Classroom Fee = 20,000 INR Online / Fast-Track Fee = 25,000 INR Fast
More informationDisaster Recovery Design Ehab Ashary University of Colorado at Colorado Springs
Disaster Recovery Design Ehab Ashary University of Colorado at Colorado Springs As a head of the campus network department in the Deanship of Information Technology at King Abdulaziz University for more
More informationNET ACCESS VOICE PRIVATE CLOUD
Page 0 2015 SOLUTION BRIEF NET ACCESS VOICE PRIVATE CLOUD A Cloud and Connectivity Solution for Hosted Voice Applications NET ACCESS LLC 9 Wing Drive Cedar Knolls, NJ 07927 www.nac.net Page 1 Table of
More informationCisco Nexus 1000V and Cisco Nexus 1110 Virtual Services Appliance (VSA) across data centers
Cisco Nexus 1000V and Cisco Nexus 1110 Virtual Services Appliance (VSA) across data centers With the improvement in storage, virtualization and L2 extension technologies, customers can now choose to have
More informationVXLAN: Scaling Data Center Capacity. White Paper
VXLAN: Scaling Data Center Capacity White Paper Virtual Extensible LAN (VXLAN) Overview This document provides an overview of how VXLAN works. It also provides criteria to help determine when and where
More informationVMware vrealize Automation
VMware vrealize Automation Reference Architecture Version 6.0 or Later T E C H N I C A L W H I T E P A P E R J U N E 2 0 1 5 V E R S I O N 1. 5 Table of Contents Overview... 4 What s New... 4 Initial Deployment
More informationVMware vcloud Automation Center 6.0
VMware 6.0 Reference Architecture TECHNICAL WHITE PAPER Table of Contents Overview... 4 Initial Deployment Recommendations... 4 General Recommendations... 4... 4 Load Balancer Considerations... 4 Database
More informationVMware vcloud Automation Center 6.1
VMware vcloud Automation Center 6.1 Reference Architecture T E C H N I C A L W H I T E P A P E R Table of Contents Overview... 4 What s New... 4 Initial Deployment Recommendations... 4 General Recommendations...
More informationFeature Comparison. Windows Server 2008 R2 Hyper-V and Windows Server 2012 Hyper-V
Comparison and Contents Introduction... 4 More Secure Multitenancy... 5 Flexible Infrastructure... 9 Scale, Performance, and Density... 13 High Availability... 18 Processor and Memory Support... 24 Network...
More informationOmniCube. SimpliVity OmniCube and Multi Federation ROBO Reference Architecture. White Paper. Authors: Bob Gropman
OmniCube SimpliVity OmniCube and Multi Federation ROBO Reference Architecture White Paper Authors: Bob Gropman Date: April 13, 2015 SimpliVity and OmniCube are trademarks of SimpliVity Corporation. All
More informationModule: Business Continuity
Upon completion of this module, you should be able to: Describe business continuity and cloud service availability Describe fault tolerance mechanisms for cloud infrastructure Discuss data protection solutions
More informationVMware@SoftLayer Cookbook Disaster Recovery (DR)
VMware@SoftLayer Cookbook Disaster Recovery (DR) IBM Global Technology Services: Khoa Huynh (khoa@us.ibm.com) Daniel De Araujo (ddearaujo@us.ibm.com) Bob Kellenberger (kellenbe@us.ibm.com) VMware: Merlin
More informationEMC Data Domain Management Center
EMC Data Domain Management Center Version 1.1 Initial Configuration Guide 302-000-071 REV 04 Copyright 2012-2015 EMC Corporation. All rights reserved. Published in USA. Published June, 2015 EMC believes
More informationWindows Geo-Clustering: SQL Server
Windows Geo-Clustering: SQL Server Edwin Sarmiento, Microsoft SQL Server MVP, Microsoft Certified Master Contents Introduction... 3 The Business Need for Geo-Clustering... 3 Single-location Clustering
More informationEMC Business Continuity for VMware View Enabled by EMC SRDF/S and VMware vcenter Site Recovery Manager
EMC Business Continuity for VMware View Enabled by EMC SRDF/S and VMware vcenter Site Recovery Manager A Detailed Review Abstract This white paper demonstrates that business continuity can be enhanced
More informationMicrosoft SharePoint 2010 on VMware Availability and Recovery Options. Microsoft SharePoint 2010 on VMware Availability and Recovery Options
This product is protected by U.S. and international copyright and intellectual property laws. This product is covered by one or more patents listed at http://www.vmware.com/download/patents.html. VMware
More informationEMC VPLEX FAMILY. Continuous Availability and data Mobility Within and Across Data Centers
EMC VPLEX FAMILY Continuous Availability and data Mobility Within and Across Data Centers DELIVERING CONTINUOUS AVAILABILITY AND DATA MOBILITY FOR MISSION CRITICAL APPLICATIONS Storage infrastructure is
More informationInformix Dynamic Server May 2007. Availability Solutions with Informix Dynamic Server 11
Informix Dynamic Server May 2007 Availability Solutions with Informix Dynamic Server 11 1 Availability Solutions with IBM Informix Dynamic Server 11.10 Madison Pruet Ajay Gupta The addition of Multi-node
More informationIntegrated Application and Data Protection. NEC ExpressCluster White Paper
Integrated Application and Data Protection NEC ExpressCluster White Paper Introduction Critical business processes and operations depend on real-time access to IT systems that consist of applications and
More informationFederated Application Centric Infrastructure (ACI) Fabrics for Dual Data Center Deployments
Federated Application Centric Infrastructure (ACI) Fabrics for Dual Data Center Deployments March 13, 2015 Abstract To provide redundancy and disaster recovery, most organizations deploy multiple data
More informationBest Practices for Monitoring Databases on VMware. Dean Richards Senior DBA, Confio Software
Best Practices for Monitoring Databases on VMware Dean Richards Senior DBA, Confio Software 1 Who Am I? 20+ Years in Oracle & SQL Server DBA and Developer Worked for Oracle Consulting Specialize in Performance
More informationValidating Long-distance VMware vmotion
Technical Brief Validating Long-distance VMware vmotion with NetApp FlexCache and F5 BIG-IP F5 BIG-IP enables long distance VMware vmotion live migration and optimizes NetApp FlexCache replication. Key
More informationVMware vsphere Data Protection 6.0
VMware vsphere Data Protection 6.0 TECHNICAL OVERVIEW REVISED FEBRUARY 2015 Table of Contents Introduction.... 3 Architectural Overview... 4 Deployment and Configuration.... 5 Backup.... 6 Application
More informationHigh Availability with Windows Server 2012 Release Candidate
High Availability with Windows Server 2012 Release Candidate Windows Server 2012 Release Candidate (RC) delivers innovative new capabilities that enable you to build dynamic storage and availability solutions
More informationCisco Nexus 1000V Switch for Microsoft Hyper-V
Data Sheet Cisco Nexus 1000V Switch for Microsoft Hyper-V Product Overview Cisco Nexus 1000V Switches provide a comprehensive and extensible architectural platform for virtual machine and cloud networking.
More informationAvailability Guide for Deploying SQL Server on VMware vsphere. August 2009
Availability Guide for Deploying SQL Server on VMware vsphere August 2009 Contents Introduction...1 SQL Server 2008 with vsphere and VMware HA/DRS...2 Log Shipping Availability Option...4 Database Mirroring...
More informationReal-time Protection for Hyper-V
1-888-674-9495 www.doubletake.com Real-time Protection for Hyper-V Real-Time Protection for Hyper-V Computer virtualization has come a long way in a very short time, triggered primarily by the rapid rate
More informationMANAGED DATABASE SOLUTIONS
Page 0 2015 SOLUTION BRIEF MANAGED DATABASE SOLUTIONS NET ACCESS LLC 9 Wing Drive Cedar Knolls, NJ 07927 www.nac.net Page 1 Table of Contents 1. Introduction... 2 2. Net Access Managed Services Solution
More informationSilver Peak Virtual Appliances
Silver Peak Virtual Appliances Frequently Asked Questions (FAQ) VX Installation This section addresses questions or problems you may encounter during Silver Peak VX installation. 1. I need to increase
More informationLeveraging Public Cloud for Affordable VMware Disaster Recovery & Business Continuity
White Paper White Paper Managing Public Cloud Computing in the Enterprise Leveraging Public Cloud for Affordable VMware Disaster Recovery & Business Continuity A Quick Start Guide By Edward Haletky Principal
More informationThe Benefits of Virtualizing
T E C H N I C A L B R I E F The Benefits of Virtualizing Aciduisismodo Microsoft SQL Dolore Server Eolore in Dionseq Hitachi Storage Uatummy Environments Odolorem Vel Leveraging Microsoft Hyper-V By Heidi
More informationBuilding Reliable, Scalable AR System Solutions. High-Availability. White Paper
Building Reliable, Scalable Solutions High-Availability White Paper Introduction This paper will discuss the products, tools and strategies available for building reliable and scalable Action Request System
More informationPart 1 - What s New in Hyper-V 2012 R2. Clive.Watson@Microsoft.com Datacenter Specialist
Part 1 - What s New in Hyper-V 2012 R2 Clive.Watson@Microsoft.com Datacenter Specialist Microsoft Cloud OS Vision Public Cloud Azure Virtual Machines Windows Azure Pack 1 Consistent Platform Windows Azure
More informationMicrosoft SMB File Sharing Best Practices Guide
Technical White Paper Microsoft SMB File Sharing Best Practices Guide Tintri VMstore, Microsoft SMB 3.0 Protocol, and VMware 6.x Author: Neil Glick Version 1.0 06/15/2016 @tintri www.tintri.com Contents
More informationA virtual SAN for distributed multi-site environments
Data sheet A virtual SAN for distributed multi-site environments What is StorMagic SvSAN? StorMagic SvSAN is a software storage solution that enables enterprises to eliminate downtime of business critical
More informationCisco HCS capacity planning
Introduction, page Network Infrastructure capacity planning, page Unified Communications Application capacity planning, page 0 Introduction This document provides capacity planning guidelines for the following
More informationSolution Brief Availability and Recovery Options: Microsoft Exchange Solutions on VMware
Introduction By leveraging the inherent benefits of a virtualization based platform, a Microsoft Exchange Server 2007 deployment on VMware Infrastructure 3 offers a variety of availability and recovery
More informationWindows Server Failover Clustering April 2010
Windows Server Failover Clustering April 00 Windows Server Failover Clustering (WSFC) is the successor to Microsoft Cluster Service (MSCS). WSFC and its predecessor, MSCS, offer high availability for critical
More informationAVAYA AND VIRTUALIZATION
AVAYA AND VIRTUALIZATION Avaya VMware Comparison VMware vcenter Administration Tool Avaya System Manager ESXi Hypervisor Application System Platform Host Physical Server Common Server Guest Virtual Server
More informationCA Cloud Overview Benefits of the Hyper-V Cloud
Benefits of the Hyper-V Cloud For more information, please contact: Email: sales@canadianwebhosting.com Ph: 888-821-7888 Canadian Web Hosting (www.canadianwebhosting.com) is an independent company, hereinafter
More informationBuilding the Virtual Information Infrastructure
Technology Concepts and Business Considerations Abstract A virtual information infrastructure allows organizations to make the most of their data center environment by sharing computing, network, and storage
More informationPivot3 Reference Architecture for VMware View Version 1.03
Pivot3 Reference Architecture for VMware View Version 1.03 January 2012 Table of Contents Test and Document History... 2 Test Goals... 3 Reference Architecture Design... 4 Design Overview... 4 The Pivot3
More informationACME Enterprises IT Infrastructure Assessment
Prepared for ACME Enterprises March 25, 2014 Table of Contents Executive Summary...2 Introduction...2 Background...2 Scope of IT Assessment...2 Findings...2 Detailed Findings for Key Areas Reviewed...3
More informationParagon Protect & Restore
Paragon Protect & Restore ver. 3 Centralized and Disaster Recovery for virtual and physical environments Tight Integration with hypervisors for agentless backups, VM replication and seamless restores Paragon
More informationVirtual Web Appliance Setup Guide
Virtual Web Appliance Setup Guide 2 Sophos Installing a Virtual Appliance Installing a Virtual Appliance This guide describes the procedures for installing a Virtual Web Appliance. If you are installing
More informationInstallation and Upgrade on Windows Server 2008/2012 When the Secondary Server is Physical VMware vcenter Server Heartbeat 6.6
Installation and Upgrade on Windows Server 2008/2012 When the Secondary Server is Physical VMware vcenter Server Heartbeat 6.6 This document supports the version of each product listed and supports all
More informationWhitepaper Continuous Availability Suite: Neverfail Solution Architecture
Continuous Availability Suite: Neverfail s Continuous Availability Suite is at the core of every Neverfail solution. It provides a comprehensive software solution for High Availability (HA) and Disaster
More informationDeltaV Virtualization High Availability and Disaster Recovery
DeltaV Distributed Control System Whitepaper October 2014 DeltaV Virtualization High Availability and Disaster Recovery This document describes High Availiability and Disaster Recovery features supported
More informationVirtual Managment Appliance Setup Guide
Virtual Managment Appliance Setup Guide 2 Sophos Installing a Virtual Appliance Installing a Virtual Appliance As an alternative to the hardware-based version of the Sophos Web Appliance, you can deploy
More informationLecture 02b Cloud Computing II
Mobile Cloud Computing Lecture 02b Cloud Computing II 吳 秀 陽 Shiow-yang Wu T. Sridhar. Cloud Computing A Primer, Part 2: Infrastructure and Implementation Topics. The Internet Protocol Journal, Volume 12,
More informationVMware vsphere Data Protection 5.8 TECHNICAL OVERVIEW REVISED AUGUST 2014
VMware vsphere Data Protection 5.8 TECHNICAL OVERVIEW REVISED AUGUST 2014 Table of Contents Introduction.... 3 Features and Benefits of vsphere Data Protection... 3 Additional Features and Benefits of
More informationDeploying F5 BIG-IP Virtual Editions in a Hyper-Converged Infrastructure
Deploying F5 BIG-IP Virtual Editions in a Hyper-Converged Infrastructure Justin Venezia Senior Solution Architect Paul Pindell Senior Solution Architect Contents The Challenge 3 What is a hyper-converged
More informationActive-Active and High Availability
Active-Active and High Availability Advanced Design and Setup Guide Perceptive Content Version: 7.0.x Written by: Product Knowledge, R&D Date: July 2015 2015 Perceptive Software. All rights reserved. Lexmark
More informationVMware vsphere Data Protection 6.1
VMware vsphere Data Protection 6.1 Technical Overview Revised August 10, 2015 Contents Introduction... 3 Architecture... 3 Deployment and Configuration... 5 Backup... 6 Application Backup... 6 Backup Data
More informationHow To Protect Your Data From Being Damaged On Vsphere Vdp Vdpa Vdpo Vdprod (Vmware) Vsphera Vdpower Vdpl (Vmos) Vdper (Vmom
VMware vsphere Data Protection 6.0 Maciej Kot Senior Systems Engineer VMware Overview vsphere Data Protection Overview Data protection for VMs and applications Agent-less VM backup and restore Agents for
More informationBosch Video Management System High Availability with Hyper-V
Bosch Video Management System High Availability with Hyper-V en Technical Service Note Bosch Video Management System Table of contents en 3 Table of contents 1 Introduction 4 1.1 General Requirements
More informationStretched Active- Active Application Centric Infrastructure (ACI) Fabric
Stretched Active- Active Application Centric Infrastructure (ACI) Fabric May 12, 2015 Abstract This white paper illustrates how the Cisco Application Centric Infrastructure (ACI) can be implemented as
More informationExpert Reference Series of White Papers. Visions of My Datacenter Virtualized
Expert Reference Series of White Papers Visions of My Datacenter Virtualized 1-800-COURSES www.globalknowledge.com Visions of My Datacenter Virtualized John A. Davis, VMware Certified Instructor (VCI),
More informationVirtual Solutions. Reliable voice performance in a virtualized environment
OVERVIEW MITEL Virtual Solutions Reliable voice performance in a virtualized environment What would your IT department do if it could spend more time on innovation and less time working just to keep the
More informationUsing Live Sync to Support Disaster Recovery
Using Live Sync to Support Disaster Recovery SIMPANA VIRTUAL SERVER AGENT FOR VMWARE Live Sync uses backup data to create and maintain a warm disaster recovery site. With backup and replication from a
More informationUpgrade and Migration Strategy
Cisco HCS Architecture, page 2 Solution Upgrade and Migration Phases, page 5 Understanding Subsystem Upgrades, page 6 1 Cisco HCS Architecture Cisco HCS Architecture The following figure provides a high-level
More informationNetworking Topology For Your System
This chapter describes the different networking topologies supported for this product, including the advantages and disadvantages of each. Select the one that best meets your needs and your network deployment.
More informationAppendix C to DIR Contract Number DIR-TSO-2736 SunGard Availability Services Discount Level: 25% Managed Data Center Services - Cloud Hosting
Resource Pool for ECS/Managed Cloud. Order in 2GB increments. Storage added separately Level: 25% Min RAM Max RAM Cloud Resource Pool (1 GB RAM /.5 vcpu) 1 2 32 26% 2 33 64 26% 3 65 128 26% 4 129 N/A 26%
More informationRestricted Document. Pulsant Technical Specification
Pulsant Technical Specification Title Pulsant Government Virtual Server IL2 Department Cloud Services Contributors RR Classification Restricted Version 1.0 Overview Pulsant offer two products based on
More informationCisco Virtual Security Gateway for Nexus 1000V Series Switch
Data Sheet Cisco Virtual Security Gateway for Nexus 1000V Series Switch Product Overview Cisco Virtual Security Gateway (VSG) for Nexus 1000V Series Switch is a virtual appliance that provides trusted
More informationvcloud Air Disaster Recovery Technical Presentation
vcloud Air Disaster Recovery Technical Presentation Agenda 1 vcloud Air Disaster Recovery Overview 2 What s New 3 Architecture 4 Setup and Configuration 5 Considerations 6 Automation Options 2 vcloud Air
More informationImplementing and Troubleshooting the Cisco Cloud Infrastructure **Part of CCNP Cloud Certification Track**
Course: Duration: Price: $ 4,295.00 Learning Credits: 43 Certification: Implementing and Troubleshooting the Cisco Cloud Infrastructure Implementing and Troubleshooting the Cisco Cloud Infrastructure**Part
More informationTop-Down Network Design
Top-Down Network Design Chapter Five Designing a Network Topology Copyright 2010 Cisco Press & Priscilla Oppenheimer Topology A map of an internetwork that indicates network segments, interconnection points,
More informationDeploying Exchange Server 2007 SP1 on Windows Server 2008
Deploying Exchange Server 2007 SP1 on Windows Server 2008 Product Group - Enterprise Dell White Paper By Ananda Sankaran Andrew Bachler April 2008 Contents Introduction... 3 Deployment Considerations...
More informationChapter 2 TOPOLOGY SELECTION. SYS-ED/ Computer Education Techniques, Inc.
Chapter 2 TOPOLOGY SELECTION SYS-ED/ Computer Education Techniques, Inc. Objectives You will learn: Topology selection criteria. Perform a comparison of topology selection criteria. WebSphere component
More informationAnnouncing the product launch of a Mitel Virtual Mitel Communication Director (Virtual MCD)
Announcing the product launch of a Mitel Virtual Mitel Communication Director (Virtual MCD) In today's global economy, businesses are - more than ever - looking to reduce capital and operating expenses.
More informationCisco and EMC Solutions for Application Acceleration and Branch Office Infrastructure Consolidation
Solution Overview Cisco and EMC Solutions for Application Acceleration and Branch Office Infrastructure Consolidation IT organizations face challenges in consolidating costly and difficult-to-manage branch-office
More informationA SWOT ANALYSIS ON CISCO HIGH AVAILABILITY VIRTUALIZATION CLUSTERS DISASTER RECOVERY PLAN
A SWOT ANALYSIS ON CISCO HIGH AVAILABILITY VIRTUALIZATION CLUSTERS DISASTER RECOVERY PLAN Eman Al-Harbi 431920472@student.ksa.edu.sa Soha S. Zaghloul smekki@ksu.edu.sa Faculty of Computer and Information
More informationVirtual Appliance Setup Guide
Virtual Appliance Setup Guide 2015 Bomgar Corporation. All rights reserved worldwide. BOMGAR and the BOMGAR logo are trademarks of Bomgar Corporation; other trademarks shown are the property of their respective
More informationAstaro Deployment Guide High Availability Options Clustering and Hot Standby
Connect With Confidence Astaro Deployment Guide Clustering and Hot Standby Table of Contents Introduction... 2 Active/Passive HA (Hot Standby)... 2 Active/Active HA (Cluster)... 2 Astaro s HA Act as One...
More informationvsphere Replication for Disaster Recovery to Cloud
vsphere Replication for Disaster Recovery to Cloud vsphere Replication 6.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced
More informationVMware vcloud Air - Disaster Recovery User's Guide
VMware vcloud Air - Disaster Recovery User's Guide vcloud Air This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition.
More informationIMPROVING VMWARE DISASTER RECOVERY WITH EMC RECOVERPOINT Applied Technology
White Paper IMPROVING VMWARE DISASTER RECOVERY WITH EMC RECOVERPOINT Applied Technology Abstract EMC RecoverPoint provides full support for data replication and disaster recovery for VMware ESX Server
More informationSymantec Cluster Server powered by Veritas
Delivers high availability and disaster recovery for your critical applications Data Sheet: High Availability Overview protects your most important applications from planned and unplanned downtime. Cluster
More informationThe Future of Computing Cisco Unified Computing System. Markus Kunstmann Channels Systems Engineer
The Future of Computing Cisco Unified Computing System Markus Kunstmann Channels Systems Engineer 2009 Cisco Systems, Inc. All rights reserved. Data Centers Are under Increasing Pressure Collaboration
More informationNutanix Tech Note. Configuration Best Practices for Nutanix Storage with VMware vsphere
Nutanix Tech Note Configuration Best Practices for Nutanix Storage with VMware vsphere Nutanix Virtual Computing Platform is engineered from the ground up to provide enterprise-grade availability for critical
More informationVirtualizing Microsoft Exchange Server 2010 with NetApp and VMware
Virtualizing Microsoft Exchange Server 2010 with NetApp and VMware Deploying Microsoft Exchange Server 2010 in a virtualized environment that leverages VMware virtualization and NetApp unified storage
More informationInstalling and Using the vnios Trial
Installing and Using the vnios Trial The vnios Trial is a software package designed for efficient evaluation of the Infoblox vnios appliance platform. Providing the complete suite of DNS, DHCP and IPAM
More informationVeritas InfoScale Availability
Veritas InfoScale Availability Delivers high availability and disaster recovery for your critical applications Overview protects your most important applications from planned and unplanned downtime. InfoScale
More informationPivot3 Desktop Virtualization Appliances. vstac VDI Technology Overview
Pivot3 Desktop Virtualization Appliances vstac VDI Technology Overview February 2012 Pivot3 Desktop Virtualization Technology Overview Table of Contents Executive Summary... 3 The Pivot3 VDI Appliance...
More informationEMC VPLEX FAMILY. Transparent information mobility within, across, and between data centers ESSENTIALS A STORAGE PLATFORM FOR THE PRIVATE CLOUD
EMC VPLEX FAMILY Transparent information mobility within, across, and between data centers A STORAGE PLATFORM FOR THE PRIVATE CLOUD In the past, users have relied on traditional physical storage to meet
More informationHow to migrate your business to the Cloud
How to migrate your business to the Cloud Introduction As discussed in Part 1 of this Cloud migration white paper series Should I migrate my business to the Cloud there are many benefits to using Cloud
More informationvsphere Networking vsphere 6.0 ESXi 6.0 vcenter Server 6.0 EN-001391-01
vsphere 6.0 ESXi 6.0 vcenter Server 6.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more
More information