VMware SDDC on IBM Cloud - Advanced Detailed Design
|
|
|
- Anastasia Hancock
- 9 years ago
- Views:
Transcription
1 VMware SDDC on IBM Cloud - Advanced Detailed Design Date: 16 th March 2016 Version: 1.1 Copyright IBM and VMware Page 1 of 132
2 Table of Contents 1 Introduction Pre-requisites Summary of Changes System Context Actors Systems Architecture Overview Physical Infrastructure Virtual Infrastructure Infrastructure Management Common Services Cloud Management Services Operational Services Business Services Logical Operational Model Logical Operational Model Structure Central Cloud Physical Infrastructure Cluster Architecture Physical Network Physical Storage Virtual Infrastructure Compute Virtualization Storage Virtualization Network Virtualization Infrastructure Management Compute Management Storage Management Network Management Common Services Identity and Access Services Domain Name Services NTP Services SMTP Services Certificate Authority Services Cloud Management Services Page 2 of 132 Copyright IBM and VMware
3 4.7.1 Service Catalog Self-Service Portal Infrastructure and Process Orchestration Software Orchestration Operational Services Backup and Restore Disaster Recovery Monitoring Log Consolidation and Analysis Patching Business Services Business Management IT Financials IT Benchmarking Cloud Region Physical Operational Model Physical Layer Compute Storage Network Virtual Infrastructure Compute Virtualization Storage Virtualization Network Virtualization Infrastructure Management vcenter Server Instances Common Services Identity and Access Services Domain Name Services DNS Configuration Requirements NTP Services SMTP Services Certificate Authority Services Cloud Management Services Cloud Management Physical Design vrealize Automation Supporting Infrastructure vrealize Automation Cloud Tenant Design vrealize Automation vsphere Integration Design Infrastructure Source Endpoints Copyright IBM and VMware Page 3 of 132
4 5.5.6 Virtualization Compute Resources Process Orchestration Software Orchestration Infrastructure Orchestration Operational Services Backup and Restore Disaster Recovery Monitoring Log Consolidation and Analysis Patching Business Services Appendix A Bare Metal Summary Management Cluster Nodes Compute Cluster Nodes Edge Cluster Nodes Appendix B Software Bill of Materials Appendix C Management Virtual Machine Summary Appendix D Maximum Configurations Appendix E Compatibility Guide Browsers Guest Operating Systems Table of Figures Figure 1 VMware SDDC on IBM Cloud Introduction 8 Figure 2 VMware SDDC on IBM Cloud System Context 10 Figure 3 VMware SDDC on IBM Cloud Architecture Overview 12 Figure 4 Logical Structure View 15 Figure 5 Component Interaction Diagram 16 Figure 6 Logical Operational Model 17 Figure 7 Logical Cluster Structure 18 Figure 8 Network Virtualization 22 Figure 9 Dual-Region Data Protection Architecture 30 Figure 10 Disaster Recovery Architecture 31 Figure 11 vrealize Operations Manager Architecture 32 Figure 12 Physical Operational Model - Virtual Servers, Networking and Clusters 36 Page 4 of 132 Copyright IBM and VMware
5 Figure 13 Network connections per physical node 39 Figure 14 VSAN concept 40 Figure 15 Network Switch Design for Management Hosts 44 Figure 16 Network Switch Design for Edge Hosts 47 Figure 17 Network Switch Design for Compute Hosts 49 Figure 18 Network Virtualization Conceptual Design 54 Figure 19 Cluster Design for NSX for vsphere 55 Figure 20 Virtual Application Network Components and Design 59 Figure 21 vra Virtual Network Design 60 Figure 22 Virtual Application Network Configuration in Central Cloud and Cloud Region 62 Figure 23 vcenter Server and PSC Deployment Model 64 Figure 24 vrealize Automation Conceptual Design 72 Figure 25 vrealize Automation Design Overview for Central Cloud 75 Figure 26 vrealize Automation Design Overview for Additional Cloud Regions 76 Figure 27 Tenant Design for Single Region 81 Figure 28 Tenant Design for Two Regions 82 Figure 29 vrealize Automation Integration with vsphere Endpoint Central Cloud 87 Figure 30 vrealize Automation Integration with vsphere Endpoint Central Cloud and a Cloud Region (Region A) 88 Figure 31 Template Synchronization 90 Figure 32 Software Orchestration Logical Design 95 Figure 33 vsphere Data Protection Logical Design 97 Figure 34 Logical Network Design for Cross-Region Deployment with Management Application Network Container 102 Figure 35 Logical Design of vrealize Operations Manager Central Cloud and a Cloud Region (Region A) Deployment 107 Figure 36 Networking Design of the vrealize Operations Manager Deployment 110 Figure 37 Application Virtual Networks in the vrealize Operations Manager Topology 111 Figure 38 Logical Design of vrealize Log Insight 114 Figure 39 Networking Design for the vrealize Log Insight Deployment 116 Figure 40 Application Virtual Networks in the vrealize Log Insight Topology 117 Figure 41 vrealize Business Logical Design 122 List of Tables Table 1 VMware SDDC on IBM Cloud Interfaced Actors 10 Table 2 VMware SDDC on IBM Cloud Interfaced Systems 11 Table 3 vrealize Operations Manager Logical Node Architecture 33 Table 4 NFS Configuration for vsphere Data Protection and vrealize Log Insight 38 Copyright IBM and VMware Page 5 of 132
6 Table 5 VSAN disk table 40 Table 6 VSAN policies 41 Table 7 VSAN object policy defaults 42 Table 8 VLAN Mapping to Traffic Types 42 Table 9 Management Cluster Distributed Switch 43 Table 10 Management Cluster Distributed Switch Port Group Configuration Settings 43 Table 11 Management Virtual Switch Port Groups and VLANs 45 Table 12 Management VMkernel Adapter 45 Table 13 Edge Cluster Distributed Switch 46 Table 14 Management Cluster Distributed Switch Port Group Configuration Settings 46 Table 15 Edge Virtual Switch Port Groups and VLANs 48 Table 16 Edge VMkernel Adapter 48 Table 17 Compute Cluster Distributed Switch 48 Table 18 Compute Cluster Distributed Switch Port Group Configuration Settings 49 Table 19 Compute Virtual Switch Port Groups and VLANs 50 Table 20 Compute VMkernel Adapter 50 Table 21 NSX Components Sizing 53 Table 22 Load Balancer Features 57 Table 23 Management Applications IP Addressing 61 Table 24 OSPF Area ID 62 Table 25 Specifications for Management vcenter Server Appliance 65 Table 26 Specifications for Platform Service Controller for Management Cluster 65 Table 27 Specifications for Compute and Edge vcenter Server Appliance 65 Table 28 Specifications for Platform Service Controller for Management Cluster 65 Table 29 Requirements for Active Directory Service 67 Table 30 Authentication types used 68 Table 31 Server Sizing 68 Table 32 Domain Naming Example 69 Table 33 SoftLayer DNS servers 70 Table 34 Time sources 70 Table 35 Root CA and Subordinate CA sizing 72 Table 36 Cloud Management Services Components 73 Table 37 Load Balancer Application Profile 78 Table 38 Load Balancer Service Monitoring Configuration 79 Table 39 Load Balancer Pool Specifications 79 Table 40 Virtual Server Characteristics 79 Table 41 Base Windows Server Blueprint 83 Page 6 of 132 Copyright IBM and VMware
7 Table 42 Base Windows Blueprint Sizing 84 Table 43 Base Linux Server Blueprint 84 Table 44 Base Linux Blueprint Sizing 84 Table 45 vrealize Integration with vsphere 85 Table 46 vrealize Orchestrator Default Configuration Ports 91 Table 47 vrealize Orchestrator Default External Communication Ports 91 Table 48 vro Service Monitor Specifications 93 Table 49 vro Service Pool Characteristics 93 Table 50 vro Virtual Server Characteristics 93 Table 51 Software Orchestration Components Sizing 96 Table 52 VMware vsphere Data Protection Performance 97 Table 53 Backup Jobs in Central Cloud 99 Table 54 Backup Jobs in Additional Cloud Region 101 Table 55 SRM Windows server sizing 105 Table 56 vsphere Replication Appliance 106 Table 57 Analytics Cluster Node Configurations 107 Table 58 DRS Cluster Anti-Affinity Rule for vrealize Operations Manager Nodes 108 Table 59 Remote Collector Node Sizes 108 Table 60 DRS Cluster Anti-Affinity Rule for vrealize Operations Remote Collector Nodes 109 Table 61 IP Subnets in the Application Virtual Network of vrealize Operations Manager 111 Table 62 DNS Names for the Application Virtual Networks 112 Table 63 Node Sizing 115 Table 64 IP Subnets in the Application Isolated Networks 117 Table 65 Example DNS names of Log Insight nodes 117 Table 66 Virtual Disk Configuration in the vrealize Log Insight Virtual Appliance 118 Table 67 Compute Resources for vum vcenter Managing the Management Cluster 121 Table 68 Compute Resources for vum vcenter Managing the Compute and Edge Clusters 121 Table 69 Management - Bare Metal Bill of Materials 124 Table 70 Compute - Bare Metal Bill of Materials 124 Table 71 Edge - Bare Metal Bill of Materials 125 Table 72 Software Bill of Materials 126 Table 73 List of Management Cluster Virtual Machines and Sizes 128 Table 74 List of Default Edge Cluster Virtual Machines 130 Copyright IBM and VMware Page 7 of 132
8 1 Introduction VMware Software Defined Data Center (SDDC) on IBM Cloud allows existing VMware virtualized datacenter clients to extend into the IBM Cloud. This permits uses like capacity expansion into the cloud (and contraction when not needed), migration to the cloud, disaster recovery to the cloud, backup into the cloud and the ability to stand up a dedicated cloud environment for development, test, training or lab. This document details the design of the Advanced version of the VMware SDDC on IBM Cloud which targets designs requiring high levels of scalability and multiple regions. Figure 1 VMware SDDC on IBM Cloud Introduction 1.1 Pre-requisites The design requires the following pre-requisites: Client is required to acquire all necessary software licenses and/or keys for all products used in this design prior to commencement of implementation Client is required to provide SoftLayer account Client is responsible for SoftLayer related charges as a result of this design s implementation Client is responsible for connectivity from this design to any on premises environment or systems Client is responsible for connectivity into this design for access by administrators and end users Client is responsible to acquire and provide domain name Client is responsible to provide hostname prefixes for the SoftLayer bare metal devices provisioned through this design Client to provide connection details and necessary credentials for any systems external to this design that are to be integrated (refer to system context for options) Client is responsible for licensing of any software products provisioned with the design Page 8 of 132 Copyright IBM and VMware
9 1.2 Summary of Changes This section records the history of significant changes to this document. Only the most significant changes are described here. Version Date Author Description of Change th Feb th March 2016 Simon Kofkin-Hansen Richard Ehrhardt Razvan Ionescu Daniel de Araujo Frank Chodacki Bob Kellenberger Bryan Buckland Christopher Moss Daniel Arrieta Alvarez Initial Release of Document Minor reported spelling and grammar corrections. Copyright IBM and VMware Page 9 of 132
10 2 System Context When depicting the VMware SDDC on IBM Cloud design as a single object, the following are the external actors and systems that interface with the design. NTP Cloud Admin Manages Cloud Service Provider Manages Services End Users Consumes Cloud Services Time Synchronisation for Cloud Services External Domain for Cloud Services Client DNS Recipes Patches VMware on IBM Cloud Patch Repo Checks for Updates SMTP Relay Sends s Client Auth. Trust Relationship for Cloud Services Provides Bare Metal Compute, Storage, and Network Builds Environment Connects To Client On Premise vsphere SoftLayer 2.1 Actors Figure 2 VMware SDDC on IBM Cloud System Context The actors that interface with the design are described in the following table. There is not a direct correlation between actors and persons. An actor role may be performed by one or more persons. Alternately, one person may perform more than one actor role. Table 1 VMware SDDC on IBM Cloud Interfaced Actors Actor Cloud Admin Service Provider Description The cloud admin or administrator is responsible for maintaining the cloud services. This includes, Assigning virtual resources to groups Maintaining Cloud Software Platform System Administrator roles Manages the cloud services that are provided to the client users. This includes, Service Catalog configuration Defining roles Defining groups Configuring user access Tenant Administrator roles Page 10 of 132 Copyright IBM and VMware
11 User Consumes the services that the cloud admin allows access to. This typically includes, Provisioning VMs De-provisioning VMs Provisioning Patterns De-provisioning Patterns Start / Stop / restart VMs and Patterns 2.2 Systems The systems that interface with the design are described in the following table. Table 2 VMware SDDC on IBM Cloud Interfaced Systems System SoftLayer Client On Premises vsphere Client SMTP Relay Client Authentication Client DNS NTP Service Patch Repo Description SoftLayer provides the bare metal, physical networking and NFS storage in addition to the automation to build the design when ordered. The design is able to connect to an existing vsphere environment on a client premises to enable Hybrid capabilities. The design connects its SMTP server to a client s SMTP relay service to provide notifications on aspects such as the process orchestration. The design is able to connect to an existing client authentication system to establish a trust relationship which extends the client s authentication system into the cloud for use by the cloud management platform. The design is able to connect to a client s domain name service (DNS) to extend the domain service into the cloud for use by the cloud management platform. The design requires an external NTP service to provide time synchronization services for use by the cloud management platform. There are a number of internet based patch repositories that the cloud management platform applications need to connect to in order to maintain the security and stability of the cloud environment. Copyright IBM and VMware Page 11 of 132
12 3 Architecture Overview VMware SDDC on IBM Cloud provides VMware automation technology on SoftLayer. This includes virtual networking, virtual storage, process orchestration, infrastructure orchestration and software orchestration. It also provides the tools for management of the services providing these functions. The architecture consists of at least one central cloud region built on SoftLayer, which provides the main portal for users and administration, plus it can include one or more cloud regions which are managed by the central cloud and provide additional functionality for remote locations. The architecture is scaled out within a region, or by adding regions. Figure 3 VMware SDDC on IBM Cloud Architecture Overview 3.1 Physical Infrastructure The physical infrastructure consists of three main components, physical compute, physical network and physical storage. The physical compute provides the physical processing and memory that is used by the virtualization infrastructure. The physical network provides the network connectivity into the environment that is then consumed by the network virtualization. The physical storage provides the raw storage capacity Page 12 of 132 Copyright IBM and VMware
13 consumed by the virtualization infrastructure. For this design the physical infrastructure components are provided by SoftLayer bare metal and all components are supported on the VMware Hardware Compatibility Guide (HCG). 3.2 Virtual Infrastructure The physical infrastructure is consumed by the virtual infrastructure. The virtual infrastructure reflects the physical with three different components, compute virtualization, storage virtualization and network virtualization. Each of these interface with the respective component in the physical infrastructure. The virtual infrastructure is installed on each physical device to form a node, for example a compute node. All the virtual resources interface to the virtual infrastructure for access to the physical infrastructure. The virtual infrastructure is accessed by either the cloud admin or the infrastructure management component Infrastructure Management Infrastructure management provides the logic to ensure the maximum benefit derived for the virtual infrastructure. This includes functions such as pooling virtual infrastructure and moving virtual resources off a node for maintenance or in the case of node failure. It controls placement of virtual resources on nodes to balance load along with placement based on business rules. It is only accessed by the cloud admin. All other access to this component is through API access from other components. 3.4 Common Services Common services provides the services which are consumed by the other cloud management services. It includes, identity and access services, SMTP services, NTP services, domain name services and certificate authority. This component is also the primary interface to external systems. Common services can connect to the client s DNS for requests outside the domain managed by the cloud services. It connects to the external NTP service to synchronize it s NTP service with an outside stratum. A trust relationship can be established between the common services and the client s authentication service for common authentication to the cloud services. 3.5 Cloud Management Services The cloud management services provide the primary interface to the users to consume cloud services in addition to the orchestration engines to process the service requests. The self-service portal is used as the primary interface to view the available cloud services (the service catalog) as well as to obtain a view of existing cloud resources that are deployed. The service catalog is the list of available services that are managed by the service provider. The service provider is able to determine which services are available to specific users or groups. The process orchestration engine controls the steps required to perform a service. This includes actions such as, obtaining an approval or connecting to an operational service system as part of the process. The process orchestration engine calls the infrastructure orchestration engine to orchestrate the build of the virtual resources for a service. The software orchestration engine builds the software that runs on the virtual resources. 3.6 Operational Services Operational services provide monitoring, patching, log consolidation, log analysis, disaster recovery and backup services for the cloud management platform. The monitoring looks for issues with the cloud management platform and notifies cloud admin via alerts on the operations console as well as s via the external SMTP relay. The patching connects to the external patch repository in order to obtain update information in support of the security or stability of the cloud management platform. Log consolidation collects the logs from the cloud management platform into a central repository which the log analysis service then operates on to provide the cloud admin with diagnostic information. The backup service keeps copies of the cloud management platform outside of the virtual infrastructure so it can be restored in the event of failure or corruption. The disaster recovery service needs at least one cloud region separate to the Copyright IBM and VMware Page 13 of 132
14 central cloud to which the cloud management platform are replicated to. In the event of failure at the primary site, the cloud management platform is restarted at the cloud region. 3.7 Business Services The business services component provides the service provider with analytics on IT financials, business management and benchmarking aspects of the cloud. The IT financials provides the service provider with details of the total cost of cloud ownership. The business management functions provide metering and chargeback capabilities for the service provider by user or group. The benchmarking functions provide the ability for service providers to analyze where the IT spend for the cloud is going, where it could be in the future and paths that need to be taken to improve. Page 14 of 132 Copyright IBM and VMware
15 4 Logical Operational Model The logical operational model provides guidance as to the design elements required to meet the functional requirements. 4.1 Logical Operational Model Structure The design consists of two distinct elements. A central cloud, through which the user and service provider manage the entire cloud, and optionally, one or more associated cloud regions. Only the central cloud contains the self-service portal. Additional regions (cloud regions) are added to provide remote sites, or additional capacity beyond that of a single central cloud within the same site. Each cloud region is configured into the central cloud for management. On premises vsphere environments are connected to SoftLayer via either a VPN connection over the internet or dedicated links to form additional cloud regions. The design of on premises vsphere environments is outside the scope of this document. Figure 4 Logical Structure View Copyright IBM and VMware Page 15 of 132
16 Within a central cloud, the components interact with each other as follows: Figure 5 Component Interaction Diagram Both the central cloud and any additional cloud regions are built on SoftLayer. Page 16 of 132 Copyright IBM and VMware
17 Microsoft Active Directory + services vum VMware vrealize Suite vrli vrops SRM vdp 4.2 Central Cloud The central cloud hosts the primary portal through which users access the cloud services. It has connections to all remote regions. The functions in a central cloud map to the following software products described in more detail in this section. vrealize Business Business Management Business Services IT Financials IT Benchmarking Common Services Identity & Access Services Cloud Management Services Self Service Portal Operational Services Backup & Restore Domain Name Services Service Catalog Disaster Recovery NTP Services Process Orchestration Monitoring SMTP Services Software Orchestration Log Consolidation & Analysis Certificate Authority Services Infrastructure Orchestration Patching Infrastructure Management vcenter Compute Management Storage Management Network Management Virtual Infrastructure ESXi VSAN NSX Compute Virtualisation Storage Virtualisation Network Virtualisation SoftLayer Physical Compute Physical Infrastructure Physical Storage Physical Network Figure 6 Logical Operational Model 4.3 Physical Infrastructure The physical Infrastructure is broken up into compute, storage and network. The compute and storage areas are combined in the cluster architecture. The network area is described in the network transport section Cluster Architecture This design splits the physical layer into clusters. A cluster represents the aggregate of the compute and memory resources of all the hosts in the cluster. All hosts in the cluster share network and storage resources. The use of clusters allows workloads (user or management) to be placed onto specific hardware. Copyright IBM and VMware Page 17 of 132
18 Each cluster is managed as a single entity, so user workloads can be managed separately from management workloads. The design differentiates between the following types of clusters: Compute cluster (one or more) Management cluster Edge cluster Storage cluster Figure 7 Logical Cluster Structure Compute Clusters Compute clusters host the VMware SDDC on IBM Cloud users virtual machines (sometimes referred to as workloads or payloads). Each compute cluster is built using SoftLayer bare metal. The environment is scaled by adding nodes to the initial compute cluster up the maximum number of nodes per cluster (refer to the physical operational model for details). Once the maximum has been reached, additional compute clusters are added to the environment Management Cluster The management cluster houses the virtual machines that manage the cloud. Like the compute clusters, the management cluster is built using SoftLayer bare metal. These servers host vcenter Server, NSX Manager, NSX Controller, vrealize Operations Management, vrealize Log Insight, vrealize Automation, and other shared management components Edge Cluster Edge clusters connect the virtual networks (overlay networks) provided by NSX for vsphere and the external networks. This includes both north-south (into the environment from outside) and east-west (between management and compute clusters) communications. Edge clusters provide the following main functions: Page 18 of 132 Copyright IBM and VMware
19 Support on-ramp and off-ramp connectivity to physical networks Connect to client on premises environments Storage Cluster A storage cluster provides network-accessible storage via NFS. This is used for backup and log archive purposes. The compute, management and edge clusters utilize Virtual Storage Area Network (VSAN) which aggregates disks located in each node of the clusters Physical Network The physical and layer 2 networking is handled by SoftLayer. The SoftLayer physical fabric provides a robust IP transport layer with the following characteristics: Simplicity Scalability High bandwidth Fault-tolerant transport Simplicity The network infrastructure at SoftLayer is simplified and standardized on three physical networks containing public, private, and out of band management (IPMI) traffic. Both the private and public networks are deployed to utilize up to 20Gbps bandwidth per physical host. The out of band management network is connected with a 1Gbps link per host. Upon ordering infrastructure components within SoftLayer, VLANs are provisioned for each of the three networks mentioned. If existing VLANs exist within an environment and there is enough space for a bare metal device to be placed in the same pod, SoftLayer will automatically assign the new device an IP on the same VLAN. The design incorporates SoftLayer portable subnets to provide IP addressing for virtual machines as well as IP addresses for the bare metal hosts. SoftLayer has also standardized the networking infrastructure using best-of-breed networking vendors. As a result, SoftLayer is able to implement and reuse automation patterns to setup, configure, and monitor the network infrastructure. Some of this automation has been exposed via API and is used by this design to simplify management tasks Scalability The SoftLayer network is designed in a multi-tier model. Each rack in a SoftLayer datacenter contains 2 frontend customer switches (FCS) and 2 backend customer switches (BCS) connected to the public and private networks, respectively. These customer switches then connect to separate, peered aggregation switches; the aggregated switches are then attached to a pair of separate routers for L3 networking. This multi-tier design allows the network to scale across racks, rows, and pods within the SoftLayer datacenter High Bandwidth Every upstream network port in the SoftLayer datacenter has multiple 10Gbps or 40Gbps connections. Every rack is terminated with multiple 10Gbps or 40Gbps connections to the public Internet and multiple 10Gbps or 40Gbps connections to the private network Fault-tolerant transport Redundancy is provided at the server level using 2x10Gbps NICs. Additionally, the backend server, frontend server, aggregation switches and routers are redundantly connected Physical Storage There are two types of storage used within this design: VSAN and NFS. Copyright IBM and VMware Page 19 of 132
20 VSAN VMware VSAN is a distributed storage technology which enables the local drives on each vsphere host to be aggregated and pooled into a shared-nothing storage device accessible to all hosts in a VSAN cluster. Additionally, the VSAN service is contained within the physical hosts, eliminating the need for external shared storage NFS Network File System (NFS) is a protocol for file based storage which allows a user on a client computer to access files over a network much like local storage is accessed. In this case, the client computer is an ESXi host which utilizes NFS as a backend datastore, and the storage is provided by a NFS-capable external storage array residing within the same SoftLayer site as the central cloud or cloud region. 4.4 Virtual Infrastructure The Virtual Infrastructure layer includes the software to provide compute, network and storage virtualization. VMware products in this layer are vsphere ESXi, VMware VSAN and VMware NSX. The designs virtual infrastructure consists of one Central Cloud and may contain several Cloud Regions that are managed by the Central Cloud. Each central cloud or cloud region has the following components: Management cluster. Houses compute resources for the management and operation of the cloud. Edge cluster. Provides external connections for the other clusters and user virtual network resources via VMware NSX. Compute (workload) cluster. Houses user virtual compute and storage Compute Virtualization Compute virtualization provides the virtualization of CPU and memory for consumption by virtual machines. This is provided by VMware vsphere ESXi and vcenter Server software Provisioning By virtualizing the hardware, new servers are able to be provisioned much faster than traditional methods. In addition, adding or removing elements like CPU or memory allocated to a virtual machine can all be done in software Resource scheduling The virtual compute is disengaged from the physical hardware which allows for precise control of the hardware in which the virtual machine is housed. Distributed Resource Scheduling (DRS) is used to distribute and balance the virtual machines across the cluster continually Availability As the virtual machine has been disengaged from the hardware, additional capabilities become possible for availability. Without compute virtualization, if the hardware had a failure, the Page 20 of 132 Copyright IBM and VMware
21 operating system and any applications running on it could be impacted. By using vsphere virtualization, virtual machines can be re-started on remaining hosts in a cluster in the event of the catastrophic failure of a host. This can also be used by the cloud admin to take a host offline for maintenance without affecting workloads on the cluster Performance The amount of physical resources available to virtual machines can be controlled through vcenter resource pools. This allows for different resource pools which can have higher or lower priority of physical resources based on share allocation Workload Movement By linking a central cloud with one or more cloud regions, vcenter Server allows a virtual machine to be migrated to a remote cloud region or from cloud region to cloud region. This is also possible from an on premises installation attached to the same environment. This enables workloads to be migrated from client premises into the cloud and back again Storage Virtualization Storage virtualization provides two levels of virtualization. The first is the virtualization of the storage arrays and the second is the virtualization of the block storage used by virtual machines Virtual Storage Area Network (VSAN) Virtual Storage Area Networking (VSAN) emulates a physical storage area network entirely within the virtualization layer. Each host in the cluster contains local drives that are combined in software to behave as a single disk array that is shared between all the hosts in the cluster as a shared datastore. Since there is no physical storage area network, VSAN has the advantage of fewer components (no external drive array, fiber cabling, etc.). It allows ease of scaling when adding new compute nodes with less administration like performing tasks such as, LUN allocation which are no longer necessary. Plus, VSAN provides high performance since local disk is used and disk I/O is spread out across all hosts within a cluster. Storage policies are used to define storage attributes such as performance and protection levels. The policy is set per virtual machine allowing great flexibility with the service levels available Virtual Machine Disks (VMDK) Each virtual machine has at least one virtual machine disk (VMDK). Additional disks can be added to a virtual machine. The virtual disks are provisioned on to the datastores provided by VSAN. All virtual disks are thin provisioned, so unused disk space within a single virtual disk does not take up datastore disk capacity Network Virtualization Network virtualization provides a network overlay that exist within the virtual layer. As a result, it can provide more rapid provisioning, deployment, re-configuration and destruction over physical devices. Copyright IBM and VMware Page 21 of 132
22 Network Virtualization Components The network virtualization architecture of this design utilizes VMware NSX for vsphere and vsphere Distributed Switches (vds). The virtualized network is organized hierarchically, with the following components from bottom to top: Data plane with the NSX vswitch and additional components Control plane with the NSX Controller Management plane with the NSX Manager Consumption plane with a Cloud management portal Figure 8 Network Virtualization Distributed Virtual Switches This design implements vsphere distributed switches. vsphere Distributed Switch (vds) offers several enhancements over standard virtual switches. Centralized management. Because distributed switches are created and managed centrally on a vcenter Server system, they make the switch configuration more consistent across ESXi hosts. Centralized management saves time, reduces mistakes, and lowers operational costs. Additional features. Distributed switches offer features that are not available on standard virtual switches. Some of these features can be useful to the applications and services that are running in the organization s infrastructure. For example, NetFlow and port mirroring provide monitoring and troubleshooting capabilities to the virtual infrastructure. Page 22 of 132 Copyright IBM and VMware
23 Distributed virtual switch implements health checks. The health check service helps identify and troubleshoot configuration errors in vsphere distributed switches. Health check helps identify the following common configuration errors: Mismatched VLAN trunks between a vsphere distributed switch and physical switch. Mismatched MTU settings between physical network adapters, distributed switches, and physical switch ports. Mismatched virtual switch teaming policies for the physical switch port-channel settings. Health check monitors VLAN, MTU, and teaming policies: VLANs. Checks whether the VLAN settings on the distributed switch match the trunk port configuration on the connected physical switch ports. MTU. For each VLAN, checks whether the physical access switch port MTU jumbo frame setting matches the distributed switch MTU setting. Teaming policies. Checks whether the connected access ports of the physical switch that participate in an EtherChannel are paired with distributed ports whose teaming policy is IP hash. Health check is limited to the access switch port to which the distributed switch uplink connects. With network I/O control, the distributed switch allocates bandwidth for the following system traffic types: vsphere vmotion traffic Management traffic VMware vsphere Replication traffic NFS traffic VMware Virtual SAN traffic vsphere Data Protection backup traffic Virtual machine traffic Fault tolerance traffic iscsi traffic Network I/O control details The bandwidth for each network resource pool is controlled by setting the physical adapter shares and host limits. The bandwidth for virtual machines is controlled by bandwidth reservation for an individual VM, similar to the way memory and CPU reservation is used. The physical adapter shares assigned to a network resource pool determine the share of the total available bandwidth guaranteed to the traffic that is associated with that network resource pool. The share of transmit bandwidth that is available to a network resource pool is determined by these factors: The network resource pool's shares. Other network resource pools that are actively transmitting Data Plane The NSX data plane consists of the NSX vswitch, which is based on the vsphere Distributed Switch (vds) and includes additional components. These components include kernel modules (VIBs), which run within the ESXi kernel and provide services such as virtual distributed router (VDR) and distributed firewall (DFW). The NSX kernel modules also enable Virtual Extensible LAN (VXLAN) capabilities. The NSX vswitch abstracts the physical network and provides access-level switching in the hypervisor. It is central to network virtualization because it enables logical networks that Copyright IBM and VMware Page 23 of 132
24 are independent of physical constructs such as VLAN. The NSX vswitch provides multiple benefits. Three types of overlay networking capabilities: Creation of a flexible logical Layer 2 overlay over existing IP networks on existing physical infrastructure. Support for east/west and north/south communication while maintaining isolation between tenants. Support for application workloads and virtual machines that operate as if they were connected to a physical Layer 2 network. Support for VXLAN and centralized network configuration. A comprehensive toolkit for traffic management, monitoring and troubleshooting within a virtual network which includes port mirroring, NetFlow/IPFIX, configuration backup and restore, network health check, Quality of Service (QoS), and Link Aggregation Control Protocol (LACP) In addition to the NSX vswitch, the data plane also includes gateway devices (NSX Edge gateways), which can provide Layer 2 bridging from the logical networking space (VXLAN) to the physical network (VLAN). NSX Edge Gateway devices offer Layer 2, Layer 3, perimeter firewall, load-balancing and other services such as Secure Socket Layer (SSL), Virtual Private Network (VPN), and Dynamic Host Control Protocol (DHCP) Control Plane The NSX control plane runs in the NSX Controller, which enables unicast VXLAN and controlplane programming of elements such as VDR (virtual distributed router). Unicast support is necessary as the multicast IP range per each VLAN is limited within SoftLayer. The number of multicast or unicast IPs determines the number of VXLANs that can be provisioned. In all cases the controller is part of the control plane and does not have any data plane traffic passing through it. The controller nodes are deployed in a cluster per NSX Manager to enable high availability and scalability. A failure of one or all controller nodes does not impact data plane traffic Management Plane The NSX management plane consists of the NSX Manager, which is the single point of configuration, and the REST API entry-points. NSX Manager integrates with vcenter. There is one NSX Manager per vcenter Server Consumption Plane Different actors interact with NSX for vsphere to access and manage the associated services in different ways: Cloud admin can manage the NSX environment from the vsphere Web Client. Users can consume the network virtualization capabilities of NSX for vsphere through the CMP (vrealize Automation) UI when deploying applications Network Virtualization Services Network virtualization services include logical switches, logical routers, logical firewall, and other components of NSX for vsphere. Page 24 of 132 Copyright IBM and VMware
25 Logical Switches Cloud deployments have a variety of applications that are used across multiple tenants. These applications and tenants require isolation from each other for security, fault isolation, and overlapping IP addresses. The NSX for vsphere logical switch creates logical broadcast domains or segments to which an application or tenant virtual machine can be logically wired. This allows for flexibility and speed of deployment while still providing all the characteristics of a physical network's broadcast domains (VLANs) without physical Layer 2 sprawl or spanning tree issues. A logical switch is distributed and can span arbitrarily large compute clusters. This allows for virtual machine mobility (migration with vmotion) within a region and between regions, without limitations of the physical Layer 2 (VLAN) boundary Logical Routers Dynamic routing provides the necessary forwarding information between Layer 2 broadcast domains, thereby allowing the cloud admin to decrease the size of Layer 2 broadcast domains and improve network efficiency and scale. NSX for vsphere extends this intelligence to where the workloads reside for east/west routing. This allows more direct VM-to-VM communication without the costly need to extend hops. At the same time, logical routers provide north/south connectivity, thereby enabling users to access public networks Logical Firewall NSX for vsphere Logical Firewall provides security mechanisms for dynamic virtual datacenters. The Distributed Firewall component of Logical Firewall allows a cloud admin to segment virtual datacenter entities like virtual machines based on VM names and attributes, user identity, vcenter objects like datacenters, and hosts, or based on traditional networking attributes like IP addresses, port groups, and so on. The Edge Firewall component helps a cloud admin to meet key perimeter security requirements, such as building DMZs based on IP/VLAN constructs, tenant-to-tenant isolation in multi-tenant virtual datacenters, Network Address Translation (NAT), partner (extranet) VPNs, and user-based SSL VPNs. The Flow Monitoring feature displays network activity between virtual machines at the application protocol level. The cloud admin can use this information to audit network traffic, define and refine firewall policies, and identify threats to a client s network Logical Virtual Private Networks (VPN s) SSL VPN-Plus allows remote users to access private corporate applications. IPSec VPN offers site-to-site connectivity between an NSX Edge instance and remote sites. L2 VPN allows users to extend their datacenter by allowing virtual machines to retain network identity across geographical boundaries Logical Load Balancers The NSX Edge load balancer enables network traffic to follow multiple paths to a specific destination. It distributes incoming service requests evenly among multiple servers in such a way that the load distribution is transparent to users. Load balancing thus helps in achieving optimal resource utilization, maximizing throughput, minimizing response time, and avoiding overload. NSX Edge provides load balancing up to Layer Service Composer Service Composer helps provision and assign network and security services to applications in a virtual infrastructure. The service provider maps these services to a security group, and the services are applied to the virtual machines in the security group. Copyright IBM and VMware Page 25 of 132
26 4.5 Infrastructure Management The infrastructure management element manages the compute, network and storage virtual resources provided by the lower layer. It also provides consolidation services to the upper layers for operational services. These functions are provided by VMware vcenter Server Compute Management In this design, VMware vcenter is employed to centralize the management of the compute resources within each ESXi host. While the ESXi hosts can be managed individually, placing them under vcenter control enables the following capabilities: Centralized control and visibility of all aspects within managed ESXi hosts and virtual machines. Provides the single pane of glass interface view via the vcenter web client for compute, network and storage management. Proactive Optimization. Enables allocation and optimization of resources for maximum efficiency across the ESX hosts. See section Compute Virtualization for optimization features enabled by vcenter Server. Extended management function for other integrated products and services such as VMware NSX, VMware Data Protection, VMware Update Manager and others as snap-ins extending the vcenter web interface. Monitoring, alerting, scheduling. Cloud admins can view events, alerts within the vcenter web client, and configure scheduled actions. Automation engine. VMware vcenter is the engine which performs the tasks given to it via the vsphere API web interface. VMware vrealize Automation and vrealize Orchestration are examples of applications that drive vcenter actions via the API Storage Management VMware vcenter enables centralized storage management within this design which allows for configuration and management of the following storage types: Local disk storage. Local hard disk drives (HDD) or solid state drives (SDD) that are attached to the local ESXi hosts. Storage area attached storage (SAN). Remote block storage that is attached to the ESXi host via fiber channel or TCPIP protocols. Network attached storage. (NAS) File based storage that is attached to the ESXi hosts via the NFS protocol. Virtual SAN Storage. Configured within the cluster object in vcenter, enables the aggregation of local disk storage across ESXi hosts into a shared pool of storage across all ESX hosts within a given cluster. Once configured, an outage of the vcenter server does not affect the availability of VSAN storage to the cluster. Within this design vcenter management of storage is primarily focused on NAS and VSAN storage as SAN is not employed. Only the ESXi host OS and swap space are used on local non VSAN disk storage NFS Storage management vcenter is responsible for configuring the mounting of NFS data stores to each ESXi host within a cluster. This ensures its access availability to any virtual machine with virtual disk files (VMDK) Page 26 of 132 Copyright IBM and VMware
27 residing on the NFS based data store, should a vmotion of the virtual machine from one ESXi host to another occur within a cluster VSAN Storage management Here again, the vcenter interface or web API has the capability of configuring VSAN data stores for a particular cluster at the cluster object level within the vcenter interface. Configuring VSAN within vcenter involves the following areas of configuration: Licensing. Prior to enabling VSAN, a valid license within vcenter licensing section is required. VSAN network. Used to configure the network VSAN will use for its back plane network. Virtual machines are made storage fault tolerant across ESXi hosts local disks on this network). Disk group configuration. On each ESXi host that contributes its local disks to a Virtual SAN cluster, disks are organized into disk groups. A disk group is a main unit of storage on a host. Each disk group includes one SSD and one or multiple HDDs. VSAN Policies. Storage policies define the virtual machine storage characteristics. Storage characteristics specify different levels of service for different virtual machines Network Management vcenter Server is used to create standard and distributed virtual switches. The virtual switches connect virtual machine (VM) network interfaces to portgroups allowing for communication between VM s hosted on the same host or different hosts. To establish communication between hosts, virtual switches need to be connected to physical uplinks which are the network interfaces of the ESXi hosts. VM s connected on the same virtual switch and hosted on the same host can communicate directly without the need of an external uplink. vcenter Server enables creation of distributed portgroups for virtual machines (aggregated virtual ports with a particular set of specifications). 4.6 Common Services Common services provide the services used by other services in the cloud management platform. This includes identity and access services, domain name services, NTP services, SMTP services and Certificate Authority Services Identity and Access Services In this design, Microsoft (MS) Active Directory (AD) is employed to provide authentication and directory services back end to the VMware Platform Service Controller (PSC) and VMware identity appliance. Within this design the VMware software components authenticate against the identity appliance which in turn authenticates against the MS AD service. The AD in this design can be extended to other regions by adding an additional AD server for that particular region s subdomain. Copyright IBM and VMware Page 27 of 132
28 4.6.2 Domain Name Services Domain Name Services (DNS) within this design are for the cloud management and infrastructure components only. DNS in this design serves as host name to IP resolution for the cloud management platform and service resolution for the AD components. When an instance of this design is tied to a customer on premises solution, this design s DNS servers are referenced by the on premises DNS infrastructure in addition to acting as a proxy for the customer s DNS infrastructure NTP Services This design s NTP servers are a sub stratum of the SoftLayer infrastructure NTP server. They serve to ensure that all components are in time synchronization for the needs of authentication, replication, clustering, log synchronization and certificate services. This includes physical and virtual components SMTP Services Simple Mail Transfer Protocol (SMTP), is utilized within this design by various components for the needs of outbound notification only. For inbound requirements (vrealize Automation, vrealize Business), the customer servers need to be configured Certificate Authority Services An Enterprise Certificate Authority (CA) based on Microsoft (MS) CA services built into MS Windows is employed in this solution to replace self-signed certificates for web interfaces within this design. 4.7 Cloud Management Services The cloud management services provide the service catalog, self-service portal and orchestration. This is provided by VMware vrealize Automation, vrealize Orchestrator and Rapid Deployment Services (RDS) pattern automation Service Catalog The service catalog is published through the self service catalog and allows users to request the provided services which can include provisioning new virtual machines from templates, provisioning new environments consisting of one or more virtual machines with software products as blueprints (also known as patterns), or managing existing deployed resources. Advanced services are also available through the service catalog by calling the orchestration component for process orchestration. The service provider role is able to customize the services available to users as well as publish additional services Self-Service Portal The self service portal provides a single point of access for users to the VMware SDDC on IBM Cloud solution. Authentication to the portal is performed against the Active Directory service. Page 28 of 132 Copyright IBM and VMware
29 4.7.3 Infrastructure and Process Orchestration Orchestration is provided by vrealize Orchestrator. It allows for tasks and remediation actions to be automated including integration with third party IT operations software. vrealize Orchestrator consists of: Workflow designer which incorporates an easy-to-use drag and drop interface to assemble workflows. The designer runs on Windows, Linux and Mac OS desktops. Scripting designer which allows for new building blocks to be created or imported for the vrealize Orchestrator platform. Orchestration engine which runs the workflows and associated scripts. The default implementation includes a built-in workflow library with common tasks. Workflows are able to be versioned and packaged to assist with change management Software Orchestration Software Orchestration is provided by a Rapid Deployment Services (RDS) solution with IBM Open Patterns. RDS implements a distributed file repository and the configuration management tools to deliver IBM Open Patterns on deployed workloads. IBM Open Patterns describe the pre-defined architecture of an application. For each component of the application (i.e. database, web server, etc.), the pattern defines: Pre-installation on an operating system Pre-integration across components Pre-configured & tuned Pre-configured Monitoring Pre-configured Security Lifecycle Management 4.8 Operational Services Operational services provide management of the cloud services. This includes backup & restore, disaster recovery, monitoring, log consolidation & analysis and patching functions Backup and Restore The data protection service protects the infrastructure that provides the virtualization, operations, security and cloud services. It does not protect any deployed user virtual machines. Data protection solutions provide the following functions in the design: Back up and restore virtual machines and database applications. Store data according to company retention policies. Copyright IBM and VMware Page 29 of 132
30 Inform administrators about backup and restore activities through reports. vsphere Data Protection provides the data protection service in each region. This is separate to disaster recovery and applies even if only the central cloud exists. An FTP server is used to backup NSX Manager. The FTP server supports SFTP and FTP protocols Disaster Recovery Figure 9 Dual-Region Data Protection Architecture The disaster recovery service adds to the data protection service by protecting the management services in the case of a complete site failure. It is an optional service to provide additional protection. Since this requires more than one site, it is only applicable where a central cloud and at least one cloud region has been included. VMware Site Recovery Manager (SRM) and vsphere Replication are used to provide this service, together with keeping the same IP addressing of the cloud management services at both sites. Note: Each central cloud or cloud region in this design is equivalent to the site construct in Site Recovery Manager. Since the central cloud contains the portal and manages the services in all the regions, the following applications are in scope of disaster recovery protection: Page 30 of 132 Copyright IBM and VMware
31 vrealize Automation together with VMware vrealize Orchestrator Analytics cluster of vrealize Operations Manager The services that support the services at each site do not require disaster recovery protection. This includes; vsphere, NSX and vcenter services which manage the services at the local site only. Authentication, DNS and NTP which is distributed to the cloud regions anyway. vrealize Log Insight and Software Orchestration which is replicated to all cloud regions Figure 10 Disaster Recovery Architecture Copyright IBM and VMware Page 31 of 132
32 4.8.3 Monitoring vrealize Operations Manager is used to track and analyze the operation of multiple data sources within the design by using specialized analytics algorithms. These algorithms help vrealize Operations Manager to learn and predicts the behavior of every object it monitors. Users access this information by using views, reports, and dashboards. vrealize Operations Manager contains functional elements that collaborate for data analysis and storage, and support creating clusters of nodes with different roles. Figure 11 vrealize Operations Manager Architecture For high availability and scalability, several vrealize Operations Manager instances are deployed in the management cluster where they have the following roles: Master Node. Required initial node in the cluster. In large-scale environments the master node manages all other nodes. In small-scale environments, the master node is the single standalone vrealize Operations Manager node. Master Replica Node. Enables high availability of the master node. Data Node. Enables scale-out of vrealize Operations Manager in larger environments. Data nodes have adapters installed to perform collection and analysis. Data nodes also host vrealize Operations Manager management packs. Remote Collector Node. Enables navigation through firewalls, interfaces with a remote data source, reduces bandwidth across regions, or reduces the load on the vrealize Operations Manager analytics cluster. Remote collector nodes only gather objects for the inventory and forward collected data to the data nodes. Remote collector nodes do not store data or perform analysis. In addition, they can be installed on a different operating system than the rest of the cluster nodes. The master and master replica nodes are data nodes with extended capabilities. Page 32 of 132 Copyright IBM and VMware
33 vrealize Operations Manager forms two types of cluster according to the nodes that participate in a cluster: Analytics clusters. Tracks, analyzes, and predicts the operation of monitored systems. Consists of a master node, data nodes, and master replica node. Remote collectors cluster. Only collects diagnostics data without storage or analysis. Consists only of remote collector nodes. The functional components of a vrealize Operations Manager instance interact to provide analysis of diagnostics data from the datacenter and visualize the result in the Web user interface. Table 3 vrealize Operations Manager Logical Node Architecture Architecture Component Diagram Description Admin / Product UI server. The UI server is a Web application that serves as both user and administration interface. REST API / Collector. The Collector collects data from all components in the datacenter. Controller. The Controller handles the data flow the UI server, Collector, and the analytics engine. Analytics. The Analytics engine creates all associations and correlations between various data sets, handles all super metric calculations, performs all capacity planning functions, and is responsible for triggering alerts. Persistence. The persistence layer handles the read and write operations on the underlying databases across all nodes. FSDB. The File System Database (FSDB) stores collected metrics in raw format. FSDB is available in all the nodes. xdb (HIS). The xdb stores data from the Historical Inventory Service (HIS). This component is available only on the master and master replica nodes. Global xdb. The Global xdb stores user preferences, alerts, and alarms, and customization that is related to the vrealize Operations Manager. This component is available only on the master and master replica nodes Log Consolidation and Analysis Log consolidation and analysis provides consolidation of the logs that are produced by each of the cloud services together with analysis of those logs. For this design, this function is provided by vrealize Log Insight. vrealize Log Insight provides real-time log management and log analysis with machine learning-based intelligent grouping, high-performance searching, and troubleshooting across physical, virtual, and cloud environments. Copyright IBM and VMware Page 33 of 132
34 vrealize Log Insight collects data from ESXi hosts using the syslog protocol. It connects to vcenter Server to collect events, tasks, and alarms data, and integrates with vrealize Operations Manager to send notification events and enable launch in context Patching Patching of the VMware software components is achieved with VMware Update Manager. This includes the VMware ESXi hosts, virtual appliances and management tooling in the design. It connects to the internet to obtain the latest vulnerability patches and automatically applies user-defined patches to the relevant components to eliminate the vulnerabilities. 4.9 Business Services Business services are those services that provides business functions. This includes business management, IT financials and IT benchmarking. vrealize Business (VRB) is configured to provide financial information, reporting and modeling. VRB integrates with vrealize Automation Business Management vrealize Business provides the following business management capabilities: Automatic private cloud metering costing and pricing IT Financials vrealize Business provides the following capability for financial management: Automatic service catalog pricing (Integrated with vrealize Automation) Private cloud consumption analysis Out-of-the-box reporting (Exportable data set) IT Benchmarking Additionally, vrealize Business can assist in modeling cost projections across cloud environments. Private cloud and public cloud cost comparison Page 34 of 132 Copyright IBM and VMware
35 4.10 Cloud Region The cloud region is a child instance of the design. It is not standalone and requires a central cloud to provide the cloud management services. Provisioning and management of virtual resources is done through the central cloud. The cloud management services do not exist for a cloud region since this is handled by the central cloud and the operational management services which contain collectors/relays to pass information back to the central cloud. Copyright IBM and VMware Page 35 of 132
36 5 Physical Operational Model The physical operational model elaborates by applying the non-functional requirements to the logical operational model. 5.1 Physical Layer Compute Figure 12 Physical Operational Model - Virtual Servers, Networking and Clusters The design leverages SoftLayer to provide the compute. This allows for flexibility in provisioning bare metal. Compute nodes are able to be deployed rapidly without needing orders to be delivered and the same nodes are able to be decommissioned without needing to wait for depreciation schedules or reselling. SoftLayer offers a variety of bare metal Intel based hardware from 1U to 4U chassis sizes, 2GB of memory up to 3TB and from 4 CPU cores up to 48 cores. For this design, a 2U server has been selected to allow for the lowest cost of entry, while still allowing for scaling up to 10,000 deployed VMs in a single central cloud or cloud region. For security and to isolate management, network and user workloads (resources) this design branches out to 3 vsphere cluster types with the following functions: Management Cluster Edge Cluster Compute Clusters This allows the scaling of one function independent of the other as the deployment is scaled out, making for a more effective use of resources. Page 36 of 132 Copyright IBM and VMware
37 Management Cluster The management cluster is the heart of operation and control of the design. It is sized from onset of deployment to allow for Compute cluster expansion and additional feature expansion without requiring additional sever nodes. It consists of 4 nodes of the following specification: 2 x 12 core CPUs (24 cores total), plus Hyperthreading 256GB RAM ~6.3TB Usable VSAN storage ~1TB Usable Local Disk for operating system Refer to Appendix A Bare Metal Summary for hardware details Edge Cluster VMware NSX is an integral part of the design. Edge virtual appliances are used where VPN end points or load balancing is required. Edge virtual appliances are controlled and can be dynamically provisioned as user applications are spun up with patterns or are preconfigured to support management functions. In either case they are deployed to the Edge cluster. This ensures that network connectivity and performance are not affected by varying workloads in the other clusters. It is sized from onset of deployment to allow for Compute cluster expansion without requiring additional servers. As the Edge virtual appliances are small, VSAN storage requirements are held to a minimum while maintaining redundancy and performance. It consists of 4 nodes of the following specification: 2 x 6 core CPUs (12 cores total), plus Hyperthreading 128GB RAM ~ 3.2TB Usable VSAN storage ~ 1TB Usable Local Disk for operating system Refer to Appendix A Bare Metal Summary for hardware details Compute Cluster As the users or administrators within the customer s organization deploy whatever applications they require via vrealize Automate, the compute workloads requested get deployed to this cluster. With scale up as well as out in mind, high resource intensive applications and other mixed workloads are able to be absorbed. Additional clusters are provisioned when the capacity of the each cluster is reached. It consists of 4 nodes of the following specification: 2 x 12 core CPUs (24 cores total), plus Hyperthreading 512GB RAM ~6.3TB Usable VSAN storage ~1TB Usable Local Disk for operating system Refer to Appendix A Bare Metal Summary for hardware details. Each node supports 80 VMs of 2 vcpu, 8GB RAM and 70GB Disk. Taking into account the following: CPU over-commit 7 90% CPU usage limit on ESXi Memory over-commit is % memory usage limit on ESXi Copyright IBM and VMware Page 37 of 132
38 5.1.2 Storage No disk over-commit Maximum of 48 nodes per cluster / Cluster and 3 clusters for 10,000 VM support. Cluster sizing based upon 1 node failure per 15 nodes. So up to 15 nodes, a single node is reserved in case of failure. Between 15 and 30 nodes, there are two nodes kept in reserve and so forth up to 45 nodes where 3 nodes are in reserve. Network File System (NFS) is a file system protocol that allows a user on a client computer to access files over a network much like local storage is accessed. In this case, the client computer is an ESXi host, and the storage is provided by a NFS-capable external storage array. The management cluster uses VMware Virtual SAN for primary storage and NFS for secondary storage. For compute clusters, the decision on which technology to use is based on the performance, capacity, and capabilities (replication, deduplication, compression, etc.) required by the workloads that are running in the clusters. For this design, additional storage for the ESXi management cluster is connected to SoftLayer File Storage on Performance Storage via 10 Gbps links with Jumbo Frames (MTU 9000) enabled. For the virtual machines that provide the backup service and log archiving (i.e., vsphere Data Protection and vrealize Log Insight) the NFS datastores are configured in the following manner: Table 4 NFS Configuration for vsphere Data Protection and vrealize Log Insight Product vsphere Data Protection vrealize Log Insight Configuration vsphere Data Protection is I/O intensive and is placed on its own, unique NFS datastore sized at 4TB vrealize Log Insight uses NFS datastores sized at 1TB for archive storage and can be shared will other virtual machines Network The design leverages SoftLayer physical networking and this is broken up into server connections, VLANs and MTU packet sizing Server connections Each compute node (physical server) within the design has two 10Gb Ethernet connection into each SoftLayer Top of Rack (ToR) switch (public and private) setup as individual connections (un-bonded) for a total of 4 x 10Gbps connections. This allows each networking interface card (NIC) connection to work independent of one another. Page 38 of 132 Copyright IBM and VMware
39 VLANs Figure 13 Network connections per physical node Four VLANs are included per design including the public network. They are as listed for this design: MTU Sizing Private VLAN 1 (default / untagged) Private VLAN 2 (trunked / tagged) Private VLAN 3 (trunked / tagged ) Public VLAN 4 (default / untagged) The private network connections are configured to use jumbo frame MTU size of This is maximum MTU allowed within VMware and SoftLayer limit which improves performance for large data transfers such as storage and vmotion. The public network connections use a standard Ethernet MTU frame size of This needs to be maintained as any changes may affect packet fragmentation over the internet. 5.2 Virtual Infrastructure Virtual infrastructure consists of compute, storage and network virtualization Compute Virtualization This design uses VMware vsphere ESXi version 6.0 u1 to virtualize the management, compute, and edge servers. The ESXi hypervisor is installed on the 2x1TB RAID-1 disk array contained on each server. RAID-1 is used in this design to provide redundancy for the vsphere hypervisor Storage Virtualization For this design, VMware Virtual Storage Area Network (VSAN) storage is employed for all storage needs with the exception of vrealize Log Insight log archival storage and VMware Data Protection backup storage. VSAN allows for the local storage across multiple ESXi hosts within a vsphere cluster to be represented as a single virtual machine datastore. VSAN supports only SATA, SAS HDD, and PCIe storage. Two 1TB SATA drives are excluded in each node regardless of which cluster it belongs to for the purpose of housing the ESXi installation. Copyright IBM and VMware Page 39 of 132
40 Figure 14 VSAN concept RAID Controller As of this version of the design, only the Avago MegaRAID i RAID controller within SoftLayer hardware is supported by VMware VSAN. Disk caching need is disabled and the controller is set in JBOD mode for all VSAN drives. Disks and disk groups Depending on the cluster function (Management, Edge, Compute) the number and sizes of the disks changes. Each VSAN disk group requires a solid state disk (SSD) for a cache layer. Within all clusters, a 2U server type is employed with a total of 12 drive slots maximum capacity. Excluding the ESXi OS drives, the following is the drive layout for each cluster type: Table 5 VSAN disk table Cluster type Number of VSAN disk groups Number of SSDs # of SSD + HDD per disk group Size of SSDs Number of HDDs Size of HDDs Management ,200GB 8 2,000GB Edge ,200GB 4 2,000GB Compute ,200GB 8 2,000GB Virtual network setup For this design the VSAN traffic will traverse between ESXi hosts on a dedicated private VLAN. No other traffic will occupy the VSAN VLAN. The two network adapters attached to the private network switch and configured in within vsphere as a virtual distributed switch (vds) with both network adapters as uplinks. A dedicated VSAN kernel port configured for the VSAN VLAN resides within the vds. Jumbo frame are enabled for the private vds Virtual SAN Policy Design Once VMware Virtual SAN is enabled and configured, create storage policies that define the virtual machine storage characteristics. Storage characteristics specify different levels of service Page 40 of 132 Copyright IBM and VMware
41 for different virtual machines. The default storage policy tolerates a single failure and has a single disk stripe. Use the default unless a client s environment requires policies with non-default behavior. If a custom policy is configured, Virtual SAN will guarantee it; however, if Virtual SAN cannot guarantee a policy, it is not possible to provision a virtual machine that uses the policy unless it is enabled to force provisioning. This design will use the default policy. Table 6 VSAN policies Capability Use Case Value Comments Number of failures to tolerate Number of disk stripes per object Flash read cache reservation (%) Object space reservation (%) Force provisioning Redundancy Performance Performance Thick provisioning Override policy Default 1 Max 3 Default 1 Max 12 Default 0 Max 100% Default 0 Max 100% Default: No A standard RAID 1 mirrored configuration that provides redundancy for a virtual machine disk. The higher the value, the more failures can be tolerated. For n failures tolerated, n+1 copies of the disk are created, and 2n+1 hosts contributing storage are required. A higher n value indicates that more replicas of virtual machines are made, which can consume more disk space than expected. A standard RAID 0 stripe configuration used to increase performance for a virtual machine disk. This setting defines the number of HDDs on which each replica of a storage object is striped. If the value is higher than 1, increased performance can result. However, an increase in system resource usage might also result. Flash capacity reserved as read cache for the storage is a percentage of the logical object size that will be reserved for that object. Only use this setting for workloads if a client must address read performance issues. The downside of this setting is that other objects cannot use a reserved cache. VMware recommends not using these reservations unless it is absolutely necessary because unreserved flash is shared fairly among all objects. The percentage of the storage object that will be thick provisioned upon VM creation. The remainder of the storage will be thin provisioned. This setting is useful if a predictable amount of storage will always be filled by an object, cutting back on repeatable disk growth operations for all but new or non-predictable storage use. Force provisioning allows for provisioning to occur even if the currently available cluster resources cannot satisfy the current policy. Force provisioning is useful in case of a planned expansion of the Virtual SAN cluster, during which provisioning of VMs must continue. Virtual SAN Copyright IBM and VMware Page 41 of 132
42 Capability Use Case Value Comments automatically tries to bring the object into compliance as resources become available. By default, policies are configured based on application requirements. However, they are applied differently depending on the object. Table 7 VSAN object policy defaults Object Policy Comments Virtual machine namespace Failures-to-Tolerate: 1 Configurable. Changes are not recommended. Swap Failures-to-Tolerate: 1 Configurable. Changes are not recommended. Virtual disk(s) Virtual disk snapshot(s) User-Configured Storage Policy Uses virtual disk policy Can be any storage policy configured on the system. Same as virtual disk policy by default. Changes are not recommended. If a user-configured policy is not specified, the default system policy of 1 failure to tolerate and 1 disk stripe is used for virtual disk(s) and virtual disk snapshot(s). Policy defaults for the VM namespace and swap are set statically and are not configurable to ensure appropriate protection for these critical virtual machine components. Policies must be configured based on the application s business requirements. Policies give Virtual SAN its power because it can adjust how a disk performs on the fly based on the policies configured Network Virtualization This design uses vsphere virtual distributed switches and VMware NSX for vsphere to implement virtual networking. By using NSX, this design implements software-defined networking Virtual Distributed Switch Design The design uses a minimum number of switches. Clusters connected to public network is configured with two distributed virtual switches. Clusters restricted to private network have only one distributed virtual switch. Separating different types of traffic is required to reduce contention and latency. Separate networks are also required for access security. VLANs are used to segment physical network functions. This design uses four (4) VLANs. Three (3) for private network traffic and one (1) for public network traffic. Traffic separation is detailed in the section below. Table 8 VLAN Mapping to Traffic Types VLAN VLAN1 VLAN2 VLAN3 VLAN4 Traffic Type ESXi Management, VXLAN (VTEP) VSAN vmotion, NFS, vsphere Replication All Internet access Traffic from workloads will travel on VXLAN backed logical switches. Page 42 of 132 Copyright IBM and VMware
43 Management Cluster Distributed Switches The management cluster uses a dual vsphere Distributed Switch with the configuration settings shown in this section. Table 9 Management Cluster Distributed Switch vsphere Distributed Switch Name vds-mgmt- Priv Function ESXi management Network IP Storage (NFS) Virtual SAN vsphere vmotion VXLAN Tunnel Endpoint (VTEP) vsphere Replication/vSphere Replication NFC Network I/O Control Enabled Load Balancing Mode Based on source MAC hash Number of Physical NIC Ports MTU 2 9,000 (Jumbo Frame) vds-mgmt- Pub External management traffic (North-South) Enabled Based on source MAC hash 2 1,500 (default) Table 10 Management Cluster Distributed Switch Port Group Configuration Settings Parameter Setting Load balancing Route based on the source MAC hash Failover detection Link status only Notify switches Enabled Failback No Failover order Active uplinks: Uplink1, Uplink2 Management cluster hosts are connected to both private and public networks. There are 2 distributed virtual switches: one for private network and one for public network. Each switch has a dedicated pair of 10Gbps network adapters. Copyright IBM and VMware Page 43 of 132
44 Figure 15 Network Switch Design for Management Hosts Page 44 of 132 Copyright IBM and VMware
45 Table 11 Management Virtual Switch Port Groups and VLANs vsphere Distributed Switch vds-mgmt-priv Port Group Name vds-mgmt-priv- Management vds-mgmt-priv vds- Mgmt-Priv - vmotion vds-mgmt-priv vds- Mgmt-Priv - VSAN vds-mgmt-priv vds- Mgmt-Priv - VTEP vds-mgmt-priv vds- Mgmt-Priv - NFS vds-mgmt-priv vds- Mgmt-Priv - Replication vds-mgmt-pub vds- Mgmt-Pub - External Teaming Uplinks VLAN ID Source MAC hash Source MAC hash Source MAC hash Source MAC hash Source MAC hash Source MAC hash Source MAC hash Active: 0, 1 Active: 0, 1 Active: 0, 1 Active: 0, 1 Active: 0, 1 Active: 0, 1 Active: 0, 1 VLAN1 VLAN3 VLAN2 VLAN1 VLAN3 VLAN3 VLAN4 Table 12 Management VMkernel Adapter vsphere Distributed Switch Network Label Connected Port Group vds-mgmt-priv Management vds- Mgmt-Priv - Management vds- Mgmt-Priv vmotion vds- Mgmt-Priv - vmotion vds- Mgmt-Priv VTEP vds- Mgmt-Priv - VTEP vds- Mgmt-Priv VSAN vds- Mgmt-Priv - VSAN Enabled Services Management Traffic vmotion Traffic 9,000-9,000 VSAN 9,000 MTU 1,500 (default) Copyright IBM and VMware Page 45 of 132
46 Edge Cluster Distributed Switches The edge cluster uses a dual vsphere Distributed Switch with the configuration settings shown in this section. Table 13 Edge Cluster Distributed Switch vsphere Distributed Switch Name Function vds-edge-priv ESXi management Virtual SAN vsphere vmotion VXLAN Tunnel Endpoint (VTEP) Network I/O Control Enabled Load Balancing Mode Based on source MAC hash Number of Physical NIC Ports MTU 2 9,000 (Jumbo Frame) vds-edge-pub External user traffic (North- South) Enabled Based on source MAC hash (default) Table 14 Management Cluster Distributed Switch Port Group Configuration Settings Parameter Setting Load balancing Route based on the source MAC hash Failover detection Link status only Notify switches Enabled Failback No Failover order Active uplinks: Uplink1, Uplink2 Page 46 of 132 Copyright IBM and VMware
47 Edge cluster hosts are connected to both private and public networks. There are 2 distributed virtual switches: one for private network and one for public network. Each switch has a dedicated pair of 10Gbps network adapters. Figure 16 Network Switch Design for Edge Hosts Copyright IBM and VMware Page 47 of 132
48 Table 15 Edge Virtual Switch Port Groups and VLANs vsphere Distributed Switch vds-edge-priv Port Group Name vds-edge-priv- Management vds-edge-priv vds-edge-priv - vmotion vds-edge-priv vds-edge-priv vds-edge-pub vds-edge -Priv- VSAN vds-edge-priv- VTEP vds-edge-pub- External Teaming Uplinks VLAN ID Source MAC hash Active: 0, 1 VLAN1 Source MAC hash Active: 0, 1 VLAN3 Source MAC hash Active: 0, 1 VLAN2 Source MAC hash Active: 0, 1 VLAN1 Source MAC hash Active: 0, 1 VLAN4 Table 16 Edge VMkernel Adapter vsphere Distributed Switch Network Label Connected Port Group vds-edge-priv Management vds-edge-priv - Management vds-edge-priv vmotion vds-edge-priv - vmotion vds-edge-priv VTEP vds-edge-priv - VTEP vds-edge-priv VSAN vds-edge-priv - VSAN Enabled Services Management Traffic vmotion Traffic 9,000-9,000 VSAN 9,000 MTU 1,500 (default) Compute Cluster Distributed Switches The compute cluster uses a single vsphere Distributed Switch with the configuration settings shown in this section. Table 17 Compute Cluster Distributed Switch vsphere Distributed Switch Name vds-compute- Priv Function ESXi management Virtual SAN vsphere vmotion VXLAN Tunnel Network I/O Control Enabled Load Balancing Mode Based on source MAC hash Number of Physical NIC Ports MTU 2 9,000 (Jumbo Frame) Page 48 of 132 Copyright IBM and VMware
49 vsphere Distributed Switch Name Function Network I/O Control Load Balancing Mode Number of Physical NIC Ports MTU Endpoint (VTEP) Table 18 Compute Cluster Distributed Switch Port Group Configuration Settings Parameter Setting Load balancing Route based on the source MAC hash Failover detection Link status only Notify switches Enabled Failback No Failover order Active uplinks: Uplink1, Uplink2 Compute cluster hosts are connected to the private network. The switch has a dedicated pair of 10Gbps network adapters. Figure 17 Network Switch Design for Compute Hosts Copyright IBM and VMware Page 49 of 132
50 Table 19 Compute Virtual Switch Port Groups and VLANs vsphere Distributed Switch vds-compute- Priv vds- Compute -Priv vds- Compute -Priv Port Group Name vds-compute- Priv- Management vds-compute- Priv-vMotion vds-compute- Priv-VSAN Teaming Uplinks VLAN ID Source MAC hash Source MAC hash Source MAC hash Active: 0, 1 Active: 0, 1 Active: 0, 1 VLAN1 VLAN3 VLAN2 vds- Compute -Priv vds-compute- Priv-VTEP Source MAC hash Active: 0, 1 VLAN1 Table 20 Compute VMkernel Adapter vsphere Distributed Switch vds-compute- Priv vds-compute- Priv vds-compute- Priv vds-compute- Priv Network Label Management vmotion VTEP VSAN Connected Port Group vds-compute- Priv- Management vds-compute- Priv-vMotion vds-compute- Priv-VTEP vds-compute- Priv-VSAN Enabled Services Management Traffic vmotion Traffic 9,000-9,000 VSAN 9,000 MTU 1,500 (default) NIC Teaming With the exception of VSAN, this design uses NIC teaming to avoid single point of failure and provide load balancing. To accomplish this, the design uses an active-active configuration with a route that is based on source MAC hash for teaming Network I/O Control This design uses Network I/O control enabled on all distributed switches with default shares. Network I/O control increases resiliency and performance of the network. Utilizing network I/O control, this design uses the distributed switch to allocate bandwidth for the following system traffic types: vsphere vmotion traffic Management traffic VMware vsphere Replication traffic NFS traffic VMware Virtual SAN traffic vsphere Data Protection backup traffic Virtual machine traffic Page 50 of 132 Copyright IBM and VMware
51 VXLAN This design uses VXLAN to create isolated, multi-tenant broadcast domains across datacenter fabrics and to enable customers to create elastic, logical networks that span physical network boundaries. VXLAN works by creating Layer 2 logical networks that are encapsulated in standard Layer 3 IP packets. A Segment ID in every frame differentiates the VXLAN logical networks from each other without any need for VLAN tags. As a result, large numbers of isolated Layer 2 VXLAN networks can coexist on a common Layer 3 infrastructure Software-Defined Network Design NSX offers the following Software-Defined Network (SDN) capabilities crucial to support the cloud management platform operations. load balancing firewalls routing logical switches VPN access Because NSX for vsphere is tied to a vcenter Server domain, this design uses two separate NSX instances. One instance is tied to the management vcenter Server, and the other instance is tied to the compute and edge vcenter Server. SDN capabilities are consumed via the cloud management platform, vsphere web client and API. The design uses API s to automate the deployment and configuration of NSX components by users and cloud admin actors NSX for vsphere Components This section describes the NSX for vsphere component configuration NSX Manager NSX Manager provides the centralized management plane for NSX for vsphere and has a one-toone mapping to a vcenter Server instance. This design uses two NSX managers (one for each vcenter Server in the design). NSX Manager performs the following functions. Provides the single point of configuration and the REST API entry-points in a vsphere environment for NSX for vsphere. Is responsible for deploying NSX Controller clusters, Edge distributed routers, and Edge service gateways in the form of OVF appliances, guest introspection services, and so on. Is responsible for preparing ESXi hosts for NSX for vsphere by installing VXLAN, distributed routing and firewall kernel modules, and the User World Agent (UWA). Communicates with NSX Controller clusters via REST and with hosts via the RabbitMQ message bus. The internal message bus is specific to NSX for vsphere and does not require setup of additional services. Generates certificates for the NSX Controller instances and ESXi hosts to secure control plane communications with mutual authentication NSX Controller The NSX Controllers perform the following functions: Provide the control plane to distribute VXLAN and logical routing information to ESXi hosts. Include nodes that are clustered for scale-out and high availability. Slice network information across cluster nodes for redundancy. Copyright IBM and VMware Page 51 of 132
52 Remove requirement of VXLAN Layer 3 multicast in the physical network. Provide ARP suppression of broadcast traffic in VXLAN networks. NSX for vsphere control plane communication occurs over the management network. This design implements a cluster of 3 NSX controllers for each NSX manager enabling high availability for the controllers NSX vswitch The NSX for vsphere data plane consists of the NSX vswitch. This vswitch is based on the vsphere Distributed Switch (vds) with additional components for add-on services. The add-on NSX for vsphere components include kernel modules which run within the hypervisor kernel and provide services such as distributed logical router (DLR) and distributed firewall (DFW), and enable VXLAN capabilities. This design uses NSX vswitch NSX Logical Switching NSX for vsphere logical switches create logically abstracted segments to which tenant virtual machines can be connected. A single logical switch is mapped to a unique VXLAN segment and is distributed across the ESXi hypervisors within a transport zone. It allows line-rate switching in the hypervisor without the constraints of VLAN sprawl or spanning tree issues. This design uses NSX Logical Switching for handling compute workloads and connectivity between different network zones Distributed Logical Router The NSX for vsphere Distributed Logical Router (DLR) is optimized for forwarding in the virtualized space, that is, forwarding between VMs on VXLAN- or VLAN-backed port groups. This design does not use distributed logical routers User World Agent The User World Agent (UWA) is a TCP (SSL) client that facilitates communication between the ESXi hosts and the NSX Controller instances as well as the retrieval of information from the NSX Manager via interaction with the message bus agent. UWA is installed on each ESXi host VXLAN Tunnel Endpoint VXLAN Tunnel Endpoints (VTEPs) are responsible for encapsulating VXLAN traffic as frames in UDP packets and for the corresponding de-encapsulation. VTEPs take the form of VMkernel ports with IP addresses and are used both to exchange packets with other VTEPs. This design uses a single VTEP per host Edge Services Gateway The primary function of the NSX for vsphere Edge services gateway is north-south communication, but it also offers support for Layer 2, Layer 3, perimeter firewall, load balancing and other services such as SSL-VPN and DHCP-relay. In this design, the edges also ensure east-west communication Distributed Firewall NSX for vsphere includes a distributed kernel-level firewall known as a distributed firewall. Security enforcement is done at the kernel and VM network adapter level. This enables firewall Page 52 of 132 Copyright IBM and VMware
53 rule enforcement in a highly scalable manner without creating bottlenecks on physical appliances. The distributed firewall has minimal CPU overhead and can perform at line rate. This design does not automatically implement distributed firewall. The cloud admin actor is able to enable this feature post implementation if required Logical Load Balancer The NSX for vsphere logical load balancer provides load balancing services up to Layer 7 (application), allowing distribution of traffic across multiple servers to achieve optimal resource utilization and availability. The logical load balancer is a service provided by the NSX Edge services gateway. This design implements load balancing for management virtual machines NSX for vsphere Physical Network Requirements VXLAN packets cannot be fragmented. Since VXLAN adds its own header information the MTU needs to be 1,600. SoftLayer has a limitation of 136 VXLAN addresses. As such, the VXLAN control plane in this design uses unicast mode to circumvent this limitation. The NSX Manager synchronizes with the same NTP server as the rest of the vsphere environment. This avoids time drift which can cause problems with authentication. The NSX Manager must be in sync with the vcenter Single Sign-On server NSX for vsphere Specifications The following table lists the components involved in the NSX for vsphere solution and the requirements for installing and running them. The compute and storage requirements have been taken into account when sizing resources to support the NSX for vsphere solution. Table 21 NSX Components Sizing VM vcpu Memory Storage Quantity per Deployment by Type NSX Manager 4 12 GB 60 GB 1 NSX Controller 4 4 GB 20 GB 3 NSX Edge services gateway - Quad Large NSX Edge services gateway X- Large 4 1 GB 512 MB GB 4.5 GB (+4 GB for SWAP) 2 The Quad Large model is utilized for high performance firewall. X-Large is utilized for both high performance load balancing and routing. Copyright IBM and VMware Page 53 of 132
54 Network Virtualization Conceptual Design The following diagram depicts the conceptual tenant architecture components and their relationship. Figure 18 Network Virtualization Conceptual Design The conceptual design has the following key components. External Networks. Connectivity to and from external networks is through a perimeter firewall. The main external network is the Internet. Perimeter Firewall. The logical firewall exists at the perimeter of the datacenter. Each tenant receives either a full instance or partition of an instance to filter external traffic. This is the primary access method for user data. Provider Logical Router (PLR). The PLR exists behind the perimeter firewall and handles north/south traffic that is entering and leaving a tenant. Tenant Edge services gateways. Edge services gateways that provide routing and firewalling capabilities. Internal Non-Tenant Networks. A single management network, which sits behind a perimeter firewall but not behind the PLR. Enables cloud admin to manage the cloud environment. Internal Tenant Networks. Connectivity for the main tenant workload. Page 54 of 132 Copyright IBM and VMware
55 Cluster Design for NSX for vsphere Management Stack In the management stack, the underlying hosts are prepared for NSX for vsphere. The management stack has these components: NSX Manager instances for both stacks (management stack and compute/edge stack), NSX Controller cluster for the management stack, NSX Edge service gateways for the management stack. Compute/Edge Stack In the compute/edge stack, the underlying hosts are prepared for NSX for vsphere. The compute/edge stack has these components: NSX Controller cluster for the compute stack All NSX Edge service gateways of the compute stack that are dedicated to handling the north/south traffic in the datacenter. A separate edge stack helps prevent VLAN sprawl because any external VLANs need only be trunked to the hosts in this cluster. Multiple compute clusters that run the tenant workloads and have the underlying hosts prepared for NSX for vsphere. Figure 19 Cluster Design for NSX for vsphere Copyright IBM and VMware Page 55 of 132
56 High Availability of NSX for vsphere Components The NSX Manager instances of both stacks run on the management cluster. vsphere HA protects the NSX Manager instances by ensuring that the VM is restarted on a different host in the event of primary host failure. The NSX Controller nodes of the management stack run on the management cluster and the NSX for vsphere Controller nodes of the compute stack run on the edge cluster. In both clusters, vsphere Distributed Resource Scheduler (DRS) rules ensure that NSX for vsphere Controller nodes do not run on the same host. The data plane remains active during outages in the management and control planes although the provisioning and modification of virtual networks is impaired until those planes become available again. The NSX Edge services gateways and distributed logical router control VMs of the compute stack are deployed on the edge cluster. The NSX Edge service gateways and distributed logical router control VMs of the management stack run on the management cluster. All NSX Edge components are deployed in NSX for vsphere HA pairs. NSX for vsphere HA provides better availability than vsphere HA. By default, the VMs fail over within 15 seconds versus a potential 5 minutes for a restart on another host under vsphere HA Logical Switch Control Plane Mode Design The control plane decouples NSX for vsphere from the physical network and handles the broadcast, unknown unicast, and multicast (BUM) traffic within the logical switches. The control plane is on top of the transport zone and is inherited by all logical switches that are created within it, although it is possible to override aspects of the control plane. The following options are available: Multicast Mode. The control plane uses multicast IP addresses on the physical network. Use multicast mode only when upgrading from existing VXLAN deployments. In this mode, PIM/IGMP must be configured on the physical network. Unicast Mode. The control plane is handled by the NSX Controllers and all replication occurs locally on the host. This mode does not require multicast IP addresses or physical network configuration. Hybrid Mode. This mode is an optimized version of the unicast mode where local traffic replication for the subnet is offloaded to the physical network. Hybrid mode requires IGMP snooping on the first-hop switch and access to an IGMP querier in each VTEP subnet. Hybrid mode does not require PIM. Due to SoftLayer constraints on the number of available multicast addresses, this design uses unicast mode Transport Zone Design A transport zone is used to define the scope of a VXLAN overlay network which can span one or more clusters within a vcenter Server domain. One or more transport zones can be configured in an NSX for vsphere solution. A transport zone is not meant to delineate a security boundary. The design implements a single transport zone per NSX instance. Each zone will be scaled to respect the limit of 100 DLRs per ESXi host. A single transport zone per NSX instance supports extending networks across resource stacks. Page 56 of 132 Copyright IBM and VMware
57 Routing Logical Design The routing logical design has to consider different levels of routing in the environment: North-south. The Provider Logical Router (PLR) handles the north/south traffic to and from a tenant East-west. Internal east-west routing at the layer beneath the PLR deals with the application workloads. NSX Edge services gateways is deployed for each management application to provide routing for the application. Virtual application network is attached directly to the edge services gateway. No distributed logical routers are deployed beneath Firewall Logical Design In this design, firewall functions are controlled at NSX Edge services gateway firewall. The Edge services gateway virtual machines act as the entry point to a tenant s virtual network space. NSX Edge services gateways are configured with the firewall enabled to support functionality that is required by virtual application networks. External access via load balancer virtual IPs when public services are provided by redundant management application VMs. External access is via a DNAT rule when service is provided by a single VM Load Balancer Design The Edge services gateway implements load balancing within NSX for vsphere; it has both a Layer 4 and a Layer 7 engine that offer different features, summarized in the following table. Table 22 Load Balancer Features Feature Layer 4 Engine Layer 7 Engine Protocols TCP TCP Load Balancing Method Round Robin Source IP Hash Least Connection HTTP HTTPS (SSL Pass-through) HTTPS (SSL Offload) Round Robin Source IP Hash Least Connection URI Health Checks TCP TCP Persistence (keeping client connections to the same back-end server) TCP: SourceIP HTTP (GET, OPTION, POST) HTTPS (GET, OPTION, POST) TCP: SourceIP, MSRDP HTTP: SourceIP, Cookie HTTPS: SourceIP, Cookie, ssl_session_id Connection Throttling No Client Side: Maximum concurrent connections, Maximum new connections per second Server Side: Maximum concurrent connections Copyright IBM and VMware Page 57 of 132
58 Feature Layer 4 Engine Layer 7 Engine High Availability Yes Yes Monitoring View VIP (Virtual IP), Pool and Server objects and stats via CLI and API View global stats for VIP sessions from the vsphere Web Client View VIP, Pool and Server objects and statistics by using CLI and API View global statistics about VIP sessions from the vsphere Web Client Layer 7 Manipulation No URL block, URL rewrite, content rewrite An NSX Load balancer is utilized to support the needs of management applications and for workloads Bridging Physical Workloads NSX for vsphere offers VXLAN to Layer 2 VLAN bridging capabilities with the data path contained entirely in the ESXi hypervisor. The bridge runs on the ESXi host where the distributed logical router control VM is located. Multiple bridges per distributed logical router are supported. This design places all virtual machines on VXLAN-backed networks. No bridging is implemented in the solution VPN Connectivity NSX for vsphere offers SSL VPN-Plus, IPSec, and Layer 2 VPN. SSL VPN-Plus. SSL VPN-Plus allows remote users to make secure client connections to private networks behind an NSX Edge service gateway and to access servers and applications. Users can be authenticated locally or via an external source such as Active Directory, LDAP, RADIUS, or RSA. The client can connect using either network or Web access modes. Network access requires a client side installation, for which a package can be created. Web access is entirely browser based, and requires no special hardware or software. IPSec VPN. NSX Edge service gateway supports site-to-site IPSec VPN between an NSX Edge instance and remote sites. It supports certificate-based authentication, preshared key mode and IP unicast traffic. Multiple subnets can be connected to behind the remote router over an IPsec tunnel but their network ranges must not overlap. The following constraint might affect potential use cases for the IPSec VPN, so is highlighted at the point of design. Note: Currently no dynamic routing protocols are supported between the NSX Edge service gateway and remote VPN routers. Layer 2 VPN. Layer 2 VPN allows creating a tunnel between sites. This VPN type requires the configuration of a Layer 2 VPN Server (the source) and Layer 2 VPN Client (the destination) with NSX Edge service gateway devices used for both of these roles. This design does not implement layer 2 VPN. Inter-region routing allows for management network reachability. SSL-VPN or IPSec VPN are available for use by users or the cloud admin actors for access to the user workload or cloud management services respectively Application Virtual Network Cloud management applications, such as VMware vrealize Automation, VMware vrealize Operations Manager, VMware vrealize Orchestrator, vrealize Log Insight, leverage a traditional Page 58 of 132 Copyright IBM and VMware
59 3-tier client/server architecture with a presentation tier (user interface), functional process logic tier, and data tier. This architecture requires a load balancer for presenting end-user facing services. This design implements each of these management applications as their own trust zone and isolates management applications from each other. Management applications are placed on isolated networks (VXLAN backed networks). The virtual application network is fronted by an NSX Edge services gateway to keep applications isolated. The use of load balancer interfaces is required for inbound access. The use of SNAT is required for outbound access. Direct access to virtual application networks is by connecting to Windows machines that are connected to the management networks directly. Unique addressing is required for all management applications. This approach to network virtualization service design improves security and mobility of the management applications, and reduces the integration effort with existing customer networks. From a software-defined networking design perspective, each management application is treated as a separate management tenant that requires one or more logical networks via VXLAN segment. The VXLAN segments are connected to the outside world by using a pair of NSX Edge services gateways. Figure 20 Virtual Application Network Components and Design The NSX Edge services gateway associated with a management application is connected via an uplink connection to a public accessible network and contains at least one IPv4 address on this network. As a result, this device offers the following capabilities. If the IPv4 range that is used for the application-internal network does not overlap with any other existing IPv4 range, the central DNS service can create DNS entries for nodes on this internal network. Split DNS is not necessary. Inbound access to services, such as Web UI, is supported by the load balancer capabilities of the NSX Edge services gateway. Application nodes can access the corporate network or the Internet via Source NAT (SNAT) on the NSX Edge services gateway or via the vsphere management network. Copyright IBM and VMware Page 59 of 132
60 Routed (static or dynamic) access to the vsphere management network is available for access to the vcenter Server instances Virtual Network Design Example The following figure presents the design for vrealize Automation that is implemented by the architecture. This design is utilized for other management applications and can be implemented for workloads deployed on compute cluster. The design is set up as follows: Figure 21 vra Virtual Network Design vrealize Automation is deployed onto a single Layer 2 segment, which is provided by a VXLAN virtual wire. Micro segmentation between NSX components is not required and therefore not used. The network used by vrealize Automation connects to external networks through NSX for vsphere. NSX Edge services gateways route traffic between management application virtual networks and the public network. Access to the isolated vsphere-mgmt network is available through the MgmtCentral- Edge services gateway that provides region-specific routing. Page 60 of 132 Copyright IBM and VMware
61 All Edge services gateways are connected over the networkexchange network that acts as a transit network and as an interface to exchange routing information between the Edge services gateways. To provide easy mobility of the network used by vrealize Automation during recovery in another region, this network uses an RFC 1918 isolated IPv4 subnet and uses Source NAT (SNAT) to access external networks such as the Internet. Services such as a Web GUI, which must be available to the users of vrealize Automation, are accessible via the NSX Edge load balancer on the IPv4 address residing on the external network. Each application must use a unique IPv4 range for the application internal network(s). The unique IPv4 range supports use of the central DNS service for creating DNS entries for nodes on this internal network. The following table shows an example of how a mapping from management applications to IPv4 subnets might look. Table 23 Management Applications IP Addressing Management Application Internal IPv4 Subnet vrealize Automation (includes vrealize Orchestrator) /24 vrealize Automation Proxy Agents / /24 vrealize Operations Manager /24 vrealize Operations Remote Collector / /24 vrealize Log Insight / /24 The management applications vrealize Automation, vrealize Operations Manager, and vrealize Log Insight divert from the above described setup slightly: vrealize Automation uses three network containers. o One container is for the main vrealize Automation application cluster that can be failed over by using Site Recovery Manager to a secondary region. o Two additional network containers - one in each region - hold vrealize Automation proxy agents. vrealize Operations Manager uses three network containers. o One container is for the main vrealize Operations Manager analytics cluster that can be failed over to a secondary region by using Site Recovery Manager. Two additional network containers - one in each region - are for connecting remote collectors. o vrealize Log Insight does not use Site Recovery Manager to fail over between regions. Instead, a dedicated instance of Log Insight is deployed in each region. To support this configuration, an additional cloud region is required. Copyright IBM and VMware Page 61 of 132
62 Routing and Region Connectivity Design This multi-region design is an extension of the single-region design that takes into account management component failover. The figure below is an example of how virtual application networks are built for vrealize Automation in the central cloud and cloud regions. The same virtual application network configuration is valid for all the other virtual application networks. Figure 22 Virtual Application Network Configuration in Central Cloud and Cloud Region Dynamic Routing Routing is handled by NSX Edge. Management network Edge services gateways, MgmtCentral-Edge and MgmtRegionA-Edge, work in conjunction with Edge services gateways that are configured to create virtual application network for each central cloud or cloud region management component. A dedicated network, networkexchange, facilitates the exchange of transit network traffic and the exchange of routing tables. All NSX Edge services gateways that are attached to the networkexchange network segment run OSPF to exchange routing information. OSPF route information exchange depends on the area definition, which is based on the local vsphere-mgmt network: Table 24 OSPF Area ID Network Region IP Addressing Area ID vsphere-mgmt Central Cloud /24 16 vsphere-mgmt Region A (Cloud Region) /24 17 Page 62 of 132 Copyright IBM and VMware
63 Virtual application network Edge services gateway receives an OSPF default route from the Management Edge services gateway which has access to the public network. Components directly attached to the vsphere-mgmt network use the Management Edge services gateway as the default gateway to reach all components in the environment. All VMs within virtual application networks can route to VMs in other virtual application networks. Base OSPF Area IDs on the second octet of the vsphere management network, and use MD5 encryption for authentication. NSX Edge appliance size is X-Large and is deployed in High Availability (HA) mode. A dedicated network for exchange of routing information and traffic between different gateways is configured. Access to public networks uses SNAT Region Connectivity and Dynamic Routing There is one Management Edge services gateway in each region. The role of the Management Edge services gateway is to provide connectivity between regions and to exchange region-specific routes with other connected Management Edge services gateways. The Management Edge services gateway is configured to exchange OSPF routing data with the entire region-specific virtual application network. Edge services gateways are attached to their respective networkexchange networks. These routes are consolidated and exchanged with all connected Management Edge services gateways using ibgp. All Management Edge services gateways belong to the same AS (Autonomous System) with regards to BGP routing. In this design, connectivity between two regions uses the direct backend connectivity available between sites Virtual Application Networks and Dynamic Routing Across Regions For the components that fail over to another region in the event of a disaster, route redistribution is disabled in NSX Edge. This prevents the same routes from being advertised in both regions for their virtual application networks. This makes it possible for the virtual application network Edge services gateway to participate in region-specific OSPF route exchanges without announcing its virtual application network route. To facilitate failover of the virtual application network from one region to the other for some components, this design creates a shadow virtual application network in the recovery region, an additional cloud region in this design. This shadow virtual application network is configured identically, with the same SNAT and DNAT rules and load balancer VIP as needed. The only difference is the IP addressing used on the networkexchange network and the optional public interface. Component virtual application networks can be moved between regions either for testing or during failover. When a component is moved the location of that component s virtual application network needs to be changed in the VPN configuration. The virtual application network Edge services gateway also needs to be reconfigured to either start or stop redistributing connected virtual application networks. The decision to start or stop redistribution depends on whether the virtual application network Edge services gateway is now the active Edge services gateway or the failed over Edge services gateway. 5.3 Infrastructure Management Infrastructure management is provided by VMware vcenter Server. The configuration of vcenter Server is described in the following sections vcenter Server Instances The vcenter Server design includes both the vcenter Server instances and the VMware Platform Services Controller instances. Copyright IBM and VMware Page 63 of 132
64 A VMware Platform Services Controller (PSC) groups a set of infrastructure services including vcenter Single Sign-On, License service, Lookup Service, and VMware Certificate Authority. Although the Platform Services controller and the associated vcenter Server system can be deployed on the same virtual machine (embedded Platform Services Controller), this design using a separate PSC and vcenter Server. This design includes two vcenter Server instances per central cloud or cloud region; one managing the management cluster, the other managing the compute and edge clusters. This separation of the management and compute stack not only provides additional security, but also supports scalability. The vcenter Server is deployed as the Linux-based vcenter Server Appliance which is functionally equivalent to the Windows based application, but easier to deploy and maintain. The two external Platform Services Controller (PSC) instances and two vcenter Server instances are both housed in the management cluster. The external PSC instances are used to enable replication of data between PSC instances. To achieve redundancy, the design joins the two Platform Services Controller instances to the same vcenter Single Sign-On domain, and connects each vcenter Server instance to their respective Platform Services Controller instance. Joining all Platform Services Controller instances into a single vcenter Single-Sign-On domain provides this design the capability to share authentication and license data across all components and regions. Additionally, the vcenter Servers in each region are configured to utilize Enhanced Linked Mode so that all inventories and actions are available from a each vcenter Server in the region. In terms of configuration, each vcenter Server system is configured with a static IP address and host name from the management private portable SoftLayer subnet. The IP address will have a valid (internal) DNS registration including reverse name resolution. The vcenter Server systems will also maintain network connections to the following components: All VMware vsphere Client and vsphere Web Client user interfaces. Systems running vcenter Server add-on modules, such as NSX or vum. Each ESXi host. Figure 23 vcenter Server and PSC Deployment Model Page 64 of 132 Copyright IBM and VMware
65 vcenter Server Resiliency Protecting the vcenter Server system is important because it is the central point of management and monitoring for the environment. In this design, the vcenter Servers and PSCs are protected via vsphere HA Size of vcenter Server Appliances The size of the vcenter Service appliances is described in the following sections vcenter Server Appliance for Management Cluster The number of hosts that are managed by the Management vcenter Server is less than 100 and the number of virtual machines to be managed is less than 1,000. As a result, this design includes a small vcenter Server configured with the following specifications and still allows for cloud admin s to include additional services within the cluster: Table 25 Specifications for Management vcenter Server Appliance vcpus Memory Disk Space Disk Type 4 16 GB 136 GB Thin Platform Services Controller for Management Cluster The Platform Services controller is deployed onto the management cluster with the following configuration: Table 26 Specifications for Platform Service Controller for Management Cluster vcpus Memory Disk Space Disk Type 2 2 GB 30 GB Thin Additionally, the PSC connects to the Active Directory server within this design for common authentication vcenter Server Appliance for Compute and Edge Clusters The Compute and Edge vcenter Server in this design manages up to 1,000 hosts or 10,000 virtual machines whichever comes first. As a result, this design includes a large vcenter Server configured with the following specifications: Table 27 Specifications for Compute and Edge vcenter Server Appliance vcpus Memory Disk Space Disk Type GB 295 GB Thin Platform Services Controller for Compute and Edge Clusters The Platform Services controller for vcenter managing the Compute and Edge clusters are deployed onto the management cluster and contain the following configuration: Table 28 Specifications for Platform Service Controller for Management Cluster vcpus Memory Disk Space Disk Type 2 2 GB 30 GB Thin Additionally, the PSC conencts to the Active Directory server within this design for common authentication. Copyright IBM and VMware Page 65 of 132
66 vcenter Database Design A vcenter Server Appliance can use either a built-in local PostgreSQL database or an external Oracle database. Both configurations support up to 1,000 hosts or 10,000 virtual machines. This design will utilize the built-in PostgreSQL to reduce both overhead and Microsoft or Oracle licensing costs. This also avoids problems with upgrades and support since the ability to attach to an external database is deprecated for vcenter Server Appliance in the next release vcenter Networking Design The vcenter Servers and Platform Services Controllers in each region will be placed onto the management network to access physical ESXi hosts within the region Cluster Configuration Each of the clusters will be configured per the following sections vsphere DRS This design will utilize vsphere Distributed Resource Scheduling (DRS) in each cluster to initially place and dynamically migrate virtual machines to achieve balanced compute clusters. The automation level is set to fully automated so that initial placement and migration recommendations are executed automatically by vsphere. Note that power management via the Distributed Power Management feature is not used in this design. Additionally, the migration threshold is set to partially aggressive so that vcenter will apply priority 1, 2, 3, and 4 recommendations to achieve at least a moderate improvement in the cluster s load balance Affinity Rules In this design, affinity and anti-affinity rules are used to ensure placement of virtual machines that are in a high availability cluster are not placed on the same ESXi host vsphere HA This design will use vsphere HA in each cluster to detect compute failures and recover virtual machines running within a cluster. The vsphere HA feature in this design is configured with both host monitoring and admission control enabled within the cluster. Additionally, each cluster will reserve 25% of cluster resources as spare capacity for the admission control policy. By default, the VM restart priority is set to medium and the host isolation response is set to leave powered on. Additionally, VM monitoring is disabled and the datastore heart-beating feature is configured to include any of the cluster datastores. In this design, the datastore is a VSAN-backed datastore vsphere vmotion vsphere vmotion is enabled in this design by the use of shared storage and network configuration vsphere FT vsphere FT is not utilized or available in this design. 5.4 Common Services Identity and Access Services Microsoft Active Directory (AD) is used as the identity provider (IDP) in this design. In a single region deployment of this design, two MS Windows 2012 R2 Active Directory server VMs are configured for a Page 66 of 132 Copyright IBM and VMware
67 single AD forest. Naming of the domain is derived from input data prior to design provisioning. Each server has the MS Windows DNS service installed and configured with all forward and reverse lookup zones as type AD integrated multi master enabled. Subsequent regions contain only one AD / DNS server in each region setup as a peer multi master within its own MS AD site configured within the AD Sites management console Forrest and subdomains This design uses Microsoft Active Directory for authentication and authorization to resources within the tornado.local domain. For a multi-region deployment, the design utilizes a domain and forest structure to store and manage Active Directory objects per region. Table 29 Requirements for Active Directory Service Requirement Active Directory configuration Active Directory users and groups Domain Instance Parent Active Directory Central child Active Directory Region-A child Active Directory Domain Name tornado.local central.tornado.local regiona.tornado.local Description Contains Domain Name System (DNS) server, time server, and universal groups that contain global groups from the child domains and are members of local groups in the child domains. Contains DNS records which replicate to all DNS servers in the forest. This child domain contains all design management users, and global and local groups. Contains DNS records which replicate to all DNS servers in the forest. This child domain contains all design management users, and global and local groups. - The default AD internal LDAP directory structure will be used for users and computer objects. ou=computers, DC=central, DC=tornado, DC=local ou=users, DC=central, DC=tornado, DC=local AD Functional level This design s AD functional level will be set at Windows 2008, which is the lowest level which is allowed by the other components in this design AD Trusts External trusts to customer AD environments are configured as required Authentication Within this design, authentication is provided by a mix of the following: direct AD integration SAML authentication via the VMware Platform Services Controller (PSC), backed by AD Copyright IBM and VMware Page 67 of 132
68 local user accounts within the application. The VMware Platform Services Controller is configured to use the AD servers within this design for its identity management directory. The following chart describes which authentication method is used for user / administrator access for each component Table 30 Authentication types used Component Authentication Context Authentication authority vra Administration Active Directory Tenant Active Directory vrops Administration / Roles Active Directory vcenter Server Administration / Roles Active Directory vro Administration / Roles Active Directory ESXi Administration Active Directory NSX Administration Active Directory vri Administration Active Directory vrb Administration Active Directory Tenant Active Directory vum Administration Active Directory NSX Administration Active Directory VPN Active Directory vdp Administration Active Directory AD VM server sizing Table 31 Server Sizing Attribute Number of CPUs 4 Memory Disk size Number of servers in central cloud Number of servers per additional cloud regions Specification 8 GB 80 GB Domain Name Services Domain name services (DNS) is a fundamental function of any multi-component network infrastructure design. DNS not only stores host name to IP address information but also service type information required by Microsoft Active Directory and other applications. Dynamic DNS capability is also required for MS AD deployments for the registration of the AD service record types as subzone creation. The AD integrated MS DNS server is utilized for this design. Page 68 of 132 Copyright IBM and VMware
69 DNS Design In a single region deployment of this design, four MS Windows 2012 R2 Active Directory server VMs are configured. Two for the root of the forest domain and 2 for the region subdomain. Each server has the MS Windows DNS service installed and configured with all forward and reverse lookup zones as type AD integrated multi master enabled. Subsequent regions contain only one AD / DNS server in each region setup as a peer multi master within its own MS AD site configured within the AD Sites management console. The following applies to all services and hosts within this design. All nodes must use static IP addresses. As a DNS best practice, the IP address, FQDNs and short names of all nodes must be forward and reverse resolvable. All nodes must be accessible from the vcenter Server instances and from the machine that hosts the vsphere Client (if vsphere Client is used instead of vsphere Web Client). All nodes in the vrealize Operations Manager analytics cluster must have unique host names. NTP source must be used for all cluster nodes DNS is an important component for the operation of this design. For a multi-region deployment, other regions must provide a root and child domains which contain separate DNS records DNS Configuration Requirements The following table shows an example of the domain naming that can be implemented by the design. The specific domain naming is determined by the client requirements. Table 32 Domain Naming Example Requirement Domain Instance Description DNS host entries tornado.local Resides in the tornado.local domain. central.tornado.local and regiona.tornado.local DNS servers reside in the central.tornado.local and regiona.tornado.local domains. Configure both DNS servers with the following settings: Dynamic updates for the domain set to Nonsecure and secure. Zone replication scope for the domain set to All DNS server in this forest. Create all hosts listed in the DNS Names documentation. Once properly configured, all nodes within this design are resolvable by FQDN. Copyright IBM and VMware Page 69 of 132
70 DNS Forwarding For name resolution outside of the design, the forest root DNS domain servers will have forwarders configured to the internal SoftLayer DNS servers if no client on premises DNS is available. If a client on premises DNS is available, then that becomes the DNS forwarding address or addresses. Table 33 SoftLayer DNS servers DNS hostname IP v4 address rs1.service.softlayer.com rs2.service.softlayer.com NTP Services Time synchronization is critical for the functionality of many of the services that comprise this design. For this purpose, Microsoft Windows Active Directory (AD) servers within the solution are configured as NTP sources for all services within the design with the exception of VMware Data Protection (vdp), vrealize Log Insight (vri) and the ESXi hosts. The Windows AD servers are configured to utilize the internal SoftLayer time server servertime.service.softlayer.com as the authoritative time source. The vdp application server, is configured to utilize VMware tools as its time source per the vdp documentation recommendation. The ESXi hosts and vri are configured to point to the SoftLayer NTP server. Table 34 Time sources Service / Application Active Directory servers vdp ESXi vrealize Log Insight All else Time Source servertime.service.softlayer.com VMware tools servertime.service.softlayer.com servertime.service.softlayer.com AD Servers SMTP Services SMTP services are required for outbound notifications and may be used for inbound communications to vrealize Automation. The following sections details the configuration of SMTP in this design SMTP outbound Within this design, notifications are sent using SMTP for the following products: vrealize Automation vrealize Business vrealize Operations vrealize Log Insight vrealize Orchestrator vcenter Server VMware Data Protection Page 70 of 132 Copyright IBM and VMware
71 An SMTP server service is configured on each windows AD server which is utilized as an SMTP relay for all outbound SMTP notifications. This service is configured to use the default SMTP port (25) where the connection to the customer s servers is over direct or VPN connection. As SoftLayer blocks port 25 outbound, a secure SSL port (465, 587), or custom port is configured on the SMTP server based on the destination provider, provided by the customer. This is only required where a direct connection or VPN from the customer site into the management network is not available SMTP inbound For vrealize Automation, an inbound server to handle inbound notifications, such as approval responses, is required to be configured. This is typically the customer server with accounts created for vra. Only one, global inbound server, which appears as the default for all tenants, is needed. The server provides accounts that are able to be customized for each user, providing separate accounts, usernames, and passwords. Each tenant can configure an override to change these settings. If service provider actors do not override these settings before enabling notifications, vrealize Automation uses the globally configured server. The connection to the customer system requires that this system be network reachable by the vra system Certificate Authority Services By default, vsphere 6.0 uses TLS/SSL certificates that are signed by VMCA (VMware Certificate Authority) residing on the VMware Platform Services Controller appliance (PSC). By default, these certificates are not trusted by end-user devices or browsers. It is a security best practice to replace at least user-facing certificates with certificates that are signed by a third-party or enterprise Certificate Authority (CA). Certificates for machine-to-machine communication can remain as VMCA-signed certificates. For this design, a Microsoft Active Directory Enterprise integrated certificate authority is used. A two tier MS certificate authority design is employed with 1 windows server as the offline root CA and an additional windows server as the online subordinate CA Offline root CA server VM As per Microsoft best practice, the root CA windows server does not participate in this design s AD domain. Setup the root CA per the following Microsoft information page: Note that while it is not recommended to connect this machine to a network, it will be network connected until the root CA is generated. Once generated, transfer the root certificate files to the subordinate CA server, remove the network access from the root CA server and shut down the OS. It is recommended that client s export this VM as an OVA to a secure location and remove it from the vcenter Server s inventory Online subordinate CA server VM Setup the subordinate CA per the following MS link: Note that this system will also be configured for handling certificate requests and distributing certificate revocation lists (CRL). The VMCA is configured with a certificate issued via the MS Active directory integrated CA such that all subsequent browser access to VMware services, are validated by any clients that a members of the designs AD domain or have imported the domains trusted root certificate. Copyright IBM and VMware Page 71 of 132
72 Table 35 Root CA and Subordinate CA sizing Attribute Specification OS Windows 2012 R2 Number of CPUs 2 Memory 4 GB Disk size 60 GB 5.5 Cloud Management Services The Cloud Management Services provide the management components for the cloud solution. This layer includes the Service Catalog, which houses the facilities to be deployed, Orchestration which provides the workflows to get the catalog items deployed, and the Self-Service Portal that empowers the users to take full advantage of the Software Defined Datacenter. vrealize Automation provides the Portal and the Catalog, and vrealize Orchestrator takes care of the process orchestration. The conceptual design of the vrealize Automation Cloud Management Services is illustrated in the following diagram. Key design components and their descriptions are also provided. Figure 24 vrealize Automation Conceptual Design The Cloud Management Services consist of the following elements and components: Users: Cloud administrators Tenant, group, fabric, infrastructure, service, and other administrators as defined by business policies and organizational structure. Page 72 of 132 Copyright IBM and VMware
73 Cloud (or tenant) users Provide direct access to virtual machine to perform operating systemlevel operations provided by vrealize Automation IaaS services. Cloud Management Portal: vrealize Automation portal, Admin access The default root tenant portal URL used to set up and administer tenants and global configuration options. vrealize Automation portal, Tenant access Refers to a subtenant and is accessed using a unique URL, with an appended tenant identifier. It is also possible for a tenant portal to refer to the default tenant portal in some configurations. In this case, the URLs are the same and the user interface is contextually controlled by the assigned RBAC permissions of that user. Tools and supporting infrastructure: VM Templates and Blueprints - These are the templates used in authoring the blueprints that tenants (users) use to provision their cloud workloads. Provisioning infrastructure - the following are the on-premises and off-premises resources which together form a hybrid cloud Virtual Supported hypervisors and associated management tools. Cloud Supported cloud providers and associated API interfaces. In the above diagram illustrating the conceptual design of the Cloud Management Platform, these resources are located in the Internal Virtual Resources and the External Cloud Resources components. The cloud management services deliver multi-platform and multi-vendor cloud services. The services available include the following items. Comprehensive and purpose-built capabilities that provide standardized resources to global customers in a short time span. Multi-platform and multi-vendor delivery methods that integrate with existing enterprise management systems. Central user-centric and business-aware governance for all physical, virtual, private, and public cloud services. Design that meets the customer and business needs and is extensible. Table 36 Cloud Management Services Components Component Services Provided vrealize Automation Identity Appliance vcenter Single Sign-On vrealize Automation virtual appliance vrealize Automation Portal Web/App Server vrealize Automation vpostgresql Database vrealize Automation Service Catalog vrealize Automation IaaS components vrealize Automation IaaS Web Server vrealize Automation IaaS Manager Services Copyright IBM and VMware Page 73 of 132
74 Component Services Provided Distributed execution components vrealize Automation Distributed Execution Managers: Orchestrator Workers Integration components vrealize Automation Agent machines Provisioning infrastructure vsphere environment vrealize Orchestrator environment Other supported physical, virtual, or cloud environments. Supporting infrastructure Microsoft SQL database environment Active Directory environment SMTP NTP DNS Cloud Management Physical Design This design uses NSX logical switches to abstract the vrealize Automation application and its supporting services. This abstraction allows the application to be hosted in any given region regardless of the underlying physical infrastructure such as network subnets, compute hardware, or storage types. This design hosts the vrealize Automation application and its supporting services in the central cloud. The same instance of the application manages both central cloud and any additional cloud regions. Page 74 of 132 Copyright IBM and VMware
75 Figure 25 vrealize Automation Design Overview for Central Cloud Copyright IBM and VMware Page 75 of 132
76 Figure 26 vrealize Automation Design Overview for Additional Cloud Regions The configuration of these elements is described in the following sections vrealize Identity Appliance vrealize Identity Appliance provides an infrastructure-independent failover capability. A single appliance is deployed in the solution. vsphere HA is used to ensure high availability for Identity Appliance. The appliance is configured with 1 vcpu and 2 GB of RAM vrealize Automation Appliance The vrealize Automation virtual appliance includes the Web portal and database services. The vrealize Automation portal allows self-service provisioning and management of cloud services, as well as authoring blueprints, administration, and governance. The vrealize Automation virtual appliance uses an embedded PostgreSQL database for catalog persistence and database replication. Page 76 of 132 Copyright IBM and VMware
77 The solution deploys two instances of the vrealize Automation appliance to achieve redundancy. The database is configured between two vrealize Automation appliances for high availability. Data is replicated between the embedded PostgreSQL database instances. Database instances are configured in an active/passive. In this configuration, manual failover between the two database instances is required. Each appliance is configured with 4 vcpu and 16 GB of RAM vrealize Automation IaaS Web Server vrealize Automation IaaS Web server provides a user interface within the vrealize Automation portal web site for the administration and consumption of IaaS components. Two vrealize Automation IaaS web servers are installed on virtual machines. Each virtual machine runs Microsoft Windows Server 2012 R2 and it performs Model Manager (Web) and IaaS Web functions. Each virtual machine is sized to 4 vcpu, 4 GB of RAM and 60 GB HDD vrealize Automation IaaS Manager Service and DEM Orchestrator Server The vrealize Automation IaaS Manager Service and Distributed Execution Management (DEM) server are at the core of the vrealize Automation IaaS platform. The vrealize Automation IaaS Manager Service and DEM server supports several functions. Manages the integration of vrealize Automation IaaS with external systems and databases. Provides multi-tenancy. Provides business logic to the DEMs. Manages business logic and execution policies. Maintains all workflows and their supporting constructs. A Distributed Execution Manager (DEM) runs the business logic of custom models, interacting with the database and with external databases and systems as required. DEMs also manage cloud and physical machines. The DEM Orchestrator monitors the status of the DEM workers. DEM worker manages the scheduled workflows by creating new workflow instances at the scheduled time and allows only one instance of a particular scheduled workflow to run at a given time. It also preprocesses workflows before execution. Preprocessing includes checking preconditions for workflows and creating the workflow's execution history. The vrealize Automation IaaS Manager Service and DEM server are separate servers, but are installed on the same virtual machine. Two virtual machines are deployed to run both IaaS Manager Service and DEM Orchestrator. The two servers share the same active/passive application model. Only one manager service can be active at a time. Each virtual machine runs Microsoft Windows Server 2012 R2. Each virtual machine is sized to 2 vcpu, 4 GB of RAM and 60 GB HDD vrealize Automation IaaS DEM Worker Server vrealize Automation IaaS DEM workers are responsible for the provisioning and deprovisioning tasks initiated by the vrealize Automation portal. DEM workers communicate with vrealize Automation endpoints. In this instance, the endpoint is vcenter Server. Each DEM Worker can process up to 15 concurrent workflows. Beyond this limit, workflows are queued for execution. The current design implements 2 DEM workers for a total of 30 concurrent workflows. DEM Workers are installed on two virtual machines running Microsoft Windows Server 2012 R2. Each virtual machine is sized to 4 vcpu, 8 GB of RAM and 60 GB HDD. Copyright IBM and VMware Page 77 of 132
78 vrealize Automation IaaS Proxy Agent The vrealize Automation IaaS Proxy Agent is a Windows program that caches and forwards information gathering from vcenter Server back to vrealize Automation. The IaaS Proxy Agent server provides the following functions. vrealize Automation IaaS Proxy Agent can interact with different types of hypervisors and public cloud services, such as Hyper-V and AWS. For this design, only the vsphere agent is used. vrealize Automation does not virtualize resources by itself, but works with vcenter Server to provision and manage the virtual machines. It uses vsphere agents to send commands to and collect data from vcenter Server. Two vrealize Automation vsphere Proxy Agent virtual machines are deployed in the current architecture. The virtual machines are deployed on a dedicated virtual network to decouple them from the main vrealize Automation infrastructure allowing. Each virtual machine runs Microsoft Windows Server 2012 R2. Each virtual machine is sized to 2 vcpu, 4 GB of RAM and 60 GB HDD Load Balancer Session persistence of a load balancer allows the same server to serve all requests after a session is established with that server. The session persistence is enabled on the load balancer to direct subsequent requests from each unique session to the same vrealize Automation server in the load balancer pool. The load balancer also handles failover for the vrealize Automation Server (Manager Service). Only one Manager Service is active at any one time. Manual failover of Manager Service is necessary. Session persistence is not enabled because it is not a required component for the Manager Service. The following tables describe load balancer implementation for vrealize Automation components: Table 37 Load Balancer Application Profile Server Role Type Enable SSL Pass-through vrealize Automation vpostgres vrealize Automation vrealize Automation IaaS Web vrealize Automation IaaS Manager vrealize Automation Orchestrator Persistence TCP n/a None n/a HTTPS (443) Enabled Source IP 120 HTTPS (443) Enabled Source IP 120 HTTPS (443) Enabled Source IP 120 HTTPS (443) Enabled Source IP 120 Expires in (Seconds) Page 78 of 132 Copyright IBM and VMware
79 Table 38 Load Balancer Service Monitoring Configuration Monitor vrealize Automation vpostgres vrealize Automation vrealize Automation IaaS Web vrealize Automation IaaS Manager vrealize Automation Orchestrator Interv Timeou Retries Type Method URL Receive al t TCP N/A N/A N/A HTTPS (443) HTTPS (443) HTTPS (443) HTTPS (443) GET /vcac/ser vices/api /status REGIST ERED GET N/A N/A GET /VMPS2 BasicHtt pbinding _VMPSP roxyage nt_policy GET /vco/api/ status REGIST ERED Table 39 Load Balancer Pool Specifications Server Role Algorithm Monitors Members Port Monitor Port vrealize Automation vpostgres Round Robin <vrealize Automation vpostgres vrealize Automation Postgres vrealize Automation vrealize Automation IaaS Web vrealize Automation IaaS Manager vrealize Automation Orchestrator Least Connection Least Connection Least Connection Least Connection monitor> <vrealize Automation monitor> <vrealize Automation IaaS Web monitor> <vrealize Automation IaaS Manager monitor> <vrealize Automation Orchestrator monitor> nodes vrealize Automation nodes IaaS web nodes IaaS Manager nodes vrealize Orchestrator nodes Table 40 Virtual Server Characteristics Protocol Port Default Pool Application Profile TCP 5432 vrealize Automation vpostgres Pool vrealize Automation vpostgres Profile HTTPS 443 vrealize Automation Pool vrealize Automation Profile HTTPS 443 vrealize Automation IaaS Web Pool vrealize Automation IaaS Web Profile Copyright IBM and VMware Page 79 of 132
80 Protocol Port Default Pool Application Profile HTTPS 443 vrealize Automation IaaS Manager Pool vrealize Automation IaaS Manager Profile HTTPS 8281 vrealize Automation Orchestrator Pool vrealize Automation Orchestrator Profile vrealize Automation Supporting Infrastructure A number of supporting elements are required for vrealize Automation. The following sections describe their configuration Microsoft SQL Server Database vrealize Automation uses a Microsoft SQL Server database to maintain the vrealize Automation IaaS elements and the policies. The database also maintains information about the machines it manages. For simple failover of the entire vrealize Automation instance from one site to another, the Microsoft SQL server is running in a virtual machine inside the vrealize Automation virtual network. The virtual machine is running Microsoft Windows Server 2012 R2 and configured with 8 vcpu, 16 GB of RAM, 80 GB HDD. The Microsoft SQL Server Database version is SQL Server Postgres SQL Server Database The vrealize Automation appliance uses a PostgreSQL database server to maintain the vrealize Automation portal elements and services, and the information about the catalog items it manages. Embedded PostgreSQL within each virtual appliance is utilized Notifications System administrators configure default settings for both the outbound and inbound s servers used to send system notifications. Current solution implements outbound SMTP server only. The automation creates a global outbound server to process outbound notifications. The server appears as the default for all tenants vrealize Automation Cloud Tenant Design A tenant is an organizational unit within a vrealize Automation deployment, and can represent a business unit within an enterprise, or a company that subscribes to cloud services from a service provider. Each tenant has its own dedicated configuration, although some system-level configuration is shared across tenants Single-Tenant and Multi-Tenant Deployments vrealize Automation supports deployments with a single tenant or multiple tenants. System-wide configuration is always performed using the default tenant, and can then be applied to one or more tenants. System-wide configuration specifies defaults for branding and notification providers. Infrastructure configuration, including the infrastructure sources that are available for provisioning, can be configured in any tenant and is shared among all tenants. The infrastructure resources, such as cloud or virtual compute resources or physical machines, can be divided into fabric groups managed by fabric administrators. The resources in each fabric group can be allocated to business groups within each tenant by using reservations. Single-Tenant Deployment In a single-tenant deployment, all configuration occurs in the default tenant. Service provider actors can manage users and groups, and configure tenant-specific branding, notifications, business policies, and catalog offerings. All users Page 80 of 132 Copyright IBM and VMware
81 log in to the vrealize Automation console at the same URL, but the features available to them are determined by their roles. Multi-Tenant Deployment In a multi-tenant deployment, the system administrator creates new tenants for each organization that uses the same vrealize Automation instance. Tenant users log in to the vrealize Automation console at a URL specific to their tenant. Tenant-level configuration is segregated from other tenants and from the default tenant, although users with system-wide roles can view and manage configuration across multiple tenants Tenant Design This design deploys a single tenant containing two business groups. The first business group is designated for production workloads provisioning. The second business group is designated for development workloads. Service provider actors manage users and groups, configure tenant-specific branding, notifications, business policies, and catalog offerings. All users log in to the vrealize Automation console at the same URL, but the features available to them are determined by their roles. The following diagram illustrates the single region tenant design. Figure 27 Tenant Design for Single Region Copyright IBM and VMware Page 81 of 132
82 The following diagram illustrates the dual-region tenant design. Figure 28 Tenant Design for Two Regions The tenant has two business groups. A separate fabric group for each region is created. Each business group can consume resources in both regions. Access to the default tenant is allowed only by the system administrator and for the purposes of managing tenants and modifying systemwide configurations. The solution automatically configures vrealize Automation based on the requested deployment type single region or dual region Service Design The service catalog provides a common interface for consumers of IT services to use to request and manage the services and resources they need. A service provider actor or service architect can specify information about the service catalog, such as the service hours, support team, and change window. The solution implements a service catalog provides the following services: Central Cloud. Service catalog that is dedicated to the central cloud. Additional Cloud Region. Service catalog that is dedicated to an additional cloud region. The solution is preinstalled with several catalog items. For a single site configuration, only the central cloud service catalog is implemented. Page 82 of 132 Copyright IBM and VMware
83 Catalog Items Users can browse the service catalog for catalog items they are entitled to request. Several generic users will be automatically created and entitled to all items in the catalog. The users can be disabled at a later stage or their permissions modified as appropriate. For some catalog items, a request results in the provisioning of an item that the user can manage. For example, the user can request a virtual machine with Windows 2012 preinstalled, and then manage that virtual machine after it has been provisioned. The service provider actor defines new catalog items and publish them to the service catalog. The service provider actor can then manage the presentation of catalog items to the consumer and entitle new items to consumers. To make the catalog item available to users, a service provider actor must entitle the item to the users and groups who should have access to it. A catalog item is defined in a blueprint, which provides a complete specification of the resource to be provisioned and the process to initiate when the item is requested. It also defines the options available to a requester of the item, such as virtual machine specifications or lease duration, or any additional information that the requester is prompted to provide when submitting the request. The blueprint also specifies custom properties that are applied to the requested resource Machine Blueprints A machine blueprint is the complete specification for a virtual machine. A machine blueprint determines the machine's attributes, how it is provisioned, and its policy and management settings. Machine blueprints are published as catalog items in the service catalog. Machine blueprints can be specific to a business group or shared among groups within a tenant. In this design the preloaded machine blueprints are shared among business groups. Service provider actors create shared blueprints that can be entitled to users in any business group within the tenant. Business group managers create group blueprints that can only be entitled to users within a specific business group. A business group manager cannot modify or delete shared blueprints. Service provider actors cannot view or modify group blueprints unless they also have the business group manager role for the appropriate group. If a service provider actor sets a shared blueprint's properties so that it can be copied, the business group manager can also copy the shared blueprint for use as a starting point to create a new group blueprint Blueprint Design The following sections provide details of each service definition that has been included as part of the current phase of cloud platform deployment. Table 41 Base Windows Server Blueprint Service Name Provisioning Method Entitlement Approval Process Operating System and Version Details Configuration Lease and Archival Details Description When users select this blueprint, vrealize Automation clones a vsphere virtual machine template with preconfigured vcenter customizations. Both Production and Development business group members. No approval (pre-approval assumed based on approved access to platform) Windows Server 2012 R2 Disk: Single disk drive Lease: Production Blueprints: No expiration date Copyright IBM and VMware Page 83 of 132
84 Service Name Pre- and Post- Deployment Requirements Description Development Blueprints: Minimum 30 days Maximum 270 days Archive: 15 days sent to manager confirming service request (include description details) Table 42 Base Windows Blueprint Sizing vcpu Memory (GB) Storage (GB) Table 43 Base Linux Server Blueprint Service Name Provisioning Method Entitlement Approval Process Operating System and Version Details Configuration Lease and Archival Details Description When users select this blueprint, vrealize Automation clones a vsphere virtual machine template with preconfigured vcenter customizations. Both Production and Development business group members No approval (pre-approval assumed based on approved access to platform) Red Hat Enterprise Server 6 Disk: Single disk drive Lease: Production Blueprints: No expiration date Development Blueprints: Minimum 30 days Maximum 270 days Pre- and Post- Deployment Requirements Archive: 15 days sent to manager confirming service request (include description details) Table 44 Base Linux Blueprint Sizing vcpu Memory (GB) Storage (GB) Branding The solution branding is preconfigured. The cloud admin can change the appearance of the vrealize Automation console to meet site-specific branding guidelines by changing the logo, the background color, or information in the header and footer. Page 84 of 132 Copyright IBM and VMware
85 5.5.4 vrealize Automation vsphere Integration Design The following terms apply to vrealize Automation integrated with vsphere. These terms and their meaning may vary from the way they are used when referring only to vsphere. Table 45 vrealize Integration with vsphere Element vsphere (vcenter Server) endpoint Compute resource Fabric groups Fabric administrators Compute reservation Storage reservation Business groups Reservation policy Build profile Description Provides information required by vrealize Automation IaaS to access vsphere compute resources. It requires the appropriate permissions for the vsphere proxy agent to manage the vcenter Server instance. Virtual object within vrealize Automation that represents a vcenter Server cluster or resource pool, and datastores or datastore clusters. vrealize Automation provisions the virtual machines requested by business group members on the compute resource. Note: Compute resources are CPU, memory, storage and networks. Datastores and datastore clusters are part of the overall storage resources. vrealize Automation IaaS organizes compute resources into fabric groups. Fabric administrators manage compute resources, which are organized into fabric groups. A share of compute resources (vsphere cluster, resource pool, datastores, or datastore clusters), such as CPU and memory reserved for use by a particular business group for provisioning virtual machines. Note: vrealize Automation uses the term reservation to define resources (be they memory, storage or networks) in a cluster. This is different than the use of reservation in vcenter Server, where a share is a percentage of total resources, and reservation is a fixed amount. Similar to compute reservation (see above), but pertaining only to a share of the available storage resources. In this context, a storage reservation in terms of gigabytes is specified from an existing LUN or Datastore. A collection of virtual machine consumers, usually corresponding to an organization's business units or departments. Only users in the business group can request virtual machines. vrealize Automation IaaS determines its reservation (also called virtual reservation) from which a particular virtual machine is provisioned. The reservation policy is a logical label or a pointer to the original reservation. Each virtual reservation can be added to one reservation policy. A set of user defined properties a user is able to apply to a virtual machine when it is provisioned. For example, the operating system used in a blueprint, or the available networks to use for connectivity at the time of provisioning the virtual machine. Build profile properties determine the specification of the virtual machine, the manner in which it is provisioned, operations to perform after it is provisioned, or management information maintained within vrealize Automation. Copyright IBM and VMware Page 85 of 132
86 Element Blueprint Description The complete specification for a virtual machine, determining the machine attributes, the manner in which it is provisioned, and its policy and management settings. Blueprint allows the users of a business group to create virtual machines on a virtual reservation (compute resource) based on the reservation policy, and using platform and cloning types. It also lets a user specify or add machine resources and build profiles. Page 86 of 132 Copyright IBM and VMware
87 The following figure shows the logical design constructs discussed in the previous section as they apply to the deployment of vrealize Automation integrated with vsphere in a single region design Figure 29 vrealize Automation Integration with vsphere Endpoint Central Cloud Copyright IBM and VMware Page 87 of 132
88 The following figure shows the logical design constructs discussed in the previous section as they apply to the deployment of vrealize Automation integrated with vsphere in a dual region design Figure 30 vrealize Automation Integration with vsphere Endpoint Central Cloud and a Cloud Region (Region A) The solution automatically implements the design presented in the picture above for a dual site configuration. When single site deployment is requested, the automation deploys only the central cloud configuration. Page 88 of 132 Copyright IBM and VMware
89 5.5.5 Infrastructure Source Endpoints An infrastructure source endpoint is a connection to the infrastructure that provides a set (or multiple sets) of resources, which can then be made available by IaaS administrators for consumption by users. vrealize Automation IaaS regularly collects information about known endpoint resources and the virtual resources provisioned therein. Endpoint resources are referred to as compute resources (or as compute pods the terms are often used interchangeably). Infrastructure data is collected through proxy agents that manage and communicate with the endpoint resources. This information about the compute resources on each infrastructure endpoint and the machines provisioned on each computer resource is collected at regular intervals. During solution deployment the proxy agents and define their associated endpoints are configured automatically Virtualization Compute Resources A virtualization compute resource is a vrealize Automation object that represents an ESXi host or a cluster of ESXi hosts (vsphere cluster). When a group member requests a virtual machine, the virtual machine is provisioned on these compute resources. vrealize Automation regularly collects information about known compute resources and the virtual machines provisioned on them through the proxy agents. Each region has one compute cluster. The compute cluster is selected automatically during deployment Fabric Groups A fabric group is a logical container of several compute resources, and can be managed by fabric administrators. A fabric group for each region is created and it includes all the compute resources and edge resources in that region Business Groups A Business group is a collection of machine consumers (users), often corresponding to a line of business, department, or other organizational unit. To request machines, a vrealize Automation user must belong to at least one Business group. Each group has access to a set of local blueprints used to request machines. Business groups have the following characteristics. A group must have at least one business group manager, who maintains blueprints for the group and approves machine requests. Groups can contain support users, who can request and manage machines on behalf of other group members. A vrealize Automation user can be a member of more than one Business group, and can have different roles in each group. Two business groups are created, one for production users and one for the development users Reservations A reservation is a share of one compute resource's available memory, CPU and storage reserved for use by a particular fabric group. Each reservation is for one fabric group only but the relationship is many-to-many. A fabric group might have multiple reservations on one compute resource or reservations on multiple compute resources or both. The solution implements only one fabric group per region. Each resource cluster has two reservations, one for production and one for development, allowing both production and development workloads to be provisioned. An edge reservation in each region is created and allows NSX to deploy edge services gateways on demand and place them on the edge cluster. Copyright IBM and VMware Page 89 of 132
90 Reservation Policies Each virtual reservation is added to one reservation policy. The reservation from which a particular virtual machine is provisioned is determined by vrealize Automation based on the reservation policy specified in the blueprint (if any), the priorities and current usage of the fabric group's reservations, and other custom properties. Two reservation policies are configured in each region, one for production and the other for development. One edge reservation in each region is created for placement of the edge service gateway Template Synchronization In case of a single region, no template synchronization is made. Dual-region deployment allows provisioning workloads across regions from the same portal using the same single-machine blueprints. vsphere Content Library is the synchronization mechanism for templates across regions this design uses. Figure 31 Template Synchronization Process Orchestration VMware vrealize Orchestrator is a development and process automation and orchestration platform that provides a library of extensible workflows to allow a cloud admin to create and run automated, configurable processes to manage the VMware vsphere infrastructure as well as other VMware and thirdparty technologies Directory Services vrealize Orchestrator instances will use Active Directory LDAP authentication. The only configuration supported for multi-domain Active Directory is domain tree. Forest and external trusts are not supported for process orchestration. Multiple domains that have two-way trust, but are not in the same tree, are not supported and do not work with vrealize Orchestrator Network Ports vrealize Orchestrator uses specific network ports to communicate with other systems. The ports are configured with a default value, which is set by the automation at build time. It is recommended that these values remain unchanged to ensure supportability of the system in the future. Firewall ports within the solution will be opened to ensure communication to the components and intra components. Firewalls not deployed by the solution need to be configured appropriately. Page 90 of 132 Copyright IBM and VMware
91 Table 46 vrealize Orchestrator Default Configuration Ports Port Number Protocol Source Target Description HTTPS Server port Web configuration HTTPS access port 8281 TCP End-user external system 8283 TCP End-user Web browser vrealize Orchestrator server vrealize Orchestrator configuration The SSL secured HTTP protocol used to connect to the vrealize Orchestrator REST API. The SSL access port for the Web UI for vrealize Orchestrator configuration. Table 47 vrealize Orchestrator Default External Communication Ports Port Number Protocol Source Target Description LDAP using SSL LDAP using Global Catalog 636 TCP vrealize Orchestrator server 3268 TCP vrealize Orchestrator server DNS 53 TCP vrealize Orchestrator server VMware vcenter Single Sign-On server (PSC) 443 TCP vrealize Orchestrator server LDAP server Global Catalog server DNS server vcenter Single Sign-On server Lookup port of the Active Directory server for secure LDAP authentication server. Port to which Microsoft Global Catalog server queries are directed. Name resolution Port used to communicate with the vcenter Single Sign-On server. Copyright IBM and VMware Page 91 of 132
92 Port Number Protocol Source Target Description SQL Server SMTP Server port vcenter Server API port vcenter Server VMware ESXi 1433 TCP vrealize Orchestrator server 25 TCP vrealize Orchestrator server 443 TCP vrealize Orchestrator server 80 TCP vrealize Orchestrator server 443 TCP vrealize Orchestrator server Microsoft SQL server SMTP Server VMware vcenter server vcenter Server ESXi hosts Port used to communicate with the Microsoft SQL Server or SQL Server Express instances that are configured as the vrealize Orchestrator database. Port used for notifications. The vcenter Server API communication port used by vrealize Orchestrator to obtain virtual infrastructure and virtual machine information from the orchestrated vcenter Server instances. Port used to tunnel HTTPS communication. Workflows using the vcenter Guest Operations API need direct connection between vrealize Orchestrator and the ESXi hosts the VM is running on vrealize Orchestrator Deployment Two vrealize Orchestrator appliance instances are required within this solution with 2 CPUs, 4 GB memory, and 16 GB of hard disk each. This solution uses MSSQL database already installed Page 92 of 132 Copyright IBM and VMware
93 to support other components. In cluster mode, multiple vrealize Orchestrator instances with identical server and plug-in configurations work together as a cluster, and share a single database. The instances are installed behind a load balancer. Although there are 2 instances for availability, the failover required in a disaster recovery scenario is manual. Please refer to the vrealize user guide for the process to manually failover these components. All vrealize Orchestrator server instances communicate with each other by exchanging heartbeats at a certain time interval. Only active vrealize Orchestrator server instances respond to client requests and run workflows. If an active vrealize Orchestrator server instance fails to send heartbeats, it is considered to be non-responsive, and one of the inactive instances takes over to resume all workflows from the point at which they were interrupted. The heartbeat is implemented through the shared database, so there are no implications in the network design for a vrealize Orchestrator cluster. If there are more than one active vrealize Orchestrator nodes in a cluster, concurrency problems can occur if different users use the different vrealize Orchestrator nodes to modify the same resource. The implementation uses an active-active cluster. The following tables outline characteristics for this vrealize Orchestrator active-active cluster design Table 48 vro Service Monitor Specifications Monitor Interval Timeout Retries Type Send String Receive String vco-https HTTPS (443) GET /vco/api/status\r\n REGISTERED Table 49 vro Service Pool Characteristics Pool Name Algorithm Monitors Members Port Monitor Port vco-pool Leastconn vco-https vrealize Orchestrator nodes Table 50 vro Virtual Server Characteristics Name Type Service Port Source Address Translation Default Pool Name vco-lb-8281 Performance (Layer 4) 8281 Automap vco-pool SSL Certificates The vrealize Orchestrator configuration interface uses a secure connection to communicate with vcenter Server, relational database management systems (RDBMS), LDAP, vcenter Single Sign- On, and other servers. The required SSL certificates are generated by the certification authority deployed within the solution vrealize Orchestrator Plug-Ins Plug-ins allow vrealize Orchestrator to access and control external technologies and applications. Exposing an external technology in a vrealize Orchestrator plug-in allows incorporating objects and functions in workflows that access the objects and functions of the external technology. The external technologies that can be accessed using plug-ins include virtualization management tools, systems, databases, directory services, and remote control interfaces. vrealize Orchestrator provides a set of standard plug-ins. The following plug-ins are configured in this design: Copyright IBM and VMware Page 93 of 132
94 vrealize Orchestrator NSX plug-in vrealize Orchestrator vrealize Automation plug-in vrealize Orchestrator vcenter Server plug-in Multi-node plugin vrealize Orchestrator comes as a single-site topology product. The multi-node plug-in creates a primary-secondary relation between vrealize Orchestrator servers that extends the package management and workflow execution features. This is only enabled when deploying a multiregion topology. The plug-in contains a set of standard workflows for hierarchical orchestration, management of vrealize Orchestrator instances, and the scale-out of vrealize Orchestrator activities vrealize Orchestrator Client The vrealize Orchestrator client is a desktop application that lets users import packages, create, run, and schedule workflows, and manage user permissions. vrealize Orchestrator Client can be installed standalone on a desktop system. Download the vrealize Orchestrator Client installation files from the vrealize Orchestrator appliance page: Alternatively, vrealize Orchestrator Client can be run using Java WebStart directly from the homepage of the vrealize Orchestrator appliance console vrealize Orchestrator Scalability A single vrealize Orchestrator instance allows up to 300 concurrent workflow instances in the running state. Workflow instances that are in the waiting or waiting-event states do not count toward that number. You can design long running workflows in a way that preserves resources by using the wait elements of the workflow palette. A single vrealize Orchestrator instance supports up to 35,000 managed virtual machines in its inventory. This architecture depicts a clustered vrealize Orchestrator environment. In a clustered environment, workflows cannot be changed while other vrealize Orchestrator instances are running. Stop all other vrealize Orchestrator instances before connecting the vrealize Orchestrator client and changing or developing a new workflow. Failure to do so will result in inconsistencies within the environment. This architecture scale s out a vrealize Orchestrator environment by having multiple independent vrealize Orchestrator instances (each with their own database instance). This allows for the increase in the number of managed inventory objects. This solution implements an active-active cluster with two nodes Software Orchestration This solution provides a centralized repository for software binaries and software orchestration templates that are implemented on deployed resources. Main software orchestration engines installed are Chef Server and Salt Stack. Each region hosts its own dedicated repository server and software orchestration stack. Software binaries are replicated between regions using rsync tool. Software orchestration components use internal specific replication mechanisms. Page 94 of 132 Copyright IBM and VMware
95 The following diagram shows a high level view of the software orchestration components. Figure 32 Software Orchestration Logical Design A single region design implements a central orchestration and a region orchestration Central Software Orchestration Central software orchestration has no direct interaction with the resources, but provides a central management point for maintaining the latest binaries and templates by the cloud administrator. It is responsible for, Maintaining the latest versions of software binaries Maintaining the latest versions of software orchestration templates Keeping the regions up to date with the latest software versions and templates Inputs and Outputs A central component pushes software binaries to the region component. A Central component is accessed by the cloud administrator to update software and templates. Design Rationale In a hybrid cloud there may be many regions each with their own software and template requirements. Plus, to avoid delays during deployment, the software binaries need to be as close to the resource cluster as possible. This drives the need for a distributed repository of software binaries. Maintaining these separately and keeping software versions on each up to date with the latest versions can add considerable support cost. A central repository can regular or on change update each of these regions without the need to update each one individually. Implementation Approach The central software orchestration components are co-located with the other central services to minimize network access configuration for administrator access. Other Functions The central repository is also used as the backup location for the NSX Managers in the central cloud Region Software Orchestration Responsibilities The region component provides the software orchestration engine that implements software to deployed resources using defined software orchestration templates. The implementation includes mounting the software installation media (software binaries), then installing and configuring the software. The implementation may also involve other actions to configure the deployed resource to the software orchestration template requirements such as setting password rules. Copyright IBM and VMware Page 95 of 132
96 It is responsible for, being a repository for software binaries being a repository for software orchestration templates providing software orchestration engine to implement software on deployed resources Inputs and Outputs A region pushes software and configurations to deployed resources. A region is called from a cloud region to deploy the software. Design Rationale A cloud needs to be able to implement software on deployed resources. This requires software binaries and software orchestration templates (patterns). The region is to provide these functions. It is separate to the cloud region in order to allow for different flavors without impacting the cloud region design. Implementation Approach The region should be co-located with the other region services to minimize network traffic and response times when deploying software to resources. Other Functions The region repository is also used as the backup location for the NSX Managers in the cloud region Software Orchestration Components Sizing The following table presents sizing for central and region Table 51 Software Orchestration Components Sizing Server Role vcpu RAM (GB) Disk (GB) Central repository and Chef server Central Salt master Region repository and Chef server Region Salt master Infrastructure Orchestration Infrastructure orchestration is handled by vrealize Orchestration and vrealize Automation which is covered in the prior sections. 5.6 Operational Services Operational services provide the services to enable the support and maintenance of the cloud management platform. This design does not include operational services for user related virtual resources Backup and Restore Data backup protects the data of this design against data loss, hardware failure, accidental deletion, or other disaster for each region. For consistent image-level backups, this design uses backup software based upon the VMware Virtual Disk Development Kit (VDDK), such as vsphere Data Protection (VDP). For this design, VDP is used to back up the infrastructure VMs with the exception of the NSX components, which have their own backup and recovery procedure documented in the networking section. Page 96 of 132 Copyright IBM and VMware
97 Logical Design vsphere Data Protection protects the virtual infrastructure at the VMware vcenter Server layer. Because vsphere Data Protection is connected to the Management vcenter Server, it can access all management ESXi hosts, and can detect the virtual machines that require backups Backup Datastore Figure 33 vsphere Data Protection Logical Design vsphere Data Protection uses deduplication technology to back up virtual environments at data block level, which enables efficient disk utilization. To optimize backups and leverage the VMware vsphere Storage APIs, all ESXi hosts must have access to the NFS datastore. The backup datastore stores all the data that is required to recover services according to a Recovery Point Objective (RPO) Performance vsphere Data Protection generates a significant amount of I/O operations, especially when performing multiple concurrent backups. The storage platform must be able to handle this I/O. If the storage platform does not meet the performance requirements, it might miss backup windows. Backup failures and error messages might occur. Run the vsphere Data Protection performance analysis feature during virtual appliance deployment or after deployment to assess performance. For this design a dedicated volume on performance storage in SoftLayer is used as a backup target. The backup volume is NFS mounted to all ESXi hosts in the management cluster. Table 52 VMware vsphere Data Protection Performance Total Backup Size Avg Mbps in 4 hours 0.5 TB 306 Mbps 1 TB 611 Mbps 2 TB 1223 Mbps Copyright IBM and VMware Page 97 of 132
98 Volume Sizing vsphere Data Protection can dynamically expand the destination backup store from 2 TB to 8 TB. Using an extended backup storage requires additional memory on the vsphere Data Protection appliance. For this design, an initial fixed NFS datastore size will be utilized. Additional NFS space can be provisioned via the SoftLayer customer portal as required. Note that an existing SoftLayer Endurance NFS storage allocation cannot be expanded once provisioned. If additional storage is required, then a new export must be requested and migrated to. For this design, 4TB of storage is initially sized with a 4 CPU and 12GB of memory for the vdp appliance virtual machine for each region Other Considerations vsphere Data Protection can protect virtual machines that reside on VMware Virtual SAN from host failures. The virtual machine storage policy is not backed up with the virtual machine, but a user is able to restore the storage policy after restoring the virtual machine. Note: The default Virtual SAN storage policy is configured and includes Number Of Failures To Tolerate = 1, which means that virtual machine data will be mirrored. vsphere Data Protection is used to restore virtual machines that failed or need their data reverted to a previous state. The FTP server required for backing up NSX Manager will reside on the RDS central repository server. Backups for NSX Manager setup with a daily schedule by the NSX Manager Backup Policies Use vsphere Data Protection backup policies to specify virtual machine backup options, the schedule window, and retention policies Virtual Machine Backup Options vsphere Data Protection provides the following options for performing a backup of a virtual machine: Hot Add. Provides full image backups of virtual machines, regardless of the guest operating system. o The virtual machine base disk is attached directly to vsphere Data Protection to back up data. vsphere Data Protection uses Changed Block Tracking to detect and back up blocks that are altered. o The backup and restore performance is faster because the data flow is through the VMkernel layer instead of over a network connection. o A quiesced snapshot can be used to redirect the I/O of a virtual machine disk.vmdk file. o Hot Add does not work in multi-writer disk mode. Network Block Device (NBD). Transfers virtual machine data across the network to allow vsphere Data Protection to back up the data. o o The performance of the virtual machine network traffic might be lower. NBD takes a quiesced snapshot. As a result, it might interrupt the I/O operations of the virtual machine to swap the.vmdk file or consolidate the data after the backup is complete. o The time to complete the virtual machine backup might be longer than the backup window. o NBD does not work in multi-writer disk mode. vsphere Data Protection Agent Inside Guest OS. Provides backup of certain applications that are running in the guest operating system through an installed backup agent. Page 98 of 132 Copyright IBM and VMware
99 o o Enables application-consistent backup and recovery with Microsoft SQL Server, Microsoft SharePoint, and Microsoft Exchange support. Provides more granularity and flexibility to restore on the file level. For this design, Hot Add will be used for all backup policies with the exception of the infrastructure SQL server backups with will utilize the Guest OS Agent, in addition to local SQL backups to disk Schedule Window Even though vsphere Data Protection uses the Changed Block Tracking technology to optimize the backup data, do not use a backup window when the production storage is in high demand to avoid any business impact. Warning: Do not perform any backup or other administrative activities during the vsphere Data Protection maintenance window. Restore operations are allowed. By default, the vsphere Data Protection maintenance window begins at 8 AM local server time and continues uninterrupted until 8 PM or until the backup jobs are complete. Configure maintenance windows according to IT organizational policy requirements. The backup window is set to default and can be modified, based on customer requirements Retention Policies Retention policies are properties of a backup job. If virtual machines are grouped by business priority, it is possible to set the retention requirements according to the business priority. This design requires that a dedicated NFS datastore be allocated for the vsphere Data Protection appliance for the backup data in each region. The NFS datastore is allocated from IBM SoftLayer Performance storage option with an initial allocation of 4TB Component Backup Jobs Backups are configured for each of this design s management components separately. No requirement to back up the entire design exists, and this design does not imply such an operation. Some products can perform internal configuration backups. Separate from this, NSX is configured to backup to the FTP server within the design. The SQL server will also be configured for local backups to disk on a daily basis Backup Jobs in Central Cloud If multiple regions are deployed, create a single backup job for the components of a management application according to the node configuration of the application in central cloud. Table 53 Backup Jobs in Central Cloud ESXi Product Image VM Backup Jobs in Central Cloud Application VM Backup Jobs in Central Cloud Platform Services Controller N/A - No Backup Part of the vcenter Server backup job Copyright IBM and VMware Page 99 of 132
100 vcenter Server Management Job vrealize Automation vrealize Log Insight vrealize Operations Manager vrealize Orchestrator mgmt01vc01.central.tornado.loc al mgmt01psc01.central.tornado.lo cal Compute Job comp01vc01.central.tornado.loc al comp01psc01.central.tornado.lo cal vra01vro01a.tornado.local vra01vro01b.tornado.local vra01dem01.tornado.local vra01dem02.tornado.local vra01ias01.tornado.local vra01ias02.tornado.local vra01ims01a.tornado.local vra01ims01b.tornado.local vra01iws01a.tornado.local vra01iws01b.tornado.local vra01svr01a.tornado.local vra01svr01b.tornado.local vra01mssql01.tornado.local vra01ids01a.tornado.local vrli-mstr-01 vrli-wrkr-01 vrli-wrkr-02 vrops-mstrn-01 vrops-repln-02 vrops-datan-03 vrops-datan-04 vrops-rmtcol-01 vrops-rmtcol-02 Part of the vrealize Automation backup job vra01mssql01.torna do.local Page 100 of 132 Copyright IBM and VMware
101 Backup Jobs in Additional Cloud Region Create a single backup job for the components of a management application according to the node configuration of the application in an additional cloud region. Table 54 Backup Jobs in Additional Cloud Region ESXi Product Platform Services Controller vcenter Server Image VM Backups in Additional Cloud Region N/A - No Backup Part of the vcenter Server backup job Management Job Application VM Backup Jobs in Additional Cloud Region mgmt01vc51.regiona.tornado.local mgmt01psc51.regiona.tornado.local None Compute Job NSX for vsphere comp01vc51.regiona.tornado.local comp01psc51.regiona.tornado.local Management Job mgmt01nsxm51.regiona.tornado.local Compute Job vrealize Automation vrealize Log Insight vrealize Operations Manager vrealize Orchestrator comp01nsxm51.regiona.tornado.local vra01ias51.tornado.local vra01ias52.tornado.local vrli-mstr-51 vrli-wrkr-51 vrli-wrkr-52 vrops-rmtcol-51 vrops-rmtcol-52 Part of the vrealize Automation backup job Disaster Recovery A Site Recovery Manager instance is required for both the protected region and the recovery region. Site Recovery Manager is installed after the installation and configuration of vcenter Server and the Platform Services Controller in the region. Site Recovery Manager takes advantage of vcenter Server and Platform Services Controller services such as storage management, authentication, authorization, and guest Copyright IBM and VMware Page 101 of 132
102 customization. Site Recovery Manager uses the standard set of vsphere administrative tools to manage these services. Site Recovery Manager is installed on a dedicated window host VM within each region the design Networking Design for Disaster Recovery Moving a service physically from one region to another represents a networking challenge, especially if applications have hard-coded IP addresses. Network address space and IP address assignment considerations require that either the same IP address or a different IP address at the recovery region is used. In many situations, new IP addresses are assigned, because VLANs do not typically stretch between regions. While protecting the management applications, it is possible to simplify the problem of IP address assignment. This design leverages a load balancer to separate a public network segment and a private network segment. The private network can remain unchanged and only the external load balancer interface has to be reassigned. On the public network segment, the management application is accessible via one or more virtual IP (VIP) addresses On the isolated private network segment, the application's virtual machines are isolated After a failover, the recovered application is available under a different IPv4 address (VIP). The use of the new IP address requires changes to the DNS records. DNS records are either changed manually or by using a script in the Site Recovery Manager recovery plan. Figure 34 Logical Network Design for Cross-Region Deployment with Management Application Network Container The IPv4 subnets (orange networks) are routed within the vsphere management network of each Page 102 of 132 Copyright IBM and VMware
103 region. Nodes on these network segments are reachable from within the SDDC. IPv4 subnets, such as the subnet for the vrealize Automation primary components, overlap across a region. Make sure that only the active IPv4 subnet is propagated in the region and beyond. The public facing Ext-Mgmt network of both regions (grey networks) are reachable by users and provide connection to external resources, such as Active Directory or DNS. Load balancing functionality is provided by NSX Edge services gateways. In each region, the same configuration for the management applications and their Site Recovery Manager shadow will be used. Active Directory and DNS services must be running in both the protected and recovery regions vsphere Replication In a VMware Virtual SAN environment, array-based replication cannot be used. This design utilizes vsphere Replication instead to transfer VMs between regions. Within the design, vsphere Replication uses a VMkernel management interface on the ESXi host to send replication traffic to the replication site's vsphere Replication appliance. To isolate vsphere Replication traffic so that it does not impact other vsphere management traffic, the vsphere Replication network is configured in the following way. vsphere Replication appliances and vsphere Replication servers are the target for the replication traffic that originates from the vsphere Replication VMkernel ports Placeholder Virtual Machines Site Recovery Manager creates a placeholder virtual machine on the recovery region for every machine from the Site Recovery Manager protection group. Placeholder virtual machine files contain virtual machine configuration metadata but not virtual machine disks, and the files are very small. Site Recovery Manager adds the placeholder virtual machines as recovery region objects in the Management vcenter Server Snapshot Space To perform failover tests, additional storage is provided for the snapshots of the replicated VMs. This storage is minimal in the beginning, but grows as test VMs write to their disks. Replication from the protected region to the recovery region continues during this time. The snapshots created during testing are deleted after the failover test is complete Messages and Commands for Site Recovery Manager Site Recovery Manager has options to present users with messages that provide notification and accept acknowledgement. Site Recovery Manager also provides a mechanism to run commands and scripts as necessary when executing a recovery plan. Pre-power-on or post-power-on messages and commands can be inserted to the recovery plans. These messages and commands are not specific to Site Recovery Manager, but support pausing the execution of the recovery plan to complete other procedures, or executing customer-specific commands or scripts to enable automation of recovery tasks Site Recovery Manager Messages Additional steps are required to be configured when building more than the central cloud (i.e. adding cloud regions). For example, the environment should be setup such that a message appears when a recovery plan is initiated, and that the cloud admin must acknowledge the message before the recovery plan continues. Messages are specific to each IT organization. Consider the following example messages and confirmation steps: Verify that IP address changes are made on the DNS server and that the changes are propagated. Copyright IBM and VMware Page 103 of 132
104 Verify that the Active Directory services are available. After the management applications are recovered, perform application tests to verify that the applications are recovered correctly. Additionally, confirmation steps can be inserted after every group of services that have a dependency on other services. These conformations can be used to pause the recovery plan so that appropriate verification and testing be performed before subsequent steps are taken. These services are defined as follows: Infrastructure services Core services Database services Middleware services Application services Web services Details on each message are specified in the workflow definition of the individual recovery plan Site Recovery Manager Commands In this initial phase of the design, custom scripts are out of scope. However, custom scripts can be running to perform infrastructure configuration updates or configuration changes on the virtual machine environment. The scripts that a recovery plan executes are located on the Site Recovery Manager server. The scripts can run against the Site Recovery Manager server or can impact a virtual machine. If a script must run in the virtual machine, Site Recovery Manager does not run it directly, but instructs the virtual machine to do it. The audit trail that Site Recovery Manager provides does not record the execution of the script because the operation is on the target virtual machine. Scripts or commands must be available in the path on the virtual machine according to the following guidelines: Use full paths to all executables. For example, c:\windows\system32\cmd.exe instead of cmd.exe. Call only.exe or.com files from the scripts. Command-line scripts can call only executables. To run a batch file, start the shell command with c:\windows\system32\cmd.exe. The scripts that are run after powering on a virtual machine are executed under the Local Security Authority of the Site Recovery Manager server. Store post-power-on scripts on the Site Recovery Manager virtual machine. Do not store such scripts on a remote network share Recovery Plans for Site Recovery Manager A recovery plan is the automated plan (runbook) for full or partial failover from the central cloud to a cloud region Startup Order and Response Time Virtual machine priority determines virtual machine startup order. Page 104 of 132 Copyright IBM and VMware
105 All priority 1 virtual machines are started before priority 2 virtual machines. All priority 2 virtual machines are started before priority 3 virtual machines. All priority 3 virtual machines are started before priority 4 virtual machines. All priority 4 virtual machines are started before priority 5 virtual machines. Additionally, a startup order of virtual machines within each priority group can be set. The following timeout parameters are set: Response time, which defines the time to wait after the first virtual machine powers on before proceeding to the next virtual machine in the plan. Maximum time to wait if the virtual machine fails to power on before proceeding to the next virtual machine. The response time values can be adjusted as necessary during execution of the recovery plan test to determine the appropriate response time values Recovery Plan Test Network When a recovery plan is created, the test network options must be configured. The following options are available. Isolated Network (Automatically Created). An isolated private network is created automatically on each ESXi host in the cluster for a virtual machine that is being recovered. Site Recovery Manager creates a standard switch and a port group on it. A limitation of this automatic configuration is that a virtual machine connected to the isolated port group on one ESXi host cannot communicate with a virtual machine on another ESXi host. This option limits testing scenarios and provides an isolated test network only for basic virtual machine testing. Port Group. Selecting an existing port group provides a more granular configuration to meet the client s testing requirements. For virtual machines across ESXi hosts to communicate, distributed switches with uplinks to the production network are used and a port group is created on the switch that is tagged with a non-routable VLAN. In this way, the network is isolated and cannot communicate with other production networks. Because the isolated application networks are fronted by a load balancer, the recovery plan test network is equal to the recovery plan production network and provides realistic verification of a recovered management application Sizing For each region, 1 SRM server and one vsphere Replication appliance are deployed. The SRM application is installed on a Windows 2012R2 virtual machine and utilizes the built in PostgreSQL DB. The vsphere Replication application is deployed as a virtual appliance. Table 55 SRM Windows server sizing VM size Attribute Number of CPUs 4 Memory Disk size Specification Medium 8 GB 60 GB Copyright IBM and VMware Page 105 of 132
106 Table 56 vsphere Replication Appliance Attribute Specification VM size Medium Number of CPUs 2 Memory 4 GB Disk size 18G Monitoring Monitoring and Operations Management is a required element of a software-defined datacenter. Monitoring operations support in vrealize Operations Manager provides capabilities for performance and capacity management of related infrastructure and cloud management components Single-Region Logical Design In this single-region design, vrealize Operations Manager is deployed with the following configuration: 4-node (large-size) vrealize Operations Manager analytics cluster that is highly available (HA). This topology provides high availability, scale-out capacity up to eight nodes, and failover. 2-node (large-size) remote collector cluster. The remote collectors communicate directly with the data nodes in the vrealize Operations Manager analytics cluster. For a multiregion design, deploy two remote collectors in each region. Note that a single region has its own remote collectors whose role is to ease scalability by performing the data collection from the applications that are not a subject of failover and periodically sending collected data to the analytics cluster. In cases of multiple regions, the analytics cluster can be failed over because the analytics cluster is the construct that analyzes and stores monitoring data. The multi-region configuration supports failover of the analytics cluster by using Site Recovery Manager. In the event of a disaster, Site Recovery Manager migrates the analytics cluster nodes to the failover region. Page 106 of 132 Copyright IBM and VMware
107 Figure 35 Logical Design of vrealize Operations Manager Central Cloud and a Cloud Region (Region A) Deployment Physical Design The vrealize Operations Manager nodes run on the management cluster in each region of this design Data Sources vrealize Operations Manager collects data from the following virtual infrastructure and cloud management components: Management vcenter Server o Platform Services Controller o vcenter Server Compute vcenter Server o Platform Services Controller o vcenter Server Management, Edge and Compute ESXi hosts NSX for vsphere for the management and compute clusters o NSX Manager o NSX Controller Instances o NSX Edge instances vrealize Automation o vrealize Orchestrator o vrealize Automation Components vrealize Log Insight vrealize Operations Manager (Self Health Monitoring) vrealize Operations Manager Nodes The analytics cluster of the vrealize Operations Manager deployment contains the nodes that analyze and store data from the monitored components Compute for vrealize Operations Manager Nodes The four-node vrealize Operations Manager analytics is deployed in the management cluster with an application virtual network. The analytics cluster consists of one master node, one master replica node, and two data nodes to enable scale out and high availability. The size of the vrealize Operations Manager analytics cluster is sized according to VMware KB article "vrealize Operations Manager 6.1 Sizing Guidelines" and includes the following management packs: Management Pack for VMware vcenter Server (installed by default) Management Pack for NSX for vsphere Management Pack for Storage Devices Management Pack for vrealize Log Insight Management Pack for vrealize Automation Note that each node in the analytics cluster will be sized identically to support scale-out, high availability, and design guidance for central cloud or cloud region infrastructure. As a result, the configuration for each node is as follows: Table 57 Analytics Cluster Node Configurations Node vcpu Memory Master GB Master Replica GB Copyright IBM and VMware Page 107 of 132
108 Data GB Data GB Storage for vrealize Operations Manager Nodes Each vrealize Operations Manager node in this design require 266GB of free space for data. To collect the required number of metrics, a 1 TB VMDK will be added to each analytics cluster node Network for vrealize Operations Manager Nodes In this design, the clusters of vrealize Operations Manager will be placed in application isolated networks for secure access, load balancing, portability, and functionality-specific subnet allocation High Availability for vrealize Operations Manager Nodes To protect the vrealize Operations Manager virtual machines from a host-level failure, this design configures vsphere DRS to run the virtual machines for the analytics cluster and for the remote collectors on different hosts in the management cluster. Table 58 DRS Cluster Anti-Affinity Rule for vrealize Operations Manager Nodes Rule Attribute Name Enable rule Type Members Value vropscluster-antiaffinity-rule Yes Separate Virtual Machines vrops master node vrops master replica node vrops data node 1 vrops-data node vrealize Operations Remote Collector Nodes Unlike the analytics cluster nodes, remote collector nodes have only the collector role. Deploying two remote collector nodes in each region does not increase the number of monitored objects, but removes the load from the analytics cluster by collecting metrics from applications that do not fail over between regions in a multi-region environment. This means that collectors are assigned when configuring the monitoring solution Compute for vrealize Operations Remote Collector Nodes In this design, the remote collector nodes are deployed on the management cluster. The nodes are identically sized and consist of the following compute resources in each region: Table 59 Remote Collector Node Sizes Node vcpu Memory Remote Collector GB Remote Collector GB Page 108 of 132 Copyright IBM and VMware
109 Storage for Remote Collector Nodes In this design the remote collector nodes are with thin-provisioned disks. Because remote collectors do not perform analytics operations or store data, the default VMDK size is sufficient for this design High Availability for vrealize Operations Manager Nodes To protect the vrealize Operations Manager virtual machines from a host-level failure, this design configures vsphere DRS to run the virtual machines for the analytics cluster and for the remote collectors on different hosts in the management cluster. Table 60 DRS Cluster Anti-Affinity Rule for vrealize Operations Remote Collector Nodes Rule Attribute Name Enable rule Type Value vropscollector-antiaffinity-rule Yes Separate Virtual Machines Members vrops remote collector 1 vrops remote collector Networking for vrealize Operations Manager In this design, the clusters of vrealize Operations Manager will be placed in application isolated networks for secure access, load balancing, portability, and functionality-specific subnet allocation. Copyright IBM and VMware Page 109 of 132
110 Figure 36 Networking Design of the vrealize Operations Manager Deployment Application Isolated Network Design In the single-region design, the two logical entities of the vrealize Operations Manager deployment, the analytics cluster and remote collectors in the central cloud is installed in an application isolated network. For multi-region, a third entity (i.e., remote collectors in an additional cloud region) is also deployed on an application isolated network. As part of this configuration, an NSX Edge services gateway will be placed in front of the application isolated network to provide routing and load balancing. This networking design contains the following features: Page 110 of 132 Copyright IBM and VMware
111 Each application virtual network of vrealize Operations Manager has connection to the application virtual networks of vrealize Automation and vrealize Log Insight through a dedicated network called networkexchange. The role of networkexchage is to support transit traffic and the exchange of routing tables. All nodes have routed access to the vsphere management network through the Management NSX Edge for the home region. Routing to the vsphere management network and the external network is dynamic, and is based on the Open Shortest Path First (OSPF) protocol. The NSX Edge instances for the analytics cluster are configured to use Source NAT (SNAT) address translation when the analytics nodes access the public network. Figure 37 Application Virtual Networks in the vrealize Operations Manager Topology IP Subnets In this design, the following table lists some example subnets for each cluster in the vrealize Operations Manager deployment. Note that placing remote collectors on their own subnet enables them to communicate with the analytics cluster and not be a part of the failover group. Table 61 IP Subnets in the Application Virtual Network of vrealize Operations Manager vrealize Operations Manager Cluster Type IP Subnet Analytics cluster in the central cloud (also valid for any additional cloud /24 region for failover) Remote collectors in the central cloud /24 Remote collectors in any additional cloud region /24 Copyright IBM and VMware Page 111 of 132
112 DNS Names vrealize Operations Manager node name resolution uses a region-specific suffix, such as central.tornado.local or regiona.tornado.local, the analytics nodes IP addresses and the load balancer virtual IP address (VIP) are mapped to the root domain suffix tornado.local. Access from the public network is provided through a VIP, the traffic to which is handled by the NSX Edge service gateway. The following table shows example names that can be used in a single- and multi-region design. Table 62 DNS Names for the Application Virtual Networks vrealize Operations Manager DNS Name vrops-cluster-01.tornado.local vrops-mstrn-01.tornado.local vrops-repln-02.tornado.local vrops-datan-03.tornado.local vrops-datan-04.tornado.local vrops-rmtcol-01.central.tornado.local vrops-rmtcol-02.central.tornado.local vrops-rmtcol-51.regiona.tornado.local vrops-rmtcol-52.regiona.tornado.local Networking for Failover and Load Balancing Node Type Virtual IP of the analytics cluster Master node in the analytics cluster Master replica node in the analytics cluster First data node in the analytics cluster Second data node in the analytics cluster First remote collector node in the central cloud Second remote collector node in the central cloud First remote collector node in any additional cloud region Second remote collector node in any additional cloud region Each node in the vrealize Operations Manager analytics cluster runs a Tomcat server instance for access to the product user interface. By default, vrealize Operations Manager does not provide a solution for load-balanced UI users sessions across nodes in the cluster. The lack of load balancing for users sessions results in the following limitations: Cloud admin must know the URL of each node to access the UI. As a result, a single node might be overloaded if all cloud admin actors access it at the same time. Each node supports up to four simultaneous cloud admin sessions. Taking a node offline for maintenance might cause an outage. Cloud admin s cannot access the UI of the node when the node is offline. To avoid such problems, analytics cluster is placed behind an NSX load balancer that is configured to allow up to four connections per node. The load balancer must distribute the load evenly to all cluster nodes. In addition, the load balancer is configured to redirect service requests from the UI on port 80 to port 443. Load balancing and access to and from the public network is not required for the remote collector nodes Security and Authentication vrealize Operations Manager can use several sources for authentication. These sources include an Active Directory service, vcenter Server, and local user inventory. Active Directory is used as the primary authentication and authorization method in this design. Page 112 of 132 Copyright IBM and VMware
113 Identity Sources The cloud admin will authenticate in vrealize Operations Manager using Active Directory authentication. This provides access to vrealize Operations Manager by using standard Active Directory accounts and ensures that authentication is available even if vcenter Server becomes unavailable Encryption Access to all vrealize Operations Manager Web interfaces requires an SSL connection. By default, vrealize Operations Manager uses a self-signed certificate. In this design, the default selfsigned certificates is replaced with a Certificate Authority (CA) signed certificate to provide secure access to the vrealize Operations Manager user interface Monitoring and Alerting vrealize Operations Manager can monitor itself and display the following administrative alerts: System alert. A component of the vrealize Operations Manager application has failed. Environment alert. vrealize Operations Manager has stopped receiving data from one or more resources. Such an alert might indicate a problem with system resources or network infrastructure. Log Insight log event. The infrastructure on which vrealize Operations Manager is running has low-level issues. You can also use the log events for root cause analysis. Custom dashboard. vrealize Operations Manager can show super metrics for datacenter monitoring, capacity trends and single pane of glass overview. In order to enable the aforementioned alerts and events, the defaults must be configured in vrealize Operations Manager to enable SMTP outbound alerts. An SMTP service is included in the design which utilizes the SoftLayer SMTP service if no client on premises SMTP service is provided. If the client has an existing SMTP service, then this will be configured. Additionally, the design will incorporate deeper root cause analysis and infrastructure alerting by containing the management pack for vrealize Log Insight Management Packs This design contains several VMware products for network, storage, and cloud management. In order to monitor and perform diagnostics on all these items, the following management packs are used: Management Pack for VMware vcenter Server (installed by default) Management Pack for NSX for vsphere Management Pack for Storage Devices Management Pack for vrealize Log Insight Management Pack for vrealize Automation Copyright IBM and VMware Page 113 of 132
114 5.6.4 Log Consolidation and Analysis In each region of this design, a vrealize Log Insight cluster is configured with three nodes. This allows for continued availability and increased log ingestion rates. Figure 38 Logical Design of vrealize Log Insight Sources of Log Data vrealize Log Insight collects logs from the following virtual infrastructure and cloud management components: Management vcenter Server o Platform Services Controller o vcenter Server Compute vcenter Server o Platform Services Controller o vcenter Server Management, Edge and Compute ESXi hosts NSX for vsphere for the management and for compute and edge clusters o NSX Manager o NSX Controller instances o NSX Edge instances vrealize Automation o vrealize Orchestrator o vrealize Automation components vrealize Operations Manager o Analytics cluster nodes Cluster Nodes The vrealize Log Insight cluster consists of one master node and two worker nodes. The Integrated Load Balancer (ILB) on the cluster is configured to have vrealize Log Insight to balance incoming traffic fairly among available nodes. vrealize Log Insight clients, using both Page 114 of 132 Copyright IBM and VMware
115 the Web user interface, and ingestion through syslog or the Ingestion API, connect to vrealize Log Insight at the ILB address Sizing In this design, the vrealize Log Insight virtual appliance has 2 vcpus, 4 GB of virtual memory, and 144 GB of disk space provisioned. vrealize Log Insight uses 100 GB of the disk space to store raw, index and metadata Sizing Nodes To accommodate all of log data from the products in this design, the Log Insight nodes must be sized properly. Table 63 Node Sizing Appliance size Attribute Number of CPUs 8 Memory IOPS Amount of processed log data Medium 16 GB 1,000 IOPS 38 GB/day Number of process log messages 7,500 Environment Disk size Sizing Storage Specification Up to 250 syslog connections per node 450 GB (see below) For this design we will retain 7 days of data. Using the following calculations for the disk space needs: For 250 syslog sources at a rate of 150 MB of logs ingested per-day per-source over 7 days: 250 sources * 150 MB of log data 37 GB log data per-day 37 GB * 7 days 260 GB log data per vcrarealize Log Insight node 260 GB * 1.7 overhead index 450 GB Note: vrealize Log Insight supports virtual hard disks of up to 2 TB. If more capacity is needed, add another virtual hard disk. Do not extend existing retention virtual disks. Copyright IBM and VMware Page 115 of 132
116 Networking Design In a multi-region deployment of the design, the vrealize Log Insight instances are connected to both the vsphere management (gray in the network diagram) and the external management (blue in the network diagram) networks. Each vrealize Log Insight instance is deployed within its own application isolated network (gray boxes in the network diagram). Figure 39 Networking Design for the vrealize Log Insight Deployment Application Isolated Network Design Each of the two instances of the vrealize Log Insight deployment is installed in its own isolated application network. An NSX Edge appliance is configured at the front of each isolated application network to provide the network isolation. NSX Edge services gateway is deployed in front of the application isolated network to provide routing and load balancing. This networking design has the following features: Each application virtual network of vrealize Log Insight has connection to the application virtual networks of vrealize Automation and vrealize Operations Manager through a Page 116 of 132 Copyright IBM and VMware
117 dedicated network called networkexchange. The role of networkexchange is to support transit traffic and the exchange of routing tables. All nodes have routed access to the vsphere management network through the Management NSX Edge for the home region. Routing to the vsphere management network and the external network is dynamic, and is based on the Open Shortest Path First (OSPF) protocol. The NSX Edge instances for the vrealize Log Insight are configured to use Source NAT (SNAT) address translation when the vrealize Log Insight nodes access the public network. The NSX Edge instances for the vrealize Log Insight provide access to vrealize Log Insight from the public network over Destination NAT (DNAT) IP Subnets Figure 40 Application Virtual Networks in the vrealize Log Insight Topology The following example subnets are allocated to the vrealize Log Insight deployment: Table 64 IP Subnets in the Application Isolated Networks vrealize Log Insight Cluster IP Subnet Central Cloud /24 Cloud Region / DNS Names vrealize Log Insight node name resolution uses a region-specific suffix, such as central.tornado.local or lax01.tornado.local, including the load balancer virtual IP addresses (VIPs). The Log Insight components in both regions have the following node names: Table 65 Example DNS names of Log Insight nodes DNS Name Role Region vrli-cluster-01.central.tornado.local Log Insight ILB VIP A Copyright IBM and VMware Page 117 of 132
118 vrli-mstr01.central.tornado.local Master node A vrli-wrkr01.central.tornado.local Worker node A vrli-wrkr02.central.tornado.local Worker node A vrli-cluster-51.regiona.tornado.local Log Insight ILB VIP B vrli-mstr51.regiona.tornado.local Master node B vrli-wrkr51.regiona.tornado.local Worker node B vrli-wrkr52.regiona.tornado.local Worker node B Retention and Archiving In vrealize Log Insight, configure log retention for one week and archiving on storage sized for 90 days according to the vrealize Log Insight Design document Retention vrealize Log Insight virtual appliances contain three default virtual disks and can use addition virtual disks for storage, for example, hard disk 4. Table 66 Virtual Disk Configuration in the vrealize Log Insight Virtual Appliance Hard Disk Size Usage Hard disk GB Root file system Hard disk GB for medium-size deployment Contains two partitions: Hard disk MB First boot only Hard disk 4 (additional virtual disk) /storage/var System logs /storage/core Storage for Collected logs. 190 GB Storage for collected logs. The capacity from this disk is added to /storage/core. Calculate the storage space that is available for log data in the following way: /storage/core = hard disk 2 space + hard disk 4 space - system logs space on hard disk 2 Based on the size of the default and additional virtual disks, the storage core is equal to /storage/core = 270 GB GB - 20 GB = 440 GB Retention = /storage/core 3% * /storage/core If /storage/core is 425 GB, vrealize Log Insight can use 413 GB for retention. Retention = 440 GB - 3% * GB Configure a retention period of 7 days for the medium-size vrealize Log Insight appliance. Page 118 of 132 Copyright IBM and VMware
119 Archiving vrealize Log Insight archives log messages as soon as possible. In the same time, they remain retained on the virtual appliance until the free local space is almost over. Data exists on both the vrealize Log Insight appliance and the archive location for most of the retention period. The archiving period must be longer than the retention period. The archive location must be on an NFS version 3 shared storage. The NFS export is mounted directly to the vrealize Log Insight nodes, configured via the management web interface. This design will apply an archive policy of 90 days for the medium-size vrealize Log Insight appliance which takes up approximately about 1 TB of shared storage Alerting vrealize Log Insight supports alerts that trigger notifications about its health. The following types of alerts exist in vrealize Log Insight: System Alerts. vrealize Log Insight generates notifications when an important system event occurs, for example when the disk space is almost exhausted and vrealize Log Insight must start deleting or archiving old log files. Content Pack Alerts. Content packs contain default alerts that can be configured to send notifications, these alerts are specific to the content pack and are disabled by default. User-Defined Alerts. Administrators and users can define their own alerts based on data ingested by vrealize Log Insight. vrealize Log Insight handles alerts in two ways: Send an over SMTP Send to vrealize Operations Manager SMTP Notification notifications are enabled for alerts in vrealize Log Insight to point to the SMTP relay service residing on the active directory server VMs Integration with vrealize Operations Manager vrealize Log Insight integrates with vrealize Operations Manager to provide a central location for monitoring and diagnostics. vrealize Log Insight integrates with vrealize Operations Manager in the following ways: Notification Events. Forward notification events from vrealize Log Insight to vrealize Operations Manager. Launch in Context. Launch vrealize Log Insight from the vrealize Operation Manager user interface. This requires the vrealize Log Insight management pack installed in vrealize Operations Manager Security and Authentication vrealize Log Insight deployment utilizes a centralized role-based authentication using Microsoft Active Directory as the authority Authentication Role-based access control is enabled in vrealize Log Insight by using the existing tornado.local Active Directory domain. Copyright IBM and VMware Page 119 of 132
120 Encryption The default self-signed certificates are replaced with a CA-signed certificate to provide secure access to the vrealize Log Insight Web user interface from the designs AD CA generated certs Configuration for Collecting Logs Client applications send logs to vrealize Log Insight in one of the following ways: Directly to vrealize Log Insight over the syslog protocol. By using vrealize Log Insight agents. Both of these are supported in this design for the different applications that provide the cloud management platform Time Synchronization Time synchronization is critical for the core functionality of vrealize Log Insight. vrealize Log Insight synchronizes time with the SoftLayer NTP service. Consistent NTP sources on all systems are configured that send log data (vcenter Server, ESXi, vrealize Operation Manager). See Time Synchronization under Common Services Connectivity in the Cluster This design requires that all vrealize Log Insight cluster nodes within a region are connected to the same LAN with no firewall or NAT between the nodes External Communication vrealize Log Insight receives log data over the syslog TCP, syslog TLS/SSL, or syslog UDP protocol. The default syslog UDP protocol is used in this design, because security is already designed at the level of the management network Event Forwarding between Regions vrealize Log Insight supports event forwarding to other clusters and standalone instances. While forwarding events, the vrealize Log Insight instance still ingests, stores and archives events locally Event Forwarding Protocol Forwarding of syslog data in vrealize Log Insight is achieved by using the Ingestion API or a native syslog implementation. The vrealize Log Insight Ingestion API uses TCP communication. In contrast to syslog, the forwarding module supports the following features for the Ingestion API: Forwarding to other vrealize Log Insight instances Both structured and unstructured data, that is, multi-line messages. Metadata in the form of tags Client-side compression Configurable disk-backed queue to save events until the server acknowledges the ingestion Disaster Recovery Each region is configured to forward log information to the vrealize Log Insight instance in the other region. Configuration of failover is not required. Page 120 of 132 Copyright IBM and VMware
121 5.6.5 Patching VMware vsphere Update Manager (vum) automates patch management and eliminates manual tracking and patching of vsphere hosts and virtual machines. It compares the state of vsphere hosts with baselines, then updates and patches to enforce compliance. Although vsphere Update Manager can be installed on the same server as a vcenter Server, this design utilizes a vum server installed on its own Windows Server virtual machine due to the size of the environment and the use of vcenter Server Appliance. Additionally, this design deploys two unique vum servers in each region that is associated and registered to a unique vcenter Server. This means that one vum server is registered to the vcenter Server managing the management clusters and another vum server registered to the vcenter Server managing the capacity and edge clusters. In addition to associating vsphere Update Manager with a unique vcenter Server, vum requires the use of a dedicated database. In this design, the vum associated with updating the capacity and edge clusters connect to an instance of the Microsoft SQL Server 2012 database running on the same virtual machine. The vum associated with the management cluster uses the Microsoft SQL Server 2012 database that is installed on the same machine as Update Manager vum for vcenter Managing the Management Cluster A vum instance is linked to a vcenter Server instance. The following section describes the configuration the vum instance for the vcenter that manages the Management Cluster Compute and Storage Design In this design, vum is installed on a virtual machine running Windows 2012 Server with an instance of Microsoft SQL Server 2012 installed. The virtual machine resides on the management cluster and utilize a VSAN-backed datastore for storage. The resources for this virtual machine are as follows in Table 67 Compute Resources for vum vcenter Managing the Management Cluster. Note that disk space was calculated using the vsphere Update Manager Sizing Estimator and includes a 1.5 GB monthly utilization rate. Table 67 Compute Resources for vum vcenter Managing the Management Cluster vcpu Memory Disk Space Disk Type 2 4 GB 60 GB Thin Network Design In this design, vum is placed on the management network which will provide appropriate access for it to upgrade and patch ESXi hosts, install and update third-party software on hosts, and upgrade virtual machine hardware, VMware tools, and virtual appliances vum for vcenter Managing Compute and Edge Clusters A vum instance is linked to a vcenter Server instance. The following section describes the configuration the vum instance for the vcenter that manages the Edge and Compute Clusters Compute and Storage Design In this design, vum is installed on a virtual machine running Windows 2012 Server and is connected to instance of Microsoft SQL Server 2012 installed on the same VM. The virtual machine resides on the management cluster and utilize a VSAN-backed datastore for storage. The resources for this virtual machine are as follows in Table 68 Compute Resources for vum vcenter Managing the Compute and Edge Clusters. Note that disk space was calculated using the vsphere Update Manager Sizing Estimator and includes a 1.5 GB monthly utilization rate. Table 68 Compute Resources for vum vcenter Managing the Compute and Edge Clusters vcpu Memory Disk Space Disk Type Copyright IBM and VMware Page 121 of 132
122 2 4 GB 60 GB Thin Network Design In this design, vum is placed on the management network which provides appropriate access for it to upgrade and patch ESXi hosts, install and update third-party software on hosts, and upgrade virtual machine hardware, VMware tools, and virtual appliances. 5.7 Business Services This design uses vrealize Business Standard Edition to provide metering and chargeback functionality for the cloud. vrealize Business provides director of cloud operations capabilities that allow them to gain greater visibility into financial aspects of their IaaS services delivery and enable them to optimize and improve these operations vrealize Business Standard Design The following figure presents the design of vrealize Business Standard: Figure 41 vrealize Business Logical Design Data Collector This component is responsible for connecting to vcenter Server instances and retrieving both inventory information (servers, virtual machines, clusters, storage devices, and associations between them) and usage (CPU and memory) statistics. Reference Database This component is responsible for providing default, out-of-the-box costs for each of the supported cost drivers. References are updated periodically. For this solution, the reference database is downloaded and installed manually. The new values affect cost calculation. Reference data that is used depends on currency selected during installation. USD is used by default. Communication between Data Collector and the Server Data collector and the server communicate through a database. Data collector writes to the database and the server reads the data. Data collector keeps inventory information time stamped, so it is possible to retrieve and view the inventory back in time. The architecture of data collector tables is flexible, and stores properties retrieved from the vcenter Server in key-value pairs. Page 122 of 132 Copyright IBM and VMware
123 vrealize Business Scalability vrealize Business Standard Edition scales up to 20,000 virtual machines across four VMware vcenter Server instances. This design implements one vrealize Business appliance for each deployed vcenter. vrealize Business Standard is deployed as a single virtual appliance hosted in the management network of vra. The appliance is configured with 2 vcpu, 4 GB of RAM and 50 GB disk vrealize Business Standard Integration vrealize Business Standard Edition is integrated with vcenter Server and extracts the inventory list from it. The inventory list contains information related to virtual machines configuration information, ESXi hosts and cluster capacity, storage profiles and capacity, vcenter Server attributes and tags. vrealize Business Standard Edition has tight integration with vrealize Automation. It uses the common services of vrealize Automation framework like SSO authentication and authorization. The Infrastructure as a Service (IaaS) component of vrealize Automation consumes the base rate APIs of vrealize Business Standard Edition to compute blueprint price of virtual machine. Copyright IBM and VMware Page 123 of 132
124 Appendix A Bare Metal Summary The bare metal servers used in this design consist of the following components found in the tables for management, compute, and edge clusters below. Management Cluster Nodes The following table shows the bill of materials for the SoftLayer bare metal servers in the management cluster. Table 69 Management - Bare Metal Bill of Materials Component Chassis Motherboard Processor Memory Network Interface Card Manufacturer / Model SuperMicro PIO-628U-TR4T+-ST031 SuperMicro X10DRU-i+_R1.02b 2 x 2.6GHz Intel Xeon-Haswell (E V3-DodecaCore) 16 x 16GB 4 x Intel Ethernet Controller 10 Gigabit X640-AT2 Disk Controller Avago Technologies MegaRAID SAS i Disks 2 x 1 TB Seagate Constellation ES 2 x 1.2 TB Intel S x 2 TB Western Digital Compute Cluster Nodes The following table shows the bill of materials for the SoftLayer bare metal servers in the compute cluster. Table 70 Compute - Bare Metal Bill of Materials Component Manufacturer / Model Chassis SuperMicro PIO-628U-TR4T+-ST031 Motherboard SuperMicro X10DRU-i+_R1.02b Processor 2 x 2.6GHz Intel Xeon-Haswell (E V3-DodecaCore) Memory 16 x 32GB Network Interface Card 4 x Intel Ethernet Controller 10 Gigabit X640-AT2 Disk Controller Avago Technologies MegaRAID SAS i Disks 2 x 1 TB Seagate Constellation ES 2 x 1.2 TB Intel S x 2 TB Western Digital Page 124 of 132 Copyright IBM and VMware
125 Edge Cluster Nodes The following table shows the bill of materials for the SoftLayer bare metal servers in the Edge cluster. Table 71 Edge - Bare Metal Bill of Materials Component Manufacturer / Model Chassis SuperMicro PIO-628U-TR4T+-ST031 Motherboard SuperMicro X10DRU-i+_R1.02b Processor 2 x 2.6GHz Intel Xeon-Haswell (E V3-DodecaCore) Memory 8 x 16GB Network Interface Card 4 x Intel Ethernet Controller 10 Gigabit X640-AT2 Disk Controller Avago Technologies MegaRAID SAS i Disks 2 x 1 TB Seagate Constellation ES 1 x 1.2 TB Intel S x 2 TB Western Digital Copyright IBM and VMware Page 125 of 132
126 Appendix B Software Bill of Materials The following software products and versions are used in this design for the cloud management platform. This does not include any software products that are deployed by users when utilizing the cloud management platform. Table 72 Software Bill of Materials Cloud Component Product Item Version Virtual Infrastructure ESXi 6.0 U1b vcenter Server Appliance (VIMISO) Virtual SAN 6.0 U1 6.0 U1 vsphere Replication 6.1 VMware vcenter Site Recovery Manager 6.1 NSX for vsphere Cloud Management vrealize Automation Appliance vrealize Automation Identity Appliance vrealize Orchestrator vrealize Orchestrator Plug-in for NSX Service Management vrealize Operations Manager Appliance Management Pack for NSX for vsphere 2.0 Management Pack for vrealize Log Insight 1.0 Manager Management Pack for vrealize Automation 1.0 Management Pack for Storage Devices 1.0 vrealize Log Insight 3.0 Business Continuity vsphere Data Protection 6.1 Business Management vrealize Business Standard Page 126 of 132 Copyright IBM and VMware
127 Cloud Component Product Item Version Patching vsphere Update Manager 6.0U1b Infrastructure Microsoft SQL Server 2012 R2 Microsoft Windows Server Ubuntu Server 2012 R LTS Software Orchestration Chef Server 12 Salt Stack Copyright IBM and VMware Page 127 of 132
128 Appendix C Management Virtual Machine Summary The following virtual machines are configured in the management cluster for the cloud management platform by default. Table 73 List of Management Cluster Virtual Machines and Sizes Function vcpu vram (GB) Analytics Edge # Analytics Edge # vdisk (GB) Analytics Master Analytics Replica Certificate Authority Server - Master CA Certificate Authority Server - Subordinate Chef Server and Software Binaries Collector Edge # Collector Edge # Data Node # Data Node # DEM Worker # DEM Worker # IaaS Web Server Appliance # IaaS Web Server Appliance # Identity Appliance Log Insight Edge # Log Insight Edge # Log Insight - Master Log Insight - Slave # Log Insight - Slave # Management North-South Edge # Management North-South Edge # Model Manager and DEM Orchestrator MS SQL Server Primary AD, Master DNS and Master NTP RDS Edge # RDS Edge # Remote Collector # Remote Collector # Page 128 of 132 Copyright IBM and VMware
129 Function vcpu vram (GB) Salt Stack Master Secondary AD, Master DNS and Master NTP Site Recovery Manager vdisk (GB) vcenter - Compute and Edge vcenter - Management VMware Data Protection Appliance vrealize Business Service vrealize Automation Appliance # vrealize Automation Appliance # vrealize Automation Proxy Agent # vrealize Automation Proxy Agent # vrealize Edge # vrealize Edge # vrealize Orchestrator # vrealize Orchestrator # vsphere Disk Replicator VMware Update Manager Edge # VMware Update Manager Edge # VMware Update Manager Management VMware Update Manager Compute & Edge NSX Manager - Management NSX Controller #1 Management NSX Controller #2 - Management NSX Controller #3 - Management NSX Manager Compute and Edge PSC Management PSC Compute and Edge TOTAL Copyright IBM and VMware Page 129 of 132
130 The following virtual machines are configured in the Edge cluster by default; Table 74 List of Default Edge Cluster Virtual Machines Function vcpu vram (GB) vdisk (GB) Edge North-South Compute # Edge North-South Compute # Edge East-West # Edge East-West # NSX Controller #1 Compute NSX Controller #2 Compute NSX Controller #3 Compute Grand Total Page 130 of 132 Copyright IBM and VMware
131 Appendix D Maximum Configurations The Advanced VMware SDDC on IBM Cloud solution supports the following; One Central Cloud One business entity (although multiple tenants are supported, the design is not intended for resale purposes) Up to four Cloud Regions A Central Cloud of up to 10,000 VMs or 1,000 compute nodes Each Cloud Region up to 10,000 VMs or 1,000 compute nodes Up to a total of 50,000 VM s can be supported by a single central cloud portal. This includes any virtual machines on a user s premise and management VMs. Up to 100 concurrent transactions Up to 2,500 catalog items on the self service portal across all users and groups Up to 48 nodes per compute cluster per central cloud or cloud region Up to 3 compute clusters per central cloud or cloud region Copyright IBM and VMware Page 131 of 132
132 Appendix E Compatibility Guide Browsers The following web browsers are supported by the design, Internet Explorer 10 Internet Explorer 11 Google Chrome Mozilla Firefox Guest Operating Systems The following operating systems are supported as provisioned virtual machines, Windows 7 Windows 8 Windows 8.1 Windows Server 2008 R2 Windows Server 2012 Windows Server 2012 R2 RHEL 5.9 RHEL 5.10 RHEL 6.1 Server RHEL 6.4 Server RHEL 6.5 Server RHEL 7.0 Server SLES 11 SP 2 SLES 11 SP 3 CentOS 5.10 CentOS 6.4 CentOS 6.5 CentOS 7.0 Ubuntu LTS Ubuntu ESX 4.1 Update 2 ESXi 4.1 Update 2 ESXi 5.1 and updates ESXi 5.5 and updates ESXi 6.0 Page 132 of 132 Copyright IBM and VMware
VMware vsphere-6.0 Administration Training
VMware vsphere-6.0 Administration Training Course Course Duration : 20 Days Class Duration : 3 hours per day (Including LAB Practical) Classroom Fee = 20,000 INR Online / Fast-Track Fee = 25,000 INR Fast
VMware NSX @SoftLayer!!
A VMware@SoftLayer CookBook v1.1 April 30, 2014 VMware NSX @SoftLayer Author(s) & Contributor(s) (IBM) Shane B. Mcelligott Dani Roisman (VMware) Merlin Glynn, [email protected] Chris Wall Geoff Wing Marcos
Expert Reference Series of White Papers. VMware vsphere Distributed Switches
Expert Reference Series of White Papers VMware vsphere Distributed Switches [email protected] www.globalknowledge.net VMware vsphere Distributed Switches Rebecca Fitzhugh, VCAP-DCA, VCAP-DCD, VCAP-CIA,
vsphere Private Cloud RAZR s Edge Virtualization and Private Cloud Administration
Course Details Level: 1 Course: V6PCRE Duration: 5 Days Language: English Delivery Methods Instructor Led Training Instructor Led Online Training Participants: Virtualization and Cloud Administrators,
VMware vsphere: Install, Configure, Manage [V5.0]
VMware vsphere: Install, Configure, Manage [V5.0] Gain hands-on experience using VMware ESXi 5.0 and vcenter Server 5.0. In this hands-on, VMware -authorized course based on ESXi 5.0 and vcenter Server
NSX Installation and Upgrade Guide
NSX 6.0 for vsphere This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of
Vmware VSphere 6.0 Private Cloud Administration
To register or for more information call our office (208) 898-9036 or email [email protected] Vmware VSphere 6.0 Private Cloud Administration Class Duration 5 Days Introduction This fast paced,
VMware vsphere 4.1 with ESXi and vcenter
VMware vsphere 4.1 with ESXi and vcenter This powerful 5-day class is an intense introduction to virtualization using VMware s vsphere 4.1 including VMware ESX 4.1 and vcenter. Assuming no prior virtualization
On-Demand Infrastructure with Secure Networks REFERENCE ARCHITECTURE
REFERENCE ARCHITECTURE Table of Contents Executive Summary.... 3 Audience.... 3 Overview.... 3 What Is an On-Demand Infrastructure?.... 4 Architecture Overview.... 5 Cluster Overview.... 8 Management Cluster...
Table of Contents. vsphere 4 Suite 24. Chapter Format and Conventions 10. Why You Need Virtualization 15 Types. Why vsphere. Onward, Through the Fog!
Table of Contents Introduction 1 About the VMware VCP Program 1 About the VCP Exam 2 Exam Topics 3 The Ideal VCP Candidate 7 How to Prepare for the Exam 9 How to Use This Book and CD 10 Chapter Format
VMware. NSX Network Virtualization Design Guide
VMware NSX Network Virtualization Design Guide Table of Contents Intended Audience... 3 Overview... 3 Components of the VMware Network Virtualization Solution... 4 Data Plane... 4 Control Plane... 5 Management
Nutanix Tech Note. VMware vsphere Networking on Nutanix
Nutanix Tech Note VMware vsphere Networking on Nutanix Nutanix Virtual Computing Platform is engineered from the ground up for virtualization and cloud environments. This Tech Note describes vsphere networking
Cross-vCenter NSX Installation Guide
NSX 6.2 for vsphere This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of
VMware vsphere 5.1 Advanced Administration
Course ID VMW200 VMware vsphere 5.1 Advanced Administration Course Description This powerful 5-day 10hr/day class is an intensive introduction to VMware vsphere 5.0 including VMware ESX 5.0 and vcenter.
VMware vsphere 5.0 Boot Camp
VMware vsphere 5.0 Boot Camp This powerful 5-day 10hr/day class is an intensive introduction to VMware vsphere 5.0 including VMware ESX 5.0 and vcenter. Assuming no prior virtualization experience, this
vcloud Suite Architecture Overview and Use Cases
vcloud Suite Architecture Overview and Use Cases vcloud Suite 5.8 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new
vsphere Networking vsphere 6.0 ESXi 6.0 vcenter Server 6.0 EN-001391-01
vsphere 6.0 ESXi 6.0 vcenter Server 6.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more
vsphere Networking ESXi 5.0 vcenter Server 5.0 EN-000599-01
ESXi 5.0 vcenter Server 5.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions
Expert Reference Series of White Papers. vcloud Director 5.1 Networking Concepts
Expert Reference Series of White Papers vcloud Director 5.1 Networking Concepts 1-800-COURSES www.globalknowledge.com vcloud Director 5.1 Networking Concepts Rebecca Fitzhugh, VMware Certified Instructor
VMWARE VSPHERE 5.0 WITH ESXI AND VCENTER
VMWARE VSPHERE 5.0 WITH ESXI AND VCENTER CORPORATE COLLEGE SEMINAR SERIES Date: April 15-19 Presented by: Lone Star Corporate College Format: Location: Classroom instruction 8 a.m.-5 p.m. (five-day session)
White Paper. Juniper Networks. Enabling Businesses to Deploy Virtualized Data Center Environments. Copyright 2013, Juniper Networks, Inc.
White Paper Juniper Networks Solutions for VMware NSX Enabling Businesses to Deploy Virtualized Data Center Environments Copyright 2013, Juniper Networks, Inc. 1 Table of Contents Executive Summary...3
VMware Virtual SAN Network Design Guide TECHNICAL WHITE PAPER
TECHNICAL WHITE PAPER Table of Contents Intended Audience.... 3 Overview.... 3 Virtual SAN Network... 3 Physical Network Infrastructure... 4 Data Center Network... 4 Host Network Adapter.... 5 Virtual
vsphere Networking vsphere 5.5 ESXi 5.5 vcenter Server 5.5 EN-001074-02
vsphere 5.5 ESXi 5.5 vcenter Server 5.5 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more
CompTIA Cloud+ 9318; 5 Days, Instructor-led
CompTIA Cloud+ 9318; 5 Days, Instructor-led Course Description The CompTIA Cloud+ certification validates the knowledge and best practices required of IT practitioners working in cloud computing environments,
Software Defined Environments
November 2015 Software Defined Environments 2015 Cloud Lecture, University of Stuttgart Jochen Breh, Director Architecture & Consulting Cognizant Global Technology Office Agenda Introduction New Requirements
Khóa học dành cho các kỹ sư hệ thống, quản trị hệ thống, kỹ sư vận hành cho các hệ thống ảo hóa ESXi, ESX và vcenter Server
1. Mục tiêu khóa học. Khóa học sẽ tập trung vào việc cài đặt, cấu hình và quản trị VMware vsphere 5.1. Khóa học xây dựng trên nền VMware ESXi 5.1 và VMware vcenter Server 5.1. 2. Đối tượng. Khóa học dành
VXLAN: Scaling Data Center Capacity. White Paper
VXLAN: Scaling Data Center Capacity White Paper Virtual Extensible LAN (VXLAN) Overview This document provides an overview of how VXLAN works. It also provides criteria to help determine when and where
CompTIA Cloud+ Course Content. Length: 5 Days. Who Should Attend:
CompTIA Cloud+ Length: 5 Days Who Should Attend: Project manager, cloud computing services Cloud engineer Manager, data center SAN Business analyst, cloud computing Summary: The CompTIA Cloud+ certification
Implementing and Troubleshooting the Cisco Cloud Infrastructure **Part of CCNP Cloud Certification Track**
Course: Duration: Price: $ 4,295.00 Learning Credits: 43 Certification: Implementing and Troubleshooting the Cisco Cloud Infrastructure Implementing and Troubleshooting the Cisco Cloud Infrastructure**Part
VMware vcloud Director for Service Providers
Architecture Overview TECHNICAL WHITE PAPER Table of Contents Scope of Document....3 About VMware vcloud Director....3 Platform for Infrastructure Cloud...3 Architecture Overview....3 Constructs of vcloud
How Network Virtualization can improve your Data Center Security
How Network Virtualization can improve your Data Center Security Gilles Chekroun SDDC, NSX Team EMEA [email protected] 2014 VMware Inc. All rights reserved. Security IT spending Security spending is
NSX TM for vsphere with Arista CloudVision
ARISTA DESIGN GUIDE NSX TM for vsphere with Arista CloudVision Version 1.0 August 2015 ARISTA DESIGN GUIDE NSX FOR VSPHERE WITH ARISTA CLOUDVISION Table of Contents 1 Executive Summary... 4 2 Extending
VMware vsphere: Fast Track [V5.0]
VMware vsphere: Fast Track [V5.0] Experience the ultimate in vsphere 5 skills-building and VCP exam-preparation training. In this intensive, extended-hours course, you will focus on installing, configuring,
Creating a VMware Software-Defined Data Center REFERENCE ARCHITECTURE VERSION 1.5
Software-Defined Data Center REFERENCE ARCHITECTURE VERSION 1.5 Table of Contents Executive Summary....4 Audience....4 Overview....4 VMware Software Components....6 Architectural Overview... 7 Cluster...
Remote PC Guide Series - Volume 1
Introduction and Planning for Remote PC Implementation with NETLAB+ Document Version: 2016-02-01 What is a remote PC and how does it work with NETLAB+? This educational guide will introduce the concepts
JOB ORIENTED VMWARE TRAINING INSTITUTE IN CHENNAI
JOB ORIENTED VMWARE TRAINING INSTITUTE IN CHENNAI Job oriented VMWARE training is offered by Peridot Systems in Chennai. Training in our institute gives you strong foundation on cloud computing by incrementing
VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014
VMware SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Table of Contents Introduction.... 3 vsphere Architectural Overview... 4 SAN Backup
HAWAII TECH TALK SDN. Paul Deakin Field Systems Engineer
HAWAII TECH TALK SDN Paul Deakin Field Systems Engineer SDN What Is It? SDN stand for Software Defined Networking SDN is a fancy term for: Using a controller to tell switches where to send packets SDN
VMware@SoftLayer Cookbook Disaster Recovery (DR)
VMware@SoftLayer Cookbook Disaster Recovery (DR) IBM Global Technology Services: Khoa Huynh ([email protected]) Daniel De Araujo ([email protected]) Bob Kellenberger ([email protected]) VMware: Merlin
(R)Evolution im Software Defined Datacenter Hyper-Converged Infrastructure
(R)Evolution im Software Defined Datacenter Hyper-Converged Infrastructure David Kernahan Senior Systems Engineer VMware Switzerland GmbH 2014 VMware Inc. All rights reserved. Agenda 1 VMware Strategy
VMUG - vcloud Air Deep Dive. 2014 VMware Inc. All rights reserved.
VMUG - vcloud Air Deep Dive 2014 VMware Inc. All rights reserved. Agenda 1 Overview of vcloud Air 2 Advanced Networking Capabilities 3 Use Cases 4 Overview of Disaster Recovery Service 5 Questions 2 VMware
Setup for Failover Clustering and Microsoft Cluster Service
Setup for Failover Clustering and Microsoft Cluster Service Update 1 ESX 4.0 ESXi 4.0 vcenter Server 4.0 This document supports the version of each product listed and supports all subsequent versions until
NSX Installation Guide
NSX 6.2 for vsphere This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of
Virtual SAN Design and Deployment Guide
Virtual SAN Design and Deployment Guide TECHNICAL MARKETING DOCUMENTATION VERSION 1.3 - November 2014 Copyright 2014 DataCore Software All Rights Reserved Table of Contents INTRODUCTION... 3 1.1 DataCore
Deployment Guide. How to prepare your environment for an OnApp Cloud deployment.
Deployment Guide How to prepare your environment for an OnApp Cloud deployment. Document version 1.07 Document release date 28 th November 2011 document revisions 1 Contents 1. Overview... 3 2. Network
VMware vrealize Automation
VMware vrealize Automation Reference Architecture Version 6.0 and Higher T E C H N I C A L W H I T E P A P E R Table of Contents Overview... 4 What s New... 4 Initial Deployment Recommendations... 4 General
VMware vcloud Networking and Security Overview
VMware vcloud Networking and Security Overview Networks and Security for Virtualized Compute Environments WHITE PAPER Overview Organizations worldwide have gained significant efficiency and flexibility
VMware EVO SDDC. General. Q. Is VMware selling and supporting hardware for EVO SDDC?
FREQUENTLY ASKED QUESTIONS VMware EVO SDDC General Q. What is VMware A. VMware EVO SDDC is the easiest way to build and run an SDDC private cloud on an integrated system. Based on an elastic, highly scalable,
Reference Architecture for Dell VIS Self-Service Creator and VMware vsphere 4
Reference Architecture for Dell VIS Self-Service Creator and VMware vsphere 4 Solutions for Large Environments Virtualization Solutions Engineering Ryan Weldon and Tom Harrington THIS WHITE PAPER IS FOR
Networking Topology For Your System
This chapter describes the different networking topologies supported for this product, including the advantages and disadvantages of each. Select the one that best meets your needs and your network deployment.
VMware vsphere Basics
ware vsphere Basics ESXi 5.0 vcenter Server 5.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check
RED HAT ENTERPRISE VIRTUALIZATION FOR SERVERS: COMPETITIVE FEATURES
RED HAT ENTERPRISE VIRTUALIZATION FOR SERVERS: COMPETITIVE FEATURES RED HAT ENTERPRISE VIRTUALIZATION FOR SERVERS Server virtualization offers tremendous benefits for enterprise IT organizations server
CA Virtual Assurance/ Systems Performance for IM r12 DACHSUG 2011
CA Virtual Assurance/ Systems Performance for IM r12 DACHSUG 2011 Happy Birthday Spectrum! On this day, exactly 20 years ago (4/15/1991) Spectrum was officially considered meant - 2 CA Virtual Assurance
Evaluation of Enterprise Data Protection using SEP Software
Test Validation Test Validation - SEP sesam Enterprise Backup Software Evaluation of Enterprise Data Protection using SEP Software Author:... Enabling you to make the best technology decisions Backup &
GRAVITYZONE HERE. Deployment Guide VLE Environment
GRAVITYZONE HERE Deployment Guide VLE Environment LEGAL NOTICE All rights reserved. No part of this document may be reproduced or transmitted in any form or by any means, electronic or mechanical, including
ESX Server 3 Configuration Guide Update 2 and later for ESX Server 3.5 and VirtualCenter 2.5
ESX Server 3 Configuration Guide Update 2 and later for ESX Server 3.5 and VirtualCenter 2.5 This document supports the version of each product listed and supports all subsequent versions until the document
E-SPIN's Virtualization Management, System Administration Technical Training with VMware vsphere Enterprise (7 Day)
Class Schedule E-SPIN's Virtualization Management, System Administration Technical Training with VMware vsphere Enterprise (7 Day) Date: Specific Pre-Agreed Upon Date Time: 9.00am - 5.00pm Venue: Pre-Agreed
VXLAN Bridging & Routing
VXLAN Bridging & Routing Darrin Machay [email protected] CHI-NOG 05 May 2015 1 VXLAN VM-1 10.10.10.1/24 Subnet A ESX host Subnet B ESX host VM-2 VM-3 VM-4 20.20.20.1/24 10.10.10.2/24 20.20.20.2/24 Load
Installation Guide Avi Networks Cloud Application Delivery Platform Integration with Cisco Application Policy Infrastructure
Installation Guide Avi Networks Cloud Application Delivery Platform Integration with Cisco Application Policy Infrastructure August 2015 Table of Contents 1 Introduction... 3 Purpose... 3 Products... 3
Brocade One Data Center Cloud-Optimized Networks
POSITION PAPER Brocade One Data Center Cloud-Optimized Networks Brocade s vision, captured in the Brocade One strategy, is a smooth transition to a world where information and applications reside anywhere
HP Matrix Operating Environment Federated CMS Overview
HP Matrix Operating Environment Federated CMS Overview HP Matrix OE infrastructure orchestration 7.0 Technical white paper Table of contents Introduction... 2 Overview... 2 Installation and configuration...
VMware Software Defined Network. Dejan Grubić VMware Systems Engineer for Adriatic
VMware Software Defined Network Dejan Grubić VMware Systems Engineer for Adriatic The Transformation of Infrastructure Infrastructure Servers Clouds Be more responsive to business, change economics of
vsphere Replication for Disaster Recovery to Cloud
vsphere Replication for Disaster Recovery to Cloud vsphere Replication 6.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced
VMware vrealize Automation
VMware vrealize Automation Reference Architecture Version 6.0 or Later T E C H N I C A L W H I T E P A P E R J U N E 2 0 1 5 V E R S I O N 1. 5 Table of Contents Overview... 4 What s New... 4 Initial Deployment
vcloud Air Disaster Recovery Technical Presentation
vcloud Air Disaster Recovery Technical Presentation Agenda 1 vcloud Air Disaster Recovery Overview 2 What s New 3 Architecture 4 Setup and Configuration 5 Considerations 6 Automation Options 2 vcloud Air
Cisco Nexus 1000V Virtual Ethernet Module Software Installation Guide, Release 4.0(4)SV1(1)
Cisco Nexus 1000V Virtual Ethernet Module Software Installation Guide, Release 4.0(4)SV1(1) September 17, 2010 Part Number: This document describes how to install software for the Cisco Nexus 1000V Virtual
VMware vsphere Design. 2nd Edition
Brochure More information from http://www.researchandmarkets.com/reports/2330623/ VMware vsphere Design. 2nd Edition Description: Achieve the performance, scalability, and ROI your business needs What
Cisco Prime Network Services Controller. Sonali Kalje Sr. Product Manager Cloud and Virtualization, Cisco Systems
Cisco Prime Network Services Controller Sonali Kalje Sr. Product Manager Cloud and Virtualization, Cisco Systems Agenda Cloud Networking Challenges Prime Network Services Controller L4-7 Services Solutions
VirtualclientTechnology 2011 July
WHAT S NEW IN VSPHERE VirtualclientTechnology 2011 July Agenda vsphere Platform Recap vsphere 5 Overview Infrastructure Services Compute, Storage, Network Applications Services Availability, Security,
vsphere Replication for Disaster Recovery to Cloud
vsphere Replication for Disaster Recovery to Cloud vsphere Replication 5.8 This document supports the version of each product listed and supports all subsequent versions until the document is replaced
VMware Virtual SAN 6.2 Network Design Guide
VMware Virtual SAN 6.2 Network Design Guide TECHNICAL WHITE PAPER APRIL 2016 Contents Intended Audience... 2 Overview... 2 Virtual SAN Network... 2 Physical network infrastructure... 3 Data center network...
Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study
White Paper Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study 2012 Cisco and/or its affiliates. All rights reserved. This
Next Gen Data Center. KwaiSeng Consulting Systems Engineer [email protected]
Next Gen Data Center KwaiSeng Consulting Systems Engineer [email protected] Taiwan Update Feb 08, kslai 2006 Cisco 2006 Systems, Cisco Inc. Systems, All rights Inc. reserved. All rights reserved. 1 Agenda
vcloud Director User's Guide
vcloud Director 5.5 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of
VMware and Brocade Network Virtualization Reference Whitepaper
VMware and Brocade Network Virtualization Reference Whitepaper Table of Contents EXECUTIVE SUMMARY VMWARE NSX WITH BROCADE VCS: SEAMLESS TRANSITION TO SDDC VMWARE'S NSX NETWORK VIRTUALIZATION PLATFORM
Virtualization, SDN and NFV
Virtualization, SDN and NFV HOW DO THEY FIT TOGETHER? Traditional networks lack the flexibility to keep pace with dynamic computing and storage needs of today s data centers. In order to implement changes,
vsphere Upgrade Update 1 ESXi 6.0 vcenter Server 6.0 EN-001804-02
Update 1 ESXi 6.0 vcenter Server 6.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent
Veritas Storage Foundation High Availability for Windows by Symantec
Veritas Storage Foundation High Availability for Windows by Symantec Simple-to-use solution for high availability and disaster recovery of businesscritical Windows applications Data Sheet: High Availability
Securely Architecting the Internal Cloud. Rob Randell, CISSP Senior Security and Compliance Specialist VMware, Inc.
Securely Architecting the Internal Cloud Rob Randell, CISSP Senior Security and Compliance Specialist VMware, Inc. Securely Building the Internal Cloud Virtualization is the Key How Virtualization Affects
RSA Authentication Manager 8.1 Setup and Configuration Guide. Revision 2
RSA Authentication Manager 8.1 Setup and Configuration Guide Revision 2 Contact Information Go to the RSA corporate website for regional Customer Support telephone and fax numbers: www.emc.com/domains/rsa/index.htm
VMware Network Virtualization Design Guide. January 2013
ware Network Virtualization Technical WHITE PAPER January 2013 ware Network Virtualization Table of Contents Intended Audience.... 3 Overview.... 3 Components of the ware Network Virtualization Solution....
VMware vsphere 5.0 Evaluation Guide
VMware vsphere 5.0 Evaluation Guide Advanced Networking Features TECHNICAL WHITE PAPER Table of Contents About This Guide.... 4 System Requirements... 4 Hardware Requirements.... 4 Servers.... 4 Storage....
DMZ Virtualization Using VMware vsphere 4 and the Cisco Nexus 1000V Virtual Switch
DMZ Virtualization Using VMware vsphere 4 and the Cisco Nexus 1000V Virtual Switch What You Will Learn A demilitarized zone (DMZ) is a separate network located in the neutral zone between a private (inside)
Monitoring Hybrid Cloud Applications in VMware vcloud Air
Monitoring Hybrid Cloud Applications in ware vcloud Air ware vcenter Hyperic and ware vcenter Operations Manager Installation and Administration Guide for Hybrid Cloud Monitoring TECHNICAL WHITE PAPER
Simplified Private Cloud Management
BUSINESS PARTNER ClouTor Simplified Private Cloud Management ClouTor ON VSPEX by LOCUZ INTRODUCTION ClouTor on VSPEX for Enterprises provides an integrated software solution for extending your existing
Best Practices for Monitoring Databases on VMware. Dean Richards Senior DBA, Confio Software
Best Practices for Monitoring Databases on VMware Dean Richards Senior DBA, Confio Software 1 Who Am I? 20+ Years in Oracle & SQL Server DBA and Developer Worked for Oracle Consulting Specialize in Performance
VMware vsphere Data Protection
VMware vsphere Data Protection Replication Target TECHNICAL WHITEPAPER 1 Table of Contents Executive Summary... 3 VDP Identities... 3 vsphere Data Protection Replication Target Identity (VDP-RT)... 3 Replication
Building the Virtual Information Infrastructure
Technology Concepts and Business Considerations Abstract A virtual information infrastructure allows organizations to make the most of their data center environment by sharing computing, network, and storage
VMware vcloud Air Networking Guide
vcloud Air This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of this document,
Nutanix Tech Note. Configuration Best Practices for Nutanix Storage with VMware vsphere
Nutanix Tech Note Configuration Best Practices for Nutanix Storage with VMware vsphere Nutanix Virtual Computing Platform is engineered from the ground up to provide enterprise-grade availability for critical
Reference Design: Deploying NSX for vsphere with Cisco UCS and Nexus 9000 Switch Infrastructure TECHNICAL WHITE PAPER
Reference Design: Deploying NSX for vsphere with Cisco UCS and Nexus 9000 Switch Infrastructure TECHNICAL WHITE PAPER Table of Contents 1 Executive Summary....3 2 Scope and Design Goals....3 2.1 NSX VMkernel
Restricted Document. Pulsant Technical Specification
Pulsant Technical Specification Title Pulsant Government Virtual Server IL2 Department Cloud Services Contributors RR Classification Restricted Version 1.0 Overview Pulsant offer two products based on
Course Syllabus. Fundamentals of Windows Server 2008 Network and Applications Infrastructure. Key Data. Audience. Prerequisites. At Course Completion
Key Data Product #: 3380 Course #: 6420A Number of Days: 5 Format: Certification Exams: Instructor-Led None This course syllabus should be used to determine whether the course is appropriate for the students,
CloudLink - The On-Ramp to the Cloud Security, Management and Performance Optimization for Multi-Tenant Private and Public Clouds
- The On-Ramp to the Cloud Security, Management and Performance Optimization for Multi-Tenant Private and Public Clouds February 2011 1 Introduction Today's business environment requires organizations
Virtual Appliance Setup Guide
The Barracuda SSL VPN Vx Virtual Appliance includes the same powerful technology and simple Web based user interface found on the Barracuda SSL VPN hardware appliance. It is designed for easy deployment
Configuration Maximums VMware vsphere 4.1
Topic Configuration s VMware vsphere 4.1 When you select and configure your virtual and physical equipment, you must stay at or below the maximums supported by vsphere 4.1. The limits presented in the
VMDC 3.0 Design Overview
CHAPTER 2 The Virtual Multiservice Data Center architecture is based on foundation principles of design in modularity, high availability, differentiated service support, secure multi-tenancy, and automated
VM-Series Firewall Deployment Tech Note PAN-OS 5.0
VM-Series Firewall Deployment Tech Note PAN-OS 5.0 Revision A 2012, Palo Alto Networks, Inc. www.paloaltonetworks.com Contents Overview... 3 Supported Topologies... 3 Prerequisites... 4 Licensing... 5
BUILDING A NEXT-GENERATION DATA CENTER
BUILDING A NEXT-GENERATION DATA CENTER Data center networking has changed significantly during the last few years with the introduction of 10 Gigabit Ethernet (10GE), unified fabrics, highspeed non-blocking
vcloud Air - Virtual Private Cloud OnDemand Networking Guide
vcloud Air - Virtual Private Cloud OnDemand Networking Guide vcloud Air This document supports the version of each product listed and supports all subsequent versions until the document is replaced by
