VMware SDDC on IBM Cloud - Advanced Detailed Design

Size: px
Start display at page:

Download "VMware SDDC on IBM Cloud - Advanced Detailed Design"

Transcription

1 VMware SDDC on IBM Cloud - Advanced Detailed Design Date: 16 th March 2016 Version: 1.1 Copyright IBM and VMware Page 1 of 132

2 Table of Contents 1 Introduction Pre-requisites Summary of Changes System Context Actors Systems Architecture Overview Physical Infrastructure Virtual Infrastructure Infrastructure Management Common Services Cloud Management Services Operational Services Business Services Logical Operational Model Logical Operational Model Structure Central Cloud Physical Infrastructure Cluster Architecture Physical Network Physical Storage Virtual Infrastructure Compute Virtualization Storage Virtualization Network Virtualization Infrastructure Management Compute Management Storage Management Network Management Common Services Identity and Access Services Domain Name Services NTP Services SMTP Services Certificate Authority Services Cloud Management Services Page 2 of 132 Copyright IBM and VMware

3 4.7.1 Service Catalog Self-Service Portal Infrastructure and Process Orchestration Software Orchestration Operational Services Backup and Restore Disaster Recovery Monitoring Log Consolidation and Analysis Patching Business Services Business Management IT Financials IT Benchmarking Cloud Region Physical Operational Model Physical Layer Compute Storage Network Virtual Infrastructure Compute Virtualization Storage Virtualization Network Virtualization Infrastructure Management vcenter Server Instances Common Services Identity and Access Services Domain Name Services DNS Configuration Requirements NTP Services SMTP Services Certificate Authority Services Cloud Management Services Cloud Management Physical Design vrealize Automation Supporting Infrastructure vrealize Automation Cloud Tenant Design vrealize Automation vsphere Integration Design Infrastructure Source Endpoints Copyright IBM and VMware Page 3 of 132

4 5.5.6 Virtualization Compute Resources Process Orchestration Software Orchestration Infrastructure Orchestration Operational Services Backup and Restore Disaster Recovery Monitoring Log Consolidation and Analysis Patching Business Services Appendix A Bare Metal Summary Management Cluster Nodes Compute Cluster Nodes Edge Cluster Nodes Appendix B Software Bill of Materials Appendix C Management Virtual Machine Summary Appendix D Maximum Configurations Appendix E Compatibility Guide Browsers Guest Operating Systems Table of Figures Figure 1 VMware SDDC on IBM Cloud Introduction 8 Figure 2 VMware SDDC on IBM Cloud System Context 10 Figure 3 VMware SDDC on IBM Cloud Architecture Overview 12 Figure 4 Logical Structure View 15 Figure 5 Component Interaction Diagram 16 Figure 6 Logical Operational Model 17 Figure 7 Logical Cluster Structure 18 Figure 8 Network Virtualization 22 Figure 9 Dual-Region Data Protection Architecture 30 Figure 10 Disaster Recovery Architecture 31 Figure 11 vrealize Operations Manager Architecture 32 Figure 12 Physical Operational Model - Virtual Servers, Networking and Clusters 36 Page 4 of 132 Copyright IBM and VMware

5 Figure 13 Network connections per physical node 39 Figure 14 VSAN concept 40 Figure 15 Network Switch Design for Management Hosts 44 Figure 16 Network Switch Design for Edge Hosts 47 Figure 17 Network Switch Design for Compute Hosts 49 Figure 18 Network Virtualization Conceptual Design 54 Figure 19 Cluster Design for NSX for vsphere 55 Figure 20 Virtual Application Network Components and Design 59 Figure 21 vra Virtual Network Design 60 Figure 22 Virtual Application Network Configuration in Central Cloud and Cloud Region 62 Figure 23 vcenter Server and PSC Deployment Model 64 Figure 24 vrealize Automation Conceptual Design 72 Figure 25 vrealize Automation Design Overview for Central Cloud 75 Figure 26 vrealize Automation Design Overview for Additional Cloud Regions 76 Figure 27 Tenant Design for Single Region 81 Figure 28 Tenant Design for Two Regions 82 Figure 29 vrealize Automation Integration with vsphere Endpoint Central Cloud 87 Figure 30 vrealize Automation Integration with vsphere Endpoint Central Cloud and a Cloud Region (Region A) 88 Figure 31 Template Synchronization 90 Figure 32 Software Orchestration Logical Design 95 Figure 33 vsphere Data Protection Logical Design 97 Figure 34 Logical Network Design for Cross-Region Deployment with Management Application Network Container 102 Figure 35 Logical Design of vrealize Operations Manager Central Cloud and a Cloud Region (Region A) Deployment 107 Figure 36 Networking Design of the vrealize Operations Manager Deployment 110 Figure 37 Application Virtual Networks in the vrealize Operations Manager Topology 111 Figure 38 Logical Design of vrealize Log Insight 114 Figure 39 Networking Design for the vrealize Log Insight Deployment 116 Figure 40 Application Virtual Networks in the vrealize Log Insight Topology 117 Figure 41 vrealize Business Logical Design 122 List of Tables Table 1 VMware SDDC on IBM Cloud Interfaced Actors 10 Table 2 VMware SDDC on IBM Cloud Interfaced Systems 11 Table 3 vrealize Operations Manager Logical Node Architecture 33 Table 4 NFS Configuration for vsphere Data Protection and vrealize Log Insight 38 Copyright IBM and VMware Page 5 of 132

6 Table 5 VSAN disk table 40 Table 6 VSAN policies 41 Table 7 VSAN object policy defaults 42 Table 8 VLAN Mapping to Traffic Types 42 Table 9 Management Cluster Distributed Switch 43 Table 10 Management Cluster Distributed Switch Port Group Configuration Settings 43 Table 11 Management Virtual Switch Port Groups and VLANs 45 Table 12 Management VMkernel Adapter 45 Table 13 Edge Cluster Distributed Switch 46 Table 14 Management Cluster Distributed Switch Port Group Configuration Settings 46 Table 15 Edge Virtual Switch Port Groups and VLANs 48 Table 16 Edge VMkernel Adapter 48 Table 17 Compute Cluster Distributed Switch 48 Table 18 Compute Cluster Distributed Switch Port Group Configuration Settings 49 Table 19 Compute Virtual Switch Port Groups and VLANs 50 Table 20 Compute VMkernel Adapter 50 Table 21 NSX Components Sizing 53 Table 22 Load Balancer Features 57 Table 23 Management Applications IP Addressing 61 Table 24 OSPF Area ID 62 Table 25 Specifications for Management vcenter Server Appliance 65 Table 26 Specifications for Platform Service Controller for Management Cluster 65 Table 27 Specifications for Compute and Edge vcenter Server Appliance 65 Table 28 Specifications for Platform Service Controller for Management Cluster 65 Table 29 Requirements for Active Directory Service 67 Table 30 Authentication types used 68 Table 31 Server Sizing 68 Table 32 Domain Naming Example 69 Table 33 SoftLayer DNS servers 70 Table 34 Time sources 70 Table 35 Root CA and Subordinate CA sizing 72 Table 36 Cloud Management Services Components 73 Table 37 Load Balancer Application Profile 78 Table 38 Load Balancer Service Monitoring Configuration 79 Table 39 Load Balancer Pool Specifications 79 Table 40 Virtual Server Characteristics 79 Table 41 Base Windows Server Blueprint 83 Page 6 of 132 Copyright IBM and VMware

7 Table 42 Base Windows Blueprint Sizing 84 Table 43 Base Linux Server Blueprint 84 Table 44 Base Linux Blueprint Sizing 84 Table 45 vrealize Integration with vsphere 85 Table 46 vrealize Orchestrator Default Configuration Ports 91 Table 47 vrealize Orchestrator Default External Communication Ports 91 Table 48 vro Service Monitor Specifications 93 Table 49 vro Service Pool Characteristics 93 Table 50 vro Virtual Server Characteristics 93 Table 51 Software Orchestration Components Sizing 96 Table 52 VMware vsphere Data Protection Performance 97 Table 53 Backup Jobs in Central Cloud 99 Table 54 Backup Jobs in Additional Cloud Region 101 Table 55 SRM Windows server sizing 105 Table 56 vsphere Replication Appliance 106 Table 57 Analytics Cluster Node Configurations 107 Table 58 DRS Cluster Anti-Affinity Rule for vrealize Operations Manager Nodes 108 Table 59 Remote Collector Node Sizes 108 Table 60 DRS Cluster Anti-Affinity Rule for vrealize Operations Remote Collector Nodes 109 Table 61 IP Subnets in the Application Virtual Network of vrealize Operations Manager 111 Table 62 DNS Names for the Application Virtual Networks 112 Table 63 Node Sizing 115 Table 64 IP Subnets in the Application Isolated Networks 117 Table 65 Example DNS names of Log Insight nodes 117 Table 66 Virtual Disk Configuration in the vrealize Log Insight Virtual Appliance 118 Table 67 Compute Resources for vum vcenter Managing the Management Cluster 121 Table 68 Compute Resources for vum vcenter Managing the Compute and Edge Clusters 121 Table 69 Management - Bare Metal Bill of Materials 124 Table 70 Compute - Bare Metal Bill of Materials 124 Table 71 Edge - Bare Metal Bill of Materials 125 Table 72 Software Bill of Materials 126 Table 73 List of Management Cluster Virtual Machines and Sizes 128 Table 74 List of Default Edge Cluster Virtual Machines 130 Copyright IBM and VMware Page 7 of 132

8 1 Introduction VMware Software Defined Data Center (SDDC) on IBM Cloud allows existing VMware virtualized datacenter clients to extend into the IBM Cloud. This permits uses like capacity expansion into the cloud (and contraction when not needed), migration to the cloud, disaster recovery to the cloud, backup into the cloud and the ability to stand up a dedicated cloud environment for development, test, training or lab. This document details the design of the Advanced version of the VMware SDDC on IBM Cloud which targets designs requiring high levels of scalability and multiple regions. Figure 1 VMware SDDC on IBM Cloud Introduction 1.1 Pre-requisites The design requires the following pre-requisites: Client is required to acquire all necessary software licenses and/or keys for all products used in this design prior to commencement of implementation Client is required to provide SoftLayer account Client is responsible for SoftLayer related charges as a result of this design s implementation Client is responsible for connectivity from this design to any on premises environment or systems Client is responsible for connectivity into this design for access by administrators and end users Client is responsible to acquire and provide domain name Client is responsible to provide hostname prefixes for the SoftLayer bare metal devices provisioned through this design Client to provide connection details and necessary credentials for any systems external to this design that are to be integrated (refer to system context for options) Client is responsible for licensing of any software products provisioned with the design Page 8 of 132 Copyright IBM and VMware

9 1.2 Summary of Changes This section records the history of significant changes to this document. Only the most significant changes are described here. Version Date Author Description of Change th Feb th March 2016 Simon Kofkin-Hansen Richard Ehrhardt Razvan Ionescu Daniel de Araujo Frank Chodacki Bob Kellenberger Bryan Buckland Christopher Moss Daniel Arrieta Alvarez Initial Release of Document Minor reported spelling and grammar corrections. Copyright IBM and VMware Page 9 of 132

10 2 System Context When depicting the VMware SDDC on IBM Cloud design as a single object, the following are the external actors and systems that interface with the design. NTP Cloud Admin Manages Cloud Service Provider Manages Services End Users Consumes Cloud Services Time Synchronisation for Cloud Services External Domain for Cloud Services Client DNS Recipes Patches VMware on IBM Cloud Patch Repo Checks for Updates SMTP Relay Sends s Client Auth. Trust Relationship for Cloud Services Provides Bare Metal Compute, Storage, and Network Builds Environment Connects To Client On Premise vsphere SoftLayer 2.1 Actors Figure 2 VMware SDDC on IBM Cloud System Context The actors that interface with the design are described in the following table. There is not a direct correlation between actors and persons. An actor role may be performed by one or more persons. Alternately, one person may perform more than one actor role. Table 1 VMware SDDC on IBM Cloud Interfaced Actors Actor Cloud Admin Service Provider Description The cloud admin or administrator is responsible for maintaining the cloud services. This includes, Assigning virtual resources to groups Maintaining Cloud Software Platform System Administrator roles Manages the cloud services that are provided to the client users. This includes, Service Catalog configuration Defining roles Defining groups Configuring user access Tenant Administrator roles Page 10 of 132 Copyright IBM and VMware

11 User Consumes the services that the cloud admin allows access to. This typically includes, Provisioning VMs De-provisioning VMs Provisioning Patterns De-provisioning Patterns Start / Stop / restart VMs and Patterns 2.2 Systems The systems that interface with the design are described in the following table. Table 2 VMware SDDC on IBM Cloud Interfaced Systems System SoftLayer Client On Premises vsphere Client SMTP Relay Client Authentication Client DNS NTP Service Patch Repo Description SoftLayer provides the bare metal, physical networking and NFS storage in addition to the automation to build the design when ordered. The design is able to connect to an existing vsphere environment on a client premises to enable Hybrid capabilities. The design connects its SMTP server to a client s SMTP relay service to provide notifications on aspects such as the process orchestration. The design is able to connect to an existing client authentication system to establish a trust relationship which extends the client s authentication system into the cloud for use by the cloud management platform. The design is able to connect to a client s domain name service (DNS) to extend the domain service into the cloud for use by the cloud management platform. The design requires an external NTP service to provide time synchronization services for use by the cloud management platform. There are a number of internet based patch repositories that the cloud management platform applications need to connect to in order to maintain the security and stability of the cloud environment. Copyright IBM and VMware Page 11 of 132

12 3 Architecture Overview VMware SDDC on IBM Cloud provides VMware automation technology on SoftLayer. This includes virtual networking, virtual storage, process orchestration, infrastructure orchestration and software orchestration. It also provides the tools for management of the services providing these functions. The architecture consists of at least one central cloud region built on SoftLayer, which provides the main portal for users and administration, plus it can include one or more cloud regions which are managed by the central cloud and provide additional functionality for remote locations. The architecture is scaled out within a region, or by adding regions. Figure 3 VMware SDDC on IBM Cloud Architecture Overview 3.1 Physical Infrastructure The physical infrastructure consists of three main components, physical compute, physical network and physical storage. The physical compute provides the physical processing and memory that is used by the virtualization infrastructure. The physical network provides the network connectivity into the environment that is then consumed by the network virtualization. The physical storage provides the raw storage capacity Page 12 of 132 Copyright IBM and VMware

13 consumed by the virtualization infrastructure. For this design the physical infrastructure components are provided by SoftLayer bare metal and all components are supported on the VMware Hardware Compatibility Guide (HCG). 3.2 Virtual Infrastructure The physical infrastructure is consumed by the virtual infrastructure. The virtual infrastructure reflects the physical with three different components, compute virtualization, storage virtualization and network virtualization. Each of these interface with the respective component in the physical infrastructure. The virtual infrastructure is installed on each physical device to form a node, for example a compute node. All the virtual resources interface to the virtual infrastructure for access to the physical infrastructure. The virtual infrastructure is accessed by either the cloud admin or the infrastructure management component Infrastructure Management Infrastructure management provides the logic to ensure the maximum benefit derived for the virtual infrastructure. This includes functions such as pooling virtual infrastructure and moving virtual resources off a node for maintenance or in the case of node failure. It controls placement of virtual resources on nodes to balance load along with placement based on business rules. It is only accessed by the cloud admin. All other access to this component is through API access from other components. 3.4 Common Services Common services provides the services which are consumed by the other cloud management services. It includes, identity and access services, SMTP services, NTP services, domain name services and certificate authority. This component is also the primary interface to external systems. Common services can connect to the client s DNS for requests outside the domain managed by the cloud services. It connects to the external NTP service to synchronize it s NTP service with an outside stratum. A trust relationship can be established between the common services and the client s authentication service for common authentication to the cloud services. 3.5 Cloud Management Services The cloud management services provide the primary interface to the users to consume cloud services in addition to the orchestration engines to process the service requests. The self-service portal is used as the primary interface to view the available cloud services (the service catalog) as well as to obtain a view of existing cloud resources that are deployed. The service catalog is the list of available services that are managed by the service provider. The service provider is able to determine which services are available to specific users or groups. The process orchestration engine controls the steps required to perform a service. This includes actions such as, obtaining an approval or connecting to an operational service system as part of the process. The process orchestration engine calls the infrastructure orchestration engine to orchestrate the build of the virtual resources for a service. The software orchestration engine builds the software that runs on the virtual resources. 3.6 Operational Services Operational services provide monitoring, patching, log consolidation, log analysis, disaster recovery and backup services for the cloud management platform. The monitoring looks for issues with the cloud management platform and notifies cloud admin via alerts on the operations console as well as s via the external SMTP relay. The patching connects to the external patch repository in order to obtain update information in support of the security or stability of the cloud management platform. Log consolidation collects the logs from the cloud management platform into a central repository which the log analysis service then operates on to provide the cloud admin with diagnostic information. The backup service keeps copies of the cloud management platform outside of the virtual infrastructure so it can be restored in the event of failure or corruption. The disaster recovery service needs at least one cloud region separate to the Copyright IBM and VMware Page 13 of 132

14 central cloud to which the cloud management platform are replicated to. In the event of failure at the primary site, the cloud management platform is restarted at the cloud region. 3.7 Business Services The business services component provides the service provider with analytics on IT financials, business management and benchmarking aspects of the cloud. The IT financials provides the service provider with details of the total cost of cloud ownership. The business management functions provide metering and chargeback capabilities for the service provider by user or group. The benchmarking functions provide the ability for service providers to analyze where the IT spend for the cloud is going, where it could be in the future and paths that need to be taken to improve. Page 14 of 132 Copyright IBM and VMware

15 4 Logical Operational Model The logical operational model provides guidance as to the design elements required to meet the functional requirements. 4.1 Logical Operational Model Structure The design consists of two distinct elements. A central cloud, through which the user and service provider manage the entire cloud, and optionally, one or more associated cloud regions. Only the central cloud contains the self-service portal. Additional regions (cloud regions) are added to provide remote sites, or additional capacity beyond that of a single central cloud within the same site. Each cloud region is configured into the central cloud for management. On premises vsphere environments are connected to SoftLayer via either a VPN connection over the internet or dedicated links to form additional cloud regions. The design of on premises vsphere environments is outside the scope of this document. Figure 4 Logical Structure View Copyright IBM and VMware Page 15 of 132

16 Within a central cloud, the components interact with each other as follows: Figure 5 Component Interaction Diagram Both the central cloud and any additional cloud regions are built on SoftLayer. Page 16 of 132 Copyright IBM and VMware

17 Microsoft Active Directory + services vum VMware vrealize Suite vrli vrops SRM vdp 4.2 Central Cloud The central cloud hosts the primary portal through which users access the cloud services. It has connections to all remote regions. The functions in a central cloud map to the following software products described in more detail in this section. vrealize Business Business Management Business Services IT Financials IT Benchmarking Common Services Identity & Access Services Cloud Management Services Self Service Portal Operational Services Backup & Restore Domain Name Services Service Catalog Disaster Recovery NTP Services Process Orchestration Monitoring SMTP Services Software Orchestration Log Consolidation & Analysis Certificate Authority Services Infrastructure Orchestration Patching Infrastructure Management vcenter Compute Management Storage Management Network Management Virtual Infrastructure ESXi VSAN NSX Compute Virtualisation Storage Virtualisation Network Virtualisation SoftLayer Physical Compute Physical Infrastructure Physical Storage Physical Network Figure 6 Logical Operational Model 4.3 Physical Infrastructure The physical Infrastructure is broken up into compute, storage and network. The compute and storage areas are combined in the cluster architecture. The network area is described in the network transport section Cluster Architecture This design splits the physical layer into clusters. A cluster represents the aggregate of the compute and memory resources of all the hosts in the cluster. All hosts in the cluster share network and storage resources. The use of clusters allows workloads (user or management) to be placed onto specific hardware. Copyright IBM and VMware Page 17 of 132

18 Each cluster is managed as a single entity, so user workloads can be managed separately from management workloads. The design differentiates between the following types of clusters: Compute cluster (one or more) Management cluster Edge cluster Storage cluster Figure 7 Logical Cluster Structure Compute Clusters Compute clusters host the VMware SDDC on IBM Cloud users virtual machines (sometimes referred to as workloads or payloads). Each compute cluster is built using SoftLayer bare metal. The environment is scaled by adding nodes to the initial compute cluster up the maximum number of nodes per cluster (refer to the physical operational model for details). Once the maximum has been reached, additional compute clusters are added to the environment Management Cluster The management cluster houses the virtual machines that manage the cloud. Like the compute clusters, the management cluster is built using SoftLayer bare metal. These servers host vcenter Server, NSX Manager, NSX Controller, vrealize Operations Management, vrealize Log Insight, vrealize Automation, and other shared management components Edge Cluster Edge clusters connect the virtual networks (overlay networks) provided by NSX for vsphere and the external networks. This includes both north-south (into the environment from outside) and east-west (between management and compute clusters) communications. Edge clusters provide the following main functions: Page 18 of 132 Copyright IBM and VMware

19 Support on-ramp and off-ramp connectivity to physical networks Connect to client on premises environments Storage Cluster A storage cluster provides network-accessible storage via NFS. This is used for backup and log archive purposes. The compute, management and edge clusters utilize Virtual Storage Area Network (VSAN) which aggregates disks located in each node of the clusters Physical Network The physical and layer 2 networking is handled by SoftLayer. The SoftLayer physical fabric provides a robust IP transport layer with the following characteristics: Simplicity Scalability High bandwidth Fault-tolerant transport Simplicity The network infrastructure at SoftLayer is simplified and standardized on three physical networks containing public, private, and out of band management (IPMI) traffic. Both the private and public networks are deployed to utilize up to 20Gbps bandwidth per physical host. The out of band management network is connected with a 1Gbps link per host. Upon ordering infrastructure components within SoftLayer, VLANs are provisioned for each of the three networks mentioned. If existing VLANs exist within an environment and there is enough space for a bare metal device to be placed in the same pod, SoftLayer will automatically assign the new device an IP on the same VLAN. The design incorporates SoftLayer portable subnets to provide IP addressing for virtual machines as well as IP addresses for the bare metal hosts. SoftLayer has also standardized the networking infrastructure using best-of-breed networking vendors. As a result, SoftLayer is able to implement and reuse automation patterns to setup, configure, and monitor the network infrastructure. Some of this automation has been exposed via API and is used by this design to simplify management tasks Scalability The SoftLayer network is designed in a multi-tier model. Each rack in a SoftLayer datacenter contains 2 frontend customer switches (FCS) and 2 backend customer switches (BCS) connected to the public and private networks, respectively. These customer switches then connect to separate, peered aggregation switches; the aggregated switches are then attached to a pair of separate routers for L3 networking. This multi-tier design allows the network to scale across racks, rows, and pods within the SoftLayer datacenter High Bandwidth Every upstream network port in the SoftLayer datacenter has multiple 10Gbps or 40Gbps connections. Every rack is terminated with multiple 10Gbps or 40Gbps connections to the public Internet and multiple 10Gbps or 40Gbps connections to the private network Fault-tolerant transport Redundancy is provided at the server level using 2x10Gbps NICs. Additionally, the backend server, frontend server, aggregation switches and routers are redundantly connected Physical Storage There are two types of storage used within this design: VSAN and NFS. Copyright IBM and VMware Page 19 of 132

20 VSAN VMware VSAN is a distributed storage technology which enables the local drives on each vsphere host to be aggregated and pooled into a shared-nothing storage device accessible to all hosts in a VSAN cluster. Additionally, the VSAN service is contained within the physical hosts, eliminating the need for external shared storage NFS Network File System (NFS) is a protocol for file based storage which allows a user on a client computer to access files over a network much like local storage is accessed. In this case, the client computer is an ESXi host which utilizes NFS as a backend datastore, and the storage is provided by a NFS-capable external storage array residing within the same SoftLayer site as the central cloud or cloud region. 4.4 Virtual Infrastructure The Virtual Infrastructure layer includes the software to provide compute, network and storage virtualization. VMware products in this layer are vsphere ESXi, VMware VSAN and VMware NSX. The designs virtual infrastructure consists of one Central Cloud and may contain several Cloud Regions that are managed by the Central Cloud. Each central cloud or cloud region has the following components: Management cluster. Houses compute resources for the management and operation of the cloud. Edge cluster. Provides external connections for the other clusters and user virtual network resources via VMware NSX. Compute (workload) cluster. Houses user virtual compute and storage Compute Virtualization Compute virtualization provides the virtualization of CPU and memory for consumption by virtual machines. This is provided by VMware vsphere ESXi and vcenter Server software Provisioning By virtualizing the hardware, new servers are able to be provisioned much faster than traditional methods. In addition, adding or removing elements like CPU or memory allocated to a virtual machine can all be done in software Resource scheduling The virtual compute is disengaged from the physical hardware which allows for precise control of the hardware in which the virtual machine is housed. Distributed Resource Scheduling (DRS) is used to distribute and balance the virtual machines across the cluster continually Availability As the virtual machine has been disengaged from the hardware, additional capabilities become possible for availability. Without compute virtualization, if the hardware had a failure, the Page 20 of 132 Copyright IBM and VMware

21 operating system and any applications running on it could be impacted. By using vsphere virtualization, virtual machines can be re-started on remaining hosts in a cluster in the event of the catastrophic failure of a host. This can also be used by the cloud admin to take a host offline for maintenance without affecting workloads on the cluster Performance The amount of physical resources available to virtual machines can be controlled through vcenter resource pools. This allows for different resource pools which can have higher or lower priority of physical resources based on share allocation Workload Movement By linking a central cloud with one or more cloud regions, vcenter Server allows a virtual machine to be migrated to a remote cloud region or from cloud region to cloud region. This is also possible from an on premises installation attached to the same environment. This enables workloads to be migrated from client premises into the cloud and back again Storage Virtualization Storage virtualization provides two levels of virtualization. The first is the virtualization of the storage arrays and the second is the virtualization of the block storage used by virtual machines Virtual Storage Area Network (VSAN) Virtual Storage Area Networking (VSAN) emulates a physical storage area network entirely within the virtualization layer. Each host in the cluster contains local drives that are combined in software to behave as a single disk array that is shared between all the hosts in the cluster as a shared datastore. Since there is no physical storage area network, VSAN has the advantage of fewer components (no external drive array, fiber cabling, etc.). It allows ease of scaling when adding new compute nodes with less administration like performing tasks such as, LUN allocation which are no longer necessary. Plus, VSAN provides high performance since local disk is used and disk I/O is spread out across all hosts within a cluster. Storage policies are used to define storage attributes such as performance and protection levels. The policy is set per virtual machine allowing great flexibility with the service levels available Virtual Machine Disks (VMDK) Each virtual machine has at least one virtual machine disk (VMDK). Additional disks can be added to a virtual machine. The virtual disks are provisioned on to the datastores provided by VSAN. All virtual disks are thin provisioned, so unused disk space within a single virtual disk does not take up datastore disk capacity Network Virtualization Network virtualization provides a network overlay that exist within the virtual layer. As a result, it can provide more rapid provisioning, deployment, re-configuration and destruction over physical devices. Copyright IBM and VMware Page 21 of 132

22 Network Virtualization Components The network virtualization architecture of this design utilizes VMware NSX for vsphere and vsphere Distributed Switches (vds). The virtualized network is organized hierarchically, with the following components from bottom to top: Data plane with the NSX vswitch and additional components Control plane with the NSX Controller Management plane with the NSX Manager Consumption plane with a Cloud management portal Figure 8 Network Virtualization Distributed Virtual Switches This design implements vsphere distributed switches. vsphere Distributed Switch (vds) offers several enhancements over standard virtual switches. Centralized management. Because distributed switches are created and managed centrally on a vcenter Server system, they make the switch configuration more consistent across ESXi hosts. Centralized management saves time, reduces mistakes, and lowers operational costs. Additional features. Distributed switches offer features that are not available on standard virtual switches. Some of these features can be useful to the applications and services that are running in the organization s infrastructure. For example, NetFlow and port mirroring provide monitoring and troubleshooting capabilities to the virtual infrastructure. Page 22 of 132 Copyright IBM and VMware

23 Distributed virtual switch implements health checks. The health check service helps identify and troubleshoot configuration errors in vsphere distributed switches. Health check helps identify the following common configuration errors: Mismatched VLAN trunks between a vsphere distributed switch and physical switch. Mismatched MTU settings between physical network adapters, distributed switches, and physical switch ports. Mismatched virtual switch teaming policies for the physical switch port-channel settings. Health check monitors VLAN, MTU, and teaming policies: VLANs. Checks whether the VLAN settings on the distributed switch match the trunk port configuration on the connected physical switch ports. MTU. For each VLAN, checks whether the physical access switch port MTU jumbo frame setting matches the distributed switch MTU setting. Teaming policies. Checks whether the connected access ports of the physical switch that participate in an EtherChannel are paired with distributed ports whose teaming policy is IP hash. Health check is limited to the access switch port to which the distributed switch uplink connects. With network I/O control, the distributed switch allocates bandwidth for the following system traffic types: vsphere vmotion traffic Management traffic VMware vsphere Replication traffic NFS traffic VMware Virtual SAN traffic vsphere Data Protection backup traffic Virtual machine traffic Fault tolerance traffic iscsi traffic Network I/O control details The bandwidth for each network resource pool is controlled by setting the physical adapter shares and host limits. The bandwidth for virtual machines is controlled by bandwidth reservation for an individual VM, similar to the way memory and CPU reservation is used. The physical adapter shares assigned to a network resource pool determine the share of the total available bandwidth guaranteed to the traffic that is associated with that network resource pool. The share of transmit bandwidth that is available to a network resource pool is determined by these factors: The network resource pool's shares. Other network resource pools that are actively transmitting Data Plane The NSX data plane consists of the NSX vswitch, which is based on the vsphere Distributed Switch (vds) and includes additional components. These components include kernel modules (VIBs), which run within the ESXi kernel and provide services such as virtual distributed router (VDR) and distributed firewall (DFW). The NSX kernel modules also enable Virtual Extensible LAN (VXLAN) capabilities. The NSX vswitch abstracts the physical network and provides access-level switching in the hypervisor. It is central to network virtualization because it enables logical networks that Copyright IBM and VMware Page 23 of 132

24 are independent of physical constructs such as VLAN. The NSX vswitch provides multiple benefits. Three types of overlay networking capabilities: Creation of a flexible logical Layer 2 overlay over existing IP networks on existing physical infrastructure. Support for east/west and north/south communication while maintaining isolation between tenants. Support for application workloads and virtual machines that operate as if they were connected to a physical Layer 2 network. Support for VXLAN and centralized network configuration. A comprehensive toolkit for traffic management, monitoring and troubleshooting within a virtual network which includes port mirroring, NetFlow/IPFIX, configuration backup and restore, network health check, Quality of Service (QoS), and Link Aggregation Control Protocol (LACP) In addition to the NSX vswitch, the data plane also includes gateway devices (NSX Edge gateways), which can provide Layer 2 bridging from the logical networking space (VXLAN) to the physical network (VLAN). NSX Edge Gateway devices offer Layer 2, Layer 3, perimeter firewall, load-balancing and other services such as Secure Socket Layer (SSL), Virtual Private Network (VPN), and Dynamic Host Control Protocol (DHCP) Control Plane The NSX control plane runs in the NSX Controller, which enables unicast VXLAN and controlplane programming of elements such as VDR (virtual distributed router). Unicast support is necessary as the multicast IP range per each VLAN is limited within SoftLayer. The number of multicast or unicast IPs determines the number of VXLANs that can be provisioned. In all cases the controller is part of the control plane and does not have any data plane traffic passing through it. The controller nodes are deployed in a cluster per NSX Manager to enable high availability and scalability. A failure of one or all controller nodes does not impact data plane traffic Management Plane The NSX management plane consists of the NSX Manager, which is the single point of configuration, and the REST API entry-points. NSX Manager integrates with vcenter. There is one NSX Manager per vcenter Server Consumption Plane Different actors interact with NSX for vsphere to access and manage the associated services in different ways: Cloud admin can manage the NSX environment from the vsphere Web Client. Users can consume the network virtualization capabilities of NSX for vsphere through the CMP (vrealize Automation) UI when deploying applications Network Virtualization Services Network virtualization services include logical switches, logical routers, logical firewall, and other components of NSX for vsphere. Page 24 of 132 Copyright IBM and VMware

25 Logical Switches Cloud deployments have a variety of applications that are used across multiple tenants. These applications and tenants require isolation from each other for security, fault isolation, and overlapping IP addresses. The NSX for vsphere logical switch creates logical broadcast domains or segments to which an application or tenant virtual machine can be logically wired. This allows for flexibility and speed of deployment while still providing all the characteristics of a physical network's broadcast domains (VLANs) without physical Layer 2 sprawl or spanning tree issues. A logical switch is distributed and can span arbitrarily large compute clusters. This allows for virtual machine mobility (migration with vmotion) within a region and between regions, without limitations of the physical Layer 2 (VLAN) boundary Logical Routers Dynamic routing provides the necessary forwarding information between Layer 2 broadcast domains, thereby allowing the cloud admin to decrease the size of Layer 2 broadcast domains and improve network efficiency and scale. NSX for vsphere extends this intelligence to where the workloads reside for east/west routing. This allows more direct VM-to-VM communication without the costly need to extend hops. At the same time, logical routers provide north/south connectivity, thereby enabling users to access public networks Logical Firewall NSX for vsphere Logical Firewall provides security mechanisms for dynamic virtual datacenters. The Distributed Firewall component of Logical Firewall allows a cloud admin to segment virtual datacenter entities like virtual machines based on VM names and attributes, user identity, vcenter objects like datacenters, and hosts, or based on traditional networking attributes like IP addresses, port groups, and so on. The Edge Firewall component helps a cloud admin to meet key perimeter security requirements, such as building DMZs based on IP/VLAN constructs, tenant-to-tenant isolation in multi-tenant virtual datacenters, Network Address Translation (NAT), partner (extranet) VPNs, and user-based SSL VPNs. The Flow Monitoring feature displays network activity between virtual machines at the application protocol level. The cloud admin can use this information to audit network traffic, define and refine firewall policies, and identify threats to a client s network Logical Virtual Private Networks (VPN s) SSL VPN-Plus allows remote users to access private corporate applications. IPSec VPN offers site-to-site connectivity between an NSX Edge instance and remote sites. L2 VPN allows users to extend their datacenter by allowing virtual machines to retain network identity across geographical boundaries Logical Load Balancers The NSX Edge load balancer enables network traffic to follow multiple paths to a specific destination. It distributes incoming service requests evenly among multiple servers in such a way that the load distribution is transparent to users. Load balancing thus helps in achieving optimal resource utilization, maximizing throughput, minimizing response time, and avoiding overload. NSX Edge provides load balancing up to Layer Service Composer Service Composer helps provision and assign network and security services to applications in a virtual infrastructure. The service provider maps these services to a security group, and the services are applied to the virtual machines in the security group. Copyright IBM and VMware Page 25 of 132

26 4.5 Infrastructure Management The infrastructure management element manages the compute, network and storage virtual resources provided by the lower layer. It also provides consolidation services to the upper layers for operational services. These functions are provided by VMware vcenter Server Compute Management In this design, VMware vcenter is employed to centralize the management of the compute resources within each ESXi host. While the ESXi hosts can be managed individually, placing them under vcenter control enables the following capabilities: Centralized control and visibility of all aspects within managed ESXi hosts and virtual machines. Provides the single pane of glass interface view via the vcenter web client for compute, network and storage management. Proactive Optimization. Enables allocation and optimization of resources for maximum efficiency across the ESX hosts. See section Compute Virtualization for optimization features enabled by vcenter Server. Extended management function for other integrated products and services such as VMware NSX, VMware Data Protection, VMware Update Manager and others as snap-ins extending the vcenter web interface. Monitoring, alerting, scheduling. Cloud admins can view events, alerts within the vcenter web client, and configure scheduled actions. Automation engine. VMware vcenter is the engine which performs the tasks given to it via the vsphere API web interface. VMware vrealize Automation and vrealize Orchestration are examples of applications that drive vcenter actions via the API Storage Management VMware vcenter enables centralized storage management within this design which allows for configuration and management of the following storage types: Local disk storage. Local hard disk drives (HDD) or solid state drives (SDD) that are attached to the local ESXi hosts. Storage area attached storage (SAN). Remote block storage that is attached to the ESXi host via fiber channel or TCPIP protocols. Network attached storage. (NAS) File based storage that is attached to the ESXi hosts via the NFS protocol. Virtual SAN Storage. Configured within the cluster object in vcenter, enables the aggregation of local disk storage across ESXi hosts into a shared pool of storage across all ESX hosts within a given cluster. Once configured, an outage of the vcenter server does not affect the availability of VSAN storage to the cluster. Within this design vcenter management of storage is primarily focused on NAS and VSAN storage as SAN is not employed. Only the ESXi host OS and swap space are used on local non VSAN disk storage NFS Storage management vcenter is responsible for configuring the mounting of NFS data stores to each ESXi host within a cluster. This ensures its access availability to any virtual machine with virtual disk files (VMDK) Page 26 of 132 Copyright IBM and VMware

27 residing on the NFS based data store, should a vmotion of the virtual machine from one ESXi host to another occur within a cluster VSAN Storage management Here again, the vcenter interface or web API has the capability of configuring VSAN data stores for a particular cluster at the cluster object level within the vcenter interface. Configuring VSAN within vcenter involves the following areas of configuration: Licensing. Prior to enabling VSAN, a valid license within vcenter licensing section is required. VSAN network. Used to configure the network VSAN will use for its back plane network. Virtual machines are made storage fault tolerant across ESXi hosts local disks on this network). Disk group configuration. On each ESXi host that contributes its local disks to a Virtual SAN cluster, disks are organized into disk groups. A disk group is a main unit of storage on a host. Each disk group includes one SSD and one or multiple HDDs. VSAN Policies. Storage policies define the virtual machine storage characteristics. Storage characteristics specify different levels of service for different virtual machines Network Management vcenter Server is used to create standard and distributed virtual switches. The virtual switches connect virtual machine (VM) network interfaces to portgroups allowing for communication between VM s hosted on the same host or different hosts. To establish communication between hosts, virtual switches need to be connected to physical uplinks which are the network interfaces of the ESXi hosts. VM s connected on the same virtual switch and hosted on the same host can communicate directly without the need of an external uplink. vcenter Server enables creation of distributed portgroups for virtual machines (aggregated virtual ports with a particular set of specifications). 4.6 Common Services Common services provide the services used by other services in the cloud management platform. This includes identity and access services, domain name services, NTP services, SMTP services and Certificate Authority Services Identity and Access Services In this design, Microsoft (MS) Active Directory (AD) is employed to provide authentication and directory services back end to the VMware Platform Service Controller (PSC) and VMware identity appliance. Within this design the VMware software components authenticate against the identity appliance which in turn authenticates against the MS AD service. The AD in this design can be extended to other regions by adding an additional AD server for that particular region s subdomain. Copyright IBM and VMware Page 27 of 132

28 4.6.2 Domain Name Services Domain Name Services (DNS) within this design are for the cloud management and infrastructure components only. DNS in this design serves as host name to IP resolution for the cloud management platform and service resolution for the AD components. When an instance of this design is tied to a customer on premises solution, this design s DNS servers are referenced by the on premises DNS infrastructure in addition to acting as a proxy for the customer s DNS infrastructure NTP Services This design s NTP servers are a sub stratum of the SoftLayer infrastructure NTP server. They serve to ensure that all components are in time synchronization for the needs of authentication, replication, clustering, log synchronization and certificate services. This includes physical and virtual components SMTP Services Simple Mail Transfer Protocol (SMTP), is utilized within this design by various components for the needs of outbound notification only. For inbound requirements (vrealize Automation, vrealize Business), the customer servers need to be configured Certificate Authority Services An Enterprise Certificate Authority (CA) based on Microsoft (MS) CA services built into MS Windows is employed in this solution to replace self-signed certificates for web interfaces within this design. 4.7 Cloud Management Services The cloud management services provide the service catalog, self-service portal and orchestration. This is provided by VMware vrealize Automation, vrealize Orchestrator and Rapid Deployment Services (RDS) pattern automation Service Catalog The service catalog is published through the self service catalog and allows users to request the provided services which can include provisioning new virtual machines from templates, provisioning new environments consisting of one or more virtual machines with software products as blueprints (also known as patterns), or managing existing deployed resources. Advanced services are also available through the service catalog by calling the orchestration component for process orchestration. The service provider role is able to customize the services available to users as well as publish additional services Self-Service Portal The self service portal provides a single point of access for users to the VMware SDDC on IBM Cloud solution. Authentication to the portal is performed against the Active Directory service. Page 28 of 132 Copyright IBM and VMware

29 4.7.3 Infrastructure and Process Orchestration Orchestration is provided by vrealize Orchestrator. It allows for tasks and remediation actions to be automated including integration with third party IT operations software. vrealize Orchestrator consists of: Workflow designer which incorporates an easy-to-use drag and drop interface to assemble workflows. The designer runs on Windows, Linux and Mac OS desktops. Scripting designer which allows for new building blocks to be created or imported for the vrealize Orchestrator platform. Orchestration engine which runs the workflows and associated scripts. The default implementation includes a built-in workflow library with common tasks. Workflows are able to be versioned and packaged to assist with change management Software Orchestration Software Orchestration is provided by a Rapid Deployment Services (RDS) solution with IBM Open Patterns. RDS implements a distributed file repository and the configuration management tools to deliver IBM Open Patterns on deployed workloads. IBM Open Patterns describe the pre-defined architecture of an application. For each component of the application (i.e. database, web server, etc.), the pattern defines: Pre-installation on an operating system Pre-integration across components Pre-configured & tuned Pre-configured Monitoring Pre-configured Security Lifecycle Management 4.8 Operational Services Operational services provide management of the cloud services. This includes backup & restore, disaster recovery, monitoring, log consolidation & analysis and patching functions Backup and Restore The data protection service protects the infrastructure that provides the virtualization, operations, security and cloud services. It does not protect any deployed user virtual machines. Data protection solutions provide the following functions in the design: Back up and restore virtual machines and database applications. Store data according to company retention policies. Copyright IBM and VMware Page 29 of 132

30 Inform administrators about backup and restore activities through reports. vsphere Data Protection provides the data protection service in each region. This is separate to disaster recovery and applies even if only the central cloud exists. An FTP server is used to backup NSX Manager. The FTP server supports SFTP and FTP protocols Disaster Recovery Figure 9 Dual-Region Data Protection Architecture The disaster recovery service adds to the data protection service by protecting the management services in the case of a complete site failure. It is an optional service to provide additional protection. Since this requires more than one site, it is only applicable where a central cloud and at least one cloud region has been included. VMware Site Recovery Manager (SRM) and vsphere Replication are used to provide this service, together with keeping the same IP addressing of the cloud management services at both sites. Note: Each central cloud or cloud region in this design is equivalent to the site construct in Site Recovery Manager. Since the central cloud contains the portal and manages the services in all the regions, the following applications are in scope of disaster recovery protection: Page 30 of 132 Copyright IBM and VMware

31 vrealize Automation together with VMware vrealize Orchestrator Analytics cluster of vrealize Operations Manager The services that support the services at each site do not require disaster recovery protection. This includes; vsphere, NSX and vcenter services which manage the services at the local site only. Authentication, DNS and NTP which is distributed to the cloud regions anyway. vrealize Log Insight and Software Orchestration which is replicated to all cloud regions Figure 10 Disaster Recovery Architecture Copyright IBM and VMware Page 31 of 132

32 4.8.3 Monitoring vrealize Operations Manager is used to track and analyze the operation of multiple data sources within the design by using specialized analytics algorithms. These algorithms help vrealize Operations Manager to learn and predicts the behavior of every object it monitors. Users access this information by using views, reports, and dashboards. vrealize Operations Manager contains functional elements that collaborate for data analysis and storage, and support creating clusters of nodes with different roles. Figure 11 vrealize Operations Manager Architecture For high availability and scalability, several vrealize Operations Manager instances are deployed in the management cluster where they have the following roles: Master Node. Required initial node in the cluster. In large-scale environments the master node manages all other nodes. In small-scale environments, the master node is the single standalone vrealize Operations Manager node. Master Replica Node. Enables high availability of the master node. Data Node. Enables scale-out of vrealize Operations Manager in larger environments. Data nodes have adapters installed to perform collection and analysis. Data nodes also host vrealize Operations Manager management packs. Remote Collector Node. Enables navigation through firewalls, interfaces with a remote data source, reduces bandwidth across regions, or reduces the load on the vrealize Operations Manager analytics cluster. Remote collector nodes only gather objects for the inventory and forward collected data to the data nodes. Remote collector nodes do not store data or perform analysis. In addition, they can be installed on a different operating system than the rest of the cluster nodes. The master and master replica nodes are data nodes with extended capabilities. Page 32 of 132 Copyright IBM and VMware

33 vrealize Operations Manager forms two types of cluster according to the nodes that participate in a cluster: Analytics clusters. Tracks, analyzes, and predicts the operation of monitored systems. Consists of a master node, data nodes, and master replica node. Remote collectors cluster. Only collects diagnostics data without storage or analysis. Consists only of remote collector nodes. The functional components of a vrealize Operations Manager instance interact to provide analysis of diagnostics data from the datacenter and visualize the result in the Web user interface. Table 3 vrealize Operations Manager Logical Node Architecture Architecture Component Diagram Description Admin / Product UI server. The UI server is a Web application that serves as both user and administration interface. REST API / Collector. The Collector collects data from all components in the datacenter. Controller. The Controller handles the data flow the UI server, Collector, and the analytics engine. Analytics. The Analytics engine creates all associations and correlations between various data sets, handles all super metric calculations, performs all capacity planning functions, and is responsible for triggering alerts. Persistence. The persistence layer handles the read and write operations on the underlying databases across all nodes. FSDB. The File System Database (FSDB) stores collected metrics in raw format. FSDB is available in all the nodes. xdb (HIS). The xdb stores data from the Historical Inventory Service (HIS). This component is available only on the master and master replica nodes. Global xdb. The Global xdb stores user preferences, alerts, and alarms, and customization that is related to the vrealize Operations Manager. This component is available only on the master and master replica nodes Log Consolidation and Analysis Log consolidation and analysis provides consolidation of the logs that are produced by each of the cloud services together with analysis of those logs. For this design, this function is provided by vrealize Log Insight. vrealize Log Insight provides real-time log management and log analysis with machine learning-based intelligent grouping, high-performance searching, and troubleshooting across physical, virtual, and cloud environments. Copyright IBM and VMware Page 33 of 132

34 vrealize Log Insight collects data from ESXi hosts using the syslog protocol. It connects to vcenter Server to collect events, tasks, and alarms data, and integrates with vrealize Operations Manager to send notification events and enable launch in context Patching Patching of the VMware software components is achieved with VMware Update Manager. This includes the VMware ESXi hosts, virtual appliances and management tooling in the design. It connects to the internet to obtain the latest vulnerability patches and automatically applies user-defined patches to the relevant components to eliminate the vulnerabilities. 4.9 Business Services Business services are those services that provides business functions. This includes business management, IT financials and IT benchmarking. vrealize Business (VRB) is configured to provide financial information, reporting and modeling. VRB integrates with vrealize Automation Business Management vrealize Business provides the following business management capabilities: Automatic private cloud metering costing and pricing IT Financials vrealize Business provides the following capability for financial management: Automatic service catalog pricing (Integrated with vrealize Automation) Private cloud consumption analysis Out-of-the-box reporting (Exportable data set) IT Benchmarking Additionally, vrealize Business can assist in modeling cost projections across cloud environments. Private cloud and public cloud cost comparison Page 34 of 132 Copyright IBM and VMware

35 4.10 Cloud Region The cloud region is a child instance of the design. It is not standalone and requires a central cloud to provide the cloud management services. Provisioning and management of virtual resources is done through the central cloud. The cloud management services do not exist for a cloud region since this is handled by the central cloud and the operational management services which contain collectors/relays to pass information back to the central cloud. Copyright IBM and VMware Page 35 of 132

36 5 Physical Operational Model The physical operational model elaborates by applying the non-functional requirements to the logical operational model. 5.1 Physical Layer Compute Figure 12 Physical Operational Model - Virtual Servers, Networking and Clusters The design leverages SoftLayer to provide the compute. This allows for flexibility in provisioning bare metal. Compute nodes are able to be deployed rapidly without needing orders to be delivered and the same nodes are able to be decommissioned without needing to wait for depreciation schedules or reselling. SoftLayer offers a variety of bare metal Intel based hardware from 1U to 4U chassis sizes, 2GB of memory up to 3TB and from 4 CPU cores up to 48 cores. For this design, a 2U server has been selected to allow for the lowest cost of entry, while still allowing for scaling up to 10,000 deployed VMs in a single central cloud or cloud region. For security and to isolate management, network and user workloads (resources) this design branches out to 3 vsphere cluster types with the following functions: Management Cluster Edge Cluster Compute Clusters This allows the scaling of one function independent of the other as the deployment is scaled out, making for a more effective use of resources. Page 36 of 132 Copyright IBM and VMware

37 Management Cluster The management cluster is the heart of operation and control of the design. It is sized from onset of deployment to allow for Compute cluster expansion and additional feature expansion without requiring additional sever nodes. It consists of 4 nodes of the following specification: 2 x 12 core CPUs (24 cores total), plus Hyperthreading 256GB RAM ~6.3TB Usable VSAN storage ~1TB Usable Local Disk for operating system Refer to Appendix A Bare Metal Summary for hardware details Edge Cluster VMware NSX is an integral part of the design. Edge virtual appliances are used where VPN end points or load balancing is required. Edge virtual appliances are controlled and can be dynamically provisioned as user applications are spun up with patterns or are preconfigured to support management functions. In either case they are deployed to the Edge cluster. This ensures that network connectivity and performance are not affected by varying workloads in the other clusters. It is sized from onset of deployment to allow for Compute cluster expansion without requiring additional servers. As the Edge virtual appliances are small, VSAN storage requirements are held to a minimum while maintaining redundancy and performance. It consists of 4 nodes of the following specification: 2 x 6 core CPUs (12 cores total), plus Hyperthreading 128GB RAM ~ 3.2TB Usable VSAN storage ~ 1TB Usable Local Disk for operating system Refer to Appendix A Bare Metal Summary for hardware details Compute Cluster As the users or administrators within the customer s organization deploy whatever applications they require via vrealize Automate, the compute workloads requested get deployed to this cluster. With scale up as well as out in mind, high resource intensive applications and other mixed workloads are able to be absorbed. Additional clusters are provisioned when the capacity of the each cluster is reached. It consists of 4 nodes of the following specification: 2 x 12 core CPUs (24 cores total), plus Hyperthreading 512GB RAM ~6.3TB Usable VSAN storage ~1TB Usable Local Disk for operating system Refer to Appendix A Bare Metal Summary for hardware details. Each node supports 80 VMs of 2 vcpu, 8GB RAM and 70GB Disk. Taking into account the following: CPU over-commit 7 90% CPU usage limit on ESXi Memory over-commit is % memory usage limit on ESXi Copyright IBM and VMware Page 37 of 132

38 5.1.2 Storage No disk over-commit Maximum of 48 nodes per cluster / Cluster and 3 clusters for 10,000 VM support. Cluster sizing based upon 1 node failure per 15 nodes. So up to 15 nodes, a single node is reserved in case of failure. Between 15 and 30 nodes, there are two nodes kept in reserve and so forth up to 45 nodes where 3 nodes are in reserve. Network File System (NFS) is a file system protocol that allows a user on a client computer to access files over a network much like local storage is accessed. In this case, the client computer is an ESXi host, and the storage is provided by a NFS-capable external storage array. The management cluster uses VMware Virtual SAN for primary storage and NFS for secondary storage. For compute clusters, the decision on which technology to use is based on the performance, capacity, and capabilities (replication, deduplication, compression, etc.) required by the workloads that are running in the clusters. For this design, additional storage for the ESXi management cluster is connected to SoftLayer File Storage on Performance Storage via 10 Gbps links with Jumbo Frames (MTU 9000) enabled. For the virtual machines that provide the backup service and log archiving (i.e., vsphere Data Protection and vrealize Log Insight) the NFS datastores are configured in the following manner: Table 4 NFS Configuration for vsphere Data Protection and vrealize Log Insight Product vsphere Data Protection vrealize Log Insight Configuration vsphere Data Protection is I/O intensive and is placed on its own, unique NFS datastore sized at 4TB vrealize Log Insight uses NFS datastores sized at 1TB for archive storage and can be shared will other virtual machines Network The design leverages SoftLayer physical networking and this is broken up into server connections, VLANs and MTU packet sizing Server connections Each compute node (physical server) within the design has two 10Gb Ethernet connection into each SoftLayer Top of Rack (ToR) switch (public and private) setup as individual connections (un-bonded) for a total of 4 x 10Gbps connections. This allows each networking interface card (NIC) connection to work independent of one another. Page 38 of 132 Copyright IBM and VMware

39 VLANs Figure 13 Network connections per physical node Four VLANs are included per design including the public network. They are as listed for this design: MTU Sizing Private VLAN 1 (default / untagged) Private VLAN 2 (trunked / tagged) Private VLAN 3 (trunked / tagged ) Public VLAN 4 (default / untagged) The private network connections are configured to use jumbo frame MTU size of This is maximum MTU allowed within VMware and SoftLayer limit which improves performance for large data transfers such as storage and vmotion. The public network connections use a standard Ethernet MTU frame size of This needs to be maintained as any changes may affect packet fragmentation over the internet. 5.2 Virtual Infrastructure Virtual infrastructure consists of compute, storage and network virtualization Compute Virtualization This design uses VMware vsphere ESXi version 6.0 u1 to virtualize the management, compute, and edge servers. The ESXi hypervisor is installed on the 2x1TB RAID-1 disk array contained on each server. RAID-1 is used in this design to provide redundancy for the vsphere hypervisor Storage Virtualization For this design, VMware Virtual Storage Area Network (VSAN) storage is employed for all storage needs with the exception of vrealize Log Insight log archival storage and VMware Data Protection backup storage. VSAN allows for the local storage across multiple ESXi hosts within a vsphere cluster to be represented as a single virtual machine datastore. VSAN supports only SATA, SAS HDD, and PCIe storage. Two 1TB SATA drives are excluded in each node regardless of which cluster it belongs to for the purpose of housing the ESXi installation. Copyright IBM and VMware Page 39 of 132

40 Figure 14 VSAN concept RAID Controller As of this version of the design, only the Avago MegaRAID i RAID controller within SoftLayer hardware is supported by VMware VSAN. Disk caching need is disabled and the controller is set in JBOD mode for all VSAN drives. Disks and disk groups Depending on the cluster function (Management, Edge, Compute) the number and sizes of the disks changes. Each VSAN disk group requires a solid state disk (SSD) for a cache layer. Within all clusters, a 2U server type is employed with a total of 12 drive slots maximum capacity. Excluding the ESXi OS drives, the following is the drive layout for each cluster type: Table 5 VSAN disk table Cluster type Number of VSAN disk groups Number of SSDs # of SSD + HDD per disk group Size of SSDs Number of HDDs Size of HDDs Management ,200GB 8 2,000GB Edge ,200GB 4 2,000GB Compute ,200GB 8 2,000GB Virtual network setup For this design the VSAN traffic will traverse between ESXi hosts on a dedicated private VLAN. No other traffic will occupy the VSAN VLAN. The two network adapters attached to the private network switch and configured in within vsphere as a virtual distributed switch (vds) with both network adapters as uplinks. A dedicated VSAN kernel port configured for the VSAN VLAN resides within the vds. Jumbo frame are enabled for the private vds Virtual SAN Policy Design Once VMware Virtual SAN is enabled and configured, create storage policies that define the virtual machine storage characteristics. Storage characteristics specify different levels of service Page 40 of 132 Copyright IBM and VMware

41 for different virtual machines. The default storage policy tolerates a single failure and has a single disk stripe. Use the default unless a client s environment requires policies with non-default behavior. If a custom policy is configured, Virtual SAN will guarantee it; however, if Virtual SAN cannot guarantee a policy, it is not possible to provision a virtual machine that uses the policy unless it is enabled to force provisioning. This design will use the default policy. Table 6 VSAN policies Capability Use Case Value Comments Number of failures to tolerate Number of disk stripes per object Flash read cache reservation (%) Object space reservation (%) Force provisioning Redundancy Performance Performance Thick provisioning Override policy Default 1 Max 3 Default 1 Max 12 Default 0 Max 100% Default 0 Max 100% Default: No A standard RAID 1 mirrored configuration that provides redundancy for a virtual machine disk. The higher the value, the more failures can be tolerated. For n failures tolerated, n+1 copies of the disk are created, and 2n+1 hosts contributing storage are required. A higher n value indicates that more replicas of virtual machines are made, which can consume more disk space than expected. A standard RAID 0 stripe configuration used to increase performance for a virtual machine disk. This setting defines the number of HDDs on which each replica of a storage object is striped. If the value is higher than 1, increased performance can result. However, an increase in system resource usage might also result. Flash capacity reserved as read cache for the storage is a percentage of the logical object size that will be reserved for that object. Only use this setting for workloads if a client must address read performance issues. The downside of this setting is that other objects cannot use a reserved cache. VMware recommends not using these reservations unless it is absolutely necessary because unreserved flash is shared fairly among all objects. The percentage of the storage object that will be thick provisioned upon VM creation. The remainder of the storage will be thin provisioned. This setting is useful if a predictable amount of storage will always be filled by an object, cutting back on repeatable disk growth operations for all but new or non-predictable storage use. Force provisioning allows for provisioning to occur even if the currently available cluster resources cannot satisfy the current policy. Force provisioning is useful in case of a planned expansion of the Virtual SAN cluster, during which provisioning of VMs must continue. Virtual SAN Copyright IBM and VMware Page 41 of 132

42 Capability Use Case Value Comments automatically tries to bring the object into compliance as resources become available. By default, policies are configured based on application requirements. However, they are applied differently depending on the object. Table 7 VSAN object policy defaults Object Policy Comments Virtual machine namespace Failures-to-Tolerate: 1 Configurable. Changes are not recommended. Swap Failures-to-Tolerate: 1 Configurable. Changes are not recommended. Virtual disk(s) Virtual disk snapshot(s) User-Configured Storage Policy Uses virtual disk policy Can be any storage policy configured on the system. Same as virtual disk policy by default. Changes are not recommended. If a user-configured policy is not specified, the default system policy of 1 failure to tolerate and 1 disk stripe is used for virtual disk(s) and virtual disk snapshot(s). Policy defaults for the VM namespace and swap are set statically and are not configurable to ensure appropriate protection for these critical virtual machine components. Policies must be configured based on the application s business requirements. Policies give Virtual SAN its power because it can adjust how a disk performs on the fly based on the policies configured Network Virtualization This design uses vsphere virtual distributed switches and VMware NSX for vsphere to implement virtual networking. By using NSX, this design implements software-defined networking Virtual Distributed Switch Design The design uses a minimum number of switches. Clusters connected to public network is configured with two distributed virtual switches. Clusters restricted to private network have only one distributed virtual switch. Separating different types of traffic is required to reduce contention and latency. Separate networks are also required for access security. VLANs are used to segment physical network functions. This design uses four (4) VLANs. Three (3) for private network traffic and one (1) for public network traffic. Traffic separation is detailed in the section below. Table 8 VLAN Mapping to Traffic Types VLAN VLAN1 VLAN2 VLAN3 VLAN4 Traffic Type ESXi Management, VXLAN (VTEP) VSAN vmotion, NFS, vsphere Replication All Internet access Traffic from workloads will travel on VXLAN backed logical switches. Page 42 of 132 Copyright IBM and VMware

43 Management Cluster Distributed Switches The management cluster uses a dual vsphere Distributed Switch with the configuration settings shown in this section. Table 9 Management Cluster Distributed Switch vsphere Distributed Switch Name vds-mgmt- Priv Function ESXi management Network IP Storage (NFS) Virtual SAN vsphere vmotion VXLAN Tunnel Endpoint (VTEP) vsphere Replication/vSphere Replication NFC Network I/O Control Enabled Load Balancing Mode Based on source MAC hash Number of Physical NIC Ports MTU 2 9,000 (Jumbo Frame) vds-mgmt- Pub External management traffic (North-South) Enabled Based on source MAC hash 2 1,500 (default) Table 10 Management Cluster Distributed Switch Port Group Configuration Settings Parameter Setting Load balancing Route based on the source MAC hash Failover detection Link status only Notify switches Enabled Failback No Failover order Active uplinks: Uplink1, Uplink2 Management cluster hosts are connected to both private and public networks. There are 2 distributed virtual switches: one for private network and one for public network. Each switch has a dedicated pair of 10Gbps network adapters. Copyright IBM and VMware Page 43 of 132

44 Figure 15 Network Switch Design for Management Hosts Page 44 of 132 Copyright IBM and VMware

45 Table 11 Management Virtual Switch Port Groups and VLANs vsphere Distributed Switch vds-mgmt-priv Port Group Name vds-mgmt-priv- Management vds-mgmt-priv vds- Mgmt-Priv - vmotion vds-mgmt-priv vds- Mgmt-Priv - VSAN vds-mgmt-priv vds- Mgmt-Priv - VTEP vds-mgmt-priv vds- Mgmt-Priv - NFS vds-mgmt-priv vds- Mgmt-Priv - Replication vds-mgmt-pub vds- Mgmt-Pub - External Teaming Uplinks VLAN ID Source MAC hash Source MAC hash Source MAC hash Source MAC hash Source MAC hash Source MAC hash Source MAC hash Active: 0, 1 Active: 0, 1 Active: 0, 1 Active: 0, 1 Active: 0, 1 Active: 0, 1 Active: 0, 1 VLAN1 VLAN3 VLAN2 VLAN1 VLAN3 VLAN3 VLAN4 Table 12 Management VMkernel Adapter vsphere Distributed Switch Network Label Connected Port Group vds-mgmt-priv Management vds- Mgmt-Priv - Management vds- Mgmt-Priv vmotion vds- Mgmt-Priv - vmotion vds- Mgmt-Priv VTEP vds- Mgmt-Priv - VTEP vds- Mgmt-Priv VSAN vds- Mgmt-Priv - VSAN Enabled Services Management Traffic vmotion Traffic 9,000-9,000 VSAN 9,000 MTU 1,500 (default) Copyright IBM and VMware Page 45 of 132

46 Edge Cluster Distributed Switches The edge cluster uses a dual vsphere Distributed Switch with the configuration settings shown in this section. Table 13 Edge Cluster Distributed Switch vsphere Distributed Switch Name Function vds-edge-priv ESXi management Virtual SAN vsphere vmotion VXLAN Tunnel Endpoint (VTEP) Network I/O Control Enabled Load Balancing Mode Based on source MAC hash Number of Physical NIC Ports MTU 2 9,000 (Jumbo Frame) vds-edge-pub External user traffic (North- South) Enabled Based on source MAC hash (default) Table 14 Management Cluster Distributed Switch Port Group Configuration Settings Parameter Setting Load balancing Route based on the source MAC hash Failover detection Link status only Notify switches Enabled Failback No Failover order Active uplinks: Uplink1, Uplink2 Page 46 of 132 Copyright IBM and VMware

47 Edge cluster hosts are connected to both private and public networks. There are 2 distributed virtual switches: one for private network and one for public network. Each switch has a dedicated pair of 10Gbps network adapters. Figure 16 Network Switch Design for Edge Hosts Copyright IBM and VMware Page 47 of 132

48 Table 15 Edge Virtual Switch Port Groups and VLANs vsphere Distributed Switch vds-edge-priv Port Group Name vds-edge-priv- Management vds-edge-priv vds-edge-priv - vmotion vds-edge-priv vds-edge-priv vds-edge-pub vds-edge -Priv- VSAN vds-edge-priv- VTEP vds-edge-pub- External Teaming Uplinks VLAN ID Source MAC hash Active: 0, 1 VLAN1 Source MAC hash Active: 0, 1 VLAN3 Source MAC hash Active: 0, 1 VLAN2 Source MAC hash Active: 0, 1 VLAN1 Source MAC hash Active: 0, 1 VLAN4 Table 16 Edge VMkernel Adapter vsphere Distributed Switch Network Label Connected Port Group vds-edge-priv Management vds-edge-priv - Management vds-edge-priv vmotion vds-edge-priv - vmotion vds-edge-priv VTEP vds-edge-priv - VTEP vds-edge-priv VSAN vds-edge-priv - VSAN Enabled Services Management Traffic vmotion Traffic 9,000-9,000 VSAN 9,000 MTU 1,500 (default) Compute Cluster Distributed Switches The compute cluster uses a single vsphere Distributed Switch with the configuration settings shown in this section. Table 17 Compute Cluster Distributed Switch vsphere Distributed Switch Name vds-compute- Priv Function ESXi management Virtual SAN vsphere vmotion VXLAN Tunnel Network I/O Control Enabled Load Balancing Mode Based on source MAC hash Number of Physical NIC Ports MTU 2 9,000 (Jumbo Frame) Page 48 of 132 Copyright IBM and VMware

49 vsphere Distributed Switch Name Function Network I/O Control Load Balancing Mode Number of Physical NIC Ports MTU Endpoint (VTEP) Table 18 Compute Cluster Distributed Switch Port Group Configuration Settings Parameter Setting Load balancing Route based on the source MAC hash Failover detection Link status only Notify switches Enabled Failback No Failover order Active uplinks: Uplink1, Uplink2 Compute cluster hosts are connected to the private network. The switch has a dedicated pair of 10Gbps network adapters. Figure 17 Network Switch Design for Compute Hosts Copyright IBM and VMware Page 49 of 132

50 Table 19 Compute Virtual Switch Port Groups and VLANs vsphere Distributed Switch vds-compute- Priv vds- Compute -Priv vds- Compute -Priv Port Group Name vds-compute- Priv- Management vds-compute- Priv-vMotion vds-compute- Priv-VSAN Teaming Uplinks VLAN ID Source MAC hash Source MAC hash Source MAC hash Active: 0, 1 Active: 0, 1 Active: 0, 1 VLAN1 VLAN3 VLAN2 vds- Compute -Priv vds-compute- Priv-VTEP Source MAC hash Active: 0, 1 VLAN1 Table 20 Compute VMkernel Adapter vsphere Distributed Switch vds-compute- Priv vds-compute- Priv vds-compute- Priv vds-compute- Priv Network Label Management vmotion VTEP VSAN Connected Port Group vds-compute- Priv- Management vds-compute- Priv-vMotion vds-compute- Priv-VTEP vds-compute- Priv-VSAN Enabled Services Management Traffic vmotion Traffic 9,000-9,000 VSAN 9,000 MTU 1,500 (default) NIC Teaming With the exception of VSAN, this design uses NIC teaming to avoid single point of failure and provide load balancing. To accomplish this, the design uses an active-active configuration with a route that is based on source MAC hash for teaming Network I/O Control This design uses Network I/O control enabled on all distributed switches with default shares. Network I/O control increases resiliency and performance of the network. Utilizing network I/O control, this design uses the distributed switch to allocate bandwidth for the following system traffic types: vsphere vmotion traffic Management traffic VMware vsphere Replication traffic NFS traffic VMware Virtual SAN traffic vsphere Data Protection backup traffic Virtual machine traffic Page 50 of 132 Copyright IBM and VMware

51 VXLAN This design uses VXLAN to create isolated, multi-tenant broadcast domains across datacenter fabrics and to enable customers to create elastic, logical networks that span physical network boundaries. VXLAN works by creating Layer 2 logical networks that are encapsulated in standard Layer 3 IP packets. A Segment ID in every frame differentiates the VXLAN logical networks from each other without any need for VLAN tags. As a result, large numbers of isolated Layer 2 VXLAN networks can coexist on a common Layer 3 infrastructure Software-Defined Network Design NSX offers the following Software-Defined Network (SDN) capabilities crucial to support the cloud management platform operations. load balancing firewalls routing logical switches VPN access Because NSX for vsphere is tied to a vcenter Server domain, this design uses two separate NSX instances. One instance is tied to the management vcenter Server, and the other instance is tied to the compute and edge vcenter Server. SDN capabilities are consumed via the cloud management platform, vsphere web client and API. The design uses API s to automate the deployment and configuration of NSX components by users and cloud admin actors NSX for vsphere Components This section describes the NSX for vsphere component configuration NSX Manager NSX Manager provides the centralized management plane for NSX for vsphere and has a one-toone mapping to a vcenter Server instance. This design uses two NSX managers (one for each vcenter Server in the design). NSX Manager performs the following functions. Provides the single point of configuration and the REST API entry-points in a vsphere environment for NSX for vsphere. Is responsible for deploying NSX Controller clusters, Edge distributed routers, and Edge service gateways in the form of OVF appliances, guest introspection services, and so on. Is responsible for preparing ESXi hosts for NSX for vsphere by installing VXLAN, distributed routing and firewall kernel modules, and the User World Agent (UWA). Communicates with NSX Controller clusters via REST and with hosts via the RabbitMQ message bus. The internal message bus is specific to NSX for vsphere and does not require setup of additional services. Generates certificates for the NSX Controller instances and ESXi hosts to secure control plane communications with mutual authentication NSX Controller The NSX Controllers perform the following functions: Provide the control plane to distribute VXLAN and logical routing information to ESXi hosts. Include nodes that are clustered for scale-out and high availability. Slice network information across cluster nodes for redundancy. Copyright IBM and VMware Page 51 of 132

52 Remove requirement of VXLAN Layer 3 multicast in the physical network. Provide ARP suppression of broadcast traffic in VXLAN networks. NSX for vsphere control plane communication occurs over the management network. This design implements a cluster of 3 NSX controllers for each NSX manager enabling high availability for the controllers NSX vswitch The NSX for vsphere data plane consists of the NSX vswitch. This vswitch is based on the vsphere Distributed Switch (vds) with additional components for add-on services. The add-on NSX for vsphere components include kernel modules which run within the hypervisor kernel and provide services such as distributed logical router (DLR) and distributed firewall (DFW), and enable VXLAN capabilities. This design uses NSX vswitch NSX Logical Switching NSX for vsphere logical switches create logically abstracted segments to which tenant virtual machines can be connected. A single logical switch is mapped to a unique VXLAN segment and is distributed across the ESXi hypervisors within a transport zone. It allows line-rate switching in the hypervisor without the constraints of VLAN sprawl or spanning tree issues. This design uses NSX Logical Switching for handling compute workloads and connectivity between different network zones Distributed Logical Router The NSX for vsphere Distributed Logical Router (DLR) is optimized for forwarding in the virtualized space, that is, forwarding between VMs on VXLAN- or VLAN-backed port groups. This design does not use distributed logical routers User World Agent The User World Agent (UWA) is a TCP (SSL) client that facilitates communication between the ESXi hosts and the NSX Controller instances as well as the retrieval of information from the NSX Manager via interaction with the message bus agent. UWA is installed on each ESXi host VXLAN Tunnel Endpoint VXLAN Tunnel Endpoints (VTEPs) are responsible for encapsulating VXLAN traffic as frames in UDP packets and for the corresponding de-encapsulation. VTEPs take the form of VMkernel ports with IP addresses and are used both to exchange packets with other VTEPs. This design uses a single VTEP per host Edge Services Gateway The primary function of the NSX for vsphere Edge services gateway is north-south communication, but it also offers support for Layer 2, Layer 3, perimeter firewall, load balancing and other services such as SSL-VPN and DHCP-relay. In this design, the edges also ensure east-west communication Distributed Firewall NSX for vsphere includes a distributed kernel-level firewall known as a distributed firewall. Security enforcement is done at the kernel and VM network adapter level. This enables firewall Page 52 of 132 Copyright IBM and VMware

53 rule enforcement in a highly scalable manner without creating bottlenecks on physical appliances. The distributed firewall has minimal CPU overhead and can perform at line rate. This design does not automatically implement distributed firewall. The cloud admin actor is able to enable this feature post implementation if required Logical Load Balancer The NSX for vsphere logical load balancer provides load balancing services up to Layer 7 (application), allowing distribution of traffic across multiple servers to achieve optimal resource utilization and availability. The logical load balancer is a service provided by the NSX Edge services gateway. This design implements load balancing for management virtual machines NSX for vsphere Physical Network Requirements VXLAN packets cannot be fragmented. Since VXLAN adds its own header information the MTU needs to be 1,600. SoftLayer has a limitation of 136 VXLAN addresses. As such, the VXLAN control plane in this design uses unicast mode to circumvent this limitation. The NSX Manager synchronizes with the same NTP server as the rest of the vsphere environment. This avoids time drift which can cause problems with authentication. The NSX Manager must be in sync with the vcenter Single Sign-On server NSX for vsphere Specifications The following table lists the components involved in the NSX for vsphere solution and the requirements for installing and running them. The compute and storage requirements have been taken into account when sizing resources to support the NSX for vsphere solution. Table 21 NSX Components Sizing VM vcpu Memory Storage Quantity per Deployment by Type NSX Manager 4 12 GB 60 GB 1 NSX Controller 4 4 GB 20 GB 3 NSX Edge services gateway - Quad Large NSX Edge services gateway X- Large 4 1 GB 512 MB GB 4.5 GB (+4 GB for SWAP) 2 The Quad Large model is utilized for high performance firewall. X-Large is utilized for both high performance load balancing and routing. Copyright IBM and VMware Page 53 of 132

54 Network Virtualization Conceptual Design The following diagram depicts the conceptual tenant architecture components and their relationship. Figure 18 Network Virtualization Conceptual Design The conceptual design has the following key components. External Networks. Connectivity to and from external networks is through a perimeter firewall. The main external network is the Internet. Perimeter Firewall. The logical firewall exists at the perimeter of the datacenter. Each tenant receives either a full instance or partition of an instance to filter external traffic. This is the primary access method for user data. Provider Logical Router (PLR). The PLR exists behind the perimeter firewall and handles north/south traffic that is entering and leaving a tenant. Tenant Edge services gateways. Edge services gateways that provide routing and firewalling capabilities. Internal Non-Tenant Networks. A single management network, which sits behind a perimeter firewall but not behind the PLR. Enables cloud admin to manage the cloud environment. Internal Tenant Networks. Connectivity for the main tenant workload. Page 54 of 132 Copyright IBM and VMware

55 Cluster Design for NSX for vsphere Management Stack In the management stack, the underlying hosts are prepared for NSX for vsphere. The management stack has these components: NSX Manager instances for both stacks (management stack and compute/edge stack), NSX Controller cluster for the management stack, NSX Edge service gateways for the management stack. Compute/Edge Stack In the compute/edge stack, the underlying hosts are prepared for NSX for vsphere. The compute/edge stack has these components: NSX Controller cluster for the compute stack All NSX Edge service gateways of the compute stack that are dedicated to handling the north/south traffic in the datacenter. A separate edge stack helps prevent VLAN sprawl because any external VLANs need only be trunked to the hosts in this cluster. Multiple compute clusters that run the tenant workloads and have the underlying hosts prepared for NSX for vsphere. Figure 19 Cluster Design for NSX for vsphere Copyright IBM and VMware Page 55 of 132

56 High Availability of NSX for vsphere Components The NSX Manager instances of both stacks run on the management cluster. vsphere HA protects the NSX Manager instances by ensuring that the VM is restarted on a different host in the event of primary host failure. The NSX Controller nodes of the management stack run on the management cluster and the NSX for vsphere Controller nodes of the compute stack run on the edge cluster. In both clusters, vsphere Distributed Resource Scheduler (DRS) rules ensure that NSX for vsphere Controller nodes do not run on the same host. The data plane remains active during outages in the management and control planes although the provisioning and modification of virtual networks is impaired until those planes become available again. The NSX Edge services gateways and distributed logical router control VMs of the compute stack are deployed on the edge cluster. The NSX Edge service gateways and distributed logical router control VMs of the management stack run on the management cluster. All NSX Edge components are deployed in NSX for vsphere HA pairs. NSX for vsphere HA provides better availability than vsphere HA. By default, the VMs fail over within 15 seconds versus a potential 5 minutes for a restart on another host under vsphere HA Logical Switch Control Plane Mode Design The control plane decouples NSX for vsphere from the physical network and handles the broadcast, unknown unicast, and multicast (BUM) traffic within the logical switches. The control plane is on top of the transport zone and is inherited by all logical switches that are created within it, although it is possible to override aspects of the control plane. The following options are available: Multicast Mode. The control plane uses multicast IP addresses on the physical network. Use multicast mode only when upgrading from existing VXLAN deployments. In this mode, PIM/IGMP must be configured on the physical network. Unicast Mode. The control plane is handled by the NSX Controllers and all replication occurs locally on the host. This mode does not require multicast IP addresses or physical network configuration. Hybrid Mode. This mode is an optimized version of the unicast mode where local traffic replication for the subnet is offloaded to the physical network. Hybrid mode requires IGMP snooping on the first-hop switch and access to an IGMP querier in each VTEP subnet. Hybrid mode does not require PIM. Due to SoftLayer constraints on the number of available multicast addresses, this design uses unicast mode Transport Zone Design A transport zone is used to define the scope of a VXLAN overlay network which can span one or more clusters within a vcenter Server domain. One or more transport zones can be configured in an NSX for vsphere solution. A transport zone is not meant to delineate a security boundary. The design implements a single transport zone per NSX instance. Each zone will be scaled to respect the limit of 100 DLRs per ESXi host. A single transport zone per NSX instance supports extending networks across resource stacks. Page 56 of 132 Copyright IBM and VMware

57 Routing Logical Design The routing logical design has to consider different levels of routing in the environment: North-south. The Provider Logical Router (PLR) handles the north/south traffic to and from a tenant East-west. Internal east-west routing at the layer beneath the PLR deals with the application workloads. NSX Edge services gateways is deployed for each management application to provide routing for the application. Virtual application network is attached directly to the edge services gateway. No distributed logical routers are deployed beneath Firewall Logical Design In this design, firewall functions are controlled at NSX Edge services gateway firewall. The Edge services gateway virtual machines act as the entry point to a tenant s virtual network space. NSX Edge services gateways are configured with the firewall enabled to support functionality that is required by virtual application networks. External access via load balancer virtual IPs when public services are provided by redundant management application VMs. External access is via a DNAT rule when service is provided by a single VM Load Balancer Design The Edge services gateway implements load balancing within NSX for vsphere; it has both a Layer 4 and a Layer 7 engine that offer different features, summarized in the following table. Table 22 Load Balancer Features Feature Layer 4 Engine Layer 7 Engine Protocols TCP TCP Load Balancing Method Round Robin Source IP Hash Least Connection HTTP HTTPS (SSL Pass-through) HTTPS (SSL Offload) Round Robin Source IP Hash Least Connection URI Health Checks TCP TCP Persistence (keeping client connections to the same back-end server) TCP: SourceIP HTTP (GET, OPTION, POST) HTTPS (GET, OPTION, POST) TCP: SourceIP, MSRDP HTTP: SourceIP, Cookie HTTPS: SourceIP, Cookie, ssl_session_id Connection Throttling No Client Side: Maximum concurrent connections, Maximum new connections per second Server Side: Maximum concurrent connections Copyright IBM and VMware Page 57 of 132

58 Feature Layer 4 Engine Layer 7 Engine High Availability Yes Yes Monitoring View VIP (Virtual IP), Pool and Server objects and stats via CLI and API View global stats for VIP sessions from the vsphere Web Client View VIP, Pool and Server objects and statistics by using CLI and API View global statistics about VIP sessions from the vsphere Web Client Layer 7 Manipulation No URL block, URL rewrite, content rewrite An NSX Load balancer is utilized to support the needs of management applications and for workloads Bridging Physical Workloads NSX for vsphere offers VXLAN to Layer 2 VLAN bridging capabilities with the data path contained entirely in the ESXi hypervisor. The bridge runs on the ESXi host where the distributed logical router control VM is located. Multiple bridges per distributed logical router are supported. This design places all virtual machines on VXLAN-backed networks. No bridging is implemented in the solution VPN Connectivity NSX for vsphere offers SSL VPN-Plus, IPSec, and Layer 2 VPN. SSL VPN-Plus. SSL VPN-Plus allows remote users to make secure client connections to private networks behind an NSX Edge service gateway and to access servers and applications. Users can be authenticated locally or via an external source such as Active Directory, LDAP, RADIUS, or RSA. The client can connect using either network or Web access modes. Network access requires a client side installation, for which a package can be created. Web access is entirely browser based, and requires no special hardware or software. IPSec VPN. NSX Edge service gateway supports site-to-site IPSec VPN between an NSX Edge instance and remote sites. It supports certificate-based authentication, preshared key mode and IP unicast traffic. Multiple subnets can be connected to behind the remote router over an IPsec tunnel but their network ranges must not overlap. The following constraint might affect potential use cases for the IPSec VPN, so is highlighted at the point of design. Note: Currently no dynamic routing protocols are supported between the NSX Edge service gateway and remote VPN routers. Layer 2 VPN. Layer 2 VPN allows creating a tunnel between sites. This VPN type requires the configuration of a Layer 2 VPN Server (the source) and Layer 2 VPN Client (the destination) with NSX Edge service gateway devices used for both of these roles. This design does not implement layer 2 VPN. Inter-region routing allows for management network reachability. SSL-VPN or IPSec VPN are available for use by users or the cloud admin actors for access to the user workload or cloud management services respectively Application Virtual Network Cloud management applications, such as VMware vrealize Automation, VMware vrealize Operations Manager, VMware vrealize Orchestrator, vrealize Log Insight, leverage a traditional Page 58 of 132 Copyright IBM and VMware

59 3-tier client/server architecture with a presentation tier (user interface), functional process logic tier, and data tier. This architecture requires a load balancer for presenting end-user facing services. This design implements each of these management applications as their own trust zone and isolates management applications from each other. Management applications are placed on isolated networks (VXLAN backed networks). The virtual application network is fronted by an NSX Edge services gateway to keep applications isolated. The use of load balancer interfaces is required for inbound access. The use of SNAT is required for outbound access. Direct access to virtual application networks is by connecting to Windows machines that are connected to the management networks directly. Unique addressing is required for all management applications. This approach to network virtualization service design improves security and mobility of the management applications, and reduces the integration effort with existing customer networks. From a software-defined networking design perspective, each management application is treated as a separate management tenant that requires one or more logical networks via VXLAN segment. The VXLAN segments are connected to the outside world by using a pair of NSX Edge services gateways. Figure 20 Virtual Application Network Components and Design The NSX Edge services gateway associated with a management application is connected via an uplink connection to a public accessible network and contains at least one IPv4 address on this network. As a result, this device offers the following capabilities. If the IPv4 range that is used for the application-internal network does not overlap with any other existing IPv4 range, the central DNS service can create DNS entries for nodes on this internal network. Split DNS is not necessary. Inbound access to services, such as Web UI, is supported by the load balancer capabilities of the NSX Edge services gateway. Application nodes can access the corporate network or the Internet via Source NAT (SNAT) on the NSX Edge services gateway or via the vsphere management network. Copyright IBM and VMware Page 59 of 132

60 Routed (static or dynamic) access to the vsphere management network is available for access to the vcenter Server instances Virtual Network Design Example The following figure presents the design for vrealize Automation that is implemented by the architecture. This design is utilized for other management applications and can be implemented for workloads deployed on compute cluster. The design is set up as follows: Figure 21 vra Virtual Network Design vrealize Automation is deployed onto a single Layer 2 segment, which is provided by a VXLAN virtual wire. Micro segmentation between NSX components is not required and therefore not used. The network used by vrealize Automation connects to external networks through NSX for vsphere. NSX Edge services gateways route traffic between management application virtual networks and the public network. Access to the isolated vsphere-mgmt network is available through the MgmtCentral- Edge services gateway that provides region-specific routing. Page 60 of 132 Copyright IBM and VMware

61 All Edge services gateways are connected over the networkexchange network that acts as a transit network and as an interface to exchange routing information between the Edge services gateways. To provide easy mobility of the network used by vrealize Automation during recovery in another region, this network uses an RFC 1918 isolated IPv4 subnet and uses Source NAT (SNAT) to access external networks such as the Internet. Services such as a Web GUI, which must be available to the users of vrealize Automation, are accessible via the NSX Edge load balancer on the IPv4 address residing on the external network. Each application must use a unique IPv4 range for the application internal network(s). The unique IPv4 range supports use of the central DNS service for creating DNS entries for nodes on this internal network. The following table shows an example of how a mapping from management applications to IPv4 subnets might look. Table 23 Management Applications IP Addressing Management Application Internal IPv4 Subnet vrealize Automation (includes vrealize Orchestrator) /24 vrealize Automation Proxy Agents / /24 vrealize Operations Manager /24 vrealize Operations Remote Collector / /24 vrealize Log Insight / /24 The management applications vrealize Automation, vrealize Operations Manager, and vrealize Log Insight divert from the above described setup slightly: vrealize Automation uses three network containers. o One container is for the main vrealize Automation application cluster that can be failed over by using Site Recovery Manager to a secondary region. o Two additional network containers - one in each region - hold vrealize Automation proxy agents. vrealize Operations Manager uses three network containers. o One container is for the main vrealize Operations Manager analytics cluster that can be failed over to a secondary region by using Site Recovery Manager. Two additional network containers - one in each region - are for connecting remote collectors. o vrealize Log Insight does not use Site Recovery Manager to fail over between regions. Instead, a dedicated instance of Log Insight is deployed in each region. To support this configuration, an additional cloud region is required. Copyright IBM and VMware Page 61 of 132

62 Routing and Region Connectivity Design This multi-region design is an extension of the single-region design that takes into account management component failover. The figure below is an example of how virtual application networks are built for vrealize Automation in the central cloud and cloud regions. The same virtual application network configuration is valid for all the other virtual application networks. Figure 22 Virtual Application Network Configuration in Central Cloud and Cloud Region Dynamic Routing Routing is handled by NSX Edge. Management network Edge services gateways, MgmtCentral-Edge and MgmtRegionA-Edge, work in conjunction with Edge services gateways that are configured to create virtual application network for each central cloud or cloud region management component. A dedicated network, networkexchange, facilitates the exchange of transit network traffic and the exchange of routing tables. All NSX Edge services gateways that are attached to the networkexchange network segment run OSPF to exchange routing information. OSPF route information exchange depends on the area definition, which is based on the local vsphere-mgmt network: Table 24 OSPF Area ID Network Region IP Addressing Area ID vsphere-mgmt Central Cloud /24 16 vsphere-mgmt Region A (Cloud Region) /24 17 Page 62 of 132 Copyright IBM and VMware

63 Virtual application network Edge services gateway receives an OSPF default route from the Management Edge services gateway which has access to the public network. Components directly attached to the vsphere-mgmt network use the Management Edge services gateway as the default gateway to reach all components in the environment. All VMs within virtual application networks can route to VMs in other virtual application networks. Base OSPF Area IDs on the second octet of the vsphere management network, and use MD5 encryption for authentication. NSX Edge appliance size is X-Large and is deployed in High Availability (HA) mode. A dedicated network for exchange of routing information and traffic between different gateways is configured. Access to public networks uses SNAT Region Connectivity and Dynamic Routing There is one Management Edge services gateway in each region. The role of the Management Edge services gateway is to provide connectivity between regions and to exchange region-specific routes with other connected Management Edge services gateways. The Management Edge services gateway is configured to exchange OSPF routing data with the entire region-specific virtual application network. Edge services gateways are attached to their respective networkexchange networks. These routes are consolidated and exchanged with all connected Management Edge services gateways using ibgp. All Management Edge services gateways belong to the same AS (Autonomous System) with regards to BGP routing. In this design, connectivity between two regions uses the direct backend connectivity available between sites Virtual Application Networks and Dynamic Routing Across Regions For the components that fail over to another region in the event of a disaster, route redistribution is disabled in NSX Edge. This prevents the same routes from being advertised in both regions for their virtual application networks. This makes it possible for the virtual application network Edge services gateway to participate in region-specific OSPF route exchanges without announcing its virtual application network route. To facilitate failover of the virtual application network from one region to the other for some components, this design creates a shadow virtual application network in the recovery region, an additional cloud region in this design. This shadow virtual application network is configured identically, with the same SNAT and DNAT rules and load balancer VIP as needed. The only difference is the IP addressing used on the networkexchange network and the optional public interface. Component virtual application networks can be moved between regions either for testing or during failover. When a component is moved the location of that component s virtual application network needs to be changed in the VPN configuration. The virtual application network Edge services gateway also needs to be reconfigured to either start or stop redistributing connected virtual application networks. The decision to start or stop redistribution depends on whether the virtual application network Edge services gateway is now the active Edge services gateway or the failed over Edge services gateway. 5.3 Infrastructure Management Infrastructure management is provided by VMware vcenter Server. The configuration of vcenter Server is described in the following sections vcenter Server Instances The vcenter Server design includes both the vcenter Server instances and the VMware Platform Services Controller instances. Copyright IBM and VMware Page 63 of 132

64 A VMware Platform Services Controller (PSC) groups a set of infrastructure services including vcenter Single Sign-On, License service, Lookup Service, and VMware Certificate Authority. Although the Platform Services controller and the associated vcenter Server system can be deployed on the same virtual machine (embedded Platform Services Controller), this design using a separate PSC and vcenter Server. This design includes two vcenter Server instances per central cloud or cloud region; one managing the management cluster, the other managing the compute and edge clusters. This separation of the management and compute stack not only provides additional security, but also supports scalability. The vcenter Server is deployed as the Linux-based vcenter Server Appliance which is functionally equivalent to the Windows based application, but easier to deploy and maintain. The two external Platform Services Controller (PSC) instances and two vcenter Server instances are both housed in the management cluster. The external PSC instances are used to enable replication of data between PSC instances. To achieve redundancy, the design joins the two Platform Services Controller instances to the same vcenter Single Sign-On domain, and connects each vcenter Server instance to their respective Platform Services Controller instance. Joining all Platform Services Controller instances into a single vcenter Single-Sign-On domain provides this design the capability to share authentication and license data across all components and regions. Additionally, the vcenter Servers in each region are configured to utilize Enhanced Linked Mode so that all inventories and actions are available from a each vcenter Server in the region. In terms of configuration, each vcenter Server system is configured with a static IP address and host name from the management private portable SoftLayer subnet. The IP address will have a valid (internal) DNS registration including reverse name resolution. The vcenter Server systems will also maintain network connections to the following components: All VMware vsphere Client and vsphere Web Client user interfaces. Systems running vcenter Server add-on modules, such as NSX or vum. Each ESXi host. Figure 23 vcenter Server and PSC Deployment Model Page 64 of 132 Copyright IBM and VMware

65 vcenter Server Resiliency Protecting the vcenter Server system is important because it is the central point of management and monitoring for the environment. In this design, the vcenter Servers and PSCs are protected via vsphere HA Size of vcenter Server Appliances The size of the vcenter Service appliances is described in the following sections vcenter Server Appliance for Management Cluster The number of hosts that are managed by the Management vcenter Server is less than 100 and the number of virtual machines to be managed is less than 1,000. As a result, this design includes a small vcenter Server configured with the following specifications and still allows for cloud admin s to include additional services within the cluster: Table 25 Specifications for Management vcenter Server Appliance vcpus Memory Disk Space Disk Type 4 16 GB 136 GB Thin Platform Services Controller for Management Cluster The Platform Services controller is deployed onto the management cluster with the following configuration: Table 26 Specifications for Platform Service Controller for Management Cluster vcpus Memory Disk Space Disk Type 2 2 GB 30 GB Thin Additionally, the PSC connects to the Active Directory server within this design for common authentication vcenter Server Appliance for Compute and Edge Clusters The Compute and Edge vcenter Server in this design manages up to 1,000 hosts or 10,000 virtual machines whichever comes first. As a result, this design includes a large vcenter Server configured with the following specifications: Table 27 Specifications for Compute and Edge vcenter Server Appliance vcpus Memory Disk Space Disk Type GB 295 GB Thin Platform Services Controller for Compute and Edge Clusters The Platform Services controller for vcenter managing the Compute and Edge clusters are deployed onto the management cluster and contain the following configuration: Table 28 Specifications for Platform Service Controller for Management Cluster vcpus Memory Disk Space Disk Type 2 2 GB 30 GB Thin Additionally, the PSC conencts to the Active Directory server within this design for common authentication. Copyright IBM and VMware Page 65 of 132

66 vcenter Database Design A vcenter Server Appliance can use either a built-in local PostgreSQL database or an external Oracle database. Both configurations support up to 1,000 hosts or 10,000 virtual machines. This design will utilize the built-in PostgreSQL to reduce both overhead and Microsoft or Oracle licensing costs. This also avoids problems with upgrades and support since the ability to attach to an external database is deprecated for vcenter Server Appliance in the next release vcenter Networking Design The vcenter Servers and Platform Services Controllers in each region will be placed onto the management network to access physical ESXi hosts within the region Cluster Configuration Each of the clusters will be configured per the following sections vsphere DRS This design will utilize vsphere Distributed Resource Scheduling (DRS) in each cluster to initially place and dynamically migrate virtual machines to achieve balanced compute clusters. The automation level is set to fully automated so that initial placement and migration recommendations are executed automatically by vsphere. Note that power management via the Distributed Power Management feature is not used in this design. Additionally, the migration threshold is set to partially aggressive so that vcenter will apply priority 1, 2, 3, and 4 recommendations to achieve at least a moderate improvement in the cluster s load balance Affinity Rules In this design, affinity and anti-affinity rules are used to ensure placement of virtual machines that are in a high availability cluster are not placed on the same ESXi host vsphere HA This design will use vsphere HA in each cluster to detect compute failures and recover virtual machines running within a cluster. The vsphere HA feature in this design is configured with both host monitoring and admission control enabled within the cluster. Additionally, each cluster will reserve 25% of cluster resources as spare capacity for the admission control policy. By default, the VM restart priority is set to medium and the host isolation response is set to leave powered on. Additionally, VM monitoring is disabled and the datastore heart-beating feature is configured to include any of the cluster datastores. In this design, the datastore is a VSAN-backed datastore vsphere vmotion vsphere vmotion is enabled in this design by the use of shared storage and network configuration vsphere FT vsphere FT is not utilized or available in this design. 5.4 Common Services Identity and Access Services Microsoft Active Directory (AD) is used as the identity provider (IDP) in this design. In a single region deployment of this design, two MS Windows 2012 R2 Active Directory server VMs are configured for a Page 66 of 132 Copyright IBM and VMware

67 single AD forest. Naming of the domain is derived from input data prior to design provisioning. Each server has the MS Windows DNS service installed and configured with all forward and reverse lookup zones as type AD integrated multi master enabled. Subsequent regions contain only one AD / DNS server in each region setup as a peer multi master within its own MS AD site configured within the AD Sites management console Forrest and subdomains This design uses Microsoft Active Directory for authentication and authorization to resources within the tornado.local domain. For a multi-region deployment, the design utilizes a domain and forest structure to store and manage Active Directory objects per region. Table 29 Requirements for Active Directory Service Requirement Active Directory configuration Active Directory users and groups Domain Instance Parent Active Directory Central child Active Directory Region-A child Active Directory Domain Name tornado.local central.tornado.local regiona.tornado.local Description Contains Domain Name System (DNS) server, time server, and universal groups that contain global groups from the child domains and are members of local groups in the child domains. Contains DNS records which replicate to all DNS servers in the forest. This child domain contains all design management users, and global and local groups. Contains DNS records which replicate to all DNS servers in the forest. This child domain contains all design management users, and global and local groups. - The default AD internal LDAP directory structure will be used for users and computer objects. ou=computers, DC=central, DC=tornado, DC=local ou=users, DC=central, DC=tornado, DC=local AD Functional level This design s AD functional level will be set at Windows 2008, which is the lowest level which is allowed by the other components in this design AD Trusts External trusts to customer AD environments are configured as required Authentication Within this design, authentication is provided by a mix of the following: direct AD integration SAML authentication via the VMware Platform Services Controller (PSC), backed by AD Copyright IBM and VMware Page 67 of 132

68 local user accounts within the application. The VMware Platform Services Controller is configured to use the AD servers within this design for its identity management directory. The following chart describes which authentication method is used for user / administrator access for each component Table 30 Authentication types used Component Authentication Context Authentication authority vra Administration Active Directory Tenant Active Directory vrops Administration / Roles Active Directory vcenter Server Administration / Roles Active Directory vro Administration / Roles Active Directory ESXi Administration Active Directory NSX Administration Active Directory vri Administration Active Directory vrb Administration Active Directory Tenant Active Directory vum Administration Active Directory NSX Administration Active Directory VPN Active Directory vdp Administration Active Directory AD VM server sizing Table 31 Server Sizing Attribute Number of CPUs 4 Memory Disk size Number of servers in central cloud Number of servers per additional cloud regions Specification 8 GB 80 GB Domain Name Services Domain name services (DNS) is a fundamental function of any multi-component network infrastructure design. DNS not only stores host name to IP address information but also service type information required by Microsoft Active Directory and other applications. Dynamic DNS capability is also required for MS AD deployments for the registration of the AD service record types as subzone creation. The AD integrated MS DNS server is utilized for this design. Page 68 of 132 Copyright IBM and VMware

69 DNS Design In a single region deployment of this design, four MS Windows 2012 R2 Active Directory server VMs are configured. Two for the root of the forest domain and 2 for the region subdomain. Each server has the MS Windows DNS service installed and configured with all forward and reverse lookup zones as type AD integrated multi master enabled. Subsequent regions contain only one AD / DNS server in each region setup as a peer multi master within its own MS AD site configured within the AD Sites management console. The following applies to all services and hosts within this design. All nodes must use static IP addresses. As a DNS best practice, the IP address, FQDNs and short names of all nodes must be forward and reverse resolvable. All nodes must be accessible from the vcenter Server instances and from the machine that hosts the vsphere Client (if vsphere Client is used instead of vsphere Web Client). All nodes in the vrealize Operations Manager analytics cluster must have unique host names. NTP source must be used for all cluster nodes DNS is an important component for the operation of this design. For a multi-region deployment, other regions must provide a root and child domains which contain separate DNS records DNS Configuration Requirements The following table shows an example of the domain naming that can be implemented by the design. The specific domain naming is determined by the client requirements. Table 32 Domain Naming Example Requirement Domain Instance Description DNS host entries tornado.local Resides in the tornado.local domain. central.tornado.local and regiona.tornado.local DNS servers reside in the central.tornado.local and regiona.tornado.local domains. Configure both DNS servers with the following settings: Dynamic updates for the domain set to Nonsecure and secure. Zone replication scope for the domain set to All DNS server in this forest. Create all hosts listed in the DNS Names documentation. Once properly configured, all nodes within this design are resolvable by FQDN. Copyright IBM and VMware Page 69 of 132

70 DNS Forwarding For name resolution outside of the design, the forest root DNS domain servers will have forwarders configured to the internal SoftLayer DNS servers if no client on premises DNS is available. If a client on premises DNS is available, then that becomes the DNS forwarding address or addresses. Table 33 SoftLayer DNS servers DNS hostname IP v4 address rs1.service.softlayer.com rs2.service.softlayer.com NTP Services Time synchronization is critical for the functionality of many of the services that comprise this design. For this purpose, Microsoft Windows Active Directory (AD) servers within the solution are configured as NTP sources for all services within the design with the exception of VMware Data Protection (vdp), vrealize Log Insight (vri) and the ESXi hosts. The Windows AD servers are configured to utilize the internal SoftLayer time server servertime.service.softlayer.com as the authoritative time source. The vdp application server, is configured to utilize VMware tools as its time source per the vdp documentation recommendation. The ESXi hosts and vri are configured to point to the SoftLayer NTP server. Table 34 Time sources Service / Application Active Directory servers vdp ESXi vrealize Log Insight All else Time Source servertime.service.softlayer.com VMware tools servertime.service.softlayer.com servertime.service.softlayer.com AD Servers SMTP Services SMTP services are required for outbound notifications and may be used for inbound communications to vrealize Automation. The following sections details the configuration of SMTP in this design SMTP outbound Within this design, notifications are sent using SMTP for the following products: vrealize Automation vrealize Business vrealize Operations vrealize Log Insight vrealize Orchestrator vcenter Server VMware Data Protection Page 70 of 132 Copyright IBM and VMware

71 An SMTP server service is configured on each windows AD server which is utilized as an SMTP relay for all outbound SMTP notifications. This service is configured to use the default SMTP port (25) where the connection to the customer s servers is over direct or VPN connection. As SoftLayer blocks port 25 outbound, a secure SSL port (465, 587), or custom port is configured on the SMTP server based on the destination provider, provided by the customer. This is only required where a direct connection or VPN from the customer site into the management network is not available SMTP inbound For vrealize Automation, an inbound server to handle inbound notifications, such as approval responses, is required to be configured. This is typically the customer server with accounts created for vra. Only one, global inbound server, which appears as the default for all tenants, is needed. The server provides accounts that are able to be customized for each user, providing separate accounts, usernames, and passwords. Each tenant can configure an override to change these settings. If service provider actors do not override these settings before enabling notifications, vrealize Automation uses the globally configured server. The connection to the customer system requires that this system be network reachable by the vra system Certificate Authority Services By default, vsphere 6.0 uses TLS/SSL certificates that are signed by VMCA (VMware Certificate Authority) residing on the VMware Platform Services Controller appliance (PSC). By default, these certificates are not trusted by end-user devices or browsers. It is a security best practice to replace at least user-facing certificates with certificates that are signed by a third-party or enterprise Certificate Authority (CA). Certificates for machine-to-machine communication can remain as VMCA-signed certificates. For this design, a Microsoft Active Directory Enterprise integrated certificate authority is used. A two tier MS certificate authority design is employed with 1 windows server as the offline root CA and an additional windows server as the online subordinate CA Offline root CA server VM As per Microsoft best practice, the root CA windows server does not participate in this design s AD domain. Setup the root CA per the following Microsoft information page: Note that while it is not recommended to connect this machine to a network, it will be network connected until the root CA is generated. Once generated, transfer the root certificate files to the subordinate CA server, remove the network access from the root CA server and shut down the OS. It is recommended that client s export this VM as an OVA to a secure location and remove it from the vcenter Server s inventory Online subordinate CA server VM Setup the subordinate CA per the following MS link: Note that this system will also be configured for handling certificate requests and distributing certificate revocation lists (CRL). The VMCA is configured with a certificate issued via the MS Active directory integrated CA such that all subsequent browser access to VMware services, are validated by any clients that a members of the designs AD domain or have imported the domains trusted root certificate. Copyright IBM and VMware Page 71 of 132

72 Table 35 Root CA and Subordinate CA sizing Attribute Specification OS Windows 2012 R2 Number of CPUs 2 Memory 4 GB Disk size 60 GB 5.5 Cloud Management Services The Cloud Management Services provide the management components for the cloud solution. This layer includes the Service Catalog, which houses the facilities to be deployed, Orchestration which provides the workflows to get the catalog items deployed, and the Self-Service Portal that empowers the users to take full advantage of the Software Defined Datacenter. vrealize Automation provides the Portal and the Catalog, and vrealize Orchestrator takes care of the process orchestration. The conceptual design of the vrealize Automation Cloud Management Services is illustrated in the following diagram. Key design components and their descriptions are also provided. Figure 24 vrealize Automation Conceptual Design The Cloud Management Services consist of the following elements and components: Users: Cloud administrators Tenant, group, fabric, infrastructure, service, and other administrators as defined by business policies and organizational structure. Page 72 of 132 Copyright IBM and VMware

73 Cloud (or tenant) users Provide direct access to virtual machine to perform operating systemlevel operations provided by vrealize Automation IaaS services. Cloud Management Portal: vrealize Automation portal, Admin access The default root tenant portal URL used to set up and administer tenants and global configuration options. vrealize Automation portal, Tenant access Refers to a subtenant and is accessed using a unique URL, with an appended tenant identifier. It is also possible for a tenant portal to refer to the default tenant portal in some configurations. In this case, the URLs are the same and the user interface is contextually controlled by the assigned RBAC permissions of that user. Tools and supporting infrastructure: VM Templates and Blueprints - These are the templates used in authoring the blueprints that tenants (users) use to provision their cloud workloads. Provisioning infrastructure - the following are the on-premises and off-premises resources which together form a hybrid cloud Virtual Supported hypervisors and associated management tools. Cloud Supported cloud providers and associated API interfaces. In the above diagram illustrating the conceptual design of the Cloud Management Platform, these resources are located in the Internal Virtual Resources and the External Cloud Resources components. The cloud management services deliver multi-platform and multi-vendor cloud services. The services available include the following items. Comprehensive and purpose-built capabilities that provide standardized resources to global customers in a short time span. Multi-platform and multi-vendor delivery methods that integrate with existing enterprise management systems. Central user-centric and business-aware governance for all physical, virtual, private, and public cloud services. Design that meets the customer and business needs and is extensible. Table 36 Cloud Management Services Components Component Services Provided vrealize Automation Identity Appliance vcenter Single Sign-On vrealize Automation virtual appliance vrealize Automation Portal Web/App Server vrealize Automation vpostgresql Database vrealize Automation Service Catalog vrealize Automation IaaS components vrealize Automation IaaS Web Server vrealize Automation IaaS Manager Services Copyright IBM and VMware Page 73 of 132

74 Component Services Provided Distributed execution components vrealize Automation Distributed Execution Managers: Orchestrator Workers Integration components vrealize Automation Agent machines Provisioning infrastructure vsphere environment vrealize Orchestrator environment Other supported physical, virtual, or cloud environments. Supporting infrastructure Microsoft SQL database environment Active Directory environment SMTP NTP DNS Cloud Management Physical Design This design uses NSX logical switches to abstract the vrealize Automation application and its supporting services. This abstraction allows the application to be hosted in any given region regardless of the underlying physical infrastructure such as network subnets, compute hardware, or storage types. This design hosts the vrealize Automation application and its supporting services in the central cloud. The same instance of the application manages both central cloud and any additional cloud regions. Page 74 of 132 Copyright IBM and VMware

75 Figure 25 vrealize Automation Design Overview for Central Cloud Copyright IBM and VMware Page 75 of 132

76 Figure 26 vrealize Automation Design Overview for Additional Cloud Regions The configuration of these elements is described in the following sections vrealize Identity Appliance vrealize Identity Appliance provides an infrastructure-independent failover capability. A single appliance is deployed in the solution. vsphere HA is used to ensure high availability for Identity Appliance. The appliance is configured with 1 vcpu and 2 GB of RAM vrealize Automation Appliance The vrealize Automation virtual appliance includes the Web portal and database services. The vrealize Automation portal allows self-service provisioning and management of cloud services, as well as authoring blueprints, administration, and governance. The vrealize Automation virtual appliance uses an embedded PostgreSQL database for catalog persistence and database replication. Page 76 of 132 Copyright IBM and VMware

77 The solution deploys two instances of the vrealize Automation appliance to achieve redundancy. The database is configured between two vrealize Automation appliances for high availability. Data is replicated between the embedded PostgreSQL database instances. Database instances are configured in an active/passive. In this configuration, manual failover between the two database instances is required. Each appliance is configured with 4 vcpu and 16 GB of RAM vrealize Automation IaaS Web Server vrealize Automation IaaS Web server provides a user interface within the vrealize Automation portal web site for the administration and consumption of IaaS components. Two vrealize Automation IaaS web servers are installed on virtual machines. Each virtual machine runs Microsoft Windows Server 2012 R2 and it performs Model Manager (Web) and IaaS Web functions. Each virtual machine is sized to 4 vcpu, 4 GB of RAM and 60 GB HDD vrealize Automation IaaS Manager Service and DEM Orchestrator Server The vrealize Automation IaaS Manager Service and Distributed Execution Management (DEM) server are at the core of the vrealize Automation IaaS platform. The vrealize Automation IaaS Manager Service and DEM server supports several functions. Manages the integration of vrealize Automation IaaS with external systems and databases. Provides multi-tenancy. Provides business logic to the DEMs. Manages business logic and execution policies. Maintains all workflows and their supporting constructs. A Distributed Execution Manager (DEM) runs the business logic of custom models, interacting with the database and with external databases and systems as required. DEMs also manage cloud and physical machines. The DEM Orchestrator monitors the status of the DEM workers. DEM worker manages the scheduled workflows by creating new workflow instances at the scheduled time and allows only one instance of a particular scheduled workflow to run at a given time. It also preprocesses workflows before execution. Preprocessing includes checking preconditions for workflows and creating the workflow's execution history. The vrealize Automation IaaS Manager Service and DEM server are separate servers, but are installed on the same virtual machine. Two virtual machines are deployed to run both IaaS Manager Service and DEM Orchestrator. The two servers share the same active/passive application model. Only one manager service can be active at a time. Each virtual machine runs Microsoft Windows Server 2012 R2. Each virtual machine is sized to 2 vcpu, 4 GB of RAM and 60 GB HDD vrealize Automation IaaS DEM Worker Server vrealize Automation IaaS DEM workers are responsible for the provisioning and deprovisioning tasks initiated by the vrealize Automation portal. DEM workers communicate with vrealize Automation endpoints. In this instance, the endpoint is vcenter Server. Each DEM Worker can process up to 15 concurrent workflows. Beyond this limit, workflows are queued for execution. The current design implements 2 DEM workers for a total of 30 concurrent workflows. DEM Workers are installed on two virtual machines running Microsoft Windows Server 2012 R2. Each virtual machine is sized to 4 vcpu, 8 GB of RAM and 60 GB HDD. Copyright IBM and VMware Page 77 of 132

78 vrealize Automation IaaS Proxy Agent The vrealize Automation IaaS Proxy Agent is a Windows program that caches and forwards information gathering from vcenter Server back to vrealize Automation. The IaaS Proxy Agent server provides the following functions. vrealize Automation IaaS Proxy Agent can interact with different types of hypervisors and public cloud services, such as Hyper-V and AWS. For this design, only the vsphere agent is used. vrealize Automation does not virtualize resources by itself, but works with vcenter Server to provision and manage the virtual machines. It uses vsphere agents to send commands to and collect data from vcenter Server. Two vrealize Automation vsphere Proxy Agent virtual machines are deployed in the current architecture. The virtual machines are deployed on a dedicated virtual network to decouple them from the main vrealize Automation infrastructure allowing. Each virtual machine runs Microsoft Windows Server 2012 R2. Each virtual machine is sized to 2 vcpu, 4 GB of RAM and 60 GB HDD Load Balancer Session persistence of a load balancer allows the same server to serve all requests after a session is established with that server. The session persistence is enabled on the load balancer to direct subsequent requests from each unique session to the same vrealize Automation server in the load balancer pool. The load balancer also handles failover for the vrealize Automation Server (Manager Service). Only one Manager Service is active at any one time. Manual failover of Manager Service is necessary. Session persistence is not enabled because it is not a required component for the Manager Service. The following tables describe load balancer implementation for vrealize Automation components: Table 37 Load Balancer Application Profile Server Role Type Enable SSL Pass-through vrealize Automation vpostgres vrealize Automation vrealize Automation IaaS Web vrealize Automation IaaS Manager vrealize Automation Orchestrator Persistence TCP n/a None n/a HTTPS (443) Enabled Source IP 120 HTTPS (443) Enabled Source IP 120 HTTPS (443) Enabled Source IP 120 HTTPS (443) Enabled Source IP 120 Expires in (Seconds) Page 78 of 132 Copyright IBM and VMware

79 Table 38 Load Balancer Service Monitoring Configuration Monitor vrealize Automation vpostgres vrealize Automation vrealize Automation IaaS Web vrealize Automation IaaS Manager vrealize Automation Orchestrator Interv Timeou Retries Type Method URL Receive al t TCP N/A N/A N/A HTTPS (443) HTTPS (443) HTTPS (443) HTTPS (443) GET /vcac/ser vices/api /status REGIST ERED GET N/A N/A GET /VMPS2 BasicHtt pbinding _VMPSP roxyage nt_policy GET /vco/api/ status REGIST ERED Table 39 Load Balancer Pool Specifications Server Role Algorithm Monitors Members Port Monitor Port vrealize Automation vpostgres Round Robin <vrealize Automation vpostgres vrealize Automation Postgres vrealize Automation vrealize Automation IaaS Web vrealize Automation IaaS Manager vrealize Automation Orchestrator Least Connection Least Connection Least Connection Least Connection monitor> <vrealize Automation monitor> <vrealize Automation IaaS Web monitor> <vrealize Automation IaaS Manager monitor> <vrealize Automation Orchestrator monitor> nodes vrealize Automation nodes IaaS web nodes IaaS Manager nodes vrealize Orchestrator nodes Table 40 Virtual Server Characteristics Protocol Port Default Pool Application Profile TCP 5432 vrealize Automation vpostgres Pool vrealize Automation vpostgres Profile HTTPS 443 vrealize Automation Pool vrealize Automation Profile HTTPS 443 vrealize Automation IaaS Web Pool vrealize Automation IaaS Web Profile Copyright IBM and VMware Page 79 of 132

80 Protocol Port Default Pool Application Profile HTTPS 443 vrealize Automation IaaS Manager Pool vrealize Automation IaaS Manager Profile HTTPS 8281 vrealize Automation Orchestrator Pool vrealize Automation Orchestrator Profile vrealize Automation Supporting Infrastructure A number of supporting elements are required for vrealize Automation. The following sections describe their configuration Microsoft SQL Server Database vrealize Automation uses a Microsoft SQL Server database to maintain the vrealize Automation IaaS elements and the policies. The database also maintains information about the machines it manages. For simple failover of the entire vrealize Automation instance from one site to another, the Microsoft SQL server is running in a virtual machine inside the vrealize Automation virtual network. The virtual machine is running Microsoft Windows Server 2012 R2 and configured with 8 vcpu, 16 GB of RAM, 80 GB HDD. The Microsoft SQL Server Database version is SQL Server Postgres SQL Server Database The vrealize Automation appliance uses a PostgreSQL database server to maintain the vrealize Automation portal elements and services, and the information about the catalog items it manages. Embedded PostgreSQL within each virtual appliance is utilized Notifications System administrators configure default settings for both the outbound and inbound s servers used to send system notifications. Current solution implements outbound SMTP server only. The automation creates a global outbound server to process outbound notifications. The server appears as the default for all tenants vrealize Automation Cloud Tenant Design A tenant is an organizational unit within a vrealize Automation deployment, and can represent a business unit within an enterprise, or a company that subscribes to cloud services from a service provider. Each tenant has its own dedicated configuration, although some system-level configuration is shared across tenants Single-Tenant and Multi-Tenant Deployments vrealize Automation supports deployments with a single tenant or multiple tenants. System-wide configuration is always performed using the default tenant, and can then be applied to one or more tenants. System-wide configuration specifies defaults for branding and notification providers. Infrastructure configuration, including the infrastructure sources that are available for provisioning, can be configured in any tenant and is shared among all tenants. The infrastructure resources, such as cloud or virtual compute resources or physical machines, can be divided into fabric groups managed by fabric administrators. The resources in each fabric group can be allocated to business groups within each tenant by using reservations. Single-Tenant Deployment In a single-tenant deployment, all configuration occurs in the default tenant. Service provider actors can manage users and groups, and configure tenant-specific branding, notifications, business policies, and catalog offerings. All users Page 80 of 132 Copyright IBM and VMware

81 log in to the vrealize Automation console at the same URL, but the features available to them are determined by their roles. Multi-Tenant Deployment In a multi-tenant deployment, the system administrator creates new tenants for each organization that uses the same vrealize Automation instance. Tenant users log in to the vrealize Automation console at a URL specific to their tenant. Tenant-level configuration is segregated from other tenants and from the default tenant, although users with system-wide roles can view and manage configuration across multiple tenants Tenant Design This design deploys a single tenant containing two business groups. The first business group is designated for production workloads provisioning. The second business group is designated for development workloads. Service provider actors manage users and groups, configure tenant-specific branding, notifications, business policies, and catalog offerings. All users log in to the vrealize Automation console at the same URL, but the features available to them are determined by their roles. The following diagram illustrates the single region tenant design. Figure 27 Tenant Design for Single Region Copyright IBM and VMware Page 81 of 132

82 The following diagram illustrates the dual-region tenant design. Figure 28 Tenant Design for Two Regions The tenant has two business groups. A separate fabric group for each region is created. Each business group can consume resources in both regions. Access to the default tenant is allowed only by the system administrator and for the purposes of managing tenants and modifying systemwide configurations. The solution automatically configures vrealize Automation based on the requested deployment type single region or dual region Service Design The service catalog provides a common interface for consumers of IT services to use to request and manage the services and resources they need. A service provider actor or service architect can specify information about the service catalog, such as the service hours, support team, and change window. The solution implements a service catalog provides the following services: Central Cloud. Service catalog that is dedicated to the central cloud. Additional Cloud Region. Service catalog that is dedicated to an additional cloud region. The solution is preinstalled with several catalog items. For a single site configuration, only the central cloud service catalog is implemented. Page 82 of 132 Copyright IBM and VMware

83 Catalog Items Users can browse the service catalog for catalog items they are entitled to request. Several generic users will be automatically created and entitled to all items in the catalog. The users can be disabled at a later stage or their permissions modified as appropriate. For some catalog items, a request results in the provisioning of an item that the user can manage. For example, the user can request a virtual machine with Windows 2012 preinstalled, and then manage that virtual machine after it has been provisioned. The service provider actor defines new catalog items and publish them to the service catalog. The service provider actor can then manage the presentation of catalog items to the consumer and entitle new items to consumers. To make the catalog item available to users, a service provider actor must entitle the item to the users and groups who should have access to it. A catalog item is defined in a blueprint, which provides a complete specification of the resource to be provisioned and the process to initiate when the item is requested. It also defines the options available to a requester of the item, such as virtual machine specifications or lease duration, or any additional information that the requester is prompted to provide when submitting the request. The blueprint also specifies custom properties that are applied to the requested resource Machine Blueprints A machine blueprint is the complete specification for a virtual machine. A machine blueprint determines the machine's attributes, how it is provisioned, and its policy and management settings. Machine blueprints are published as catalog items in the service catalog. Machine blueprints can be specific to a business group or shared among groups within a tenant. In this design the preloaded machine blueprints are shared among business groups. Service provider actors create shared blueprints that can be entitled to users in any business group within the tenant. Business group managers create group blueprints that can only be entitled to users within a specific business group. A business group manager cannot modify or delete shared blueprints. Service provider actors cannot view or modify group blueprints unless they also have the business group manager role for the appropriate group. If a service provider actor sets a shared blueprint's properties so that it can be copied, the business group manager can also copy the shared blueprint for use as a starting point to create a new group blueprint Blueprint Design The following sections provide details of each service definition that has been included as part of the current phase of cloud platform deployment. Table 41 Base Windows Server Blueprint Service Name Provisioning Method Entitlement Approval Process Operating System and Version Details Configuration Lease and Archival Details Description When users select this blueprint, vrealize Automation clones a vsphere virtual machine template with preconfigured vcenter customizations. Both Production and Development business group members. No approval (pre-approval assumed based on approved access to platform) Windows Server 2012 R2 Disk: Single disk drive Lease: Production Blueprints: No expiration date Copyright IBM and VMware Page 83 of 132

84 Service Name Pre- and Post- Deployment Requirements Description Development Blueprints: Minimum 30 days Maximum 270 days Archive: 15 days sent to manager confirming service request (include description details) Table 42 Base Windows Blueprint Sizing vcpu Memory (GB) Storage (GB) Table 43 Base Linux Server Blueprint Service Name Provisioning Method Entitlement Approval Process Operating System and Version Details Configuration Lease and Archival Details Description When users select this blueprint, vrealize Automation clones a vsphere virtual machine template with preconfigured vcenter customizations. Both Production and Development business group members No approval (pre-approval assumed based on approved access to platform) Red Hat Enterprise Server 6 Disk: Single disk drive Lease: Production Blueprints: No expiration date Development Blueprints: Minimum 30 days Maximum 270 days Pre- and Post- Deployment Requirements Archive: 15 days sent to manager confirming service request (include description details) Table 44 Base Linux Blueprint Sizing vcpu Memory (GB) Storage (GB) Branding The solution branding is preconfigured. The cloud admin can change the appearance of the vrealize Automation console to meet site-specific branding guidelines by changing the logo, the background color, or information in the header and footer. Page 84 of 132 Copyright IBM and VMware

85 5.5.4 vrealize Automation vsphere Integration Design The following terms apply to vrealize Automation integrated with vsphere. These terms and their meaning may vary from the way they are used when referring only to vsphere. Table 45 vrealize Integration with vsphere Element vsphere (vcenter Server) endpoint Compute resource Fabric groups Fabric administrators Compute reservation Storage reservation Business groups Reservation policy Build profile Description Provides information required by vrealize Automation IaaS to access vsphere compute resources. It requires the appropriate permissions for the vsphere proxy agent to manage the vcenter Server instance. Virtual object within vrealize Automation that represents a vcenter Server cluster or resource pool, and datastores or datastore clusters. vrealize Automation provisions the virtual machines requested by business group members on the compute resource. Note: Compute resources are CPU, memory, storage and networks. Datastores and datastore clusters are part of the overall storage resources. vrealize Automation IaaS organizes compute resources into fabric groups. Fabric administrators manage compute resources, which are organized into fabric groups. A share of compute resources (vsphere cluster, resource pool, datastores, or datastore clusters), such as CPU and memory reserved for use by a particular business group for provisioning virtual machines. Note: vrealize Automation uses the term reservation to define resources (be they memory, storage or networks) in a cluster. This is different than the use of reservation in vcenter Server, where a share is a percentage of total resources, and reservation is a fixed amount. Similar to compute reservation (see above), but pertaining only to a share of the available storage resources. In this context, a storage reservation in terms of gigabytes is specified from an existing LUN or Datastore. A collection of virtual machine consumers, usually corresponding to an organization's business units or departments. Only users in the business group can request virtual machines. vrealize Automation IaaS determines its reservation (also called virtual reservation) from which a particular virtual machine is provisioned. The reservation policy is a logical label or a pointer to the original reservation. Each virtual reservation can be added to one reservation policy. A set of user defined properties a user is able to apply to a virtual machine when it is provisioned. For example, the operating system used in a blueprint, or the available networks to use for connectivity at the time of provisioning the virtual machine. Build profile properties determine the specification of the virtual machine, the manner in which it is provisioned, operations to perform after it is provisioned, or management information maintained within vrealize Automation. Copyright IBM and VMware Page 85 of 132

86 Element Blueprint Description The complete specification for a virtual machine, determining the machine attributes, the manner in which it is provisioned, and its policy and management settings. Blueprint allows the users of a business group to create virtual machines on a virtual reservation (compute resource) based on the reservation policy, and using platform and cloning types. It also lets a user specify or add machine resources and build profiles. Page 86 of 132 Copyright IBM and VMware

87 The following figure shows the logical design constructs discussed in the previous section as they apply to the deployment of vrealize Automation integrated with vsphere in a single region design Figure 29 vrealize Automation Integration with vsphere Endpoint Central Cloud Copyright IBM and VMware Page 87 of 132

88 The following figure shows the logical design constructs discussed in the previous section as they apply to the deployment of vrealize Automation integrated with vsphere in a dual region design Figure 30 vrealize Automation Integration with vsphere Endpoint Central Cloud and a Cloud Region (Region A) The solution automatically implements the design presented in the picture above for a dual site configuration. When single site deployment is requested, the automation deploys only the central cloud configuration. Page 88 of 132 Copyright IBM and VMware

89 5.5.5 Infrastructure Source Endpoints An infrastructure source endpoint is a connection to the infrastructure that provides a set (or multiple sets) of resources, which can then be made available by IaaS administrators for consumption by users. vrealize Automation IaaS regularly collects information about known endpoint resources and the virtual resources provisioned therein. Endpoint resources are referred to as compute resources (or as compute pods the terms are often used interchangeably). Infrastructure data is collected through proxy agents that manage and communicate with the endpoint resources. This information about the compute resources on each infrastructure endpoint and the machines provisioned on each computer resource is collected at regular intervals. During solution deployment the proxy agents and define their associated endpoints are configured automatically Virtualization Compute Resources A virtualization compute resource is a vrealize Automation object that represents an ESXi host or a cluster of ESXi hosts (vsphere cluster). When a group member requests a virtual machine, the virtual machine is provisioned on these compute resources. vrealize Automation regularly collects information about known compute resources and the virtual machines provisioned on them through the proxy agents. Each region has one compute cluster. The compute cluster is selected automatically during deployment Fabric Groups A fabric group is a logical container of several compute resources, and can be managed by fabric administrators. A fabric group for each region is created and it includes all the compute resources and edge resources in that region Business Groups A Business group is a collection of machine consumers (users), often corresponding to a line of business, department, or other organizational unit. To request machines, a vrealize Automation user must belong to at least one Business group. Each group has access to a set of local blueprints used to request machines. Business groups have the following characteristics. A group must have at least one business group manager, who maintains blueprints for the group and approves machine requests. Groups can contain support users, who can request and manage machines on behalf of other group members. A vrealize Automation user can be a member of more than one Business group, and can have different roles in each group. Two business groups are created, one for production users and one for the development users Reservations A reservation is a share of one compute resource's available memory, CPU and storage reserved for use by a particular fabric group. Each reservation is for one fabric group only but the relationship is many-to-many. A fabric group might have multiple reservations on one compute resource or reservations on multiple compute resources or both. The solution implements only one fabric group per region. Each resource cluster has two reservations, one for production and one for development, allowing both production and development workloads to be provisioned. An edge reservation in each region is created and allows NSX to deploy edge services gateways on demand and place them on the edge cluster. Copyright IBM and VMware Page 89 of 132

90 Reservation Policies Each virtual reservation is added to one reservation policy. The reservation from which a particular virtual machine is provisioned is determined by vrealize Automation based on the reservation policy specified in the blueprint (if any), the priorities and current usage of the fabric group's reservations, and other custom properties. Two reservation policies are configured in each region, one for production and the other for development. One edge reservation in each region is created for placement of the edge service gateway Template Synchronization In case of a single region, no template synchronization is made. Dual-region deployment allows provisioning workloads across regions from the same portal using the same single-machine blueprints. vsphere Content Library is the synchronization mechanism for templates across regions this design uses. Figure 31 Template Synchronization Process Orchestration VMware vrealize Orchestrator is a development and process automation and orchestration platform that provides a library of extensible workflows to allow a cloud admin to create and run automated, configurable processes to manage the VMware vsphere infrastructure as well as other VMware and thirdparty technologies Directory Services vrealize Orchestrator instances will use Active Directory LDAP authentication. The only configuration supported for multi-domain Active Directory is domain tree. Forest and external trusts are not supported for process orchestration. Multiple domains that have two-way trust, but are not in the same tree, are not supported and do not work with vrealize Orchestrator Network Ports vrealize Orchestrator uses specific network ports to communicate with other systems. The ports are configured with a default value, which is set by the automation at build time. It is recommended that these values remain unchanged to ensure supportability of the system in the future. Firewall ports within the solution will be opened to ensure communication to the components and intra components. Firewalls not deployed by the solution need to be configured appropriately. Page 90 of 132 Copyright IBM and VMware

VMware vsphere-6.0 Administration Training

VMware vsphere-6.0 Administration Training VMware vsphere-6.0 Administration Training Course Course Duration : 20 Days Class Duration : 3 hours per day (Including LAB Practical) Classroom Fee = 20,000 INR Online / Fast-Track Fee = 25,000 INR Fast

More information

VMware NSX @SoftLayer!!

VMware NSX @SoftLayer!! A VMware@SoftLayer CookBook v1.1 April 30, 2014 VMware NSX @SoftLayer Author(s) & Contributor(s) (IBM) Shane B. Mcelligott Dani Roisman (VMware) Merlin Glynn, mglynn@vmware.com Chris Wall Geoff Wing Marcos

More information

Expert Reference Series of White Papers. VMware vsphere Distributed Switches

Expert Reference Series of White Papers. VMware vsphere Distributed Switches Expert Reference Series of White Papers VMware vsphere Distributed Switches info@globalknowledge.net www.globalknowledge.net VMware vsphere Distributed Switches Rebecca Fitzhugh, VCAP-DCA, VCAP-DCD, VCAP-CIA,

More information

vsphere Private Cloud RAZR s Edge Virtualization and Private Cloud Administration

vsphere Private Cloud RAZR s Edge Virtualization and Private Cloud Administration Course Details Level: 1 Course: V6PCRE Duration: 5 Days Language: English Delivery Methods Instructor Led Training Instructor Led Online Training Participants: Virtualization and Cloud Administrators,

More information

VMware vsphere: Install, Configure, Manage [V5.0]

VMware vsphere: Install, Configure, Manage [V5.0] VMware vsphere: Install, Configure, Manage [V5.0] Gain hands-on experience using VMware ESXi 5.0 and vcenter Server 5.0. In this hands-on, VMware -authorized course based on ESXi 5.0 and vcenter Server

More information

NSX Installation and Upgrade Guide

NSX Installation and Upgrade Guide NSX 6.0 for vsphere This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of

More information

Vmware VSphere 6.0 Private Cloud Administration

Vmware VSphere 6.0 Private Cloud Administration To register or for more information call our office (208) 898-9036 or email register@leapfoxlearning.com Vmware VSphere 6.0 Private Cloud Administration Class Duration 5 Days Introduction This fast paced,

More information

VMware vsphere 4.1 with ESXi and vcenter

VMware vsphere 4.1 with ESXi and vcenter VMware vsphere 4.1 with ESXi and vcenter This powerful 5-day class is an intense introduction to virtualization using VMware s vsphere 4.1 including VMware ESX 4.1 and vcenter. Assuming no prior virtualization

More information

On-Demand Infrastructure with Secure Networks REFERENCE ARCHITECTURE

On-Demand Infrastructure with Secure Networks REFERENCE ARCHITECTURE REFERENCE ARCHITECTURE Table of Contents Executive Summary.... 3 Audience.... 3 Overview.... 3 What Is an On-Demand Infrastructure?.... 4 Architecture Overview.... 5 Cluster Overview.... 8 Management Cluster...

More information

Table of Contents. vsphere 4 Suite 24. Chapter Format and Conventions 10. Why You Need Virtualization 15 Types. Why vsphere. Onward, Through the Fog!

Table of Contents. vsphere 4 Suite 24. Chapter Format and Conventions 10. Why You Need Virtualization 15 Types. Why vsphere. Onward, Through the Fog! Table of Contents Introduction 1 About the VMware VCP Program 1 About the VCP Exam 2 Exam Topics 3 The Ideal VCP Candidate 7 How to Prepare for the Exam 9 How to Use This Book and CD 10 Chapter Format

More information

VMware. NSX Network Virtualization Design Guide

VMware. NSX Network Virtualization Design Guide VMware NSX Network Virtualization Design Guide Table of Contents Intended Audience... 3 Overview... 3 Components of the VMware Network Virtualization Solution... 4 Data Plane... 4 Control Plane... 5 Management

More information

Nutanix Tech Note. VMware vsphere Networking on Nutanix

Nutanix Tech Note. VMware vsphere Networking on Nutanix Nutanix Tech Note VMware vsphere Networking on Nutanix Nutanix Virtual Computing Platform is engineered from the ground up for virtualization and cloud environments. This Tech Note describes vsphere networking

More information

Cross-vCenter NSX Installation Guide

Cross-vCenter NSX Installation Guide NSX 6.2 for vsphere This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of

More information

VMware vsphere 5.1 Advanced Administration

VMware vsphere 5.1 Advanced Administration Course ID VMW200 VMware vsphere 5.1 Advanced Administration Course Description This powerful 5-day 10hr/day class is an intensive introduction to VMware vsphere 5.0 including VMware ESX 5.0 and vcenter.

More information

VMware vsphere 5.0 Boot Camp

VMware vsphere 5.0 Boot Camp VMware vsphere 5.0 Boot Camp This powerful 5-day 10hr/day class is an intensive introduction to VMware vsphere 5.0 including VMware ESX 5.0 and vcenter. Assuming no prior virtualization experience, this

More information

vcloud Suite Architecture Overview and Use Cases

vcloud Suite Architecture Overview and Use Cases vcloud Suite Architecture Overview and Use Cases vcloud Suite 5.8 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new

More information

vsphere Networking vsphere 6.0 ESXi 6.0 vcenter Server 6.0 EN-001391-01

vsphere Networking vsphere 6.0 ESXi 6.0 vcenter Server 6.0 EN-001391-01 vsphere 6.0 ESXi 6.0 vcenter Server 6.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more

More information

vsphere Networking ESXi 5.0 vcenter Server 5.0 EN-000599-01

vsphere Networking ESXi 5.0 vcenter Server 5.0 EN-000599-01 ESXi 5.0 vcenter Server 5.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions

More information

Expert Reference Series of White Papers. vcloud Director 5.1 Networking Concepts

Expert Reference Series of White Papers. vcloud Director 5.1 Networking Concepts Expert Reference Series of White Papers vcloud Director 5.1 Networking Concepts 1-800-COURSES www.globalknowledge.com vcloud Director 5.1 Networking Concepts Rebecca Fitzhugh, VMware Certified Instructor

More information

VMWARE VSPHERE 5.0 WITH ESXI AND VCENTER

VMWARE VSPHERE 5.0 WITH ESXI AND VCENTER VMWARE VSPHERE 5.0 WITH ESXI AND VCENTER CORPORATE COLLEGE SEMINAR SERIES Date: April 15-19 Presented by: Lone Star Corporate College Format: Location: Classroom instruction 8 a.m.-5 p.m. (five-day session)

More information

White Paper. Juniper Networks. Enabling Businesses to Deploy Virtualized Data Center Environments. Copyright 2013, Juniper Networks, Inc.

White Paper. Juniper Networks. Enabling Businesses to Deploy Virtualized Data Center Environments. Copyright 2013, Juniper Networks, Inc. White Paper Juniper Networks Solutions for VMware NSX Enabling Businesses to Deploy Virtualized Data Center Environments Copyright 2013, Juniper Networks, Inc. 1 Table of Contents Executive Summary...3

More information

VMware Virtual SAN Network Design Guide TECHNICAL WHITE PAPER

VMware Virtual SAN Network Design Guide TECHNICAL WHITE PAPER TECHNICAL WHITE PAPER Table of Contents Intended Audience.... 3 Overview.... 3 Virtual SAN Network... 3 Physical Network Infrastructure... 4 Data Center Network... 4 Host Network Adapter.... 5 Virtual

More information

vsphere Networking vsphere 5.5 ESXi 5.5 vcenter Server 5.5 EN-001074-02

vsphere Networking vsphere 5.5 ESXi 5.5 vcenter Server 5.5 EN-001074-02 vsphere 5.5 ESXi 5.5 vcenter Server 5.5 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more

More information

CompTIA Cloud+ 9318; 5 Days, Instructor-led

CompTIA Cloud+ 9318; 5 Days, Instructor-led CompTIA Cloud+ 9318; 5 Days, Instructor-led Course Description The CompTIA Cloud+ certification validates the knowledge and best practices required of IT practitioners working in cloud computing environments,

More information

Software Defined Environments

Software Defined Environments November 2015 Software Defined Environments 2015 Cloud Lecture, University of Stuttgart Jochen Breh, Director Architecture & Consulting Cognizant Global Technology Office Agenda Introduction New Requirements

More information

Khóa học dành cho các kỹ sư hệ thống, quản trị hệ thống, kỹ sư vận hành cho các hệ thống ảo hóa ESXi, ESX và vcenter Server

Khóa học dành cho các kỹ sư hệ thống, quản trị hệ thống, kỹ sư vận hành cho các hệ thống ảo hóa ESXi, ESX và vcenter Server 1. Mục tiêu khóa học. Khóa học sẽ tập trung vào việc cài đặt, cấu hình và quản trị VMware vsphere 5.1. Khóa học xây dựng trên nền VMware ESXi 5.1 và VMware vcenter Server 5.1. 2. Đối tượng. Khóa học dành

More information

VXLAN: Scaling Data Center Capacity. White Paper

VXLAN: Scaling Data Center Capacity. White Paper VXLAN: Scaling Data Center Capacity White Paper Virtual Extensible LAN (VXLAN) Overview This document provides an overview of how VXLAN works. It also provides criteria to help determine when and where

More information

CompTIA Cloud+ Course Content. Length: 5 Days. Who Should Attend:

CompTIA Cloud+ Course Content. Length: 5 Days. Who Should Attend: CompTIA Cloud+ Length: 5 Days Who Should Attend: Project manager, cloud computing services Cloud engineer Manager, data center SAN Business analyst, cloud computing Summary: The CompTIA Cloud+ certification

More information

Implementing and Troubleshooting the Cisco Cloud Infrastructure **Part of CCNP Cloud Certification Track**

Implementing and Troubleshooting the Cisco Cloud Infrastructure **Part of CCNP Cloud Certification Track** Course: Duration: Price: $ 4,295.00 Learning Credits: 43 Certification: Implementing and Troubleshooting the Cisco Cloud Infrastructure Implementing and Troubleshooting the Cisco Cloud Infrastructure**Part

More information

VMware vcloud Director for Service Providers

VMware vcloud Director for Service Providers Architecture Overview TECHNICAL WHITE PAPER Table of Contents Scope of Document....3 About VMware vcloud Director....3 Platform for Infrastructure Cloud...3 Architecture Overview....3 Constructs of vcloud

More information

How Network Virtualization can improve your Data Center Security

How Network Virtualization can improve your Data Center Security How Network Virtualization can improve your Data Center Security Gilles Chekroun SDDC, NSX Team EMEA gchekroun@vmware.com 2014 VMware Inc. All rights reserved. Security IT spending Security spending is

More information

NSX TM for vsphere with Arista CloudVision

NSX TM for vsphere with Arista CloudVision ARISTA DESIGN GUIDE NSX TM for vsphere with Arista CloudVision Version 1.0 August 2015 ARISTA DESIGN GUIDE NSX FOR VSPHERE WITH ARISTA CLOUDVISION Table of Contents 1 Executive Summary... 4 2 Extending

More information

VMware vsphere: Fast Track [V5.0]

VMware vsphere: Fast Track [V5.0] VMware vsphere: Fast Track [V5.0] Experience the ultimate in vsphere 5 skills-building and VCP exam-preparation training. In this intensive, extended-hours course, you will focus on installing, configuring,

More information

Creating a VMware Software-Defined Data Center REFERENCE ARCHITECTURE VERSION 1.5

Creating a VMware Software-Defined Data Center REFERENCE ARCHITECTURE VERSION 1.5 Software-Defined Data Center REFERENCE ARCHITECTURE VERSION 1.5 Table of Contents Executive Summary....4 Audience....4 Overview....4 VMware Software Components....6 Architectural Overview... 7 Cluster...

More information

Remote PC Guide Series - Volume 1

Remote PC Guide Series - Volume 1 Introduction and Planning for Remote PC Implementation with NETLAB+ Document Version: 2016-02-01 What is a remote PC and how does it work with NETLAB+? This educational guide will introduce the concepts

More information

JOB ORIENTED VMWARE TRAINING INSTITUTE IN CHENNAI

JOB ORIENTED VMWARE TRAINING INSTITUTE IN CHENNAI JOB ORIENTED VMWARE TRAINING INSTITUTE IN CHENNAI Job oriented VMWARE training is offered by Peridot Systems in Chennai. Training in our institute gives you strong foundation on cloud computing by incrementing

More information

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Table of Contents Introduction.... 3 vsphere Architectural Overview... 4 SAN Backup

More information

HAWAII TECH TALK SDN. Paul Deakin Field Systems Engineer

HAWAII TECH TALK SDN. Paul Deakin Field Systems Engineer HAWAII TECH TALK SDN Paul Deakin Field Systems Engineer SDN What Is It? SDN stand for Software Defined Networking SDN is a fancy term for: Using a controller to tell switches where to send packets SDN

More information

VMware@SoftLayer Cookbook Disaster Recovery (DR)

VMware@SoftLayer Cookbook Disaster Recovery (DR) VMware@SoftLayer Cookbook Disaster Recovery (DR) IBM Global Technology Services: Khoa Huynh (khoa@us.ibm.com) Daniel De Araujo (ddearaujo@us.ibm.com) Bob Kellenberger (kellenbe@us.ibm.com) VMware: Merlin

More information

(R)Evolution im Software Defined Datacenter Hyper-Converged Infrastructure

(R)Evolution im Software Defined Datacenter Hyper-Converged Infrastructure (R)Evolution im Software Defined Datacenter Hyper-Converged Infrastructure David Kernahan Senior Systems Engineer VMware Switzerland GmbH 2014 VMware Inc. All rights reserved. Agenda 1 VMware Strategy

More information

VMUG - vcloud Air Deep Dive. 2014 VMware Inc. All rights reserved.

VMUG - vcloud Air Deep Dive. 2014 VMware Inc. All rights reserved. VMUG - vcloud Air Deep Dive 2014 VMware Inc. All rights reserved. Agenda 1 Overview of vcloud Air 2 Advanced Networking Capabilities 3 Use Cases 4 Overview of Disaster Recovery Service 5 Questions 2 VMware

More information

Setup for Failover Clustering and Microsoft Cluster Service

Setup for Failover Clustering and Microsoft Cluster Service Setup for Failover Clustering and Microsoft Cluster Service Update 1 ESX 4.0 ESXi 4.0 vcenter Server 4.0 This document supports the version of each product listed and supports all subsequent versions until

More information

NSX Installation Guide

NSX Installation Guide NSX 6.2 for vsphere This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of

More information

Virtual SAN Design and Deployment Guide

Virtual SAN Design and Deployment Guide Virtual SAN Design and Deployment Guide TECHNICAL MARKETING DOCUMENTATION VERSION 1.3 - November 2014 Copyright 2014 DataCore Software All Rights Reserved Table of Contents INTRODUCTION... 3 1.1 DataCore

More information

Deployment Guide. How to prepare your environment for an OnApp Cloud deployment.

Deployment Guide. How to prepare your environment for an OnApp Cloud deployment. Deployment Guide How to prepare your environment for an OnApp Cloud deployment. Document version 1.07 Document release date 28 th November 2011 document revisions 1 Contents 1. Overview... 3 2. Network

More information

VMware vrealize Automation

VMware vrealize Automation VMware vrealize Automation Reference Architecture Version 6.0 and Higher T E C H N I C A L W H I T E P A P E R Table of Contents Overview... 4 What s New... 4 Initial Deployment Recommendations... 4 General

More information

VMware vcloud Networking and Security Overview

VMware vcloud Networking and Security Overview VMware vcloud Networking and Security Overview Networks and Security for Virtualized Compute Environments WHITE PAPER Overview Organizations worldwide have gained significant efficiency and flexibility

More information

VMware EVO SDDC. General. Q. Is VMware selling and supporting hardware for EVO SDDC?

VMware EVO SDDC. General. Q. Is VMware selling and supporting hardware for EVO SDDC? FREQUENTLY ASKED QUESTIONS VMware EVO SDDC General Q. What is VMware A. VMware EVO SDDC is the easiest way to build and run an SDDC private cloud on an integrated system. Based on an elastic, highly scalable,

More information

Reference Architecture for Dell VIS Self-Service Creator and VMware vsphere 4

Reference Architecture for Dell VIS Self-Service Creator and VMware vsphere 4 Reference Architecture for Dell VIS Self-Service Creator and VMware vsphere 4 Solutions for Large Environments Virtualization Solutions Engineering Ryan Weldon and Tom Harrington THIS WHITE PAPER IS FOR

More information

Networking Topology For Your System

Networking Topology For Your System This chapter describes the different networking topologies supported for this product, including the advantages and disadvantages of each. Select the one that best meets your needs and your network deployment.

More information

VMware vsphere Basics

VMware vsphere Basics ware vsphere Basics ESXi 5.0 vcenter Server 5.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check

More information

RED HAT ENTERPRISE VIRTUALIZATION FOR SERVERS: COMPETITIVE FEATURES

RED HAT ENTERPRISE VIRTUALIZATION FOR SERVERS: COMPETITIVE FEATURES RED HAT ENTERPRISE VIRTUALIZATION FOR SERVERS: COMPETITIVE FEATURES RED HAT ENTERPRISE VIRTUALIZATION FOR SERVERS Server virtualization offers tremendous benefits for enterprise IT organizations server

More information

CA Virtual Assurance/ Systems Performance for IM r12 DACHSUG 2011

CA Virtual Assurance/ Systems Performance for IM r12 DACHSUG 2011 CA Virtual Assurance/ Systems Performance for IM r12 DACHSUG 2011 Happy Birthday Spectrum! On this day, exactly 20 years ago (4/15/1991) Spectrum was officially considered meant - 2 CA Virtual Assurance

More information

Evaluation of Enterprise Data Protection using SEP Software

Evaluation of Enterprise Data Protection using SEP Software Test Validation Test Validation - SEP sesam Enterprise Backup Software Evaluation of Enterprise Data Protection using SEP Software Author:... Enabling you to make the best technology decisions Backup &

More information

GRAVITYZONE HERE. Deployment Guide VLE Environment

GRAVITYZONE HERE. Deployment Guide VLE Environment GRAVITYZONE HERE Deployment Guide VLE Environment LEGAL NOTICE All rights reserved. No part of this document may be reproduced or transmitted in any form or by any means, electronic or mechanical, including

More information

ESX Server 3 Configuration Guide Update 2 and later for ESX Server 3.5 and VirtualCenter 2.5

ESX Server 3 Configuration Guide Update 2 and later for ESX Server 3.5 and VirtualCenter 2.5 ESX Server 3 Configuration Guide Update 2 and later for ESX Server 3.5 and VirtualCenter 2.5 This document supports the version of each product listed and supports all subsequent versions until the document

More information

E-SPIN's Virtualization Management, System Administration Technical Training with VMware vsphere Enterprise (7 Day)

E-SPIN's Virtualization Management, System Administration Technical Training with VMware vsphere Enterprise (7 Day) Class Schedule E-SPIN's Virtualization Management, System Administration Technical Training with VMware vsphere Enterprise (7 Day) Date: Specific Pre-Agreed Upon Date Time: 9.00am - 5.00pm Venue: Pre-Agreed

More information

VXLAN Bridging & Routing

VXLAN Bridging & Routing VXLAN Bridging & Routing Darrin Machay darrin@arista.com CHI-NOG 05 May 2015 1 VXLAN VM-1 10.10.10.1/24 Subnet A ESX host Subnet B ESX host VM-2 VM-3 VM-4 20.20.20.1/24 10.10.10.2/24 20.20.20.2/24 Load

More information

Installation Guide Avi Networks Cloud Application Delivery Platform Integration with Cisco Application Policy Infrastructure

Installation Guide Avi Networks Cloud Application Delivery Platform Integration with Cisco Application Policy Infrastructure Installation Guide Avi Networks Cloud Application Delivery Platform Integration with Cisco Application Policy Infrastructure August 2015 Table of Contents 1 Introduction... 3 Purpose... 3 Products... 3

More information

Brocade One Data Center Cloud-Optimized Networks

Brocade One Data Center Cloud-Optimized Networks POSITION PAPER Brocade One Data Center Cloud-Optimized Networks Brocade s vision, captured in the Brocade One strategy, is a smooth transition to a world where information and applications reside anywhere

More information

HP Matrix Operating Environment Federated CMS Overview

HP Matrix Operating Environment Federated CMS Overview HP Matrix Operating Environment Federated CMS Overview HP Matrix OE infrastructure orchestration 7.0 Technical white paper Table of contents Introduction... 2 Overview... 2 Installation and configuration...

More information

VMware Software Defined Network. Dejan Grubić VMware Systems Engineer for Adriatic

VMware Software Defined Network. Dejan Grubić VMware Systems Engineer for Adriatic VMware Software Defined Network Dejan Grubić VMware Systems Engineer for Adriatic The Transformation of Infrastructure Infrastructure Servers Clouds Be more responsive to business, change economics of

More information

vsphere Replication for Disaster Recovery to Cloud

vsphere Replication for Disaster Recovery to Cloud vsphere Replication for Disaster Recovery to Cloud vsphere Replication 6.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced

More information

VMware vrealize Automation

VMware vrealize Automation VMware vrealize Automation Reference Architecture Version 6.0 or Later T E C H N I C A L W H I T E P A P E R J U N E 2 0 1 5 V E R S I O N 1. 5 Table of Contents Overview... 4 What s New... 4 Initial Deployment

More information

vcloud Air Disaster Recovery Technical Presentation

vcloud Air Disaster Recovery Technical Presentation vcloud Air Disaster Recovery Technical Presentation Agenda 1 vcloud Air Disaster Recovery Overview 2 What s New 3 Architecture 4 Setup and Configuration 5 Considerations 6 Automation Options 2 vcloud Air

More information

Cisco Nexus 1000V Virtual Ethernet Module Software Installation Guide, Release 4.0(4)SV1(1)

Cisco Nexus 1000V Virtual Ethernet Module Software Installation Guide, Release 4.0(4)SV1(1) Cisco Nexus 1000V Virtual Ethernet Module Software Installation Guide, Release 4.0(4)SV1(1) September 17, 2010 Part Number: This document describes how to install software for the Cisco Nexus 1000V Virtual

More information

VMware vsphere Design. 2nd Edition

VMware vsphere Design. 2nd Edition Brochure More information from http://www.researchandmarkets.com/reports/2330623/ VMware vsphere Design. 2nd Edition Description: Achieve the performance, scalability, and ROI your business needs What

More information

Cisco Prime Network Services Controller. Sonali Kalje Sr. Product Manager Cloud and Virtualization, Cisco Systems

Cisco Prime Network Services Controller. Sonali Kalje Sr. Product Manager Cloud and Virtualization, Cisco Systems Cisco Prime Network Services Controller Sonali Kalje Sr. Product Manager Cloud and Virtualization, Cisco Systems Agenda Cloud Networking Challenges Prime Network Services Controller L4-7 Services Solutions

More information

VirtualclientTechnology 2011 July

VirtualclientTechnology 2011 July WHAT S NEW IN VSPHERE VirtualclientTechnology 2011 July Agenda vsphere Platform Recap vsphere 5 Overview Infrastructure Services Compute, Storage, Network Applications Services Availability, Security,

More information

vsphere Replication for Disaster Recovery to Cloud

vsphere Replication for Disaster Recovery to Cloud vsphere Replication for Disaster Recovery to Cloud vsphere Replication 5.8 This document supports the version of each product listed and supports all subsequent versions until the document is replaced

More information

VMware Virtual SAN 6.2 Network Design Guide

VMware Virtual SAN 6.2 Network Design Guide VMware Virtual SAN 6.2 Network Design Guide TECHNICAL WHITE PAPER APRIL 2016 Contents Intended Audience... 2 Overview... 2 Virtual SAN Network... 2 Physical network infrastructure... 3 Data center network...

More information

Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study

Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study White Paper Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study 2012 Cisco and/or its affiliates. All rights reserved. This

More information

Next Gen Data Center. KwaiSeng Consulting Systems Engineer kslai@cisco.com

Next Gen Data Center. KwaiSeng Consulting Systems Engineer kslai@cisco.com Next Gen Data Center KwaiSeng Consulting Systems Engineer kslai@cisco.com Taiwan Update Feb 08, kslai 2006 Cisco 2006 Systems, Cisco Inc. Systems, All rights Inc. reserved. All rights reserved. 1 Agenda

More information

vcloud Director User's Guide

vcloud Director User's Guide vcloud Director 5.5 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of

More information

VMware and Brocade Network Virtualization Reference Whitepaper

VMware and Brocade Network Virtualization Reference Whitepaper VMware and Brocade Network Virtualization Reference Whitepaper Table of Contents EXECUTIVE SUMMARY VMWARE NSX WITH BROCADE VCS: SEAMLESS TRANSITION TO SDDC VMWARE'S NSX NETWORK VIRTUALIZATION PLATFORM

More information

Virtualization, SDN and NFV

Virtualization, SDN and NFV Virtualization, SDN and NFV HOW DO THEY FIT TOGETHER? Traditional networks lack the flexibility to keep pace with dynamic computing and storage needs of today s data centers. In order to implement changes,

More information

vsphere Upgrade Update 1 ESXi 6.0 vcenter Server 6.0 EN-001804-02

vsphere Upgrade Update 1 ESXi 6.0 vcenter Server 6.0 EN-001804-02 Update 1 ESXi 6.0 vcenter Server 6.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent

More information

Veritas Storage Foundation High Availability for Windows by Symantec

Veritas Storage Foundation High Availability for Windows by Symantec Veritas Storage Foundation High Availability for Windows by Symantec Simple-to-use solution for high availability and disaster recovery of businesscritical Windows applications Data Sheet: High Availability

More information

Securely Architecting the Internal Cloud. Rob Randell, CISSP Senior Security and Compliance Specialist VMware, Inc.

Securely Architecting the Internal Cloud. Rob Randell, CISSP Senior Security and Compliance Specialist VMware, Inc. Securely Architecting the Internal Cloud Rob Randell, CISSP Senior Security and Compliance Specialist VMware, Inc. Securely Building the Internal Cloud Virtualization is the Key How Virtualization Affects

More information

RSA Authentication Manager 8.1 Setup and Configuration Guide. Revision 2

RSA Authentication Manager 8.1 Setup and Configuration Guide. Revision 2 RSA Authentication Manager 8.1 Setup and Configuration Guide Revision 2 Contact Information Go to the RSA corporate website for regional Customer Support telephone and fax numbers: www.emc.com/domains/rsa/index.htm

More information

VMware Network Virtualization Design Guide. January 2013

VMware Network Virtualization Design Guide. January 2013 ware Network Virtualization Technical WHITE PAPER January 2013 ware Network Virtualization Table of Contents Intended Audience.... 3 Overview.... 3 Components of the ware Network Virtualization Solution....

More information

VMware vsphere 5.0 Evaluation Guide

VMware vsphere 5.0 Evaluation Guide VMware vsphere 5.0 Evaluation Guide Advanced Networking Features TECHNICAL WHITE PAPER Table of Contents About This Guide.... 4 System Requirements... 4 Hardware Requirements.... 4 Servers.... 4 Storage....

More information

DMZ Virtualization Using VMware vsphere 4 and the Cisco Nexus 1000V Virtual Switch

DMZ Virtualization Using VMware vsphere 4 and the Cisco Nexus 1000V Virtual Switch DMZ Virtualization Using VMware vsphere 4 and the Cisco Nexus 1000V Virtual Switch What You Will Learn A demilitarized zone (DMZ) is a separate network located in the neutral zone between a private (inside)

More information

Monitoring Hybrid Cloud Applications in VMware vcloud Air

Monitoring Hybrid Cloud Applications in VMware vcloud Air Monitoring Hybrid Cloud Applications in ware vcloud Air ware vcenter Hyperic and ware vcenter Operations Manager Installation and Administration Guide for Hybrid Cloud Monitoring TECHNICAL WHITE PAPER

More information

Simplified Private Cloud Management

Simplified Private Cloud Management BUSINESS PARTNER ClouTor Simplified Private Cloud Management ClouTor ON VSPEX by LOCUZ INTRODUCTION ClouTor on VSPEX for Enterprises provides an integrated software solution for extending your existing

More information

Best Practices for Monitoring Databases on VMware. Dean Richards Senior DBA, Confio Software

Best Practices for Monitoring Databases on VMware. Dean Richards Senior DBA, Confio Software Best Practices for Monitoring Databases on VMware Dean Richards Senior DBA, Confio Software 1 Who Am I? 20+ Years in Oracle & SQL Server DBA and Developer Worked for Oracle Consulting Specialize in Performance

More information

VMware vsphere Data Protection

VMware vsphere Data Protection VMware vsphere Data Protection Replication Target TECHNICAL WHITEPAPER 1 Table of Contents Executive Summary... 3 VDP Identities... 3 vsphere Data Protection Replication Target Identity (VDP-RT)... 3 Replication

More information

Building the Virtual Information Infrastructure

Building the Virtual Information Infrastructure Technology Concepts and Business Considerations Abstract A virtual information infrastructure allows organizations to make the most of their data center environment by sharing computing, network, and storage

More information

VMware vcloud Air Networking Guide

VMware vcloud Air Networking Guide vcloud Air This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of this document,

More information

Nutanix Tech Note. Configuration Best Practices for Nutanix Storage with VMware vsphere

Nutanix Tech Note. Configuration Best Practices for Nutanix Storage with VMware vsphere Nutanix Tech Note Configuration Best Practices for Nutanix Storage with VMware vsphere Nutanix Virtual Computing Platform is engineered from the ground up to provide enterprise-grade availability for critical

More information

Reference Design: Deploying NSX for vsphere with Cisco UCS and Nexus 9000 Switch Infrastructure TECHNICAL WHITE PAPER

Reference Design: Deploying NSX for vsphere with Cisco UCS and Nexus 9000 Switch Infrastructure TECHNICAL WHITE PAPER Reference Design: Deploying NSX for vsphere with Cisco UCS and Nexus 9000 Switch Infrastructure TECHNICAL WHITE PAPER Table of Contents 1 Executive Summary....3 2 Scope and Design Goals....3 2.1 NSX VMkernel

More information

Restricted Document. Pulsant Technical Specification

Restricted Document. Pulsant Technical Specification Pulsant Technical Specification Title Pulsant Government Virtual Server IL2 Department Cloud Services Contributors RR Classification Restricted Version 1.0 Overview Pulsant offer two products based on

More information

Course Syllabus. Fundamentals of Windows Server 2008 Network and Applications Infrastructure. Key Data. Audience. Prerequisites. At Course Completion

Course Syllabus. Fundamentals of Windows Server 2008 Network and Applications Infrastructure. Key Data. Audience. Prerequisites. At Course Completion Key Data Product #: 3380 Course #: 6420A Number of Days: 5 Format: Certification Exams: Instructor-Led None This course syllabus should be used to determine whether the course is appropriate for the students,

More information

CloudLink - The On-Ramp to the Cloud Security, Management and Performance Optimization for Multi-Tenant Private and Public Clouds

CloudLink - The On-Ramp to the Cloud Security, Management and Performance Optimization for Multi-Tenant Private and Public Clouds - The On-Ramp to the Cloud Security, Management and Performance Optimization for Multi-Tenant Private and Public Clouds February 2011 1 Introduction Today's business environment requires organizations

More information

Virtual Appliance Setup Guide

Virtual Appliance Setup Guide The Barracuda SSL VPN Vx Virtual Appliance includes the same powerful technology and simple Web based user interface found on the Barracuda SSL VPN hardware appliance. It is designed for easy deployment

More information

Configuration Maximums VMware vsphere 4.1

Configuration Maximums VMware vsphere 4.1 Topic Configuration s VMware vsphere 4.1 When you select and configure your virtual and physical equipment, you must stay at or below the maximums supported by vsphere 4.1. The limits presented in the

More information

VMDC 3.0 Design Overview

VMDC 3.0 Design Overview CHAPTER 2 The Virtual Multiservice Data Center architecture is based on foundation principles of design in modularity, high availability, differentiated service support, secure multi-tenancy, and automated

More information

VM-Series Firewall Deployment Tech Note PAN-OS 5.0

VM-Series Firewall Deployment Tech Note PAN-OS 5.0 VM-Series Firewall Deployment Tech Note PAN-OS 5.0 Revision A 2012, Palo Alto Networks, Inc. www.paloaltonetworks.com Contents Overview... 3 Supported Topologies... 3 Prerequisites... 4 Licensing... 5

More information

BUILDING A NEXT-GENERATION DATA CENTER

BUILDING A NEXT-GENERATION DATA CENTER BUILDING A NEXT-GENERATION DATA CENTER Data center networking has changed significantly during the last few years with the introduction of 10 Gigabit Ethernet (10GE), unified fabrics, highspeed non-blocking

More information

vcloud Air - Virtual Private Cloud OnDemand Networking Guide

vcloud Air - Virtual Private Cloud OnDemand Networking Guide vcloud Air - Virtual Private Cloud OnDemand Networking Guide vcloud Air This document supports the version of each product listed and supports all subsequent versions until the document is replaced by

More information