Brocade VCS Fabric Technology and NAS with NFS Validation Test



Similar documents
ADVANCED NETWORK CONFIGURATION GUIDE

Nutanix Tech Note. VMware vsphere Networking on Nutanix

Brocade One Data Center Cloud-Optimized Networks

全 新 企 業 網 路 儲 存 應 用 THE STORAGE NETWORK MATTERS FOR EMC IP STORAGE PLATFORMS

Data Center Convergence. Ahmad Zamer, Brocade

Ethernet Fabrics: An Architecture for Cloud Networking

Brocade Solution for EMC VSPEX Server Virtualization

Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study

VMware Virtual SAN 6.2 Network Design Guide

Virtual PortChannels: Building Networks without Spanning Tree Protocol

Nutanix Hyperconverged Appliance with the Brocade VDX ToR Switch Deployment Guide

How To Evaluate Netapp Ethernet Storage System For A Test Drive

How To Connect Virtual Fibre Channel To A Virtual Box On A Hyperv Virtual Machine

Building the Virtual Information Infrastructure

VIDEO SURVEILLANCE WITH SURVEILLUS VMS AND EMC ISILON STORAGE ARRAYS

Deploying Brocade VDX 6720 Data Center Switches with Brocade VCS in Enterprise Data Centers

FlexNetwork Architecture Delivers Higher Speed, Lower Downtime With HP IRF Technology. August 2011

Introduction to MPIO, MCS, Trunking, and LACP

Storage Protocol Comparison White Paper TECHNICAL MARKETING DOCUMENTATION

How To Set Up A Virtual Network On Vsphere (Vsphere) On A 2Nd Generation Vmkernel (Vklan) On An Ipv5 Vklan (Vmklan)

Brocade VCS Fabric Technology with the EMC VNX5300

VMware Virtual SAN Network Design Guide TECHNICAL WHITE PAPER

Simplifying Virtual Infrastructures: Ethernet Fabrics & IP Storage

DATA CENTER. Brocade VDX/VCS Data Center Layer 2 Fabric Design Guide for Brocade Network OS v2.1.1

DEDICATED NETWORKS FOR IP STORAGE

Solution Brief Network Design Considerations to Enable the Benefits of Flash Storage

VMware vsphere 5.1 Advanced Administration

Expert Reference Series of White Papers. VMware vsphere Distributed Switches

Building Tomorrow s Data Center Network Today

vsphere Networking vsphere 5.5 ESXi 5.5 vcenter Server 5.5 EN

How to Configure Intel Ethernet Converged Network Adapter-Enabled Virtual Functions on VMware* ESXi* 5.1

vsphere Networking vsphere 6.0 ESXi 6.0 vcenter Server 6.0 EN

The Future of Computing Cisco Unified Computing System. Markus Kunstmann Channels Systems Engineer

Dell EqualLogic Best Practices Series

Block based, file-based, combination. Component based, solution based

Data Center Networking Designing Today s Data Center

VMware and Brocade Network Virtualization Reference Whitepaper

Data Center Fabric Convergence for Cloud Computing (the Debate of Ethernet vs. Fibre Channel is Over)

iscsi Top Ten Top Ten reasons to use Emulex OneConnect iscsi adapters

Multi-Chassis Trunking for Resilient and High-Performance Network Architectures

W H I T E P A P E R. Best Practices for High Performance NFS Storage with VMware

SN A. Reference Guide Efficient Data Center Virtualization with QLogic 10GbE Solutions from HP

Private Cloud Migration

Fibre Channel over Ethernet in the Data Center: An Introduction

Chapter 3. Enterprise Campus Network Design

Intel Ethernet Switch Load Balancing System Design Using Advanced Features in Intel Ethernet Switch Family

Network Troubleshooting & Configuration in vsphere VMware Inc. All rights reserved

Enterasys Data Center Fabric

How to Create a Virtual Switch in VMware ESXi

Emulex OneConnect 10GbE NICs The Right Solution for NAS Deployments

Storage Area Network Design Overview Using Brocade DCX Backbone Switches

Virtualizing the SAN with Software Defined Storage Networks

Global Headquarters: 5 Speen Street Framingham, MA USA P F

Using NetApp Unified Connect to Create a Converged Data Center

What s New in VMware vsphere 5.5 Networking

Networking Topology For Your System

An Introduction to Brocade VCS Fabric Technology

VMDC 3.0 Design Overview

VXLAN: Scaling Data Center Capacity. White Paper

A High-Performance Storage and Ultra-High-Speed File Transfer Solution

Ethernet Storage Best Practices

vsphere Networking ESXi 5.0 vcenter Server 5.0 EN

Where IT perceptions are reality. Test Report. OCe14000 Performance. Featuring Emulex OCe14102 Network Adapters Emulex XE100 Offload Engine

> Resilient Data Center Solutions for VMware ESX Servers Technical Configuration Guide

Scalable Approaches for Multitenant Cloud Data Centers

iscsi: Accelerating the Transition to Network Storage

Best Practice of Server Virtualization Using Qsan SAN Storage System. F300Q / F400Q / F600Q Series P300Q / P400Q / P500Q / P600Q Series

SAN Conceptual and Design Basics

IP ETHERNET STORAGE CHALLENGES

SUN DUAL PORT 10GBase-T ETHERNET NETWORKING CARDS

Brocade VCS Fabric Technology with the NetApp 3170 NAS Storage Array

Dell PowerEdge Blades Outperform Cisco UCS in East-West Network Performance

TRILL for Service Provider Data Center and IXP. Francois Tallet, Cisco Systems

Juniper Networks EX Series/ Cisco Catalyst Interoperability Test Results. May 1, 2009

FIBRE CHANNEL OVER ETHERNET

Configuring iscsi Multipath

Broadcom Ethernet Network Controller Enhanced Virtualization Functionality

Isilon IQ Network Configuration Guide

JOB ORIENTED VMWARE TRAINING INSTITUTE IN CHENNAI

Enabling Technologies for Distributed Computing

Best practices when deploying VMware vsphere 5.0 connected to HP Networking Switches

Validating Long-distance VMware vmotion

Cisco Datacenter 3.0. Datacenter Trends. David Gonzalez Consulting Systems Engineer Cisco

Simplify VMware vsphere* 4 Networking with Intel Ethernet 10 Gigabit Server Adapters

VMware vsphere on NetApp. Course: 5 Day Hands-On Lab & Lecture Course. Duration: Price: $ 4, Description:

EMC SYNCPLICITY FILE SYNC AND SHARE SOLUTION

Best Practices for Implementing iscsi Storage in a Virtual Server Environment

Simplify Virtual Machine Management and Migration with Ethernet Fabrics in the Datacenter

Private cloud computing advances

Preparation Guide. How to prepare your environment for an OnApp Cloud v3.0 (beta) deployment.

Running Philips IntelliSpace Portal with VMware vmotion, DRS and HA on vsphere 5.1 and 5.5. September 2014

Server and Storage Virtualization with IP Storage. David Dale, NetApp

VCS Monitoring and Troubleshooting Using Brocade Network Advisor

RFP-MM Enterprise Storage Addendum 1

Converged Networking Solution for Dell M-Series Blades. Spencer Wheelwright

Transcription:

Brocade VCS Fabric Technology and NAS with NFS Validation Test NetApp/VMware vsphere 5.0 Red Hat Enterprise Linux This material outlines sample configurations and associated test results of Brocade VCS Fabric technology with NFS file servers.

CONTENTS Contents...2 1 Preface...4 1.1 Overview...4 1.2 Purpose of This Document...4 1.3 Audience...4 1.4 Objectives...4 1.5 Brocade VCS Features...5 1.6 Summary...6 1.7 Related Documents...6 1.8 About Brocade...7 2. Test Case #1: NetApp FAS3050...8 2.1 Test Case #1 Summary...8 2.2 Topology...9 2.3 Hardware Resources...9 2.4 Compute Resources...10 2.5 Software Resources...10 2.6 Test 1: I/O Verification...10 2.7 Test 2: Link Failure...12 2.8 Test 3: Active Path Failure...14 2.9 Test 4: Switch Failure...17 2.10 Test 5: vmotion with Link/Path Failure...22 3. Test Case #2: Red Hat Enterprise Linux NFS Server...30 3.1 Test Case #2 Results...30 3.2 Topology...31 3.3 Hardware Resources...31 3.4 Compute Resources...31 3.5 Software Resources...32 3.6 Test 1: I/O Verification...32 3.7 Test 2: Link Failure...34 3.8 Test 3: Switch Failure...36 3.8.1 Test Procedure...37 Appendix A: Test Case #1...41 Brocade VCS Deployment Considerations: Dynamic LACP vlag with NetApp...41 Brocade VDX Deployment Considerations: Static LACP vlag with ESXi Server and Brocade 1020 CNA...43 Brocade VCS Deployment Considerations: Enable Brocade VDX Jumbo Frame Support...45 Brocade VCS Deployment Considerations: Enable Ethernet Pause/Flow Control Support...45 NetApp FAS3050 Deployment Procedure: Volume and VIF Creation...46 NetApp Volume Creation...46 NetApp VIF Creation...46 Brocade VCS Fabric Technology and NAS with NFS Validation Test 2 of 52

VMware vsphere Client Deployment Procedure: ESXi Datastore Creation...46 ESX Datastore Creation...46 VMware vsphere Client Deployment Procedure: Virtual Machine Creation...46 VM Creation...46 VMware ESXi Deployment Procedure: NIC Teaming for vswitch...47 VMware ESXi NIC Teaming for vswitch...47 Appendix b: Test Case #2...48 Brocade VCS Fabric Configuration: Static LACP vlag with Red Hat Enterprise Linux NFS Server and Brocade 1020 CNA...48 Red Hat Enterprise Linux NFS Server: NIC Bonding Configuration...50 Appendix C: References...52 Brocade VCS Fabric Technology and NAS with NFS Validation Test 3 of 52

1 PREFACE 1.1 Overview As per the Gartner 2011 NAS Magic Quadrant report, the midrange and high-end Network-Attached Storage (NAS) market for 2010 experienced a growth rate of 33 percent over 2009 in terms of hardware vendor revenue. This favorable growth rate was a result of several factors: fast-growing unstructured file data, widespread availability of data de-duplication/compression in NAS storage solutions, ease of management, support of virtualized environments such as VMware, and the flexibility of unified storage. 1 Additionally, IDC predicts that by 2014 more than 83 percent of enterprise storage system capacity will be shipped for file-based data, taking the Compound Annual Growth Rate (CAGR) for file serving storage capacity to 2.5 times the CAGR for block storage capacity 2. As also discussed in the Gartner 2011 NAS Magic Quadrant report, NAS support of the VMware environment has become more prominent in the past year, as more and more NAS vendors have invested in this area to increase the appeal of their products. Additionally, NAS products use industry-standard remote file protocols, including Network File System (NFS). The Gartner report also observes that because some applications, such as Oracle database applications and VMware, are built on files instead of blocks, NAS is increasingly used as application storage for those environments, providing ease-of-use benefits to users as compared with storage arrays that use block protocols, which may offer higher performance than NAS. As a result, many midrange and high-end NAS products are used to consolidate storage for both server applications and home directories for PC clients. Given the tremendous growth rates in NAS and the large number of existing deployments, this presents an opportunity to demonstrate that Brocade VCS Fabric technology interoperates with the underlying NFS protocol used by these systems. 1.2 Purpose of This Document This document provides the validation of Brocade VCS Fabric technology with two implementations of the Network File System (NFS) Protocol, including both the NetApp FAS3050 NFS filer and Red Hat Enterprise Linux configured as an NFS server. This validation demonstrates that existing deployments using NFS will interoperate with Brocade VCS fabrics and exhibit resiliency to failure scenarios. This ensures that inputs/outputs (I/Os) between clients and servers operate in a non-disruptive manner. The testing demonstrates NFS interoperability with Brocade VCS Fabric technology, while providing sample configurations and test results associated with fabric failover scenarios. This document should provide peace of mind for both network and storage administrators and architects who are already using NFS and are considering the use of Brocade VCS Fabric technology. 1.3 Audience The content in this document is written for a technical audience, including solution architects, solutioneers, system engineers, and technical development representatives. This document assumes the audience is familiar with Brocade VCS Fabric technology. 1.4 Objectives The objectives of this document are to evaluate NFS protocol interoperability with Brocade VCS Fabric technology in the following two test cases: Test Case #1 with the NetApp FAS3050: This test consists of a 6-node Brocade VCS fabric with 2 ESXi hosts using Brocade 1020 FCoE (Fiber Channel over Ethernet) CNAs and the NetApp FAS3050. For this test, the Virtual Machine (VM) data store associated with the ESXi cluster resides on a volume of the NetApp FAS3050. Iometer is used as a measurement and characterization tool for this test. Brocade VCS Fabric Technology and NAS with NFS Validation Test 4 of 52

Test Case #2 with Red Hat Enterprise Linux NFS server: This test consists of a 4-node Brocade VCS fabric with a Red Hat Enterprise Linux NFS server. For this test, Spirent is used as a characterization tool for the NFS clients. Spirent emulates many NFS clients accessing the Red Hat Enterprise Linux NFS server share. Each of the Spirent-emulated NFS clients mounted the NFS share of the Red Hat Enterprise Linux NFS server. Both tests demonstrate that the topology of the network can change, with limited implication to the application, using NFS as an underlying protocol. Please note that this does not include a full configuration or validation of specific features on ESXi, Red Hat, NetApp, and VMware. 1.5 Brocade VCS Features The following Brocade VCS features are used in the validation testing for both test cases. These features are considered best practices when utilizing NFS over a Brocade VCS fabric. Please refer to Appendices A and B for the actual configuration procedures for these features. 1.5.1 Brocade Inter-Switch Link (ISL) Trunks For both Test Case #1 and Test Case #2, Brocade Inter-Switch Link (ISL) Trunking is used within the Brocade VCS fabric to provide additional redundancy and load balancing between the NFS clients and NFS server. Typically, multiple links between two switches are bundled together in a Link Aggregation Group (LAG) to provide redundancy and load balancing. Setting up a LAG requires lines of configuration on the switches and selecting a hash-based algorithm for load balancing based on source-destination IP or MAC addresses. All flows with the same hash traverse the same link, regardless of the total number of links in a LAG. This might result in some links within a LAG, such as those carrying flows to a storage target, being overutilized and packets being dropped, while other links in the LAG remain underutilized. Instead of LAG-based switch interconnects, Brocade VCS Ethernet fabrics automatically form ISL trunks when multiple connections are added between two Brocade VDX switches. Simply adding another cable increases bandwidth, providing linear scalability of switch-to-switch traffic, and this does not require any configuration on the switch. In addition, ISL trunks use a frame-by-frame load balancing technique, which evenly balances traffic across all members of the ISL trunk group. 1.5.2 Equal-Cost Multipath (ECMP) A standard link-state routing protocol that runs at Layer 2 determines if there are Equal-Cost Multipaths (ECMPs) between RBridges in an Ethernet fabric and load balances the traffic to make use of all available ECMPs. If a neighbor switch is reachable via several interfaces with different bandwidths, all of them are treated as equal-cost paths. While it is possible to set the link cost based on the link speed, such an algorithm complicates the operation of the fabric. Simplicity is a key value of Brocade VCS Fabric technology, so an implementation is chosen in the test cases that does not consider the bandwidth of the interface when selecting equal-cost paths. This is a key feature needed to expand network capacity, to keep ahead of customer bandwidth requirements. 1.5.3 Virtual Link Aggregation Group (vlag) For both Test Case #1 and Test Case #2, Virtual Link Aggregation Groups (vlags) are used for the ESXi hosts, the NetApp FAS3050, and the Red Hat Enterprise Linux NFS server. In the case of the NetApp FAS3050, a dynamic Link Aggregation Control Protocol (LACP) vlag is used. In the case of both ESXi hosts and the Red Hat Enterprise Linux NFS server, static LACP vlags are used. While Brocade ISLs are used as interconnects between Brocade VDX switches within a Brocade VCS fabric, industrystandard LACP LAGs are supported for connecting to other network devices outside the Brocade VCS fabric. Typically, LACP LAGs can only be created using ports from a single physical switch to a second physical switch. In a Brocade VCS fabric, a vlag can be created using ports from two Brocade VDX switches to a device to which both VDX switches are connected. This provides an additional degree of device-level redundancy, while providing active-active link-level load balancing. For additional configuration details please refer to Appendices A and B. Brocade VCS Fabric Technology and NAS with NFS Validation Test 5 of 52

1.5.4 Pause Flow Control For these test cases, Pause Flow Control is enabled on vlag-facing interfaces connected to the ESXi hosts, the NetApp FAS3050, and the Red Hat Enterprise Linux NFS server. Brocade VDX Series switches support the Pause Flow Control feature. IEEE 802.3x Ethernet pause and Ethernet Priority-Based Flow Control (PFC) are used to prevent dropped frames by slowing traffic at the source end of a link. When a port on a switch or host is not ready to receive more traffic from the source perhaps due to congestion it sends pause frames to the source to pause the traffic flow. When the congestion is cleared, the port stops requesting the source to pause traffic flow, and traffic resumes without any frame drop. When Ethernet pause is enabled, pause frames are sent to the traffic source. Similarly, when PFC is enabled, there is no frame drop; pause frames are sent to the source switch. For configuration details on Ethernet Pause and PFC on Brocade VDX Series switches, please refer to Appendix A. 1.5.5 Ultra-Low Latency The Brocade VDX series of switches provides industry-leading performance and ultra-low latency through wire-speed ports with 600 nanosecond port-to-port latency and hardware-based Brocade ISL Trunking. This is helpful for environments that require high availability, such as providing Ethernet storage connectivity for FCoE, Internet Small Computer Systems Interface (iscsi), and NAS. 1.5.6 Jumbo Frames Brocade VDX Series switches support the transport of jumbo frames. Jumbo frames are enabled by default on the Brocade ISL trunks. However, to accommodate end-to-end jumbo frame support on the network for the edge hosts, this feature can be enabled under the vlag interface connected to the ESXi hosts, the NetApp FAS3050, and Red Hat Enterprise Linux NFS server. The default Maximum Transmission Unit (MTU) on these interfaces is 2500. This MTU is set to 9216 to optimize the network for jumbo frame support. For additional configuration details of jumbo frames on Brocade VDX Series switches, please refer to Appendix A. For additional details on best practices for enabling jumbo frames on the host devices that transmit the frames, please reference the NetApp and VMware vsphere Storage Best Practices guide: http://communities.netapp.com/servlet/jiveservlet/previewbody/11657-102-1-22108/tr3749netappandvmwarevspherestoragebestpracticesjul10.pdf 1.6 Summary These results support the validation that Brocade VCS Fabric technology interoperates with two implementations of the Network File System (NFS) Protocol, including both the NetApp FAS3050 NFS filer and the Red Hat Enterprise Linux NFS server. This validation demonstrates that existing deployments using NFS interoperate with Brocade VCS fabrics and exhibit resiliency to failure scenarios. This ensures that I/Os between clients and servers operate in a non-disruptive manner. 1.7 Related Documents For more information about Brocade VCS Fabric technology, please see the Brocade VCS Fabric Technical Architecture brief: http://www.brocade.com/downloads/documents/technical_briefs/vcs-technical-architecture-tb.pdf For the Brocade Network OS (NOS) Admin Guide and NOS Command Reference: http://www.brocade.com/downloads/documents/product_manuals/b_vdx/nos_adminguide_v211.pdf http://www.brocade.com/downloads/documents/product_manuals/b_vdx/nos_commandref_v211.pdf The Brocade NOS Release notes can be found at http://my.brocade.com Brocade VCS Fabric Technology and NAS with NFS Validation Test 6 of 52

For more information about the Brocade VDX Series of switches, please see the product Data sheets: Brocade VDX 6710 Data Center Switch: http://www.brocade.com/products/all/switches/product-details/vdx-6710-dc-switches/index.page Brocade VDX 6720 Data Center Switch: http://www.brocade.com/products/all/switches/product-details/vdx-6720-dc-switches/index.page Brocade VDX 6730 Data Center Switch: http://www.brocade.com/products/all/switches/product-details/vdx-6730-dc-switches/index.page 1.8 About Brocade As information becomes increasingly mobile and distributed across the enterprise, organizations are transitioning to a highly virtualized infrastructure, which often increases overall IT complexity. To simplify this process, organizations must have reliable, flexible network solutions that utilize IT resources whenever and wherever needed enabling the full advantages of virtualization and cloud computing. As a global provider of comprehensive networking solutions, Brocade has more than 15 years of experience in delivering Ethernet, storage, and converged networking technologies that are used in the world s most missioncritical environments. Based on the Brocade One strategy, this unique approach reduces complexity and disruption by removing network layers, simplifying management, and protecting existing technology investments. As a result, organizations can utilize cloud-optimized networks to achieve their goals of non-stop operations in highly virtualized infrastructures where information and applications are available anywhere. Brocade VCS Fabric Technology and NAS with NFS Validation Test 7 of 52

2. TEST CASE #1: NETAPP FAS3050 For this test, one type of NAS storage is used. NetApp FAS3050 is used as a NAS server. This storage option is commonly deployed when using NAS storage pools with VMware vsphere. Test Case #1 with the NetApp FAS3050 consists of a 6-node Brocade VCS fabric with 2 ESXi hosts using Brocade 1020 CNAs and NetApp FAS3050. For this test, the VM data store associated with the ESXi cluster resides on a volume of the NetApp FAS3050. Iometer is used as a measurement and characterization tool for this test. The demonstration shows that the topology of the network can change with limited implication to the application using NFS as an underlying protocol. Please note that this does not include a full configuration or validation of specific features on ESXi, Red Hat, NetApp, and VMware. Lastly, please note that for this test case, the NetApp FAS3050 used is limited to Gigabit Ethernet (GbE) interfaces. Therefore, a Brocade VDX 6710 was used to extend Gigabit Ethernet to both RB1 and RB2. Assuming the NetApp FAS3050 had 10 GbE interfaces, it would simply be connected to RB1 and RB2 without the need for RB21 and RB22. 2.1 Test Case #1 Summary All tests for Test Case #1 were performed successfully, with no issues. This validation demonstrates that existing deployments using NFS do interoperate with Brocade VCS fabrics and exhibit resiliency to I/Os between clients and servers in a non-disruptive manner. The following tests were conducted. 1. Test 1 validated baseline I/O using the shortest equal-cost paths in the fabric, with NFS clients on RB3 and RB4 accessing the NetApp FAS3050 on the vlag interface on RB21 and RB22. 2. Test 2 validated that I/O continues between a VM using an NFS client accessing a storage pool on a NetApp FAS3050 when failing a link in the Brocade ISL trunk. 3. Test 3 validated that I/O continues between a VM using an NFS client accessing a storage pool on a NetApp FAS3050 when a complete path fails. 4. Test 4 validated that I/O flows are not impacted between a VM using an NFS client accessing a storage pool on a NetApp FAS3050 when a switch in the Brocade VCS fabric fails. 5. Test 5 validated that there was no impact throughout the duration of a successful vmotion between ESX servers acting as NFS clients with a link/path failure. This summarizes the test results for Test Case #1. Test Description Results 1 Baseline of shortest paths and traffic distribution in the Brocade VCS fabric Pass 2 Perform link failure within a Brocade ISL trunk Pass 3 Perform a complete active path failure within the Brocade VCS fabric 4 Perform multiple switch failures within the Brocade VCS fabric 5 Perform vmotion between ESX servers as NFS clients with switch failure Pass Pass Pass Brocade VCS Fabric Technology and NAS with NFS Validation Test 8 of 52

2.2 Topology Figure 1. NetApp FAS3050 NFS validation topology and components for Test Case #1. 2.3 Hardware Resources The following equipment was used in this configuration: Description Quantity Revisions Brocade VDX 6710-54 2 Brocade NOS 2.1.1 Brocade VDX 6720-24 3 Brocade NOS 2.1.1 Brocade VDX 6730-32 1 Brocade NOS 2.1.1 NetApp FAS3050 1 Release 7.3.6; NFS Brocade VCS Fabric Technology and NAS with NFS Validation Test 9 of 52

2.4 Compute Resources The following equipment was used in this configuration: Description Quantity Revisions VMware ESX 2 Intel Xeon X5670, 2.93 GHz 2 socket, 6 cores, 32 GB RAM VMware ESX Continued 2 Brocade 1020 VMware vsphere Management 1 Intel Xeon X3430, 2.4 GHz Quad cores, 8 GB RAM 2.5 Software Resources The following software was used in this configuration: Description Revision Brocade NOS 2.1.1 NetApp Data ONTAP 7.3.6 VMware vsphere 5.0.0 VMware ESX 5.0.0 2.6 Test 1: I/O Verification The purpose of test 1 is to validate baseline I/O is using the shortest equal-cost paths in the fabric, with the NFS clients on RB3 and RB4 accessing the NetApp FAS3050 on the vlag interface on RB21 and RB22. Brocade VCS Fabric Technology and NAS with NFS Validation Test 10 of 52

Figure 2. NetApp FAS3050 NFS validation topology for test 1. 2.6.1 Test Procedure Step 1: Run Iometer on the Windows VM. Note that NIC teaming for vswitch was configured for the Route based on IP hash policy. Please see Appendix A for details. Note that the I/O path shown with a green line shows valid ECMPs in the fabric between ESXi and the NetApp FAS3050. The actual flows that traverse these links vary, depending on the number of hosts, MAC addresses, and flows. Step 2: Execute the following command on all Brocade VDX ISLs and VDX vlag interfaces to confirm that traffic flow and I/O path is evenly distributed as expected. The following is an example: VDX6710-RB21# do show interface port-channel 33 in rate 2.6.2 Expected Results The expected results for test 1 are to show distributed I/O due to an active/active NIC teaming vlag end-to-end from ESX VM NFS clients to the NetApp server datastore. 2.6.3 Actual Results The actual results for test 1 confirm evenly distributed I/O due to an active/active NIC teaming vlag end-to-end from ESX VM NFS clients to the NetApp server datastore. Performance from Iometer on the Windows VM (maximum throughput was approximately 119 MB/sec): Brocade VCS Fabric Technology and NAS with NFS Validation Test 11 of 52

2.7 Test 2: Link Failure The purpose of test 2 is to validate that I/O flows are not impacted between the NFS client VM and NetApp FAS3050 while introducing a link failure on one link in the Brocade trunk, which is an active shortest path in the fabric. The links that are failed as part of this test are highlighted below in red. Figure 3. NetApp FAS3050 NFS validation topology for test 2. Brocade VCS Fabric Technology and NAS with NFS Validation Test 12 of 52

The NFS clients on RB3 and RB4 continue to access the NetApp FAS3050 on the active/active vlag interface on RB21 and RB22. 2.7.1 Test Procedure Step 1: Run Iometer on the Windows VM for the duration of the test. Step 2: Execute the following command on the desired Brocade VDX switches to validate baseline traffic prior to failure. The following is an example: VDX6730-RB3# show in te 3/0/1 in rate VDX6730-RB3# show in te 3/0/2 in rate Step 3: Now that active traffic on both Brocade ISLs is confirmed, an ISL in the trunk is failed. In this example, Interface TenGbE3/0/1 is failed, which is one of the active ISLs in the trunk between RB3 and RB1. VDX6730-0 RB3(config)# in te 3/0/1 VDX6730-RB3(conf-if-te-3/0/1)# shut Step 4: Execute the following command on the desired Brocade VDX switches to observe traffic levels after interface failure. The following is an example: VDX6730-RB3# show in te 3/0/1 in rate VDX6730-RB3# show in te 3/0/2 in rate Step 5: Execute the following command on the desired Brocade VDX switches to re-enable the failed ISL: VDX6730-RB3# conf Entering configuration mode terminal VDX6730-RB3(config)# in te 3/0/1 VDX6730-RB3(conf-if-te-3/0/1)# no shut Step 6: Execute the following command on the desired Brocade VDX switches to observe traffic levels after interface re-enabling. The following is an example: VDX6730-RB3# show in te 3/0/1 in rate VDX6730-RB3# show in te 3/0/2 in rate Repeat Steps 1 6: Steps 1 6 were also performed for the other Brocade ISLs that are highlighted in red under section 2.7. 2.7.2 Expected Results The expected results for test 2 are to confirm that distributed I/O is not impacted while failing an active Brocade member ISL within the fabric. The NFS clients on RB3 and RB4 continue to access the NetApp FAS3050 on the active/active vlag interface on RB21 and RB22. Brocade VCS Fabric Technology and NAS with NFS Validation Test 13 of 52

All I/Os from the NFS client on RB3 and RB4 fail over to the remaining link on the Brocade trunks going into RB21 and RB22, where the active/active vlag is located on the NFS server. Since the trunks have 20 Gbps of aggregated bandwidth, there should be no loss of throughput, unless there is more than 10 Gbps of I/O going through the fabric when a single 10 Gbps link is failed during the test. When the links in the trunk are re-enabled, the I/Os get evenly redistributed to all of the active links in the trunk. 2.7.3 Actual Results The actual results for test 2 confirm that distributed I/O is not impacted while failing an active Brocade member ISL within the fabric. During the link failure in the Brocade trunk, while I/O was evenly distributed between the two 10 GbE links on the 20 GbE trunk, all I/O was immediately switched to the remaining 10 GbE link. When the link was re-enabled, all I/O was immediately rebalanced between both links again. The same results were observed between other Brocade ISL trunk link failures in the fabric. These failures were performed for the other Brocade ISLs, highlighted in red under section 2.6.5. Again, there was no disruption of I/O from the perspective of the Windows VM/NFS client. The maximum throughput of 119 MB/sec was the same before, during, and after the link failure. 2.8 Test 3: Active Path Failure The purpose of test 3 is to validate that I/O flows are not impacted between the NFS client VM and NetApp FAS3050 while introducing a complete active path failure. This is done by failing both interfaces in the Brocade trunk, which is an active shortest path in the fabric. The links that are failed as part of this test are highlighted below in red. Observe that all I/Os get rerouted via fabric switching in the remaining paths. Brocade VCS Fabric Technology and NAS with NFS Validation Test 14 of 52

Figure 4. NetApp FAS3050 NFS validation topology for test 3. 2.8.1 Test Procedure Step 1: Run Iometer on the Windows VM for the duration of the test. Step 2: Execute the following command on RB3 to validate fabric topology prior to path failure. VDX6730-RB3# show fabric route topology Total Path Count: 8 Src Dst Out Out Nbr Nbr RB-ID RB-ID Index Interface Hops Cost Index Interface BW Trunk ----------------------------------------------------------------------------------- 3 1 2 Te 3/0/2 1 500 7 Te 1/0/7 20G Yes 2 6 Te 3/0/6 1 500 7 Te 2/0/7 20G Yes 4 2 Te 3/0/2 2 1000 7 Te 1/0/7 20G Yes 4 6 Te 3/0/6 2 1000 7 Te 2/0/7 20G Yes 21 2 Te 3/0/2 2 1000 7 Te 1/0/7 20G Yes 21 6 Te 3/0/6 2 1000 7 Te 2/0/7 20G Yes 22 6 Te 3/0/6 2 1000 7 Te 2/0/7 20G Yes 22 2 Te 3/0/2 2 1000 7 Te 1/0/7 20G Yes Brocade VCS Fabric Technology and NAS with NFS Validation Test 15 of 52

In the base testing, it is known that the NFS clients on RB3 and RB4 take the shortest path to the NFS server. In the fabric topology prior to the path failure, note that total path count is 8, and from RB3 to RB1, note that the hop count is 1 and the cost is 500. Step 3: In this example, interfaces TenGigE3/0/1 and TenGigE3/0/2 are failed, to fail the active path that includes both active ISLs in the trunk between RB3 and RB1. VDX6730-RB3(config)# in te 3/0/1 VDX6730-RB3(conf-if-te-3/0/1)# shut VDX6730-RB3(conf-if-te-3/0/1)# in te 3/0/2 VDX6730-RB3(conf-if-te-3/0/2)# shut Step 4: Execute the following command on RB3 to observe fabric topology after path failure: VDX6730-RB3# show fabric route topology Total Path Count: 5 Src Dst Out Out Nbr Nbr RB-ID RB-ID Index Interface Hops Cost Index Interface BW Trunk ----------------------------------------------------------------------------------- 3 1 6 Te 3/0/6 2 1000 7 Te 2/0/7 20G Yes 2 6 Te 3/0/6 1 500 7 Te 2/0/7 20G Yes 4 6 Te 3/0/6 2 1000 7 Te 2/0/7 20G Yes 21 6 Te 3/0/6 2 1000 7 Te 2/0/7 20G Yes 22 6 Te 3/0/6 2 1000 7 Te 2/0/7 20G Yes Step 5: Execute the following command on the desired Brocade VDX switches to observe traffic levels after interface failure. The following is an example: VDX6710-RB21# show interface Port-channel 33 in rate Step 6: Execute the following command on the desired Brocade VDX switches to re-enable the failed ISL: VDX6730-RB3# conf Entering configuration mode terminal VDX6730-RB3(config)# in te 3/0/1 VDX6730-RB3(conf-if-te-3/0/1)# no shut VDX6730-RB3(config)# in te 3/0/2 VDX6730-RB3(conf-if-te-3/0/2)# no shut Brocade VCS Fabric Technology and NAS with NFS Validation Test 16 of 52

Step 7: Execute the following command on the desired Brocade VDX switches to observe traffic levels after interface re-enabling. The following is an example: VDX6710-RB22# show interface Port-channel 33 in rate 2.8.2 Expected Results The expected results for test 3 are to confirm that distributed I/O is not impacted while introducing a complete active path failure. It should be observed that all I/Os are rerouted via fabric switching in the remaining paths. The NFS client on RB3 and RB4 should continue to access the NetApp FAS3050 on the active vlag interface on RB21 and RB22. This test will show the resiliency of the fabric. In this example, the shortest path on the left side from RB3 to RB21 will go down between RB1 and RB3. Through Brocade VCS Ethernet fabric switching, all of the I/Os are redirected through RB2 to RB1, in order to get to RB21. 2.8.3 Actual Results The actual results for test 3 confirm that distributed I/O is not impacted while introducing a complete active path failure. It is observed that all I/Os are rerouted via fabric switching in the remaining paths. As expected, during the shortest path failure on the left side from the NFS client to the NetApp FAS3050, all I/Os were immediately switched over to the alternate path. Again, there was no disruption of I/O from the perspective of the Windows VM/NFS client. From the perspective of the Iometer application, there were no errors, and all I/Os continued to flow through the fabric. Upon re-enabling the shortest path, all I/Os failed back to the shortest path between RB3 and RB1 without any disruption. 2.9 Test 4: Switch Failure The purpose of test 4 is to validate that I/O flows are not impacted between the NFS client VM and NetApp FAS3050 while introducing multiple switch failures in the active path of the Brocade VCS fabric. The first switch to be reloaded, RB2, is the principal switch in the Brocade VCS fabric. The second switch to be reloaded, RB3, is a switch directly connected to an active vlag interface of the NFS client. For this test, all MACs were hashed onto the GbE 0/c of the NetApp FAS3050. Brocade VCS Fabric Technology and NAS with NFS Validation Test 17 of 52

Figure 5. NetApp FAS3050 NFS validation topology for test 4. 2.9.1 Test Procedure Step 1: Validate that all switches in the Brocade VCS fabric are up, and the shortest paths between the NFS client and the server are being used as depicted in test 1. Step 2: Execute the following command on the desired Brocade VDX switches to validate fabric topology prior to failure: VDX6720-RB4# show fabric route topology Total Path Count: 8 Src Dst Out Out Nbr Nbr RB-ID RB-ID Index Interface Hops Cost Index Interface BW Trunk ----------------------------------------------------------------------------------- 4 1 2 Te 4/0/2 1 500 10 Te 1/0/10 20G Yes 2 5 Te 4/0/5 1 500 11 Te 2/0/11 20G Yes 3 5 Te 4/0/5 2 1000 11 Te 2/0/11 20G Yes 3 2 Te 4/0/2 2 1000 10 Te 1/0/10 20G Yes Brocade VCS Fabric Technology and NAS with NFS Validation Test 18 of 52

21 2 Te 4/0/2 2 1000 10 Te 1/0/10 20G Yes 21 5 Te 4/0/5 2 1000 11 Te 2/0/11 20G Yes 22 5 Te 4/0/5 2 1000 11 Te 2/0/11 20G Yes 22 2 Te 4/0/2 2 1000 10 Te 1/0/10 20G Yes Step 3: Execute the following command on any Brocade VDX switch to validate the existing principal switch in the fabric: VDX6720-RB2# show fabric all VCS Id: 1 Config Mode: Local-Only Rbridge-id WWN IP Address Name ---------------------------------------------------------------------------- 1 10:00:00:05:33:46:6A:8D 10.17.83.222 VDX6720-RB1 2 10:00:00:05:33:67:FE:A5 10.17.83.223 > VDX6720-RB2 * 3 10:00:00:05:33:91:3A:E4 10.17.83.224 VDX6730-RB3 4 10:00:00:05:33:67:CA:BC 10.17.82.237 VDX6720-RB4 21 10:00:00:05:33:8C:AD:91 10.17.84.31 VDX6710-RB21 22 10:00:00:05:33:8C:D3:FD 10.17.84.77 VDX6710-RB22 The Fabric has 6 Rbridge(s) Step 4: Run Iometer on the Windows VM for the duration of the test. Step 5: Now that it is confirmed that the principal switch is RB2, the disabling of the RB2 switch can proceed. VDX6720-RB2# reload Warning: Unsaved configuration will be lost. Are you sure you want to reload the switch? [y/n]:y The system is going down for reload NOW!! Broadcast message from root Wed Jan 18 18:44:57 2012... The system is going down for reboot NOW!! Step 6: Execute the following command on the desired Brocade VDX switches to validate the fabric topology after RB2 failure: VDX6720-RB4# show fabric all VCS Id: 1 Config Mode: Local-Only Rbridge-id WWN IP Address Name ---------------------------------------------------------------------------- 1 10:00:00:05:33:46:6A:8D 10.17.83.222 > VDX6720-RB1 3 10:00:00:05:33:91:3A:E4 10.17.83.224 VDX6730-RB3 4 10:00:00:05:33:67:CA:BC 10.17.82.237 VDX6720-RB4 * Brocade VCS Fabric Technology and NAS with NFS Validation Test 19 of 52

21 10:00:00:05:33:8C:AD:91 10.17.84.31 VDX6710-RB21 The Fabric has 4 Rbridge(s) Step 7: Verify that I/Os are still running after principal switch RB2 failure: Step 8: Disable the RB3 switch. VDX6730-RB3# reload Warning: Unsaved configuration will be lost. Are you sure you want to reload the switch? [y/n]:y The system is going down for reload NOW!! Broadcast message from root Wed Jan 18 18:51:54 2012... The system is going down for reboot NOW!! Step 9: Execute the following command on the desired Brocade VDX switches to validate the fabric topology after RB3 failure: VDX6720-RB4# show fabric all VCS Id: 1 Config Mode: Local-Only Rbridge-id WWN IP Address Name ---------------------------------------------------------------------------- 1 10:00:00:05:33:46:6A:8D 10.17.83.222 > VDX6720-RB1 4 10:00:00:05:33:67:CA:BC 10.17.82.237 VDX6720-RB4 * 21 10:00:00:05:33:8C:AD:91 10.17.84.31 VDX6710-RB21 Brocade VCS Fabric Technology and NAS with NFS Validation Test 20 of 52

The Fabric has 3 Rbridge(s) Step 10: Verify that I/Os are still running after principal switch RB3 failure: Step 11: Execute the following command on the desired Brocade VDX switches to observe traffic levels after interface re-enabling. The following is an example: VDX6710-RB21# show interface port-channel 33 in rate 2.9.3 Expected Results The expected results for test 4 are to confirm that distributed I/O is not impacted while introducing multiple switch failures within the active path of the Brocade VCS fabric. It should be observed that all I/Os are rerouted via fabric switching in the remaining paths. During the failure of the switches in the fabric impacting both end-device vlags, all Iometer I/Os will continue to flow. This will show little to no disruption of I/Os during the simultaneous failure of the RB2 principal switch, as well as RB3. Since the NFS client has redundant connections on RB3 and RB4 on its vlag, it gets continuous access to the NetApp FAS3050. In addition to that, the NetApp FAS3050 also has redundant connections on RB21 and RB22 on the vlag. This test shows the resiliency of Brocade VCS Fabric technology and its end devices when vlags are used. Note that in this scenario, when the principal switch RB2 is rebooted, RB22 should also exit the fabric. When RB3 gets rebooted, there should be a total of three switches leaving the fabric, and I/O will continue to flow from the NFS client to the server. 2.9.3 Actual Results The actual results for test 4 confirm that distributed I/O is not impacted while introducing multiple switch failures in the active path of the Brocade VCS fabric. It is observed that all I/Os are rerouted via fabric switching in the remaining paths. As expected, during the failure of the principal switches RB2 and RB3, all Iometer traffic continued to flow between the NFS client and server. Brocade VCS Ethernet fabric switching utilized all available paths between the end devices and failed over I/O within sub-second convergence times. The vlags between the NFS server and NFS client to the Brocade VCS Fabric Technology and NAS with NFS Validation Test 21 of 52

fabric had redundant connections without having to connect to the same physical switch. This resulted in high availability of application traffic between the NFS client and server, even during switch failures. 2.10 Test 5: vmotion with Link/Path Failure The purpose of test 5 is to validate that I/O flows were not impacted throughout the duration of a successful vmotion between ESX servers acting as NFS clients with a link/path failure. In this test scenario, there are two 10 GbE ESX servers dual connected to both RBridges 3 and 4 to the Brocade VCS fabric with NIC Teaming/vLAG, using LACP/Port-Channels 44 and 55. There are two virtual machines between the ESX servers, one RedHat Linux Enterprise 5 and the other a Windows 2008 Server, serving as NFS clients using the NetApp FAS3050 as the VM datastore. For the NFS server, there is 1 GbE NetApp storage that is dual connected to the Brocade VCS fabric with NIC Teaming/vLAG using LACP/Port-Channel 33. See Figure 6. Brocade VCS Fabric Technology and NAS with NFS Validation Test 22 of 52

Figure 6. NetApp FAS3050 NFS validation topology for test 5. 2.10.1 Test Procedure Step 1: Validate that all switches in the Brocade VCS fabric are up and that the shortest paths between the NFS client and server are being used as depicted in test 1. Step 2: Execute the following command on the desired Brocade VDX switches to validate fabric topology prior to failure: VDX6730-RB3# show fabric route topology Total Path Count: 8 Src Dst Out Out Nbr Nbr RB-ID RB-ID Index Interface Hops Cost Index Interface BW Trunk ----------------------------------------------------------------------------------- 3 1 2 Te 3/0/2 1 500 7 Te 1/0/7 20G Yes 2 6 Te 3/0/6 1 500 7 Te 2/0/7 20G Yes 4 6 Te 3/0/6 2 1000 7 Te 2/0/7 20G Yes 4 2 Te 3/0/2 2 1000 7 Te 1/0/7 20G Yes 21 6 Te 3/0/6 2 1000 7 Te 2/0/7 20G Yes 21 2 Te 3/0/2 2 1000 7 Te 1/0/7 20G Yes 22 6 Te 3/0/6 2 1000 7 Te 2/0/7 20G Yes 22 2 Te 3/0/2 2 1000 7 Te 1/0/7 20G Yes Brocade VCS Fabric Technology and NAS with NFS Validation Test 23 of 52

Step 3: Execute the following command on any Brocade VDX switch to validate the member RBridges in the fabric: VDX6730-RB3# show fabric all VCS Id: 1 Config Mode: Local-Only Rbridge-id WWN IP Address Name ---------------------------------------------------------------------------- 1 10:00:00:05:33:46:6A:8D 10.17.83.222 > VDX6720-RB1 2 10:00:00:05:33:67:FE:A5 10.17.83.223 VDX6720-RB2 3 10:00:00:05:33:91:3A:E4 10.17.83.224 VDX6730-RB3 * 4 10:00:00:05:33:67:CA:BC 10.17.82.237 VDX6720-RB4 21 10:00:00:05:33:8C:AD:91 10.17.84.31 VDX6710-RB21 22 10:00:00:05:33:8C:D3:FD 10.17.84.77 VDX6710-RB22 The Fabric has 6 Rbridge(s) Step 4: Validate the vsphere client view of the VM NFS clients and NetApp FAS3050 datastore while on ESX server #2 (on the right of the figure) before vmotion and prior to RB4 failure. Note that from the ESX server #2 on the right, vmnic2 will be disabled when RB4 is failed. This is the vsphere view prior to RB4 failure: Step 5: On the Windows VM, send I/Os throughout the duration of the vmotion and prior to RB4 failure, using Iometer to read/write to the NFS server. Brocade VCS Fabric Technology and NAS with NFS Validation Test 24 of 52

Step 6: On the Linux VM, send I/Os before the vmotion and prior to RB4 failure, using dd to read/write to the NFS server. Step 7: Begin vmotion. Brocade VCS Fabric Technology and NAS with NFS Validation Test 25 of 52

Step 8: Proceed with disabling the RB4 switch. VDX6720-RB4# reload Warning: Unsaved configuration will be lost. Are you sure you want to reload the switch? [y/n]:y The system is going down for reload NOW!! Broadcast message from root Tue Jan 23 14:22:49 2012... The system is going down for reboot NOW!! Step 9: Execute the following command on the desired Brocade VDX switches to validate fabric topology after RB4 failure: VDX6730-RB3# show fabric all VCS Id: 1 Config Mode: Local-Only Rbridge-id WWN IP Address Name ---------------------------------------------------------------------------- 1 10:00:00:05:33:46:6A:8D 10.17.83.222 > VDX6720-RB1 2 10:00:00:05:33:67:FE:A5 10.17.83.223 VDX6720-RB2 3 10:00:00:05:33:91:3A:E4 10.17.83.224 VDX6730-RB3 * 21 10:00:00:05:33:8C:AD:91 10.17.84.31 VDX6710-RB21 22 10:00:00:05:33:8C:D3:FD 10.17.84.77 VDX6710-RB22 The Fabric has 5 Rbridge(s) Step 10: Execute the following command on any Brocade VDX switch to validate the member RBridges in the fabric after RB4 failure: VDX6730-RB3# show fabric route topology Total Path Count: 6 Src Dst Out Out Nbr Nbr RB-ID RB-ID Index Interface Hops Cost Index Interface BW Trunk ----------------------------------------------------------------------------------- 3 1 2 Te 3/0/2 1 500 7 Te 1/0/7 20G Yes 2 6 Te 3/0/6 1 500 7 Te 2/0/7 20G Yes 21 6 Te 3/0/6 2 1000 7 Te 2/0/7 20G Yes 21 2 Te 3/0/2 2 1000 7 Te 1/0/7 20G Yes 22 6 Te 3/0/6 2 1000 7 Te 2/0/7 20G Yes 22 2 Te 3/0/2 2 1000 7 Te 1/0/7 20G Yes Brocade VCS Fabric Technology and NAS with NFS Validation Test 26 of 52

Step 11: Validate the vsphere client view of the VM NFS clients and NetApp FAS3050 datastore during RB4 failure and vmotion: Step 12: Validate the vsphere client view of the VM NFS clients and NetApp FAS3050 datastore during RB4 failure and after successful vmotion: Brocade VCS Fabric Technology and NAS with NFS Validation Test 27 of 52

Step 13: Verify that I/Os are still running after switch RB4 failure that the Windows VM is sending I/Os during vmotion and RB4 failure, using Iometer to read/write to the NFS server. Step 14: Verify that I/Os are still running after switch RB4 failure that the Linux VM is sending I/Os during vmotion and RB4 failure, using dd to write to the NFS server. Brocade VCS Fabric Technology and NAS with NFS Validation Test 28 of 52

2.10.2 Expected Results The expected results for test 5 are to confirm that distributed I/O is not impacted while introducing a switch failure within the active path of a Brocade VCS fabric while VM clients are sending I/Os to the NFS server; they will be moved to another ESX server using vmotion. During the vmotion, there will be a failure on RB4, but there will be no disruption of I/Os. 2.10.3 Actual Results The actual results for test 5 confirm that distributed I/O is not impacted while introducing a switch failure in the active path of the Brocade VCS fabric while VM clients are sending I/Os to the NFS server; they are moved to another ESX server using vmotion. Before, during, and after RB4 failure and vmotion of both VMs, while they were accessing the NFS server, there was no disruption of I/O. Note that one ping from the Windows VM was lost due to an inflight frame that was lost during the failure. It was observed that all I/Os were rerouted via fabric switching in the remaining paths. Brocade VCS Fabric Technology and NAS with NFS Validation Test 29 of 52

3. TEST CASE #2: RED HAT ENTERPRISE LINUX NFS SERVER The topology used for validation is shown below. For this test, one type of NAS storage was used, with Red Hat Enterprise Linux configured as an NFS server. Test Case #2 with Red Hat Enterprise Linux NFS server: This test consists of a 4-node Brocade VCS fabric with a Red Hat Enterprise Linux NFS server. For this test, Spirent is used as a characterization tool for NFS clients. Spirent will emulate many NFS clients accessing the Red Hat Enterprise Linux NFS server share. Each of the NFS clients mounted the NFS share of the Red Hat Enterprise Linux NFS server. The purpose is to demonstrate that the topology of the network can change with limited implication to the application using NFS as an underlying protocol. Please note that this does not include a full configuration or validation of specific features on ESXi, Red Hat, NetApp, and VMware. Also, please note that for this test case, the NetApp FAS3050 has Gigabit Ethernet interfaces. Therefore, a Brocade VDX 6710 was used to extend Gigabit Ethernet to both RB1 and RB2. However, for most cases, it is recommended that you use redundant links for all fabric facing interfaces. 3.1 Test Case #2 Results Test Case #2 was performed successfully, with no issues. The Brocade VCS fabric interoperates with the Network File System (NFS) Protocol using Red Hat Enterprise Linux NFS server. This validation demonstrates that existing deployments using NFS will interoperate with Brocade VCS fabrics and exhibit resiliency to input/output (I/O) between clients and servers in a non-disruptive manner. This is a summary of the tests conducted for Test Case #2. 1. Test 1 validated that baseline I/O is using the shortest equal-cost paths in the fabric. NFS clients on RB3 and RB4 will access the Linux NFS server on the active vlag interface on RB1. 2. Test 2 validated that I/O flows are not impacted between the NFS clients emulated by Spirent Avalanche and Red Hat Enterprise Linux NFS server while introducing a link failure on one link in the Brocade trunk, which is an active shortest path in the fabric. 3. Test 4 validated that I/O flows are not impacted between the NFS clients emulated by Spirent Avalanche and Red Hat Enterprise Linux NFS server while introducing a switch failure in the active path of the Brocade VCS fabric. This summarizes the test results for Test Case #2. Test Description Results 1 Baseline of shortest paths and traffic distribution in the Brocade VCS fabric Pass 2 Perform link failure within a Brocade ISL trunk Pass 3 Perform multiple switch failures within the Brocade VCS fabric Pass Brocade VCS Fabric Technology and NAS with NFS Validation Test 30 of 52

3.2 Topology Figure 7. Red Hat Enterprise Linux NFS server validation topology and components for Test Case #2. 3.3 Hardware Resources The following equipment was used in this configuration: Description Quantity Revisions Brocade VDX 6720-24 3 Brocade NOS 2.1.1 Brocade VDX 6730-32 1 Brocade NOS 2.1.1 Linux NFS Server 1 Red Hat Enterprise Linux Server Release 5.2 (Tikanga) Spirent Avalanche 1 3.6.0 3.4 Compute Resources The following equipment was used in this configuration: Description Quantity Revisions Linux NFS Server 1 HP Proliant DL380G5p 320 GB HD 4 GB RAM & 1020 CNA Brocade VCS Fabric Technology and NAS with NFS Validation Test 31 of 52

3.5 Software Resources The following software was used in this configuration: Description Revision Brocade NOS Brocade NOS 2.1.1 Linux NFS Server Red Hat Enterprise Linux Server release 5.2 (Tikanga) 3.6 Test 1: I/O Verification The purpose of test 1 is to validate baseline I/O is using the shortest equal-cost paths in the fabric. NFS clients on RB3 and RB4 will access the Linux NFS server on the active vlag interface on RB1. Figure 8. Red Hat Enterprise Linux NFS server validation topology for test 1. 3.6.1 Test Procedure Step 1: Execute the following command on all Brocade VDX ISLs and VDX vlag interfaces to confirm traffic flow and that the I/O path is distributed as expected. The following is an example: VDX6730-RB3# show in te 3/0/24 in rate Step 2: Run Spirent Test Center, which emulates the NFS clients on RB3/RB2. Brocade VCS Fabric Technology and NAS with NFS Validation Test 32 of 52

3.6.2 Expected Results The expected results for test 1 is that all I/Os from each of the NFS clients on RB3 and RB4 will be spread evenly across the Brocade trunks going into RB1, where the active vlag on the server is located. 3.6.3 Actual Results The actual results for test 1 confirm that all I/Os from each of the NFS clients on RB3 and RB4 are spread evenly across the Brocade trunks going into RB1, where the active vlag on the server is located. The total bandwidth from the 95 Spirent clients on RB3 (3/0/24) and RB4 (4/0/24) are received on the Egress port going to the NFS server (1/0/23). Brocade VCS Fabric Technology and NAS with NFS Validation Test 33 of 52

3.7 Test 2: Link Failure The purpose of test 2 is to validate that I/O flows are not impacted between the NFS clients emulated by Spirent Avalanche and Red Hat Enterprise Linux NFS server while introducing a link failure on one link in the Brocade trunk, which is an active shortest path in the fabric. The links that are failed as part of this test are highlighted below in red. Figure 9. Red Hat Enterprise Linux NFS server validation topology for test 2. The NFS clients on RB3 and RB4 continue to access the Red Hat Enterprise Linux NFS server on the active vlag interface on RB1. 3.7.1 Test Procedure Step 1: Execute the following command on the desired Brocade VDX switches to validate baseline traffic prior to failure. The following is an example: VDX6730-RB3# show in te 3/0/1 in rate VDX6730-RB3# show in te 3/0/2 in rate Step 2: Run Spirent Avalanche for the duration of the test to emulate the NFS clients. Step 3: Now that active traffic on both Brocade ISLs is confirmed, an ISL in the trunk is failed. In this example, Interface TenGigE3/0/1 on RB3 is failed, which is one of the active ISLs in the trunk between RB3 and RB1. VDX6730-RB3# conf t Entering configuration mode terminal VDX6730-RB3(config)# in te 3/0/1 VDX6730-RB3(conf-if-te-3/0/1)# shut Brocade VCS Fabric Technology and NAS with NFS Validation Test 34 of 52

Step 4: Execute the following command on the desired Brocade VDX switches to observe traffic levels after interface failure. The following is an example: VDX6730-RB3# show in te 3/0/1 in rate VDX6730-RB3# show in te 3/0/2 in rate Step 5: Execute the following command on the desired Brocade VDX switches to re-enable the failed ISL between RB3 and RB1. VDX6730-RB3# conf t Entering configuration mode terminal VDX6730-RB3(config)# in te 3/0/1 VDX6730-RB3(conf-if-te-3/0/1)# no shut Step 6: Execute the following command on the desired Brocade VDX switches to observe traffic levels after interface re-enabling. The following is an example: VDX6730-RB3# show in te 3/0/1 in rate VDX6730-RB3# show in te 3/0/2 in rate Repeat Steps 1 6: Steps 1 6 were also performed for the other Brocade ISLs, highlighted in red under section 3.5.1. 3.7.2 Expected Results The expected results for test 2 are to confirm that distributed I/O is not impacted while failing an active Brocade member ISL within the fabric. The NFS clients on RB3 and RB4 continue to access the Red Hat Enterprise Linux NFS server on the active vlag interface on RB1. All I/Os from each of the NFS clients on RB3 and RB4 fail over to the remaining link on the Brocade trunk that goes into RB1, where the active vlag on the server is located. The total bandwidth from the clients on RB3 (3/0/24) and RB4 (4/0/24) are received on the Egress port going to the NFS server (1/0/23). There should be no loss in I/O, with the exception of in-flight frames during the failure of the link. When the links in the trunk are re-enabled, the I/Os should be evenly redistributed to all of the active links in the trunk. The same results should be observed between other Brocade ISL trunk link failures in the fabric. These failures were performed for the other Brocade ISLs highlighted in red under section 3.5.1. 3.7.3 Actual Results The actual results for test 2 confirm that distributed I/O is not impacted while failing an active Brocade member ISL within the fabric. During the link failure in the Brocade trunk while I/O was evenly distributed between the two 10 GbE links on the 20 GbE trunk; all I/O was immediately switched to the remaining 10 GbE link. When the link was re-enabled, all I/O was immediately rebalanced between both links again. The same results were observed between other Brocade ISL trunk link failures in the fabric. These failures were performed for the other Brocade ISLs highlighted in red under section 3.5.1. Again, there was no disruption of I/O from the perspective of the Windows VM/NFS client. Brocade VCS Fabric Technology and NAS with NFS Validation Test 35 of 52

All I/Os from each of the NFS clients on RB3 and RB4 fail over to the remaining link on the Brocade trunk that goes into RB1, where the active vlag on the server is located. The total bandwidth from the clients on RB3 (3/0/24) and RB4 (4/0/24) are received on the Egress port going to the NFS server (1/0/23). When the links in the trunk are re-enabled, the I/O is evenly redistributed to all of the active links in the trunk. It appears that there was a loss of in-flight NFS frames during the failure, as highlighted in the output below. 3.8 Test 3: Switch Failure The purpose of test 3 is to validate that I/O flows are not impacted between the NFS clients emulated by Spirent Avalanche and Red Hat Enterprise Linux NFS server, while introducing a switch failure in the active path of the Brocade VCS fabric. The switch to be reloaded, RB1, is a switch directly connected to an active vlag interface of the Red Hat Enterprise Linux NFS server. Brocade VCS Fabric Technology and NAS with NFS Validation Test 36 of 52

Figure 10. Red Hat Enterprise Linux NFS server validation topology for test 3. 3.8.1 Test Procedure Step 1: Validate that all switches in the Brocade VCS fabric are up and that the shortest paths between the NFS client and server are being used as depicted in test 1. Step 2: Execute the following command on the desired Brocade VDX switches to validate baseline traffic prior to failure. The following is an example: VDX6730-RB3# show in te 3/0/24 in rate VDX6720-RB2# show in te 2/0/23 in rate Step 3: Execute the following command on the desired Brocade VDX switches to validate fabric topology prior to failure: VDX6730-RB3# show fabric route topology Total Path Count: 6 Src Dst Out Out Nbr Nbr RB-ID RB-ID Index Interface Hops Cost Index Interface BW Trunk ----------------------------------------------------------------------------------- 3 1 1 Te 3/0/1 1 500 6 Te 1/0/6 20G Yes 2 5 Te 3/0/5 1 500 6 Te 2/0/6 20G Yes 4 5 Te 3/0/5 2 1000 6 Te 2/0/6 20G Yes 4 1 Te 3/0/1 2 1000 6 Te 1/0/6 20G Yes 21 1 Te 3/0/1 2 1000 6 Te 1/0/6 20G Yes 22 5 Te 3/0/5 2 1000 6 Te 2/0/6 20G Yes Brocade VCS Fabric Technology and NAS with NFS Validation Test 37 of 52

VDX6720-RB4# show fabric route topology Total Path Count: 6 Src Dst Out Out Nbr Nbr RB-ID RB-ID Index Interface Hops Cost Index Interface BW Trunk ----------------------------------------------------------------------------------- 4 1 1 Te 4/0/1 1 500 11 Te 1/0/11 20G Yes 2 6 Te 4/0/6 1 500 10 Te 2/0/10 20G Yes 3 6 Te 4/0/6 2 1000 10 Te 2/0/10 20G Yes 3 1 Te 4/0/1 2 1000 11 Te 1/0/11 20G Yes 21 1 Te 4/0/1 2 1000 11 Te 1/0/11 20G Yes 22 6 Te 4/0/6 2 1000 10 Te 2/0/10 20G Yes Note that all shortest paths are active, and there are six switches in the fabric. Step 4: Disable RB1 to simulate switch failure, forcing the vlag to use the standby link on RB2 to the NFS server. VDX6720-RB1# reload Warning: Unsaved configuration will be lost. Are you sure you want to reload the switch? [y/n]:y The system is going down for reload NOW!! Broadcast message from root Thu Dec 22 00:04:10 2011... The system is going down for reboot NOW!! Step 5: Execute the following command on the desired Brocade VDX switches to validate fabric topology after RB1 failure: VDX6730-RB3# show fabric route topology Total Path Count: 3 Src Dst Out Out Nbr Nbr RB-ID RB-ID Index Interface Hops Cost Index Interface BW Trunk ----------------------------------------------------------------------------------- 3 2 5 Te 3/0/5 1 500 6 Te 2/0/6 20G Yes 4 5 Te 3/0/5 2 1000 6 Te 2/0/6 20G Yes 22 5 Te 3/0/5 2 1000 6 Te 2/0/6 20G Yes Brocade VCS Fabric Technology and NAS with NFS Validation Test 38 of 52

VDX6720-RB4# show fabric route topology Total Path Count: 3 Src Dst Out Out Nbr Nbr RB-ID RB-ID Index Interface Hops Cost Index Interface BW Trunk ----------------------------------------------------------------------------------- 4 2 6 Te 4/0/6 1 500 10 Te 2/0/10 20G Yes 3 6 Te 4/0/6 2 1000 10 Te 2/0/10 20G Yes 22 6 Te 4/0/6 2 1000 10 Te 2/0/10 20G Yes Step 6: Verify that I/Os are still running after principal switch RB2 failure. The previous standby vlag link to the NFS server on RB2 is now the active link (Egress to NFS server). The following is an example: VDX6720-RB2# show in te 2/0/23 in rate Step 7: Execute the following command on the desired Brocade VDX switches to validate fabric topology after RB1 has become active: VDX6730-RB3# show fabric route topology Total Path Count: 6 Src Dst Out Out Nbr Nbr RB-ID RB-ID Index Interface Hops Cost Index Interface BW Trunk ----------------------------------------------------------------------------------- 3 1 1 Te 3/0/1 1 500 6 Te 1/0/6 20G Yes 2 5 Te 3/0/5 1 500 6 Te 2/0/6 20G Yes 4 5 Te 3/0/5 2 1000 6 Te 2/0/6 20G Yes 4 1 Te 3/0/1 2 1000 6 Te 1/0/6 20G Yes 21 1 Te 3/0/1 2 1000 6 Te 1/0/6 20G Yes 22 5 Te 3/0/5 2 1000 6 Te 2/0/6 20G Yes VDX6720-RB4# show fabric route topology Total Path Count: 6 Src Dst Out Out Nbr Nbr RB-ID RB-ID Index Interface Hops Cost Index Interface BW Trunk ----------------------------------------------------------------------------------- 4 1 1 Te 4/0/1 1 500 11 Te 1/0/11 20G Yes 2 6 Te 4/0/6 1 500 10 Te 2/0/10 20G Yes 3 6 Te 4/0/6 2 1000 10 Te 2/0/10 20G Yes 3 1 Te 4/0/1 2 1000 11 Te 1/0/11 20G Yes 21 1 Te 4/0/1 2 1000 11 Te 1/0/11 20G Yes 22 6 Te 4/0/6 2 1000 10 Te 2/0/10 20G Yes Brocade VCS Fabric Technology and NAS with NFS Validation Test 39 of 52

3.8.2 Expected Results The expected results for test 3 are to confirm that distributed I/O is not impacted while introducing a switch failure within the active path of the Brocade VCS fabric. It should be observed that all I/Os are rerouted via fabric switching in the remaining paths. During the failure of the switches in the fabric impacting both end-device vlags, all Spirent client sessions will continue to flow. This will show little to no disruption of I/Os during the failure of RB1. Since the Spirent NFS server has redundant connections on RB1 and RB2 on its vlag, it gets continuous access to the Spirent NFS clients. This test should show the resiliency of Brocade VCS Fabric technology and its end devices when vlags are used. 3.8.3 Actual Results The actual results for test 3 confirm that distributed I/O is not impacted while introducing multiple switch failures in the active path of the Brocade VCS fabric. It is observed that all Spirent NFS clients are rerouted via fabric switching in the remaining paths. As expected, during the switch failure of RB1, all Spirent client sessions continued to flow to the NFS server. Brocade VCS Ethernet fabric switching utilized all available paths between the end devices and failed over within sub-second convergence times. This test shows the resiliency of Brocade VCS fabric technology and its end devices when vlags are used. Brocade VCS Fabric Technology and NAS with NFS Validation Test 40 of 52

APPENDIX A: TEST CASE #1 Brocade VCS Deployment Considerations: Dynamic LACP vlag with NetApp Step 1: Configure vlag Port-channel Interface on Brocade VDX 6710-RB21 to NetApp FAS3050 GbE 0/c: VDX6710-RB21# conf t Entering configuration mode terminal VDX6710-RB21(config)# interface Port-channel 33 VDX6710-RB21(config-Port-channel-33)# vlag ignore-split VDX6710-RB21(config-Port-channel-33)# speed 1000 VDX6710-RB21(config-Port-channel-33)# description NetApp3050-VDX VDX6710-RB21(config-Port-channel-33)# switchport VDX6710-RB21(config-Port-channel-33)# switchport mode access VDX6710-RB21(config-Port-channel-33)# switchport access vlan 1 VDX6710-RB21(config-Port-channel-33)# no shutdown Step 2: Configure Interface Gigabit Ethernet 21/0/1 on Brocade VDX6710-RB21: VDX6710-RB21# conf t Entering configuration mode terminal VDX6710-RB21(config)# interface GigabitEthernet 21/0/1 VDX6710-RB21(conf-if-gi-21/0/1)# description NetApp3050_e0c VDX6710-RB21(conf-if-gi-21/0/1)# channel-group 33 mode active type standard VDX6710-RB21(conf-if-gi-21/0/1)# lacp timeout long VDX6710-RB21(conf-if-gi-21/0/1)# no shutdown Step 3: Configure vlag Port-channel Interface on Brocade VDX 6710-RB22 to NetApp FAS3050 GbE 0/d: VDX6710-RB22# conf t Entering configuration mode terminal VDX6710-RB22(config)# interface Port-channel 33 VDX6710-RB22(config-Port-channel-33)# vlag ignore-split VDX6710-RB22(config-Port-channel-33)# speed 1000 VDX6710-RB22(config-Port-channel-33)# description NetApp3050-VDX VDX6710-RB22(config-Port-channel-33)# switchport VDX6710-RB22(config-Port-channel-33)# switchport mode access VDX6710-RB22(config-Port-channel-33)# switchport access vlan 1 VDX6710-RB22(config-Port-channel-33)# no shutdown Step 4: Configure Interface Gigabit Ethernet 22/0/1 on Brocade VDX6710-RB22: VDX6710-RB22# conf t Entering configuration mode terminal VDX6710-RB22(config)# interface GigabitEthernet 22/0/1 VDX6710-RB22(conf-if-gi-22/0/1)# channel-group 33 mode active type standard VDX6710-RB22(conf-if-gi-22/0/1)# lacp timeout long VDX6710-RB22(conf-if-gi-22/0/1)# no shutdown Brocade VCS Fabric Technology and NAS with NFS Validation Test 41 of 52

Step 5: Validate vlag Port-channel Interface on Brocade VDX 6710-RB21 to NetApp FAS3050 GbE 0/c: VDX6710-RB21# show interface Port-channel 33 Port-channel 33 is up, line protocol is up Hardware is AGGREGATE, address is 0005.338c.adee Current address is 0005.338c.adee Description: NetApp3050-VDX Interface index (ifindex) is 671088673 Minimum number of links to bring Port-channel up is 1 MTU 2500 bytes LineSpeed Actual : 1000 Mbit Allowed Member Speed : 1000 Mbit Step 6: Validate Interface Gigabit Ethernet 22/0/1 on Brocade VDX6710-RB22: VDX6710-RB21# show interface gigabitethernet 21/0/1 GigabitEthernet 21/0/1 is up, line protocol is up (connected) Hardware is Ethernet, address is 0005.338c.adb6 Current address is 0005.338c.adb6 Fixed Copper RJ45 Media Present Description: NetApp3050_e0c Interface index (ifindex) is 90597048320 MTU 2500 bytes LineSpeed : 1000 Mbit, Duplex: Full Flowcontrol rx: off, tx: off Step 7: Validate vlag Port-channel Interface on Brocade VDX 6710-RB21 to NetApp FAS3050 GbE 0/c: VDX6710-RB22# show interface Port-channel 33 Port-channel 33 is up, line protocol is up Hardware is AGGREGATE, address is 0005.338c.d45a Current address is 0005.338c.d45a Description: NetApp3050-VDX Interface index (ifindex) is 671088673 Minimum number of links to bring Port-channel up is 1 MTU 2500 bytes LineSpeed Actual : 1000 Mbit Allowed Member Speed : 1000 Mbit Flowcontrol rx: off, tx: off Step 8: Validate Interface Gigabit Ethernet 22/0/1 on Brocade VDX6710-RB22: VDX6710-RB22# show interface gigabitethernet 22/0/1 GigabitEthernet 22/0/1 is up, line protocol is up (connected) Hardware is Ethernet, address is 0005.338c.d422 Current address is 0005.338c.d422 Fixed Copper RJ45 Media Present Interface index (ifindex) is 94892015616 MTU 2500 bytes LineSpeed : 1000 Mbit, Duplex: Full Flowcontrol rx: off, tx: off Brocade VCS Fabric Technology and NAS with NFS Validation Test 42 of 52

Brocade VDX Deployment Considerations: Static LACP vlag with ESXi Server and Brocade 1020 CNA Step 1: Configure vlag Port-channel Interface on Brocade VDX 6730-RB3 to ESXi5-vswitch1: VDX6730-RB3# conf t Entering configuration mode terminal VDX6730-RB3(config)# interface Port-channel 44 VDX6730-RB3(config-Port-channel-44)# vlag ignore-split VDX6730-RB3(config-Port-channel-44)# description ESXi5-vSwitch1 VDX6730-RB3(config-Port-channel-44)# switchport VDX6730-RB3(config-Port-channel-44)# switchport mode access VDX6730-RB3(config-Port-channel-44)# switchport access vlan 1 VDX6730-RB3(config-Port-channel-44)# no shutdown Step 2: Configure Interface TenGigabit Ethernet 3/0/10 on Brocade VDX6720-RB3: VDX6730-RB3# conf t Entering configuration mode terminal VDX6730-RB3(config)# interface TenGigabitEthernet 3/0/10 VDX6730-RB3(conf-if-te-3/0/10)# description ESXi5-vmnic2 VDX6730-RB3(conf-if-te-3/0/10)# channel-group 44 mode on type standard VDX6730-RB3(conf-if-te-3/0/10)# no shutdown Step 3: Configure vlag Port-channel Interface on Brocade VDX 6720-RB4 to ESXi5-vswitch1: VDX6720-RB4# conf t Entering configuration mode terminal VDX6720-RB4(config)# interface Port-channel 44 VDX6720-RB4(config-Port-channel-44)# vlag ignore-split VDX6720-RB4(config-Port-channel-44)# description ESXi5-vSwitch1 VDX6720-RB4(config-Port-channel-44)# switchport VDX6720-RB4(config-Port-channel-44)# switchport mode access VDX6720-RB4(config-Port-channel-44)# switchport access vlan 1 VDX6720-RB4(config-Port-channel-44)# no shutdown Step 4: Configure Interface TenGigabit Ethernet 4/0/10 on Brocade VDX6720-RB4: VDX6720-RB4# conf t Entering configuration mode terminal VDX6720-RB4(config)# interface TenGigabitEthernet 4/0/10 VDX6720-RB4(conf-if-te-4/0/10)# description ESXi5-vmnic3 VDX6720-RB4(conf-if-te-4/0/10)# channel-group 44 mode on type standard VDX6720-RB4(conf-if-te-4/0/10)# no shutdown Brocade VCS Fabric Technology and NAS with NFS Validation Test 43 of 52

Step 5: Validate vlag Port-channel Interface on Brocade VDX 6730-RB3 to to ESXi5-vswitch1: VDX6730-RB3# show interface Port-channel 44 Port-channel 44 is up, line protocol is up Hardware is AGGREGATE, address is 0005.3391.3b24 Current address is 0005.3391.3b24 Description: ESXi5-vSwitch1 Interface index (ifindex) is 671088684 Minimum number of links to bring Port-channel up is 1 MTU 2500 bytes LineSpeed Actual : 10000 Mbit Allowed Member Speed : 10000 Mbit Flowcontrol rx: off, tx: off Step 6: Validate Interface TenGigabit Ethernet 3/0/10 on Brocade VDX6730-RB3: VDX6730-RB3# show interface tengigabitethernet 3/0/10 TenGigabitEthernet 3/0/10 is up, line protocol is up (connected) Hardware is Ethernet, address is 0005.3391.3b12 Current address is 0005.3391.3b12 Pluggable media present Description: ESXi5-vmnic2 Interface index (ifindex) is 13288210441 MTU 2500 bytes LineSpeed Actual : 10000 Mbit LineSpeed Configured : Auto, Duplex: Full Flowcontrol rx: off, tx: off Step 7: Validate vlag Port-channel Interface on Brocade VDX 6720-RB4 to to ESXi5-vswitch1: VDX6720-RB4# show interface Port-channel 44 Port-channel 44 is up, line protocol is up Hardware is AGGREGATE, address is 0005.3367.cafc Current address is 0005.3367.cafc Description: ESXi5-vSwitch1 Interface index (ifindex) is 671088684 Minimum number of links to bring Port-channel up is 1 MTU 2500 bytes LineSpeed Actual : 10000 Mbit Allowed Member Speed : 10000 Mbit Flowcontrol rx: off, tx: off Step 8: Validate Interface TenGigabit Ethernet 4/0/10 on Brocade VDX6720-RB4: VDX6720-RB4# show interface tengigabitethernet 4/0/10 TenGigabitEthernet 4/0/10 is up, line protocol is up (connected) Hardware is Ethernet, address is 0005.3367.caea Current address is 0005.3367.caea Pluggable media present Description: ESXi5-vmnic3 Brocade VCS Fabric Technology and NAS with NFS Validation Test 44 of 52

Interface index (ifindex) is 17583177737 MTU 2500 bytes LineSpeed Actual : 10000 Mbit LineSpeed Configured : Auto, Duplex: Full Flowcontrol rx: off, tx: off Brocade VCS Deployment Considerations: Enable Brocade VDX Jumbo Frame Support Note that this must be performed on all RBbridges where a given interface port-channel is located. In this example, interface port-channel 44 is on RBridge 3 and RBridge 4, so we will apply configurations from both RBridge 3 and RBridge 4. Step 1: Example to enable Jumbo Frame Support on applicable VDX interfaces for which Jumbo Frame support is required: VDX6730-RB3# conf t Entering configuration mode terminal VDX6730-RB3(config)# interface Port-channel 44 VDX6730-RB3(config-Port-channel-44)# mtu (<NUMBER:1522-9216>) (9216): 9216 Brocade VCS Deployment Considerations: Enable Ethernet Pause/Flow Control Support Note that this must be performed on all RBbridges where a given interface port-channel is located. In this example, interface port-channel 44 is on RBridge 3 and RBridge 4, so we will apply from both RBridges. Step 1: Enable QOS Flow Control for both tx and rx on vlag Port-channel 44 Interface of VDX 6730-RB4 to ESXi5- vswitch1: VDX6730-RB3# conf t Entering configuration mode terminal VDX6730-RB3(config)# interface Port-channel 44 VDX6730-RB3(config-Port-channel-44)# qos flowcontrol tx on rx on Step 2: Validate vlag Port-channel 44 Interface qos rcv-queue on Brocade VDX 6730-RB4 to to ESXi5-vswitch1 after changes have been made: VDX6730-RB3(config-Port-channel-44)# do show running-config interface Port-channel 44 interface Port-channel 44 vlag ignore-split description ESXi5-vSwitch1 switchport switchport mode access switchport access vlan 1 qos flowcontrol tx on rx on no shutdown Brocade VCS Fabric Technology and NAS with NFS Validation Test 45 of 52

NetApp FAS3050 Deployment Procedure: Volume and VIF Creation For additional details on the NetApp FAS3050 deployment procedure, please see the following: http://www.netapp.com/us/library/technical-reports.html NetApp Volume Creation Step 1: Use NetApp FilerView to Create New Volume Step 2: Use NetApp FilerView to Add a New Volume Step 3: Use NetApp FilerView Select Flexible Volume Type Step 4: Use NetApp FilerView Input Volume Parameters Step 5: Use NetApp FilerView Select Flexible Volume Parameters Step 6: Use NetApp FilerView Select Flexible Volume Size Step 7: Use NetApp FilerView Commit Changes Step 8: Use NetApp FilerView Volume Created Successfully Step 9: Use NetApp FilerView to Verify Using Manage Volumes Selection Step 10: Use NetApp FilerView Complete Final Verification NetApp VIF Creation Step 1: Create VIF with vif create lacp VDX e0c e0d VMware vsphere Client Deployment Procedure: ESXi Datastore Creation For additional details on VMware ESXi datastore creation, please see the following: http://www.vmware.com/files/pdf/vmware_esxi_management_wp.pdf http://www.vmware.com/pdf/perf_best_practices_vsphere5.0.pdf ESX Datastore Creation Step 1: Use VMware vsphere Client Select Datastores from Configuration Menu Step 2: Use VMware vsphere Client Select Storage Type as Network File System Step 3: Use VMware vsphere Client Select NFS Location for vsphere Datastore Step 4: Use VMware vsphere Client Review Network File System To Be Added as a Shared VMFS Datastore Step 5: Use VMware vsphere Client Validate Datastores with Configuration View VMware vsphere Client Deployment Procedure: Virtual Machine Creation For additional details on VMware vsphere VM creation, please see the following: http://www.vmware.com/support/pubs/vsphere-esxi-vcenter-server-pubs.html VM Creation Step 1: Use VMware vsphere Client Select Deploy OVF Template from File Menu Step 2: Use VMware vsphere Client Select Available Image Step 3: Use VMware vsphere Client Select the Source Location Step 4: Use VMware vsphere Client Verify OVF Template Step 5: Use VMware vsphere Client Specify a Name and Location for the Deployed Template Brocade VCS Fabric Technology and NAS with NFS Validation Test 46 of 52

Step 6: Use VMware vsphere Client Select Location of the Virtual Machine Files Step 7: Use VMware vsphere Client Select the Format for the Virtual Disks Step 8: Use VMware vsphere Client Map Networks Used in the OVF Template to Networks in the Inventory Step 9: Use VMware vsphere Client Finish the OVF Template Step 10: Use VMware vsphere Client Deployment Completed Successfully Step 11: Use VMware vsphere Client Verify Summary VMware ESXi Deployment Procedure: NIC Teaming for vswitch For additional details on VMware ESXi NIC teaming for vswitch, please see the following: http://www.vmware.com/files/pdf/virtual_networking_concepts.pdf http://www.vmware.com/pdf/perf_best_practices_vsphere5.0.pdf VMware ESXi NIC Teaming for vswitch Step 1: Use VMware vsphere Client Step 2: Use VMware vsphere Client Step 3: Use VMware vsphere Client Select VMkernel for Networking Hardware Step 4: Use VMware vsphere Client Create a vsphere Standard Switch Step 5: Use VMware vsphere Client Complete VMkernel Connection Settings Step 6: Use VMware vsphere Client Complete VMkernel IP Connection Settings Step 7: Use VMware vsphere Client Verify vsphere Standard Switches are Configured Step 8: Use VMware vsphere Client Select Standard vswitch1 Step 9: Use VMware vsphere Client Select Properties for Standard vswitch1 Step 10: Use VMware vsphere Client Edit vswitch Properties Step 11: Use VMware vsphere Client Select NIC Teaming Step 12: Use VMware vsphere Client Select Route Based on IP Hash for Load Balancing Brocade VCS Fabric Technology and NAS with NFS Validation Test 47 of 52

APPENDIX B: TEST CASE #2 Brocade VCS Fabric Configuration: Static LACP vlag with Red Hat Enterprise Linux NFS Server and Brocade 1020 CNA Step 1: Configure vlag Port-channel Interface on Brocade VDX 6720-RB1 to Red Hat Enterprise Linux NFS Server TenGig Eth 4: VDX6720-RB1# conf t Entering configuration mode terminal VDX6720-RB1(config)# interface Port-channel 11 VDX6720-RB1(config-Port-channel-11)# vlag ignore-split VDX6720-RB1(config-Port-channel-11)# speed 1000 VDX6720-RB1(config-Port-channel-11)# description Linux-NFS-Server-bond0 VDX6720-RB1(config-Port-channel-11)# switchport VDX6720-RB1(config-Port-channel-11)# switchport mode access VDX6720-RB1(config-Port-channel-11)# switchport access vlan 1 VDX6720-RB1(config-Port-channel-11)# no shutdown Step 2: Configure Interface TenGigabit Ethernet 1/0/23 on Brocade VDX6720-RB1: VDX6720-RB1# conf t Entering configuration mode terminal VDX6720-RB1(config)# interface TenGigabitEthernet 1/0/23 VDX6720-RB1(conf-if-te-1/0/23)# description Linux-NFS-Server-eth4 VDX6720-RB1(conf-if-te-1/0/23)# channel-group 11 mode on type standard VDX6720-RB1(conf-if-te-1/0/23)# no shutdown Step 3: Configure vlag Port-channel Interface on Brocade VDX 6720-RB2 to Redhat Linux TenGig Eth 5: VDX6720-RB2# conf t Entering configuration mode terminal VDX6720-RB2(config)# interface Port-channel 11 VDX6720-RB2(config-Port-channel-11)# vlag ignore-split VDX6720-RB2(config-Port-channel-11)# description Linux-NFS-Server-bond0 VDX6720-RB2(config-Port-channel-11)# switchport VDX6720-RB2(config-Port-channel-11)# switchport mode access VDX6720-RB2(config-Port-channel-11)# switchport access vlan 1 VDX6720-RB2(config-Port-channel-11)# no shutdown Step 4: Configure Interface TenGigabit Ethernet 2/0/23 on Brocade VDX6720-RB2: VDX6720-RB2# conf t Entering configuration mode terminal VDX6720-RB2(config)# interface TenGigabitEthernet 2/0/23 VDX6720-RB2(conf-if-te-2/0/23)# description Linux-NFS-Server-eth5 VDX6720-RB2(conf-if-te-2/0/23)# channel-group 11 mode on type standard VDX6720-RB2(conf-if-te-2/0/23)# no shutdown Brocade VCS Fabric Technology and NAS with NFS Validation Test 48 of 52

Step 5: Validate vlag Port-channel Interface on Brocade VDX 6720-RB1 to Redhat Linux TenGig Eth 4: VDX6720-RB1# show interface Port-channel 11 Port-channel 11 is up, line protocol is up Hardware is AGGREGATE, address is 0005.3346.6ad0 Current address is 0005.3346.6ad0 Description: Linux-NFS-Server-bond0 Interface index (ifindex) is 671088651 Minimum number of links to bring Port-channel up is 1 MTU 2500 bytes LineSpeed Actual : 10000 Mbit Allowed Member Speed : 10000 Mbit Flowcontrol rx: off, tx: off Step 6: Validate Interface TenGigabit Ethernet 1/0/23 on Brocade VDX6720-RB1: VDX6720-RB1# show interface tengigabitethernet 1/0/23 TenGigabitEthernet 1/0/23 is up, line protocol is up (connected) Hardware is Ethernet, address is 0005.3346.6ac8 Current address is 0005.3346.6ac8 Pluggable media present Description: Linux-NFS-Server-eth4 Interface index (ifindex) is 4699127830 MTU 2500 bytes LineSpeed Actual : 10000 Mbit LineSpeed Configured : Auto, Duplex: Full Flowcontrol rx: off, tx: off Step 7: Validate vlag Port-channel Interface on Brocade VDX 6720-RB20 to Redhat Linux TenGig Eth 5: VDX6720-RB2# show interface Port-channel 11 Port-channel 11 is up, line protocol is up Hardware is AGGREGATE, address is 0005.3367.fee8 Current address is 0005.3367.fee8 Description: Linux-NFS-Server-bond0 Interface index (ifindex) is 671088651 Minimum number of links to bring Port-channel up is 1 MTU 2500 bytes LineSpeed Actual : 10000 Mbit Allowed Member Speed : 10000 Mbit Flowcontrol rx: off, tx: off Flowcontrol rx: off, tx: off Step 8: Validate Interface Gigabit Ethernet 2/0/23 on Brocade VDX6720-RB2: VDX6720-RB2# show interface tengigabitethernet 2/0/23 TenGigabitEthernet 2/0/23 is up, line protocol is up (connected) Hardware is Ethernet, address is 0005.3367.fee0 Current address is 0005.3367.fee0 Pluggable media present Description: Linux-NFS-Server-eth5 Brocade VCS Fabric Technology and NAS with NFS Validation Test 49 of 52

Interface index (ifindex) is 8994095126 MTU 2500 bytes LineSpeed Actual : 10000 Mbit LineSpeed Configured : Auto, Duplex: Full Flowcontrol rx: off, tx: off Red Hat Enterprise Linux NFS Server: NIC Bonding Configuration For additional details on Red Hat Enterprise Linux NIC bonding, please see the following: http://docs.redhat.com/docs/en-us/red_hat_enterprise_linux/6/html/deployment_guide/sec- Using_Channel_Bonding.html Step 1: Validate Red Hat release: # cat /etc/redhat-release Red Hat Enterprise Linux Server release 5.2 (Tikanga) Step 2: Append ifcfg-bond0: # vi /etc/sysconfig/network-scripts/ifcfg-bond0 Append the following linest: DEVICE=bond0 IPADDR=192.168.1.20 NETWORK=192.168.1.0 NETMASK=255.255.255.0 USERCTL=no BOOTPROTO=none ONBOOT=yes Step 3: Append ifcfg-eth4: # vi /etc/sysconfig/network-scripts/ifcfg-eth4 Modify/append directive as follows: DEVICE=eth4 USERCTL=no ONBOOT=yes MASTER=bond0 SLAVE=yes BOOTPROTO=none Step 4: Append ifcfg-eth5: # vi /etc/sysconfig/network-scripts/ifcfg-eth5 Modify/append directive as follows: DEVICE=eth5 USERCTL=no ONBOOT=yes MASTER=bond0 SLAVE=yes BOOTPROTO=none Brocade VCS Fabric Technology and NAS with NFS Validation Test 50 of 52

Step 5: Append modprobe.conf: # vi /etc/modprobe.conf Append following two lines:alias bond0 bonding options bond0 mode=balance-alb miimon=100 Step 6: Service network restart # modprobe bonding # service network restart Step 7: Verify bond0: # ifconfig bond0 bond0 Link encap:ethernet HWaddr 8C:7C:FF:0A:EB:00 inet addr:192.168.1.2 Bcast:192.168.1.255 Mask:255.255.255.0 inet6 addr: fe80::8e7c:ffff:fe0a:eb00/64 Scope:Link UP BROADCAST RUNNING MASTER MULTICAST MTU:1500 Metric:1 RX packets:93651 errors:0 dropped:0 overruns:0 frame:0 TX packets:52213 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:70915327 (67.6 MiB) TX bytes:18760904 (17.8 MiB) Step 8: Verify eth4/eth5: # ifconfig eth4 eth4 Link encap:ethernet HWaddr 8C:7C:FF:0A:EB:00 inet6 addr: fe80::8e7c:ffff:fe0a:eb00/64 Scope:Link UP BROADCAST RUNNING SLAVE MULTICAST MTU:1500 Metric:1 RX packets:90726 errors:0 dropped:0 overruns:0 frame:0 TX packets:26115 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:70633844 (67.3 MiB) TX bytes:9399071 (8.9 MiB) Memory:fdfc0000-fdffffff # ifconfig eth5 eth5 Link encap:ethernet HWaddr 8C:7C:FF:0A:EB:00 inet6 addr: fe80::8e7c:ffff:fe0a:eb00/64 Scope:Link UP BROADCAST RUNNING SLAVE MULTICAST MTU:1500 Metric:1 RX packets:2935 errors:0 dropped:0 overruns:0 frame:0 TX packets:26101 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:282083 (275.4 KiB) TX bytes:9364213 (8.9 MiB) Memory:fdf40000-fdf7ffff Step 9: Additional validation for Spirent-Related Configuration: # vi /etc/sysconfig/nfs # Port rpc.mountd should listen on. MOUNTD_PORT=892 # exportfs /HDS1 <world> [root@mw-57 ~]# cat /etc/exports /HDS1 *(rw,no_root_squash,sync,insecure) Brocade VCS Fabric Technology and NAS with NFS Validation Test 51 of 52

APPENDIX C: REFERENCES 1 Magic Quadrant for Midrange and High-End NAS Solutions, Gartner #G00210934, March 24, 2011. 2 Worldwide File-Based Storage 2010 2014 Forecast: Consolidation, Efficiency, and Objects Shape Market, IDC #223558, June 2010. 2012 Brocade Communications Systems, Inc. All Rights Reserved. 02/12 GA-SG-433-00 Brocade, Brocade Assurance, the B-wing symbol, DCX, Fabric OS, MLX, SAN Health, VCS, and VDX are registered trademarks, and AnyIO, Brocade One, CloudPlex, Effortless Networking, ICX, NET Health, OpenScript, and The Effortless Network are trademarks of Brocade Communications Systems, Inc., in the United States and/or in other countries. Other brands, products, or service names mentioned may be trademarks of their respective owners. Notice: This document is for informational purposes only and does not set forth any warranty, expressed or implied, concerning any equipment, equipment feature, or service offered or to be offered by Brocade. Brocade reserves the right to make changes to this document at any time, without notice, and assumes no responsibility for its use. This informational document describes features that may not be currently available. Contact a Brocade sales office for information on feature and product availability. Export of technical data contained in this document may require an export license from the United States government. Brocade VCS Fabric Technology and NAS with NFS Validation Test 52 of 52