Ethernet-based Storage Configuration



Similar documents
Improving Scalability for Citrix Presentation Server

Ready Time Observations

Configuration Maximums VMware Infrastructure 3

VMware ESX Server Q VLAN Solutions W H I T E P A P E R

SAN Conceptual and Design Basics

Virtual Machine Encryption Basics

Storage Protocol Comparison White Paper TECHNICAL MARKETING DOCUMENTATION

Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1

Running VirtualCenter in a Virtual Machine

Using esxtop to Troubleshoot Performance Problems

Oracle Database Scalability in VMware ESX VMware ESX 3.5

Enabling NetFlow on Virtual Switches ESX Server 3.5

Performance Analysis Methods ESX Server 3

VMware Virtual Machine Importer User s Manual

Solution Brief Availability and Recovery Options: Microsoft Exchange Solutions on VMware

Using Raw Device Mapping

Cisco engaged Miercom to conduct an independent verification of

What s New in VMware vsphere 4.1 Storage. VMware vsphere 4.1

W H I T E P A P E R. Optimized Backup and Recovery for VMware Infrastructure with EMC Avamar

Microsoft Exchange Solutions on VMware

VirtualCenter Database Performance for Microsoft SQL Server 2005 VirtualCenter 2.5

vsphere Networking ESXi 5.0 vcenter Server 5.0 EN

Managing Remote Access

Guest Operating System. Installation Guide

VirtualCenter Database Maintenance VirtualCenter 2.0.x and Microsoft SQL Server

Configuration Maximums VMware vsphere 4.0

VMWARE Introduction ESX Server Architecture and the design of Virtual Machines

What s New in VMware vsphere 5.5 Networking

A Dell Technical White Paper Dell PowerConnect Team

Best Practices Guide: Network Convergence with Emulex LP21000 CNA & VMware ESX Server

The functionality and advantages of a high-availability file server system

Dell PowerVault MD Series Storage Arrays: IP SAN Best Practices

Integration with Active Directory

Extending SANs Over TCP/IP by Richard Froom & Erum Frahim

VMware Virtual SAN Network Design Guide TECHNICAL WHITE PAPER

Deployments and Tests in an iscsi SAN

Using VMware ESX Server With Hitachi Data Systems NSC or USP Storage ESX Server 3.0.2

vsphere Networking vsphere 5.5 ESXi 5.5 vcenter Server 5.5 EN

Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build

Fiber Channel Over Ethernet (FCoE)

Configuration Maximums VMware vsphere 4.1

SAP Solutions on VMware Infrastructure 3: Customer Implementation - Technical Case Study

IP SAN Best Practices

Using AnywhereUSB to Connect USB Devices

Using VMware ESX Server with IBM System Storage SAN Volume Controller ESX Server 3.0.2

VMware Virtual Machine File System: Technical Overview and Best Practices

Esri ArcGIS Server 10 for VMware Infrastructure

VMware High Availability (VMware HA): Deployment Best Practices

VMware vsphere 5.0 Evaluation Guide

Multipathing Configuration for Software iscsi Using Port Binding

vrealize Operations Manager Customization and Administration Guide

Setup for Failover Clustering and Microsoft Cluster Service

W H I T E P A P E R. Reducing Server Total Cost of Ownership with VMware Virtualization Software

BackupEnabler: Virtually effortless backups for VMware Environments

Building High-Performance iscsi SAN Configurations. An Alacritech and McDATA Technical Note

HP StorageWorks MPX200 Simplified Cost-Effective Virtualization Deployment

ESX Server 3 Configuration Guide Update 2 and later for ESX Server 3.5 and VirtualCenter 2.5

EMC Virtual Infrastructure for Microsoft SQL Server

IP SAN BEST PRACTICES

Best Practice of Server Virtualization Using Qsan SAN Storage System. F300Q / F400Q / F600Q Series P300Q / P400Q / P500Q / P600Q Series

Brocade Solution for EMC VSPEX Server Virtualization

Lecture 15: Congestion Control. CSE 123: Computer Networks Stefan Savage

Computer Networks. Chapter 5 Transport Protocols

Setup for Failover Clustering and Microsoft Cluster Service

Question: 3 When using Application Intelligence, Server Time may be defined as.

VMware Consolidated Backup

FIBRE CHANNEL OVER ETHERNET

Solving complex performance problems in TCP/IP and SNA environments.

Maximizing Backup and Restore Performance of Large Databases

MEASURING WIRELESS NETWORK CONNECTION QUALITY

Over the past few years organizations have been adopting server virtualization

vsphere Networking vsphere 6.0 ESXi 6.0 vcenter Server 6.0 EN

Kronos Workforce Central on VMware Virtual Infrastructure

Frequently Asked Questions: EMC UnityVSA

Network Instruments white paper

Migrating a Windows PC to Run in VMware Fusion VMware Fusion 2.0

Quality of Service Analysis of site to site for IPSec VPNs for realtime multimedia traffic.

VMWARE WHITE PAPER 1

ESXi Configuration Guide

Remote Printing VMware ACE 2

Getting Started with VMware Fusion

IP SAN Fundamentals: An Introduction to IP SANs and iscsi

HP Virtual Connect Ethernet Cookbook: Single and Multi Enclosure Domain (Stacked) Scenarios

Networking Topology For Your System

Configuration Maximums

What s New in VMware vsphere 5.0 Networking TECHNICAL MARKETING DOCUMENTATION

1. The subnet must prevent additional packets from entering the congested region until those already present can be processed.

The Problem with TCP. Overcoming TCP s Drawbacks

VMware vsphere 4.1 Networking Performance

Virtualizing Microsoft Exchange Server 2010 with NetApp and VMware

Setup for Failover Clustering and Microsoft Cluster Service

The Portable Datacenter R E F E R E N C E A R C H I T E C T U R E

Configuring iscsi Multipath

B-2 Analyzing TCP/IP Networks with Wireshark. Ray Tompkins Founder of Gearbit

Improving Effective WAN Throughput for Large Data Flows By Peter Sevcik and Rebecca Wetzel November 2008

New Features in SANsymphony -V10 Storage Virtualization Software

VXLAN Performance Evaluation on VMware vsphere 5.1

WHITE PAPER. VMware Infrastructure 3 Pricing, Packaging and Licensing Overview

Balancing CPU, Storage

Simplify VMware vsphere* 4 Networking with Intel Ethernet 10 Gigabit Server Adapters

Transcription:

VMWARE TECHNICAL NOTE VMware ESX Server Ethernet-based Storage Configuration Using Ethernet for a storage medium is not a new idea. Network file systems such as NFS and SMB/CIFS are entering their third decade of use. Network-based storage is popular because it allows data to be shared among many users using a medium and technology that already exists. Newer generations of network-based storage attempt to share existing infrastructure, while combining ideas from the storage industry, such as fast, reliable transport of data. Gigabit Ethernet, for speed, and TCP, for reliability, represent this combination for the latest versions of NFS and iscsi. iscsi uses the sharing and virtualization capabilities of a traditional SAN, but also includes the low-cost infrastructure of a traditional Ethernet. For reasons of security and performance, many companies set up isolated networks for networkbased storage. Data passed on a shared network is more susceptible to inspection and modification. Storage data will be slowed by competing with other data on the network. For iscsi configurations, VMware only supports isolated networks. This technical note examines key issues in configuring iscsi storage with virtual machines running on an ESX Server system. It covers the following topics: Storage Networks on page 2 A Potential Configuration Pitfall on page 2 Performance Tuning on page 4 Difference from Fibre Channel on page 5 Adapters That Do Not Support Fast Retransmission on page 5 1

Storage Networks A typical storage network consists of a collection of computers connected to a collection of storage devices through a network of switches. It is common to have numerous computers accessing the same storage. The following figure shows several computer systems connected to network storage through an Ethernet switch. In this case, all of the systems communicate with the storage through a single link between the switch and the storage. It's also common to have a single computer accessing numerous different storage devices, especially in the campus environment, with numerous NFS-mounted file systems. In the case of systems running ESX Server, each computer can actually represent many virtual machines connected to the storage device. In the previous figure, the number of virtual machines would be a large multiple of the three physical systems, all sharing the storage. A Potential Configuration Pitfall A common way of configuring a system is similar to the figure on the previous page. Many systems can access the storage in this case, and storage can potentially be shared across the various systems. A virtual machine can be migrated from one of these systems to another, and data stores are available to any of the systems. However, the configuration means that each system is connected through a Gigabit Ethernet link to the switch, which is also connected to the storage through one Gigabit Ethernet link. In most configurations, with modern switches and typical traffic, this won't cause a problem. When systems read data from storage, the best the storage can do is send enough data to fill the gigabit link between the storage and switch. This means it is unlikely that any single system or virtual machine will get gigabit, or wire speed read data transfers, but this is the expected situation when many systems share one storage device. When writing data to storage, multiple systems or virtual machines might attempt to fill their links. As the following figure shows, when this happens, the switch between the systems and the storage has to drop data because it has a 1Gb connection to the storage device, but more than 1Gb of traffic to send to the storage device. An excess of data on the two connections from 2

the systems on the left that can't be carried on the 1Gb connection to the storage is dropped by the switch. In this case, network packets are dropped by the switch because it's limited to the amount of data it can transmit by the speed of the link between it and storage. TCP is a reliable transmission protocol that ensures that dropped packets are retried and eventually reach their destination. Recovering from dropped network packets results in large performance degradation. In addition to time spent determining that data was dropped, the retransmission uses network bandwidth that could otherwise be used for current transactions. TCP is designed to recover from occasional dropped packets and retransmits them quickly and seamlessly. When packets are discarded by the switch with any regularity, network throughput suffers significantly. The network becomes congested with requests to resend data and with the resent packets, and less data is actually transferred than in a network without congestion. Most quality Ethernet switches have some capability of buffering, or storing, data and giving every device attempting to send data an equal chance to get to the destination. This ability to buffer some transmissions, combined with many systems limiting the number of outstanding commands, allows small bursts from several systems to be sent to a storage system in turn. If the transactions are large, though, and there are multiple systems trying to send data through a single output port, they can exceed a switch's ability to buffer one request while another is transmitted. In this case, again, the switch drops the data it cannot send, and requests to retransmit and retransmitted data packets follow. For example, if an Ethernet switch can buffer 32K on an input port, but the system connected to it thinks it can send 256K to the storage device, some of the data is dropped. Most managed switches provide information on dropped packets. *: interface is up IHQ: pkts in input hold queue IQD: pkts dropped from input queue OHQ: pkts in output hold queue OQD: pkts dropped from output queue RXBS: rx rate (bits/sec) RXPS: rx rate (pkts/sec) TXBS: tx rate (bits/sec) TXPS: tx rate (pkts/sec) TRTL: throttle count Interface IHQ IQD OHQ OQD RXBS RXPS TXBS TXPS TRTL --------------------------------------------------------------------- * GigabitEthernet0/1 3 9922 0 0 476303000 62273 477840000 63677 0 In this example from a Cisco switch, the bandwidth used is 476303000 bits/second, which is less than half of gigabit wire speed. Even so, the port is buffering incoming packets and has dropped 3

quite a few. The final line of this interface summary indicates this port has already dropped almost 10,000 inbound packets in the IQD column. If a network regularly gets into this situation and packets are frequently dropped, performance will be less than optimal. Performance Tuning Configuration changes to avoid this problem involve making sure several input Ethernet links are not funneled into one output link, resulting in an oversubscribed link. Any time a number of links transmitting near capacity are switched to a smaller number of links, oversubscription is a possibility. There usually is no problem switching many links into one. A few dropped packets are expected and do not result in serious performance degradation. The problem is a result of switching many heavily used links into one. In this situation, the oversubscribed links can't handle the traffic and packets are dropped. As mentioned in the previous example, two 1Gb links cannot be forced onto a single gigabit link, and the result is dropped packets. Whenever the number of input links exceeds the number of output links, oversubscription and performance degradation can occur. The problem exists even when links are oversubscribed momentarily, as when two systems attempt to write data to the same storage port, and the write request is larger than the switch s capability to buffer requests. In such situations, Ethernet packets are dropped. Performance degrades in this situation. As a rule of thumb, applications or systems that write a lot of data to storage, such as data acquisition or transaction logging systems, should not share Ethernet links to a storage device. These types of applications perform best with multiple connections to storage devices. In the following figure, there are multiple connections from the switch to the storage, allowing more than 1Gb of data to be sent to the storage. Using VLANs or VPNs does not provide a suitable solution to the problem of link oversubscription in shared configurations. VLANs and other virtual partitioning of a network provide a way of logically designing a network, but don't change the physical capabilities of links and trunks between switches. When storage traffic and other network traffic end up sharing physical connections, as they would with a VPN, the possibility for oversubscription and lost packets exists. The same is true of VLANs that share interswitch trunks. Performance design for a storage network must take into account the physical constraints of the network, not logical allocations. Even under heavy loads in this configuration, the storage processor might limit the total throughput. If you are working with systems in which the write load is heavy, you may need to 4

assign separate storage processors to separate systems to handle the amount of traffic bound for the storage. As the following figure shows, splitting the load across storage processors might improve performance. Difference from Fibre Channel Fibre Channel SANs do not suffer from the congestion and oversubscription problems as readily because of the Fibre Channel protocol. Fibre Channel includes a mechanism for port-to-port throttling, not just end-to-end throttling. This means a Fibre Channel switch can limit the amount of data a connected system sends and avoids situations in which information must be dropped. Therefore, Fibre Channel does not degrade into a retransmission recovery scheme that saps the storage network s performance, and is more capable of operating near wire speed. Adapters That Do Not Support Fast Retransmission Some iscsi adapters that do not support fast retransmission can exhibit slow performance in congested networks or when retransmissions are common. The symptom of this is I/O stalling, or stopping completely, for a few seconds at a time when these conditions exist. In extreme cases, this results in instances of ESX Server or the guest operating system attempting to abandon outstanding commands by issuing SCSI aborts. These aborts are logged in the VMkernel log. The aborted commands do not represent lost data or any compromise of data integrity. In the worst case, these adapters might wait more than half a second between retransmission requests, effectively freezing I/O until all requests are fulfilled. In some VMware observations, the number of retransmission requests was more than 10, meaning no I/O happened for five seconds. The following trace shows TCP activity between an adapter and a storage array, with the adapter waiting longer than half a second between each request. This is an extreme case that happens only on a congested network with frequently dropped packets. The Time column represents elapsed time in seconds in the trace. The Info column shows ACKs corresponding to each previous retransmission request. Some output has been removed from the Info column for readability. 5

No. Time Source Destination Protocol Info 52831 4.158316 <ESX adapter> <storage> TCP 52832 4.158339 <ESX adapter> <storage> TCP 52833 4.549148 <ESX adapter> <storage> TCP [TCP Retransmission] 52834 4.549404 <storage> <ESX adapter> TCP 3260 > 29812 [ACK] 52835 5.149050 <ESX adapter> <storage> TCP [TCP Retransmission] 52836 5.149252 <storage> <ESX adapter> TCP 3260 > 29812 [ACK] 52837 5.749079 <ESX adapter> <storage> TCP [TCP Retransmission] 52838 5.749083 <storage> <ESX adapter> TCP 3260 > 29812 [ACK] 52839 6.348860 <ESX adapter> <storage> TCP [TCP Retransmission] 52840 6.348972 <storage> <ESX adapter> TCP 3260 > 29812 [ACK] 52841 6.948882 <ESX adapter <storage> TCP [TCP Retransmission] 52842 6.948888 <storage> <ESX adapter> TCP 3260 > 29812 [ACK] 52843 7.548775 <ESX adapter> <storage> TCP [TCP Retransmission] 52844 7.548780 <storage> <ESX adapter> TCP 3260 > 29812 [ACK] 52845 8.148679 <ESX adapter> <storage> TCP [TCP Retransmission] 52846 8.148683 <storage> <ESX adapter> TCP 3260 > 29812 [ACK] In this example, nearly four seconds elapse while requesting data retransmissions. Because the adapter is stalled (waiting for retransmitted data), no other I/O is going on in this case. If stalls described here are seen when using iscsi with ESX Server, follow suggestions described in the Performance Tuning section earlier in this document. Ensure there are not multiple systems performing frequent, large-block write operations to a single storage port. In environments where this type of activity is likely to happen, dedicate additional links between the switch and the storage or additional storage processors to those systems. Note that the some adapters demonstrate poor characteristics that make the problem worse, but frequently dropped packets and network congestion reduce the performance of any network. VMware, Inc. 3145 Porter Drive Palo Alto, CA 94304 www.vmware.com Copyright 1998-2006 VMware, Inc. All rights reserved. Protected by one or more of U.S. Patent Nos. 6,397,242, 6,496,847, 6,704,925, 6,711,672, 6,725,289, 6,735,601, 6,785,886, 6,789,156, 6,795,966, 6,880,022 6,961,941, 6,961,806 and 6,944,699; patents pending. VMware, the VMware boxes logo and design, Virtual SMP and VMotion are registered trademarks or trademarks of VMware, Inc. in the United States and/or other jurisdictions. Microsoft, Windows and Windows NT are registered trademarks of Microsoft Corporation. Linux is a registered trademark of Linus Torvalds. All other marks and names mentioned herein may be trademarks of their respective companies. Revision 20060607 Item: ESX-ENG-Q206-245 6