Pump Up Your Network Server Performance with HP- UX



Similar documents
HP-UX 11i TCP/IP Performance White Paper

Performance and Recommended Use of AB545A 4-Port Gigabit Ethernet Cards

Application Note. Windows 2000/XP TCP Tuning for High Bandwidth Networks. mguard smart mguard PCI mguard blade

TCP Offload Engines. As network interconnect speeds advance to Gigabit. Introduction to

HP Auto Port Aggregation (APA) Release Notes

D1.2 Network Load Balancing

Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build

Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck

Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks. An Oracle White Paper April 2003

TCP tuning guide for distributed application on wide area networks 1.0 Introduction

Network Performance Optimisation and Load Balancing. Wulf Thannhaeuser

Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study

Optimizing TCP Forwarding

Building High-Performance iscsi SAN Configurations. An Alacritech and McDATA Technical Note

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago

N5 NETWORKING BEST PRACTICES

DEPLOYMENT GUIDE Version 1.1. Configuring BIG-IP WOM with Oracle Database Data Guard, GoldenGate, Streams, and Recovery Manager

The Advantages of Multi-Port Network Adapters in an SWsoft Virtual Environment

Virtualization: TCP/IP Performance Management in a Virtualized Environment Orlando Share Session 9308

Accelerating High-Speed Networking with Intel I/O Acceleration Technology

Frequently Asked Questions

From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller

B-2 Analyzing TCP/IP Networks with Wireshark. Ray Tompkins Founder of Gearbit

Mobile Communications Chapter 9: Mobile Transport Layer

Getting the most TCP/IP from your Embedded Processor

Introduction to Mainframe (z/os) Network Management

Linux NIC and iscsi Performance over 40GbE

Windows TCP Chimney: Network Protocol Offload for Optimal Application Scalability and Manageability

TCP and Wireless Networks Classical Approaches Optimizations TCP for 2.5G/3G Systems. Lehrstuhl für Informatik 4 Kommunikation und verteilte Systeme

Scalability and Performance with Apache 2.0

Deploying Riverbed wide-area data services in a LeftHand iscsi SAN Remote Disaster Recovery Solution

QuickSpecs. HP NC532m Dual Port Flex-10 10GbE Multifunction BL-c Adapter. What's New Support for BL2x220c G6 server blade

Gigabit Ethernet Design

Windows Server Performance Monitoring

vsphere Networking vsphere 6.0 ESXi 6.0 vcenter Server 6.0 EN

Troubleshooting Tools

STANDPOINT FOR QUALITY-OF-SERVICE MEASUREMENT

New!! - Higher performance for Windows and UNIX environments

TCP Labs. WACREN Network Monitoring and Measurement Workshop Antoine Delvaux perfsonar developer

Top 10 Tips for z/os Network Performance Monitoring with OMEGAMON. Ernie Gilman IBM. August 10, 2011: 1:30 PM-2:30 PM.

LESSON Networking Fundamentals. Understand TCP/IP

PERFORMANCE TUNING ORACLE RAC ON LINUX

Iperf Tutorial. Jon Dugan Summer JointTechs 2010, Columbus, OH

TFTP TRIVIAL FILE TRANSFER PROTOCOL OVERVIEW OF TFTP, A VERY SIMPLE FILE TRANSFER PROTOCOL FOR SIMPLE AND CONSTRAINED DEVICES

IP Storage: The Challenge Ahead Prasenjit Sarkar, Kaladhar Voruganti Abstract 1 Introduction

Top 10 Tips for z/os Network Performance Monitoring with OMEGAMON Ernie Gilman

Evaluation Report: Emulex OCe GbE and OCe GbE Adapter Comparison with Intel X710 10GbE and XL710 40GbE Adapters

VPN over Satellite A comparison of approaches by Richard McKinney and Russell Lambert

Integrated Network Acceleration Features.

Technote: AIX EtherChannel Load Balancing Options

Lecture 8 Performance Measurements and Metrics. Performance Metrics. Outline. Performance Metrics. Performance Metrics Performance Measurements

bbc Adobe LiveCycle Data Services Using the F5 BIG-IP LTM Introduction APPLIES TO CONTENTS

Building Enterprise-Class Storage Using 40GbE

MEASURING WORKLOAD PERFORMANCE IS THE INFRASTRUCTURE A PROBLEM?

Doubling the I/O Performance of VMware vsphere 4.1

IP SAN Best Practices

Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking

PE10G2T Dual Port Fiber 10 Gigabit Ethernet TOE PCI Express Server Adapter Broadcom based

I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology

TCP Tuning Techniques for High-Speed Wide-Area Networks. Wizard Gap

Measure wireless network performance using testing tool iperf

iscsi Top Ten Top Ten reasons to use Emulex OneConnect iscsi adapters

First Midterm for ECE374 02/25/15 Solution!!

vsphere Networking vsphere 5.5 ESXi 5.5 vcenter Server 5.5 EN

vsphere Networking ESXi 5.0 vcenter Server 5.0 EN

VMWARE WHITE PAPER 1

OneCommand NIC Teaming and VLAN Manager

Networking Driver Performance and Measurement - e1000 A Case Study

HP and IPv6 Deployment. Bill Medlin HP-UX IPv6 Project Manager

RFC 6349 Testing with TrueSpeed from JDSU Experience Your Network as Your Customers Do

Deployments and Tests in an iscsi SAN

Performance Tuning best pracitces and performance monitoring with Zabbix

ICOM : Computer Networks Chapter 6: The Transport Layer. By Dr Yi Qian Department of Electronic and Computer Engineering Fall 2006 UPRM

TCP/IP Optimization for Wide Area Storage Networks. Dr. Joseph L White Juniper Networks

Using TrueSpeed VNF to Test TCP Throughput in a Call Center Environment

Intel Data Direct I/O Technology (Intel DDIO): A Primer >

CSIS CSIS 3230 Spring Networking, its all about the apps! Apps on the Edge. Application Architectures. Pure P2P Architecture

theguard! Service Management Center (Valid for Version 6.3 and higher)

Master s Thesis Evaluation and implementation of a network performance measurement tool for analyzing network performance under heavy workloads

A Congestion Control Algorithm for Data Center Area Communications

Transport Layer Protocols

UDR: UDT + RSYNC. Open Source Fast File Transfer. Allison Heath University of Chicago

The Problem with TCP. Overcoming TCP s Drawbacks

Testing Virtual Machine Performance with VMware vsphere 4 on 10 Gigabit Networks

VXLAN Performance Evaluation on VMware vsphere 5.1

Enabling High performance Big Data platform with RDMA

TCP Performance Management for Dummies

SILVER PEAK ACCELERATION WITH EMC VSPEX PRIVATE CLOUD WITH RECOVERPOINT FOR VMWARE VSPHERE

3G Converged-NICs A Platform for Server I/O to Converged Networks

Lecture Computer Networks

Introduction to Computer Security Benoit Donnet Academic Year

Leveraging NIC Technology to Improve Network Performance in VMware vsphere

Procedure: You can find the problem sheet on Drive D: of the lab PCs. 1. IP address for this host computer 2. Subnet mask 3. Default gateway address

Boosting Data Transfer with TCP Offload Engine Technology

ZEN LOAD BALANCER EE v3.04 DATASHEET The Load Balancing made easy

Names & Addresses. Names & Addresses. Hop-by-Hop Packet Forwarding. Longest-Prefix-Match Forwarding. Longest-Prefix-Match Forwarding

Transcription:

Pump Up Your Network Server Performance with HP- UX Paul Comstock Network Performance Architect Hewlett-Packard 2004 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice

Purpose of this presentation Understand the factors affecting network performance, and what you can do about them Survey hardware and software options for HP-UX network servers Learn the network configuration and tuning parameters affecting performance 2

Benchmarking Tools SPEC benchmarks (www.specbench.org) SPECweb99: static (70%) and dynamic (30%) HTTP SPECweb99_SSL: w/ssl encryption/decryption SPECweb2004: Under development new workloads such as banking, e-commerce. Netperf (www.netperf.org) Publicly available from HP Measures maximum throughput (Stream) and transactional (Request-Response) performance Your application benchmark 3

Performance Tools HP-UX commands ifconfig/lanscan/lanadmin ndd netstat (-s) ping (for roundtrip time) top traceroute (for multi-hop networks) DSPP Developer Edge tools (www.hp.com) vsar caliper (for Itanium) HP-UX Internet Express (software.hp.com) tcpdump Glanceplus (managementsoftware.hp.com) 4

Where is the bottleneck? Application Memory Bandwidth Protocol CPU

Sample netstat s output (partial)! " #!$# # %&# '# # #" ("&# $# & #( # ) # $# #"# $## 6

Increase your bandwidth Use 1 Gigabit Ethernet NICs instead of 100BT Use a NIC with offload features Trunk multiple interfaces using HP Auto Port Aggregation (APA) (software.hp.com) One of today s CPUs can run a GigE link at full speed For scalability, use multiple NICs Spread device interrupts using HP-UX Interrupt Migration (software.hp.com) 7

Interrupt Migration intctl command * +,-.&! # # / 01 01 22222222222222222222222222222222222222222222222222222222222222222222222222,,, 3+../0,456/,,,)7#38/80/9-3:1,,,)7#38/80/); - 8( <,,,)7#38/80/); - 8( < Spread high speed network devices between CPUs Other devices, such as disks, may also be a concern depending on usage 8

Checksum Offload TCP Segmentation Offload

Checksum Offload (CKO) Performs inbound and outbound TCP/UDP checksum calculations in hardware, offloading the host CPU Available for all HP-UX Gigabit Ethernet hardware Currently done for IPv4 only on HP-UX Example: ""( " (2=;.'4>?@1/@85'>;AA0AB'C;350/@85'/D? EEE """"""EEE 10

TCP Segmentation Offload (TSO) IS: Segmentation of outbound data into IP datagrams in the NIC Required TCP/IP stack and NIC support Builds on CKO and offloads even more host processing Currently IPv4 only on HP-UX Uses a large virtual MTU (VMTU) internally, standard MTU on the wire IS NOT: Not a new protocol on the wire Not jumbo frames 11

TSO Software Transport Optional Upgrade Release (TOUR) 2.2 GigEtherEnh-01: Enhancement Software for GigEther-01 Both are free from software.hp.com Configuration through lanadmin: # lanadmin x vmtu <ppa> Driver/Hardware supports TCP Segmentation Offload. Current VMTU = 32160 12

New Offload Technologies Even more network processing may be offloaded in the future, as network speeds increase New technologies that provide network offload capability include RDMA, TCP Offload Engine (TOE), ETA, and iscsi. These include TCP and non-tcp based technologies For more information, see break-out session What Is RDMA? 13

How much do offloads boost performance? The answer is always It depends. A wise computer science instructor

Avoidance Maneuvers

Programming with Sendfile " A@C< " &"" &#(& 8FA?.808 * #=,E& G7 " ' "'""7""' G7 ' #! H& ' " (I Sendfile avoids copying between file system and network buffers for TCP socket applications that send all or part of a file across the network Used by web servers (Zeus, Apache), and ftp on all versions of HP-UX 16

Network Server Accelerator NSA HTTP available for free from software.hp.com Uses a memory based cache to handle repetitive HTTP GET requests for static content Transparent to web server Avoids multiple socket system calls needed to accept a new connection and perform a web transaction Performance boost will vary depending on how much of the workload is static web requests. Limitations: doesn t help with dynamic or encrypted content For more info see break-out session Accelerating Web Server Performance on HP-UX Using NSA HTTP 17

Configuration and Tuning

Network Stack Configuration A number of network tunables are commonly modified on big servers or in high performance environments tcphashsz (system tunable) default 2048; tune up to 64K for large configurations tcp_conn_request (ndd tunable) default 4096; good in most cases; be sure to use a large backlog when calling listen(2) socket_caching_tcp (ndd_tunable) default 1 (on); use a number greater than 512 based on number of simultaneous TCP connections in use SO_SENDBUF/SO_RCVBUF (setsockopt(2)) default 32768; SO_RCVBUF sets the TCP receive window; SO_SENDBUF helps determine when outbound flow control occurs 19

Determining the Receive Buffer For long, fat pipes (LFPs), a large receive buffer may be needed to use all of the available bandwidth. LFPs have a long round trip time (RTT), and high (fat) bandwidth, so lots of data can be in transit The minimum buffer can be determined by the formula rcvbuf = RTT * BW RTT can be determined with ping, or more accurately on actual TCP connections using tcpdump For example, on a 100 Mbit network has a 80 ms round trip time. The rcvbuf should be 100,000,000 b/s *.08 s = 8 Mbits (1 MB) 20

Parameters for Networks with Special Needs TCP Selective Acknowledgement (SACK) RFC 2018, uses option fields in TCP header Faster retransmission of multiple gaps in sequence space tcp_sack_enable (ndd) default 2 (don t initiate SACK) tcp_smoothed_rtt (ndd) default 0; can be used for networks with volatile delay behavior such as those with satellite-based and cellular links tcp_rexmit_interval_min/tcp_rexmit_interval_max (ndd) default 500 ms/60 sec; not usually changed, as timerbased retransmissions are not that common, and the actual interval is based on RTT measurements TCP_NODELAY (setsockopt) default 0; avoids delays in transmission of small segments (Nagle algorithm), but won t help system-wide performance 21

Anatomy of a SPECweb99 Result How to read a SPECweb disclosure Examples of tuning parameters from an actual benchmark http://www.specbench.org/web99/results/res2004q1/web99-20040211-00259.html 22

References Transport Optional Upgrade Release (TOUR) 2.0 FAQ (HP-UX 11i v1, HP-UX 11i v2), docs.hp.com/hpux/netcom Network Server Accelerator HTTP PerformanceWhite Paper (HP-UX 11i v1), docs.hp.com/hpux/internet PCI-X 2 Gigabit Fibre Channel and Gigabit Ethernet Performance Paper (HP-UX 11i v1, HP-UX 11i v2), docs.hp.com/hpux/netcom PCI Gigabit Ethernet Performance on HP Server rp7410 (HP-UX 11i v1), docs.hp.com/hpux/netcom Using APA to Build a Screaming Fast Network Server Connection, docs.hp.com/hpux/netcom Running SPECweb99 with Zeus, Zeus Technology, http://support.zeus.com/doc/tech/specweb99.pdf Web Servers for HP-UX, http://www.hp.com/products1/unix/webservers 23

Co-produced by: