EVALUATING NETWORK BUFFER SIZE REQUIREMENTS
|
|
|
- Tracey Higgins
- 10 years ago
- Views:
Transcription
1 EVALUATING NETWORK BUFFER SIZE REQUIREMENTS for Very Large Data Transfers Michael Smitasin Lawrence Berkeley National Laboratory (LBNL) Brian Tierney Energy Sciences Network (ESnet)
2 [ 2 ]
3 Example Workflow Step 1 Analyze raw data at Beamline in real time and adjust experiment [ 3 ]
4 Example Workflow Step 2 Store raw data [ 4 ]
5 Example Workflow Step 3 Analyze stored data locally [ 5 ]
6 Example Workflow Step 4 Transfer to supercomputers over WAN [ 6 ]
7 Example Workflow Step 5 Analyze processed results and adjust experiment [ 7 ]
8 LAN Transfer Bits per 10 millisecs 1 ~1ms RTT 1 100Mb/0.01sec = 10Gbps [ 8 ]
9 WAN 1 Transfer Bits per 10 millisecs 2 70ms RTT The WAN just sucks, right? 1 70ms simulated RTT via netem 2 100Mb/0.01sec = 10Gbps [ 9 ]
10 Uncongested vs Congested WAN Transfers Bits per 1 millisec 2 captured via passive tap Uncongested Congested 3 Data Transfer (TCP) Background Traffic (UDP) 1 70ms simulated RTT via netem 2 10Mb/0.001sec = 10Gbps 3 Adding 2Gbps UDP Traffic [ 10 ]
11 Real World 70ms RTT 35ms x2 Special thanks to Mark Lukasczyk at Brookhaven National Laboratory for providing far-end test servers [ 11 ]
12 Uncongested vs Congested WAN Transfers Real World tests California to New York 1 Optical tap / LBNL border 1 ~ 70ms real RTT [ 12 ]
13 Impact of packet loss at different distances Source: [ 13 ]
14 TCP s Congestion Control w/ insufficient buffers 50ms simulated RTT Congestion w/ 2Gbps UDP traffic HTCP / Linux [ 14 ]
15 TCP s Congestion Control w/ sufficient buffers 50ms simulated RTT Congestion w/ 2Gbps UDP traffic HTCP / Linux [ 15 ]
16 Congestion = Packet Loss = Poor Performance so what do we do? With this? Replace (something like) this Royalty checks to: Michael Smitasin PO Box Berkeley, CA Source: [ 16 ]
17 That is a hard ( and expensive) pill to swallow. (and it s not always the right choice) [ 17 ]
18 What if you could try before you buy? Easy to build test environment Open source (free) software LAN distances, WAN latencies Isolated, controlled (no saturating production links!) Quickly compare models, vendors, configs, etc. [ 18 ]
19 Hardware 4x Servers w/ 10G NICs We used Dell R320s and Intel X520s 1x Receiving switch w/ 3x 10G ports Doesn t need to be expensive, this is where data fans out so no congestion on this side. 5x SFP+ Direct-Attached Copper Cables (cheapest) OR 6x 10G Optics + 5x fiber cables (flexible options) We used SR optics + 50µm multimode fiber Wanted flexibility for testing models w/ X2, XENPAK, etc Sending switch(es) w/ 3x 10G ports (to test) [ 19 ]
20 Software / Configuration Linux (distro generally your preference) We used CentOS 6 (some Fedora too) Install test utilities (or just use perfsonar 1 ) import Internet2 repo install iperf nuttcp bwctl-client bwctl-server Host and NIC tuning per FasterData 2 recommendations: TCP Tuning - /etc/sysctl.conf TX Queue Length TX / RX Descriptors Jumbo Frames [ 20 ]
21 Test Overview Add delay between Host 1 and Host 3 using netem Host 3 runs iperf3 server Host 4 runs iperf3 server Host 2 sends UDP 2Gbps Host 1 sends TCP traffic and we measure rate Sending Switch s outbound interface is congested and buffering occurs [ 21 ]
22 Test Commands Add delay: host1 # tc qdisc add dev ethn root netem delay 25ms host3 # tc qdisc add dev ethn root netem delay 25ms Start iperf3 servers to receive data: host3 # iperf3 -s host4 # iperf3 -s Start background traffic (to cause congestion): host2 # iperf3 -c host4 -u -b2g -t3000 Start TCP traffic to simulate data transfer: host1 # iperf3 -c host3 -P2 -t30 -O5 [ 22 ]
23 Test Results [ ID] Interval Transfer Bandwidth Retr Cwnd... [ 4] sec 201 MB 1.69 Gbps MB [ 6] sec 126 MB 1.06 Gbps MB [SUM] sec 328 MB 2.75 Gbps [ ID] Interval Transfer Bandwidth Retr [ 4] sec 5.85 GB 1.68 Gbps 40 sender [ 4] sec 5.83 GB 1.67 Gbps receiver [ 6] sec 4.04 GB 1.16 Gbps 39 sender [ 6] sec 4.01 GB 1.15 Gbps receiver [SUM] sec 9.89 GB 2.83 Gbps 79 sender [SUM] sec 9.85 GB 2.82 Gbps receiver [ 23 ]
24 Swap out sending switch and repeat [ 24 ]
25 Average TCP results, various switches Buffers per 10G egress port 2x parallel TCP streams 50ms simulated RTT 2Gbps UDP background traffic 15 iterations 1MB Brocade MLXe 1 9MB Arista MB Cisco MB Brocade MLXe 1 90MB Cisco MB Cisco VOQ Arista NI-MLX-10Gx8-M 2 Over-subscription Mode 3 Performance Mode [ 25 ]
26 Tunable Buffers with a Brocade MLXe 1 Buffers per 10G egress port 2x parallel TCP streams 50ms simulated RTT 2Gbps UDP background traffic 15 iterations qos queue-type 0 max-queue-size 1 NI-MLX-10Gx8-M [ 26 ]
27 Is netem accurate? Real World RTT vs Simulated RTT 20% = ~76MB? 70ms RTT 2x parallel TCP streams 2Gbps UDP background traffic Juniper MX80 config: class-of-service scheduler buffer-size percent Special thanks to Mark Lukasczyk at Brookhaven National Laboratory for providing far-end test servers [ 27 ]
28 Tap from LBNL border CA to NY MX80 w/ 1% buffer MX80 w/ 5% buffer [ 28 ]
29 Tap from LBNL border CA to NY MX80 w/ 50% buffer [ 29 ]
30 What if there s a small buffered switch upstream? No Congestion on it? No problem Congestion on it? Many problem [ 30 ]
31 Alternate test - nuttcp [ 31 ]
32 Alternate test nuttcp commands Add delay: host1# tc qdisc add dev eth1 root netem delay 25ms host2# tc qdisc add dev eth1 root netem delay 25ms Start 2Gbps UDP flow to add congestion: host4# iperf3 s host3# iperf3 -c host4 -u -b2g -t3000 nuttcp basic test parameters 1 : host2# nuttcp -S host1# nuttcp -l8972 -T30 -u -w4m Ri300m/X i1 host2 1 [ 32 ]
33 nuttcp conclusion This will probably have no packet loss on smaller buffer switches: nuttcp -l8972 -T30 -u -w4m -Ri300m/65 -i1 While this will probably have some: nuttcp -l8972 -T30 -u -w4m -Ri300m/300 -i1 BUT only applies to where there is congestion. Small buffer switch that isn t congested won t be detectable with this method. [ 33 ]
34 Then Big Buffers = good? Only in the context of these Elephant flows Very large data transfers (Terabytes, Petabytes) Large pipes (10 Gbps & up) Long distances (50ms+) Between small numbers of hosts By big we re talking MBs per 10G port, not GBs. Important to have enough buffers to ride out micro-bursts. May need to drop 1 or 2 packets to fit available capacity, but to maintain performance we need to keep TCP from getting stuck in loss recovery mode. [ 34 ]
35 General Purpose Network Congestion, Insufficient Buffering [ 35 ]
36 General Purpose Network Science DMZ [ 36 ]
37 Workflow-specific Networks: Beamline 10G LAN Campus LAN Science DMZ [ 37 ]
38 Effects of TCP Segmentation Offload TCP Throughput on Small Buffer Switch (Congestion w/ 2Gbps UDP background traffic) ethtool -k EthN Show current offload settings ethtool -K EthN tso off Set TSO off TSO Off was negligible on our hosts with Intel 10Gbps (~ Mbps) BUT may be NIC specific, or more 40Gbps+ [ 38 ]
39 Fair Queuing and Pacing in Kernel TCP Throughput on Small Buffer Switch (Congestion w/ 2Gbps UDP background traffic) Requires newer kernel version not available in tc qdisc add dev EthN root fq Enable Fair Queuing Pacing side effect of Fair Queuing yields ~1.25Gbps increase in 10Gbps on our hosts TSO differences still negligible on our hosts w/ Intel X520 [ 39 ]
40 Additional Information A History of Buffer Sizing Jim Warner s Packet Buffer Page Faster ESnet Michael Smitasin [email protected] Brian Tierney [email protected] [ 40 ]
41 EVALUATING NETWORK BUFFER SIZE REQUIREMENTS for Very Large Data Transfers Michael Smitasin Lawrence Berkeley National Laboratory (LBNL) Brian Tierney Energy Sciences Network (ESnet)
Campus Network Design Science DMZ
Campus Network Design Science DMZ Dale Smith Network Startup Resource Center [email protected] The information in this document comes largely from work done by ESnet, the USA Energy Sciences Network see
TCP Labs. WACREN Network Monitoring and Measurement Workshop Antoine Delvaux [email protected] perfsonar developer 30.09.
TCP Labs WACREN Network Monitoring and Measurement Workshop Antoine Delvaux [email protected] perfsonar developer 30.09.2015 Hands-on session We ll explore practical aspects of TCP Checking the effect
Using Linux Traffic Control on Virtual Circuits J. Zurawski Internet2 [email protected] February 25 nd 2013
Using Linux Traffic Control on Virtual Circuits J. Zurawski Internet2 [email protected] February 25 nd 2013 1. Abstract Research and Education (R&E) networks have experimented with the concept of
TCP loss sensitivity analysis ADAM KRAJEWSKI, IT-CS-CE
TCP loss sensitivity analysis ADAM KRAJEWSKI, IT-CS-CE Original problem IT-DB was backuping some data to Wigner CC. High transfer rate was required in order to avoid service degradation. 4G out of... 10G
Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build 164009
Performance Study Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build 164009 Introduction With more and more mission critical networking intensive workloads being virtualized
Performance of VMware vcenter (VC) Operations in a ROBO Environment TECHNICAL WHITE PAPER
Performance of VMware vcenter (VC) Operations in a ROBO Environment TECHNICAL WHITE PAPER Introduction Many VMware customers have virtualized their ROBO (Remote Office Branch Office) offices in order to
1000Mbps Ethernet Performance Test Report 2014.4
1000Mbps Ethernet Performance Test Report 2014.4 Test Setup: Test Equipment Used: Lenovo ThinkPad T420 Laptop Intel Core i5-2540m CPU - 2.60 GHz 4GB DDR3 Memory Intel 82579LM Gigabit Ethernet Adapter CentOS
D1.2 Network Load Balancing
D1. Network Load Balancing Ronald van der Pol, Freek Dijkstra, Igor Idziejczak, and Mark Meijerink SARA Computing and Networking Services, Science Park 11, 9 XG Amsterdam, The Netherlands June [email protected],[email protected],
Tier3 Network Issues. Richard Carlson May 19, 2009 [email protected]
Tier3 Network Issues Richard Carlson May 19, 2009 [email protected] Internet2 overview Member organization with a national backbone infrastructure Campus & Regional network members National and International
High-Speed TCP Performance Characterization under Various Operating Systems
High-Speed TCP Performance Characterization under Various Operating Systems Y. Iwanaga, K. Kumazoe, D. Cavendish, M.Tsuru and Y. Oie Kyushu Institute of Technology 68-4, Kawazu, Iizuka-shi, Fukuoka, 82-852,
Iperf Tutorial. Jon Dugan <[email protected]> Summer JointTechs 2010, Columbus, OH
Iperf Tutorial Jon Dugan Summer JointTechs 2010, Columbus, OH Outline What are we measuring? TCP Measurements UDP Measurements Useful tricks Iperf Development What are we measuring? Throughput?
www.careercert.info Please purchase PDF Split-Merge on www.verypdf.com to remove this watermark.
2007 Cisco Systems, Inc. All rights reserved. DESGN v2.0 3-11 Enterprise Campus and Data Center Design Review Analyze organizational requirements: Type of applications, traffic volume, and traffic pattern
perfsonar: End-to-End Network Performance Verification
perfsonar: End-to-End Network Performance Verification Toby Wong Sr. Network Analyst, BCNET Ian Gable Technical Manager, Canada Overview 1. IntroducGons 2. Problem Statement/Example Scenario 3. Why perfsonar?
Fundamentals of Data Movement Hardware
Fundamentals of Data Movement Hardware Jason Zurawski ESnet Science Engagement [email protected] CC-NIE PI Workshop April 30 th 2014 With contributions from S. Balasubramanian, G. Bell, E. Dart, M. Hester,
The new frontier of the DATA acquisition using 1 and 10 Gb/s Ethernet links. Filippo Costa on behalf of the ALICE DAQ group
The new frontier of the DATA acquisition using 1 and 10 Gb/s Ethernet links Filippo Costa on behalf of the ALICE DAQ group DATE software 2 DATE (ALICE Data Acquisition and Test Environment) ALICE is a
Question: 3 When using Application Intelligence, Server Time may be defined as.
1 Network General - 1T6-521 Application Performance Analysis and Troubleshooting Question: 1 One component in an application turn is. A. Server response time B. Network process time C. Application response
Transparent Optimization of Grid Server Selection with Real-Time Passive Network Measurements. Marcia Zangrilli and Bruce Lowekamp
Transparent Optimization of Grid Server Selection with Real-Time Passive Network Measurements Marcia Zangrilli and Bruce Lowekamp Overview Grid Services Grid resources modeled as services Define interface
Deploying 10/40G InfiniBand Applications over the WAN
Deploying 10/40G InfiniBand Applications over the WAN Eric Dube ([email protected]) Senior Product Manager of Systems November 2011 Overview About Bay Founded in 2000 to provide high performance
Allocating Network Bandwidth to Match Business Priorities
Allocating Network Bandwidth to Match Business Priorities Speaker Peter Sichel Chief Engineer Sustainable Softworks [email protected] MacWorld San Francisco 2006 Session M225 12-Jan-2006 10:30 AM -
Evaluation of 40 Gigabit Ethernet technology for data servers
Evaluation of 40 Gigabit Ethernet technology for data servers Azher Mughal, Artur Barczyk Caltech / USLHCNet CHEP-2012, New York http://supercomputing.caltech.edu Agenda Motivation behind 40GE in Data
Netest: A Tool to Measure the Maximum Burst Size, Available Bandwidth and Achievable Throughput
Netest: A Tool to Measure the Maximum Burst Size, Available Bandwidth and Achievable Throughput Guojun Jin Brian Tierney Distributed Systems Department Lawrence Berkeley National Laboratory 1 Cyclotron
perfsonar Host Hardware
perfsonar Host Hardware This document is a result of work by the perfsonar Project (h@p://www.perfsonar.net) and is licensed under CC BY- SA 4.0 (h@ps://creakvecommons.org/licenses/by- sa/4.0/). Event
Open Source Bandwidth Management: Introduction to Linux Traffic Control
Open Source Bandwidth Management: Introduction to Linux Traffic Control Christian Benvenuti International Centre for Theoretical Physics (ICTP), Trieste [email protected] [http://benve.info]
Optimizing Throughput on Guaranteed-Bandwidth WAN Networks for the Large Synoptic Survey Telescope (LSST)
Optimizing Throughput on Guaranteed-Bandwidth WAN Networks for the Large Synoptic Survey Telescope (LSST) D. Michael Freemon National Center for Supercomputing Applications University of Illinois Urbana,
Comparison of 40G RDMA and Traditional Ethernet Technologies
Comparison of 40G RDMA and Traditional Ethernet Technologies Nichole Boscia, Harjot S. Sidhu 1 NASA Advanced Supercomputing Division NASA Ames Research Center Moffett Field, CA 94035-1000 [email protected]
VMWARE WHITE PAPER 1
1 VMWARE WHITE PAPER Introduction This paper outlines the considerations that affect network throughput. The paper examines the applications deployed on top of a virtual infrastructure and discusses the
PE310G4BPi40-T Quad port Copper 10 Gigabit Ethernet PCI Express Bypass Server Intel based
PE310G4BPi40-T Quad port Copper 10 Gigabit Ethernet PCI Express Bypass Server Intel based Description Silicom s quad port Copper 10 Gigabit Ethernet Bypass server adapter is a PCI-Express X8 network interface
Scaling Up Your Network Monitoring: From the Garden Hose to the Fire Hose
UNIVERSITY OF CALIFORNIA Scaling Up Your Network Monitoring: From the Garden Hose to the Fire Hose Vincent Stoffer Cyber Security Engineer Technology Exchange October 28, 2014 Agenda Intro / overview The
Gigabit Ethernet Design
Gigabit Ethernet Design Laura Jeanne Knapp Network Consultant 1-919-254-8801 [email protected] www.lauraknapp.com Tom Hadley Network Consultant 1-919-301-3052 [email protected] HSEdes_ 010 ed and
TCP tuning guide for distributed application on wide area networks 1.0 Introduction
TCP tuning guide for distributed application on wide area networks 1.0 Introduction Obtaining good TCP throughput across a wide area network usually requires some tuning. This is especially true in high-speed
High Speed Ethernet. Dr. Sanjay P. Ahuja, Ph.D. Professor School of Computing, UNF
High Speed Ethernet Dr. Sanjay P. Ahuja, Ph.D. Professor School of Computing, UNF Hubs and Switches Hubs and Switches Shared Medium Hub The total capacity in the shared medium hub configuration (figure
Network testing with iperf
Network testing with iperf Bill Farrow [email protected] Why use iperf? What does it run on? TCP example UDP example Use Case: Network Server Appliance Use Case: Embedded Video Streaming Use Case:
Overview of Network Measurement Tools
Overview of Network Measurement Tools Jon M. Dugan Energy Sciences Network Lawrence Berkeley National Laboratory NANOG 43, Brooklyn, NY June 1, 2008 Networking for the Future of Science
Performance of Host Identity Protocol on Nokia Internet Tablet
Performance of Host Identity Protocol on Nokia Internet Tablet Andrey Khurri Helsinki Institute for Information Technology HIP Research Group IETF 68 Prague March 23, 2007
Network Monitoring with the perfsonar Dashboard
Network Monitoring with the perfsonar Dashboard Andy Lake Brian Tierney ESnet Advanced Network Technologies Group TIP2013 Honolulu HI January 15, 2013 Overview perfsonar overview Dashboard history and
Small is Better: Avoiding Latency Traps in Virtualized DataCenters
Small is Better: Avoiding Latency Traps in Virtualized DataCenters SOCC 2013 Yunjing Xu, Michael Bailey, Brian Noble, Farnam Jahanian University of Michigan 1 Outline Introduction Related Work Source of
N5 NETWORKING BEST PRACTICES
N5 NETWORKING BEST PRACTICES Table of Contents Nexgen N5 Networking... 2 Overview of Storage Networking Best Practices... 2 Recommended Switch features for an iscsi Network... 2 Setting up the iscsi Network
Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking
Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking Burjiz Soorty School of Computing and Mathematical Sciences Auckland University of Technology Auckland, New Zealand
Infrastructure for active and passive measurements at 10Gbps and beyond
Infrastructure for active and passive measurements at 10Gbps and beyond Best Practice Document Produced by UNINETT led working group on network monitoring (UFS 142) Author: Arne Øslebø August 2014 1 TERENA
Cisco Bandwidth Quality Manager 3.1
Cisco Bandwidth Quality Manager 3.1 Product Overview Providing the required quality of service (QoS) to applications on a wide-area access network consistently and reliably is increasingly becoming a challenge.
Improving the Performance of TCP Using Window Adjustment Procedure and Bandwidth Estimation
Improving the Performance of TCP Using Window Adjustment Procedure and Bandwidth Estimation R.Navaneethakrishnan Assistant Professor (SG) Bharathiyar College of Engineering and Technology, Karaikal, India.
Network Diagnostic Tools. Jijesh Kalliyat Sr.Technical Account Manager, Red Hat 15th Nov 2014
Network Diagnostic Tools Jijesh Kalliyat Sr.Technical Account Manager, Red Hat 15th Nov 2014 Agenda Network Diagnostic Tools Linux Tcpdump Wireshark Tcpdump Analysis Sources of Network Issues If a system
Frequently Asked Questions
Frequently Asked Questions 1. Q: What is the Network Data Tunnel? A: Network Data Tunnel (NDT) is a software-based solution that accelerates data transfer in point-to-point or point-to-multipoint network
Challenges of Sending Large Files Over Public Internet
Challenges of Sending Large Files Over Public Internet CLICK TO EDIT MASTER TITLE STYLE JONATHAN SOLOMON SENIOR SALES & SYSTEM ENGINEER, ASPERA, INC. CLICK TO EDIT MASTER SUBTITLE STYLE OUTLINE Ø Setting
A Simulation Study of Effect of MPLS on Latency over a Wide Area Network (WAN)
A Simulation Study of Effect of MPLS on Latency over a Wide Area Network (WAN) Adeyinka A. Adewale, Samuel N. John, and Charles Ndujiuba 1 Department of Electrical and Information Engineering, Covenant
IT@Intel. Optimizing WAN Performance for the Global Enterprise
White Paper Intel Information Technology Computer Manufacturing WAN Performance and Optimization Optimizing WAN Performance for the Global Enterprise To improve throughput on Intel wide area network (WAN)
APPOSITE TECHNOLOGIES Smoothing the Transition to 10 Gbps. WAN Emulation Made Easy
APPOSITE TECHNOLOGIES Smoothing the Transition to 10 Gbps WAN Emulation Made Easy The Transition to 10 Gbps 1. THE PROBLEM - Application Performance 2. THE SOLUTION - Performance Validation Using WAN Emulation
Performance Tuning Guidelines for PowerExchange for Microsoft Dynamics CRM
Performance Tuning Guidelines for PowerExchange for Microsoft Dynamics CRM 1993-2016 Informatica LLC. No part of this document may be reproduced or transmitted in any form, by any means (electronic, photocopying,
Network Performance Optimisation and Load Balancing. Wulf Thannhaeuser
Network Performance Optimisation and Load Balancing Wulf Thannhaeuser 1 Network Performance Optimisation 2 Network Optimisation: Where? Fixed latency 4.0 µs Variable latency
ITL Lab 5 - Performance Measurements and SNMP Monitoring 1. Purpose
Lab 5 - Performance Measurements and SNMP Monitoring 1 Purpose Before the Lab Measure the performance (throughput) of TCP connections Measure the performance of UDP connections; observe an RTP flow Examine
Measuring Wireless Network Performance: Data Rates vs. Signal Strength
EDUCATIONAL BRIEF Measuring Wireless Network Performance: Data Rates vs. Signal Strength In January we discussed the use of Wi-Fi Signal Mapping technology as a sales tool to demonstrate signal strength
TCP Tuning Techniques for High-Speed Wide-Area Networks. Wizard Gap
NFNN2, 20th-21st June 2005 National e-science Centre, Edinburgh TCP Tuning Techniques for High-Speed Wide-Area Networks Distributed Systems Department Lawrence Berkeley National Laboratory http://gridmon.dl.ac.uk/nfnn/
Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago
Globus Striped GridFTP Framework and Server Raj Kettimuthu, ANL and U. Chicago Outline Introduction Features Motivation Architecture Globus XIO Experimental Results 3 August 2005 The Ohio State University
LotServer Deployment Manual
LotServer Deployment Manual Maximizing Network Performance, Responsiveness and Availability DEPLOYMENT 1. Introduction LotServer is a ZetaTCP powered software product that can be installed on origin web/application
Data Networks Summer 2007 Homework #3
Data Networks Summer Homework # Assigned June 8, Due June in class Name: Email: Student ID: Problem Total Points Problem ( points) Host A is transferring a file of size L to host B using a TCP connection.
Using TrueSpeed VNF to Test TCP Throughput in a Call Center Environment
Using TrueSpeed VNF to Test TCP Throughput in a Call Center Environment TrueSpeed VNF provides network operators and enterprise users with repeatable, standards-based testing to resolve complaints about
Quantum StorNext. Product Brief: Distributed LAN Client
Quantum StorNext Product Brief: Distributed LAN Client NOTICE This product brief may contain proprietary information protected by copyright. Information in this product brief is subject to change without
Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck
Sockets vs. RDMA Interface over 1-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Pavan Balaji Hemal V. Shah D. K. Panda Network Based Computing Lab Computer Science and Engineering
Network Probe. Figure 1.1 Cacti Utilization Graph
Network Probe Description The MCNC Client Network Engineering group will install several open source network performance management tools on a computer provided by the LEA or charter school to build a
Choosing Tap or SPAN for Data Center Monitoring
Choosing Tap or SPAN for Data Center Monitoring Technical Brief Key Points Taps are passive, silent, and deliver a perfect record of link traffic, but require additional hardware and create a point of
Measure wireless network performance using testing tool iperf
Measure wireless network performance using testing tool iperf By Lisa Phifer, SearchNetworking.com Many companies are upgrading their wireless networks to 802.11n for better throughput, reach, and reliability,
How Router Technology Shapes Inter-Cloud Computing Service Architecture for The Future Internet
How Router Technology Shapes Inter-Cloud Computing Service Architecture for The Future Internet Professor Jiann-Liang Chen Friday, September 23, 2011 Wireless Networks and Evolutional Communications Laboratory
Configuring Your Computer and Network Adapters for Best Performance
Configuring Your Computer and Network Adapters for Best Performance ebus Universal Pro and User Mode Data Receiver ebus SDK Application Note This application note covers the basic configuration of a network
Effects of Interrupt Coalescence on Network Measurements
Effects of Interrupt Coalescence on Network Measurements Ravi Prasad, Manish Jain, and Constantinos Dovrolis College of Computing, Georgia Tech., USA ravi,jain,[email protected] Abstract. Several
Windows Server 2012 R2 Hyper-V: Designing for the Real World
Windows Server 2012 R2 Hyper-V: Designing for the Real World Steve Evans @scevans www.loudsteve.com Nick Hawkins @nhawkins www.nickahawkins.com Is Hyper-V for real? Microsoft Fan Boys Reality VMware Hyper-V
Assessing the Performance of Virtualization Technologies for NFV: a Preliminary Benchmarking
Assessing the Performance of Virtualization Technologies for NFV: a Preliminary Benchmarking Roberto Bonafiglia, Ivano Cerrato, Francesco Ciaccia, Mario Nemirovsky, Fulvio Risso Politecnico di Torino,
Smart Tips. Enabling WAN Load Balancing. Key Features. Network Diagram. Overview. Featured Products. WAN Failover. Enabling WAN Load Balancing Page 1
Smart Tips Enabling WAN Load Balancing Overview Many small businesses today use broadband links such as DSL or Cable, favoring them over the traditional link such as T1/E1 or leased lines because of the
A High-Performance Storage and Ultra-High-Speed File Transfer Solution
A High-Performance Storage and Ultra-High-Speed File Transfer Solution Storage Platforms with Aspera Abstract A growing number of organizations in media and entertainment, life sciences, high-performance
Hands on Workshop. Network Performance Monitoring and Multicast Routing. Yasuichi Kitamura NICT Jin Tanaka KDDI/NICT APAN-JP NOC
Hands on Workshop Network Performance Monitoring and Multicast Routing Yasuichi Kitamura NICT Jin Tanaka KDDI/NICT APAN-JP NOC July 18th TEIN2 Site Coordination Workshop Network Performance Monitoring
Operating Systems and Networks Sample Solution 1
Spring Term 2014 Operating Systems and Networks Sample Solution 1 1 byte = 8 bits 1 kilobyte = 1024 bytes 10 3 bytes 1 Network Performance 1.1 Delays Given a 1Gbps point to point copper wire (propagation
Linux NIC and iscsi Performance over 40GbE
Linux NIC and iscsi Performance over 4GbE Chelsio T8-CR vs. Intel Fortville XL71 Executive Summary This paper presents NIC and iscsi performance results comparing Chelsio s T8-CR and Intel s latest XL71
Voice Over IP. MultiFlow 5048. IP Phone # 3071 Subnet # 10.100.24.0 Subnet Mask 255.255.255.0 IP address 10.100.24.171. Telephone.
Anritsu Network Solutions Voice Over IP Application Note MultiFlow 5048 CALL Manager Serv # 10.100.27 255.255.2 IP address 10.100.27.4 OC-48 Link 255 255 25 IP add Introduction Voice communications over
Deploying in a Distributed Environment
Deploying in a Distributed Environment Distributed enterprise networks have many remote locations, ranging from dozens to thousands of small offices. Typically, between 5 and 50 employees work at each
Leveraging NIC Technology to Improve Network Performance in VMware vsphere
Leveraging NIC Technology to Improve Network Performance in VMware vsphere Performance Study TECHNICAL WHITE PAPER Table of Contents Introduction... 3 Hardware Description... 3 List of Features... 4 NetQueue...
perfsonar Host Hardware
perfsonar Host Hardware Event Presenter, OrganizaKon, Email Date This document is a result of work by the perfsonar Project (h@p://www.perfsonar.net) and is licensed under CC BY- SA 4.0 (h@ps://creakvecommons.org/licenses/by-
Advanced Network Services Teaming
Advanced Network Services Teaming Advanced Network Services (ANS) Teaming, a feature of the Advanced Network Services component, lets you take advantage of multiple adapters in a system by grouping them
Network Instruments white paper
Network Instruments white paper ANALYZING FULL-DUPLEX NETWORKS There are a number ways to access full-duplex traffic on a network for analysis: SPAN or mirror ports, aggregation TAPs (Test Access Ports),
Layer 2 Network Encryption where safety is not an optical illusion Marko Bobinac SafeNet PreSales Engineer
Layer 2 Network Encryption where safety is not an optical illusion Marko Bobinac SafeNet PreSales Engineer Layer 2 Network Encryption where safety is not an optical illusion Todays Agenda Fibre is safe
IP SAN Best Practices
IP SAN Best Practices A Dell Technical White Paper PowerVault MD3200i Storage Arrays THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS AND TECHNICAL INACCURACIES.
100 Gigabit Ethernet is Here!
100 Gigabit Ethernet is Here! Introduction Ethernet technology has come a long way since its humble beginning in 1973 at Xerox PARC. With each subsequent iteration, there has been a lag between time of
Clearing the Way for VoIP
Gen2 Ventures White Paper Clearing the Way for VoIP An Alternative to Expensive WAN Upgrades Executive Overview Enterprises have traditionally maintained separate networks for their voice and data traffic.
Net Optics and Cisco NAM
When Cisco decided to break its Network Analysis Module (NAM) out of the box and into a stand-alone appliance, they turned to Net Optics for monitoring access connectivity. Cisco NAM 2200 Series Cisco
Requirements of Voice in an IP Internetwork
Requirements of Voice in an IP Internetwork Real-Time Voice in a Best-Effort IP Internetwork This topic lists problems associated with implementation of real-time voice traffic in a best-effort IP internetwork.
Technical Bulletin. Arista LANZ Overview. Overview
Technical Bulletin Arista LANZ Overview Overview Highlights: LANZ provides unparalleled visibility into congestion hotspots LANZ time stamping provides for precision historical trending for congestion
Performance Measurement of Wireless LAN Using Open Source
Performance Measurement of Wireless LAN Using Open Source Vipin M Wireless Communication Research Group AU KBC Research Centre http://comm.au-kbc.org/ 1 Overview General Network Why Network Performance
How To Monitor And Test An Ethernet Network On A Computer Or Network Card
3. MONITORING AND TESTING THE ETHERNET NETWORK 3.1 Introduction The following parameters are covered by the Ethernet performance metrics: Latency (delay) the amount of time required for a frame to travel
pathchar a tool to infer characteristics of Internet paths
pathchar a tool to infer characteristics of Internet paths Van Jacobson ([email protected]) Network Research Group Lawrence Berkeley National Laboratory Berkeley, CA 94720 MSRI April 21, 1997 c 1997 by Van
