Receive Descriptor Recycling for Small Packet High Speed Ethernet Traffic
|
|
|
- Amice Walton
- 10 years ago
- Views:
Transcription
1 IEEE MELECON 2006, May 6-9, Benalmádena (Málaga), Spain Receive Descriptor Recycling for Small Packet High Speed Ethernet Traffic Cedric Walravens Department of Electrical Engineering - ESAT Katholieke Universiteit Leuven, Belgium [email protected] Benjamin Gaidioz CERN Geneva, Switzerland [email protected] Abstract This paper presents the concept of Receive Descriptor Recycling to significantly reduce the performance drop associated with small packet Gigabit Ethernet traffic. Since limits and trade-offs are inherent when optimising for small packet traffic, all important aspects in this context are covered. Lowlevel measurements were performed at the CERN LHCb online Data Acquisition (DAQ) system, which is to a large extend made up of commodity equipment. Results gathered show the Ethernet Controller ( Interface Card, ) driver currently is the major bottleneck, preventing the system from reaching maximal Gigabit Ethernet performance. Receive Descriptor Recycling is implemented under Linux for Intel s e000 driver, and is shown to successfully remedy the driver inefficiency. I. INTRODUCTION Broadband technologies enable considerable amounts of data to be transferred through a large network at comfortable speeds. Many modern DAQ systems, such as the LHCb Online DAQ system [], can therefore depend on commodity hardware to reduce cost and simplify administration. Gigabit Ethernet is particularly attractive because this technology matured over the years and it has a very appealing performance to cost of ownership ratio. Apart from full load link traffic, of which the effects have been studied in [2] [3], small packets have become more and more important in many real time applications. An example of a popular application, that causes this kind of network traffic, is voice-over-ip (VoIP). For each VoIP connection, 50 network packets are sent every second with a data payload size of 60 bytes or even less [4]. All Linux based firewalls, routers and web caches, where a large amount of small network packets is traveling through, can benefit from the results of this paper. High reliability of small-sized transmissions is also crucial for the correctness of calibration runs for the LHCb [5] experiment, at the CERN LHC accelerator [6]. Here, small size packets are combined with extreme high packet rates and a near real time constraint. Some work on small packet traffic effects was conducted in [7], but no thorough low-level analysis has yet been performed. Also, Intel has released a document on small packet traffic performance [8], however, it presents merely recommendations which are not substantiated by solid measurements to proof any performance gain. This paper studies the performance of an Ethernet device with small packets and remedies the associated performance drop by implementing Receive Descriptor Recycling (RDR) for Intel s e000 Linux driver. First, Section II will elaborate on the LHCb DAQ system, where the low-level measurements are performed. Section III describes the hardware setup and equipment used, followed by the throughput performance measurements and low-level analysis in Section IV and V, respectively. Both gather proof implying the Ethernet Controller s driver as the actual bottleneck. Finally, a solution to this bottleneck is provided by the RDR mechanism described in Section VI. Section VII concludes the paper with a discussion on further work. II. LHCB DAQ ONLINE SYSTEM The LHCb Data Acquisition network [] relies on Gigabit Ethernet over copper UTP to interconnect custom made electronics with the computing farm of commodity PCs. The farm of processors is partitioned into sub-farms, each one being interfaced to the readout network switch by a gateway (called an Sub-Farm Controller, SFC). A schematic overview of the system is shown in Fig.. Performance of a SFC is critical for sizing whole the system. It can be measured by means of packet rate, packet loss, response time and throughput. Also, efficient utilisation and resource requirements are important criteria. When optimising for small packet traffic, inevitable trade-offs are encountered, e.g. and resource requirements increase in order to accommodate high rates of small packets, along with a rise in host bus utilisation. Note that transmit performance is not affected by small packet traffic to the same extent as receive performance. This asymmetry exists because the local host cannot usually overwhelm the Ethernet Controller with outgoing traffic. III. HARDWARE SETUP All benchmarks were performed on a SFC candidate system running two 32-bit Intel Xeon processors. The DAQ LAN is a Gigabit Ethernet network, connected by cat 5e copper UTP wires. The setup contained two servers and one network processor [9] as a very powerful traffic generator. All servers had hyper-threading [0] enabled. Both Intel Pro/000 MT dual and quad port Ethernet s were used. On the SRV06 host, version 6.4-k2-NAPI of the e000 network driver was used, on the SFC04 host this was version k2-NAPI. An overview of relevant system /06/$ IEEE 252
2 All local connections GbE 0GbE uplink to local storage and CERN Storage Switched ~500 data sources Core DAQ ~00 Gateways received send./20*6/20*x+.085 pci bw Intel(r) 82546EB Dual port Distribution Layer Storage Fig.. Up to 23 s per subnet Schematic view of the LHCb DAQ. TABLE I SPECIFICATIONS OF HARDWARE USED. ~2000 Farm Nodes Host Chipset Linux PCI-X bus Systembus Kernel SRV06 ServerWorks GC Dual Xeon 2.4GHz Scientific bit 33MHz 400MHz smp SFC06 Dell SC425 Dual Xeon 2.8GHz Scientific bit 33MHz 800MHz 2.6.-smp Type Ports Chipset Host bus (max) Intel Pro/000MT dual 82546EB PCI-X 64b 33MHz Intel Pro/000MT quad 82546EB PCI-X 64b 33MHz IBM PowerNP NP4GS3 tri BCM5700 PCI specifications is presented in Table I, including the network processor used for frame generation. The tests presented in this paper were performed using raw Ethernet frames using the IEEE MAC format []; they consist of the 4 byte header followed by variable length user data and a 4 byte CRC-checksum at the end, resulting in a total data overhead of 8 bytes per frame. IV. RECEIVE THROUGHPUT PERFORMANCE To measure receive throughput, the network processor (NP) acted as a source, flooding the SFC with packets. Fig. 2(a) shows the bit rate measured when receiving on one port of the SFC from one port of the NP with a variable frame size. The bit rate measured at the output of the NP well matches the, apart from regular drops. According to the NP documentation, this behaviour is to be expected when pushing this device to the Gigabit Ethernet limit. The SFC receive rate drops accordingly of course. In the range of 0 to 200, the bit rate measured at the input of the SFC is much lower (data is lost). In higher ranges, the sent bit rate is reached (no losses). Receiving on both ports of the dual card simultaneously, produces an identical plot. Fig. 2(a) also shows the theoretical limit of the PCI bus. It is established that, in the case of operating one single port, this PCI bandwidth will not pose any limit on Gigabit Ethernet transfers, even taking into account the fact that this is (a) Dual card, one port, plotted with theoretical Gigabit Ethernet limit, PCI bandwidth and slope Intel(r) 82546EB Quad port received eth2 send ->eth2 0.2 received eth3 send 2->eth (b) Quad card, two ports simultaneously, plotted with theoretical Gigabit Ethernet limit. Fig. 2.. Receive throughput from NP (port x) to SFC (eth y) for Intel a theoretical value which will never actually be reached due to numerous non-idealities. The same figure also shows a linear fit of the first part of the curve. The slope is 0.004, which is in agreement with measurements in [7]. In the continuation of this paper, however, a different conclusion with regard to what contributes to this slope will be reached. The same measurements were done on an Intel quad port, the resulting plot is shown in Fig. 2(b). The slope of the linear part now amounts to 0.002, a value insensitive to the inter-frame delay setting of the NP. This slope is clearly worse than the slope of the dual card, indicating the quad card to be not as streamlined as the single chip dual card. During this test, no PCI-X transfer bottleneck was observed. 253
3 V. LOW LEVEL MEASUREMENTS The above-mentioned benchmark measures the real performance of the network, but generally does not allow for identification of bottlenecks present. Therefore, lower level tests as PCI-X traffic analysis and kernel profiling were subsequently carried out, emphasising the hardware and software, respectively. A. PCI/PCI-X Analysis Similar to network overhead issues, PCI/PCI-X bus protocols also impose penalties on small packet traffic. In addition to the actual data transfer, each bus transaction requires extra control cycles which introduce overhead (e.g., arbitration latency, address phases, attribute phases, wait states). Bus overhead costs become more pronounced with small packet traffic as bus control and wait cycles consume a larger fraction of each bus transaction. Fig. 3. Transition between smooth and stumble PCI traffic. ( of 2 ports receiving from PCI 64bit 66MHz) Fig. 3 shows the two regimes of transmission across the PCI bus when flooding the with a NP. The dark areas indicate that the signal varies a lot, i.e. frames are transmitted across the bus. Notwithstanding that this plot was taken in PCI mode, exactly the same behaviour was observed when analysing the PCI-X traffic. This section summarises the flow of traffic on the bus, going from the smooth to the stumble regime. During the smooth regime, all frames are nicely put on the PCI bus behind each other (inter-frame delay of 54 CLKs or 0.8 μs). The Intel Pro/000 MT Ethernet Controller uses receive and transmit descriptors to keep the books [2]. Such a descriptor is basically a pointer to memory space, indicating a block of configurable size for the to store a received frame or read in a frame to send. From here on, we will concentrate on reception of frames. Receive Descriptors (RD) are made available to the in groups of 6 consecutive descriptors. The tail address of this group is provided by a Receive Descriptor Tail (RDT). When a frame is received, the Ethernet Controller uses the next RD in the current RDT to find out where to store the frame. After the DMA transfer finishes, this RD is sent back to the driver to advertise the presence of a frame at that location (delay of 7 CLKs). Fig. 4 visualises an interpreted version of Fig. 3. It shows left to right and top to bottom the advancing time axis (in clock ticks). The series of black dots are frames that are written to main memory. The following traffic pattern can be extracted for the smooth regime (0 to CLKs): i+ 343 i i i Fig. 4. Trace for the transmission from NP to SRV06 (from Fig. 3). ) Four frames (black dots) are transferred to main memory through DMA bursts with an inter-frame delay of about 45 CLKs or 0.7 μs. 2) Following afterwards, i.e. 25 CLKs or 0.4 μs, the 4 corresponding RDs are written back to the driver (one grey spot). 3) Step and 2 are repeated 6 times (containing a total of 64 frames), after which a new RDT is employed by the, allowing to fetch 6 new RDs (small gaps in between series of 64 frames). Under normal circumstances, the driver will provide the with new RDTs (marked in Fig. 4), allowing it to fetch newly allocated RDs in time. This transfer of new RDs continues until all allocated descriptors for that batch have been provided, or, the card s frame receive buffer is nearly full and the card terminates the descriptor transfer by raising the STOP# signal in order not to lose any frames. Since a NP is flooding the card during this test (with approximately million packets per second), the latter will happen more and more, as 254
4 seen on the STOP# signal line (Fig. 3). After the descriptor transfer, the frame receive buffer is quickly emptied to regain the time lost during this transfer. All frames are now put on the PCI bus with a minimum delay of 7 CLKs in between. Clearly, under persistent heavy traffic load, the card will become unable to fetch enough descriptors to keep up with the high frame rate, and at the same time sent every single frame to the system. The card will have to allow for a FIFO receive buffer overrun and silently drop frames until it has some more RDs to quickly empty the buffer and try start receiving again. The relief is, however, of short duration, as the card will soon have exhausted it (already) very limited pool of RDs, and it will have to wait again. This is what happens in the second so-called stumble regime (4 500 CLKs to end), where huge gaps of CLKs show the lack of RDs is preventing any further traffic until new RDTs are received. This results in many FIFO buffer overruns and a huge packet loss of up to 500k packets per second. Furthermore, the omnipresent STOP# indicates that any transfer that takes longer than absolutely necessary is abruptly terminated by the. In the smooth part, the system can process frame per microsecond (i.e. what the NP sends). When tumbling down into stumble regime, this number degrades to a value of merely 0.3 frames per microsecond or 0.28 Gb/s, cfr. earlier results, e.g. Fig. 2(a). This also confirms that the linear part for small packet sizes is caused by this stumble behaviour. Where one might be tempted to think the PCI bus or is the cause for this bottleneck, calculations on PCI bus utilisation proved otherwise. During the whole trace, peak data rates did not reach any higher than 250 MB/s. Since the practical limit for a PCI bus is about 50% of the theoretical maximum, which is 533 MB/s in this case, this is acceptable and it was already a clear indication that nor the PCI bus, nor the card were responsible for this stumble behaviour. Note that this trace was taken for a single port receiving frames. It is clear that a PCI 64bit 66MHz will pose a bottleneck when more than port and/or card is operating on the same bus. Compared to PCI, it was observed that PCI-X traffic contained less errors and less accompanying WAIT states. Along with higher clock frequencies, the more intelligent use of WAIT states and so-called split transactions instead of delayed transactions are among the most important features of the PCI- X protocol. A look at the kernel s memory management information, provided by the /var/slabinfo file, made clear that memory access was responsible for the large gaps in stumble regime. These gaps can be explained by the DMA slab freeings. Normally, a keeps a freed DMA slab in its bucket (one bucket per ). Occasionally, however, a large quantity of main memory is reallocated by the driver and then the main memory needs to be accessed, preventing others, e.g. DMA accesses by the, from accessing the main memory. Beware, not the memory bank technology but the way the driver seems to handle memory access is causing the bottleneck. Simple on the back-of-the-envelope calculations show that the DDR-2 400MHz (PC2-3200) provides enough raw bandwidth. It is clear that the Intel Pro/000 MT Ethernet Controller card, when used in combination with a fast PCI-X host bus, will not become a bottleneck, even for quad port operation. Therefore, the next section takes a closer look at the software side of the network infrastructure, i.e. the Linux operating system and the e000 Ethernet Controller driver. B. Kernel Profiling This last test employed the OProfile package [3], a system-wide profiler for Linux systems, capable of profiling all running code at low overhead. It reports on the number of times a certain function (in user space, kernel, driver,...) has been called and what percentage of the total samples taken this represents. Fig. 5. Summary of relative occupation of the during a 5 minute flood by the NP ( of 2 PCI-X 64bit 33MHz). Results of the OProfile test are summarised in Fig. 5. Analysing the exact content of the most frequently made calls, it is established that all are involve with freeing and reallocating RDs. This gave cause for the idea to tune the e000 driver to more cleverly handle this RD processing, and implement some kind of Receive Descriptor recycling. This idea, along with its implementation in the e000 source code, will be covered in the next section. VI. RECEIVE DESCRIPTOR RECYCLING The e000_main.c file of the e000 driver source code [4] contains most of the driver s basic functionality. Analysing this source for packet reception handling on driver level, points out that the most frequently called functions of Fig. 5 are all related to one single very time consuming operation: the freeing of large heaps of memory. Especially when talking small payload sizes, it is clear that the situation only worsens due to more descriptors and thus more memory management overhead. It is this overhead that prevents the driver from quickly sending new RDs to the Ethernet Controller, as they need to be freed and reallocated first. This is why the Receive Descriptor Recycling mechanism was implemented. The idea is to allocate a fixed number of permanent descriptors which are reused 255
5 every time, effectively taking away the need for the costly reallocation overhead. The only processing that remains to be done is resetting some fields in the descriptors. The remainder of this section will outline the details of this implementation. First it was needed to store the permanent buffers in the e000_adapter struct (see e000.h), so they became associated with each Ethernet Controller present in the system. For this, it was extended by two members: an array to store the pre-allocated socket buffers, and an array to point to the associated data fields. A fixed array size of was chosen, which was the RxDescriptors e000 driver parameter used in all previous tests. During the driver initialisation, the array is allocated in memory and the data pointers are set up. To prevent the driver from freeing the allocated array, the number of users for each socket buffer is artificially increased to 2 by calling skb_get(). As long as the number of users remains higher than 2, kfree_skb() will never free them, since it believes someone is still using the socket buffer and its data. This is realised by altering the driver function call flow in e000_alloc_rx_buffers(). Here the call to dev_alloc_skb(), which would return a fresh allocated socket buffer (skb), is replaced by returning one of our preallocated skbs in the array. Next, the skb reset is implemented by the newly added reset_skb() function. Please refer to [5] for a detailed overview of exact code implementations. To check for any performance increase for small packet sizes, the Receive Throughput test (see Section IV) was repeated. For small frame sizes, the RDR-enabled driver was able to reduce packet loss by 40%, see Fig. 6. The influence is the most clear for short delay packets. The performance of higher payloads remained unchanged with RDR. VII. CONCLUSIONS AND FURTHER WORK Several benchmarks were performed on Gigabit Ethernet hardware and the Ethernet Controller driver. Among the aspects analysed in this work were throughput, packet loss and efficient resource utilisation. Emphasis was put on small packet size network traffic, to be used for calibration of the LHCb experiment. It was shown that the current bottleneck lies in the way the e000 driver handles the Receive Descriptor memory management, which proofs to be fatal for small packet sizes. In order to remedy the situation, a Receive Descriptor Recycling mechanism was proposed and implemented in the official e000 driver. Results have shown an improvement of small packet size performance by 40% in terms of increase in throughput and reduction of packet loss. The knowledge gathered by the reported measurements, will be used to optimise the design of the DAQ infrastructure. The proposed Receive Descriptor Recycling will be tested in the real-life DAQ Prototype Farm environment. Further work also includes checking the possibility of turning off cache coherency by requesting non-cacheable buffer memory as to prevent numerous cache trashing NP->SRV06 - Intel dual card, two ports to NP - kernel original with RDR Fig. 6. Receive throughput from NP to SRV for two ports on a dual Intel using official and patched driver. VIII. ACKNOWLEDGEMENTS This work has been funded by the Summer Student Internship Committee of CERN, Geneva, within the framework of the Summer Student Programme REFERENCES [] A. Barczyk, J.-P. Dufey, B. Jost, and N. Neufeld, The new LHCb trigger and DAQ strategy: A system architecture based on gigabit-ethernet, IEEE Trans. Nucl. Sci., vol. 5, pp , June [2] A. Barczyk, A. Carbone, J.-P. Dufey, D. Galli, B. Jost, U. Marconi, N. Neufeld, G. Peco, and V. Vagnoni, Reliability of datagram transmission on gigabit ethernet at full link load, CERN, Geneva, CERN/LHCb , [3] R. Hughes-Jones, P. Clarke, and S. Dallison, Performance of and 0 gigabit ethernet cards with server quality motherboards, Future Generation Computer Systems, vol. 2, pp , Apr [4] Voice Over IP - per call bandwidth consumption. [Online]. Available: consume.html [5] LHCb, LHCb technical proposal, CERN, Geneva, CERN/LHCC 98 4, 998. [6] O. S. Bruning, P. Collier, et al., Eds., LHC Design Report. Geneva, CH: CERN, [7] A. Barczyk, D. Bortolotti, A. Carbone, J.-P. Dufey, D. Galli, B. Gaidioz, D. Gregori, B. Jost, U. Marconi, N. Neufeld, G. Peco, and V. Vagnoni, High rate packets transmission on ethernet LAN using commodity hardware, IEEE Trans. Nucl. Sci., accepted for publication. [8] Small packet traffic performance optimization for 8255x and 8254x ethernet controllers, Application Note AP-453 rev..0, Intel, Sept [9] J. R. Allen, Jr., et al., IBM PowerNP network processor: Hardware, software, and applications, IBM J. Res. and Dev., vol. 47, Mar [0] Hyper-threading technology on the Intel Xeon processor family for servers, White Paper, Intel, Oct [] Ethernet LAN Medium Access Control (MAC) Specification, IEEE Std , 985. [2] PCI/PCI-X family of gigabit ethernet controllers software developer s manual, rev. 2.5, Intel, July [3] OProfile development pages. [Online]. Available: sourceforge.net [4] e000 source code at the Linux cross reference project. [Online]. Available: [5] C. Walravens and B. Gaidioz, Low level gigabit ethernet analysis for the LHCb computing farm, CERN, Geneva, CERN/LHCb
The new frontier of the DATA acquisition using 1 and 10 Gb/s Ethernet links. Filippo Costa on behalf of the ALICE DAQ group
The new frontier of the DATA acquisition using 1 and 10 Gb/s Ethernet links Filippo Costa on behalf of the ALICE DAQ group DATE software 2 DATE (ALICE Data Acquisition and Test Environment) ALICE is a
Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck
Sockets vs. RDMA Interface over 1-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Pavan Balaji Hemal V. Shah D. K. Panda Network Based Computing Lab Computer Science and Engineering
Network Performance Optimisation and Load Balancing. Wulf Thannhaeuser
Network Performance Optimisation and Load Balancing Wulf Thannhaeuser 1 Network Performance Optimisation 2 Network Optimisation: Where? Fixed latency 4.0 µs Variable latency
D1.2 Network Load Balancing
D1. Network Load Balancing Ronald van der Pol, Freek Dijkstra, Igor Idziejczak, and Mark Meijerink SARA Computing and Networking Services, Science Park 11, 9 XG Amsterdam, The Netherlands June [email protected],[email protected],
Networking Driver Performance and Measurement - e1000 A Case Study
Networking Driver Performance and Measurement - e1000 A Case Study John A. Ronciak Intel Corporation [email protected] Ganesh Venkatesan Intel Corporation [email protected] Jesse Brandeburg
Analyzing and Optimizing the Linux Networking Stack
Analyzing and Optimizing the Linux Networking Stack Raffaele Bolla*, Roberto Bruschi*, Andrea Ranieri*, Gioele Traverso * Department of Communications, Computer and Systems Science (DIST) University of
Results of Benchmark Tests upon a Gigabit Ethernet Switch
Results of Benchmark Tests upon a Gigabit Ethernet Switch LHCb Technical Note Issue: Draft Revision: 1 Reference: LHCb DAQ 2001-nn Created: 13 th August 2001 Last modified: 17 th August 2001 Prepared By:
VMWARE WHITE PAPER 1
1 VMWARE WHITE PAPER Introduction This paper outlines the considerations that affect network throughput. The paper examines the applications deployed on top of a virtual infrastructure and discusses the
Intel DPDK Boosts Server Appliance Performance White Paper
Intel DPDK Boosts Server Appliance Performance Intel DPDK Boosts Server Appliance Performance Introduction As network speeds increase to 40G and above, both in the enterprise and data center, the bottlenecks
Performance Evaluation of Linux Bridge
Performance Evaluation of Linux Bridge James T. Yu School of Computer Science, Telecommunications, and Information System (CTI) DePaul University ABSTRACT This paper studies a unique network feature, Ethernet
The Bus (PCI and PCI-Express)
4 Jan, 2008 The Bus (PCI and PCI-Express) The CPU, memory, disks, and all the other devices in a computer have to be able to communicate and exchange data. The technology that connects them is called the
PCI Express* Ethernet Networking
White Paper Intel PRO Network Adapters Network Performance Network Connectivity Express* Ethernet Networking Express*, a new third-generation input/output (I/O) standard, allows enhanced Ethernet network
Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking
Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking Burjiz Soorty School of Computing and Mathematical Sciences Auckland University of Technology Auckland, New Zealand
Performance of Software Switching
Performance of Software Switching Based on papers in IEEE HPSR 2011 and IFIP/ACM Performance 2011 Nuutti Varis, Jukka Manner Department of Communications and Networking (COMNET) Agenda Motivation Performance
How To Test A Microsoft Vxworks Vx Works 2.2.2 (Vxworks) And Vxwork 2.4.2-2.4 (Vkworks) (Powerpc) (Vzworks)
DSS NETWORKS, INC. The Gigabit Experts GigMAC PMC/PMC-X and PCI/PCI-X Cards GigPMCX-Switch Cards GigPCI-Express Switch Cards GigCPCI-3U Card Family Release Notes OEM Developer Kit and Drivers Document
Autonomous NetFlow Probe
Autonomous Ladislav Lhotka [email protected] Martin Žádník [email protected] TF-CSIRT meeting, September 15, 2005 Outline 1 2 Specification Hardware Firmware Software 3 4 Short-term fixes Test
High-Performance IP Service Node with Layer 4 to 7 Packet Processing Features
UDC 621.395.31:681.3 High-Performance IP Service Node with Layer 4 to 7 Packet Processing Features VTsuneo Katsuyama VAkira Hakata VMasafumi Katoh VAkira Takeyama (Manuscript received February 27, 2001)
Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build 164009
Performance Study Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build 164009 Introduction With more and more mission critical networking intensive workloads being virtualized
How To Improve Performance On A Linux Based Router
Linux Based Router Over 10GE LAN Cheng Cui, Chui-hui Chiu, and Lin Xue Department of Computer Science Louisiana State University, LA USA Abstract High speed routing with 10Gbps link speed is still very
The Elements of GigE Vision
What Is? The standard was defined by a committee of the Automated Imaging Association (AIA). The committee included Basler AG and companies from all major product segments in the vision industry. The goal
pco.interface GigE & USB Installation Guide
pco.interface GigE & USB Installation Guide In this manual you find installation instructions for the GigE Vision and USB2.0 interface on Microsoft Windows platforms. Target Audience: This camera is designed
USB readout board for PEBS Performance test
June 11, 2009 Version 1.0 USB readout board for PEBS Performance test Guido Haefeli 1 Li Liang 2 Abstract In the context of the PEBS [1] experiment a readout board was developed in order to facilitate
LCMON Network Traffic Analysis
LCMON Network Traffic Analysis Adam Black Centre for Advanced Internet Architectures, Technical Report 79A Swinburne University of Technology Melbourne, Australia [email protected] Abstract The Swinburne
The Lagopus SDN Software Switch. 3.1 SDN and OpenFlow. 3. Cloud Computing Technology
3. The Lagopus SDN Software Switch Here we explain the capabilities of the new Lagopus software switch in detail, starting with the basics of SDN and OpenFlow. 3.1 SDN and OpenFlow Those engaged in network-related
Collecting Packet Traces at High Speed
Collecting Packet Traces at High Speed Gorka Aguirre Cascallana Universidad Pública de Navarra Depto. de Automatica y Computacion 31006 Pamplona, Spain [email protected] Eduardo Magaña Lizarrondo
Analysis of QoS parameters of VOIP calls over Wireless Local Area Networks
Analysis of QoS parameters of VOIP calls over Wireless Local Area Networks Ayman Wazwaz, Computer Engineering Department, Palestine Polytechnic University, Hebron, Palestine, [email protected] Duaa sweity
General Pipeline System Setup Information
Product Sheet General Pipeline Information Because of Pipeline s unique network attached architecture it is important to understand each component of a Pipeline system in order to create a system that
UPPER LAYER SWITCHING
52-20-40 DATA COMMUNICATIONS MANAGEMENT UPPER LAYER SWITCHING Gilbert Held INSIDE Upper Layer Operations; Address Translation; Layer 3 Switching; Layer 4 Switching OVERVIEW The first series of LAN switches
Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging
Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging In some markets and scenarios where competitive advantage is all about speed, speed is measured in micro- and even nano-seconds.
Meeting the Five Key Needs of Next-Generation Cloud Computing Networks with 10 GbE
White Paper Meeting the Five Key Needs of Next-Generation Cloud Computing Networks Cloud computing promises to bring scalable processing capacity to a wide range of applications in a cost-effective manner.
Technical Brief. DualNet with Teaming Advanced Networking. October 2006 TB-02499-001_v02
Technical Brief DualNet with Teaming Advanced Networking October 2006 TB-02499-001_v02 Table of Contents DualNet with Teaming...3 What Is DualNet?...3 Teaming...5 TCP/IP Acceleration...7 Home Gateway...9
Effects of Filler Traffic In IP Networks. Adam Feldman April 5, 2001 Master s Project
Effects of Filler Traffic In IP Networks Adam Feldman April 5, 2001 Master s Project Abstract On the Internet, there is a well-documented requirement that much more bandwidth be available than is used
This is a product of VCCI Class B Compliance
FCC Warning This equipment has been tested and found to comply with the regulations for a Class B digital device, pursuant to Part 15 of the FCC Rules. These limits are designed to provide reasonable protection
Latency on a Switched Ethernet Network
Application Note 8 Latency on a Switched Ethernet Network Introduction: This document serves to explain the sources of latency on a switched Ethernet network and describe how to calculate cumulative latency
Oracle Database Scalability in VMware ESX VMware ESX 3.5
Performance Study Oracle Database Scalability in VMware ESX VMware ESX 3.5 Database applications running on individual physical servers represent a large consolidation opportunity. However enterprises
Accelerating High-Speed Networking with Intel I/O Acceleration Technology
White Paper Intel I/O Acceleration Technology Accelerating High-Speed Networking with Intel I/O Acceleration Technology The emergence of multi-gigabit Ethernet allows data centers to adapt to the increasing
Improving Quality of Service
Improving Quality of Service Using Dell PowerConnect 6024/6024F Switches Quality of service (QoS) mechanisms classify and prioritize network traffic to improve throughput. This article explains the basic
Packet Capture in 10-Gigabit Ethernet Environments Using Contemporary Commodity Hardware
Packet Capture in 1-Gigabit Ethernet Environments Using Contemporary Commodity Hardware Fabian Schneider Jörg Wallerich Anja Feldmann {fabian,joerg,anja}@net.t-labs.tu-berlin.de Technische Universtität
The proliferation of the raw processing
TECHNOLOGY CONNECTED Advances with System Area Network Speeds Data Transfer between Servers with A new network switch technology is targeted to answer the phenomenal demands on intercommunication transfer
Readout Unit using Network Processors
Readout Unit using Network Processors FEM FEM FEM FEM DAQ RU GbE GbE GbE GbE IBM NP4GS3 Phy GMII Phy GMII Phy GMII Phy GMII 4 x 1Gb full duplex Ethernet MACs ECS Ethernet CC-PC PCI Switch Bus IBM NP4GS3
SIDN Server Measurements
SIDN Server Measurements Yuri Schaeffer 1, NLnet Labs NLnet Labs document 2010-003 July 19, 2010 1 Introduction For future capacity planning SIDN would like to have an insight on the required resources
Performance Characteristics of a Cost-Effective Medium-Sized Beowulf Cluster Supercomputer
Res. Lett. Inf. Math. Sci., 2003, Vol.5, pp 1-10 Available online at http://iims.massey.ac.nz/research/letters/ 1 Performance Characteristics of a Cost-Effective Medium-Sized Beowulf Cluster Supercomputer
Attenuation (amplitude of the wave loses strength thereby the signal power) Refraction Reflection Shadowing Scattering Diffraction
Wireless Physical Layer Q1. Is it possible to transmit a digital signal, e.g., coded as square wave as used inside a computer, using radio transmission without any loss? Why? It is not possible to transmit
How To Monitor And Test An Ethernet Network On A Computer Or Network Card
3. MONITORING AND TESTING THE ETHERNET NETWORK 3.1 Introduction The following parameters are covered by the Ethernet performance metrics: Latency (delay) the amount of time required for a frame to travel
Gigabit Ethernet Design
Gigabit Ethernet Design Laura Jeanne Knapp Network Consultant 1-919-254-8801 [email protected] www.lauraknapp.com Tom Hadley Network Consultant 1-919-301-3052 [email protected] HSEdes_ 010 ed and
OpenFlow with Intel 82599. Voravit Tanyingyong, Markus Hidell, Peter Sjödin
OpenFlow with Intel 82599 Voravit Tanyingyong, Markus Hidell, Peter Sjödin Outline Background Goal Design Experiment and Evaluation Conclusion OpenFlow SW HW Open up commercial network hardware for experiment
Performance and Recommended Use of AB545A 4-Port Gigabit Ethernet Cards
Performance and Recommended Use of AB545A 4-Port Gigabit Ethernet Cards From Results on an HP rx4640 Server Table of Contents June 2005 Introduction... 3 Recommended Use Based on Performance and Design...
32-bit PCI 10/100/1000Mbps Gigabit Ethernet Card. User s Guide. FCC Warning. CE Mark Warning. VCCI Warning
FCC Warning 32-bit PCI 10/100/1000Mbps Gigabit Ethernet Card User s Guide This equipment has been tested and found to comply with the regulations for a Class B digital device, pursuant to Part 15 of the
PCI Express Overview. And, by the way, they need to do it in less time.
PCI Express Overview Introduction This paper is intended to introduce design engineers, system architects and business managers to the PCI Express protocol and how this interconnect technology fits into
Exploiting Remote Memory Operations to Design Efficient Reconfiguration for Shared Data-Centers over InfiniBand
Exploiting Remote Memory Operations to Design Efficient Reconfiguration for Shared Data-Centers over InfiniBand P. Balaji, K. Vaidyanathan, S. Narravula, K. Savitha, H. W. Jin D. K. Panda Network Based
Welcome to the Dawn of Open-Source Networking. Linux IP Routers Bob Gilligan [email protected]
Welcome to the Dawn of Open-Source Networking. Linux IP Routers Bob Gilligan [email protected] Outline About Vyatta: Open source project, and software product Areas we re working on or interested in
HANIC 100G: Hardware accelerator for 100 Gbps network traffic monitoring
CESNET Technical Report 2/2014 HANIC 100G: Hardware accelerator for 100 Gbps network traffic monitoring VIKTOR PUš, LUKÁš KEKELY, MARTIN ŠPINLER, VÁCLAV HUMMEL, JAN PALIČKA Received 3. 10. 2014 Abstract
VoIP in 802.11. Mika Nupponen. S-72.333 Postgraduate Course in Radio Communications 06/04/2004 1
VoIP in 802.11 Mika Nupponen S-72.333 Postgraduate Course in Radio Communications 06/04/2004 1 Contents Introduction VoIP & WLAN Admission Control for VoIP Traffic in WLAN Voice services in IEEE 802.11
Per-Flow Queuing Allot's Approach to Bandwidth Management
White Paper Per-Flow Queuing Allot's Approach to Bandwidth Management Allot Communications, July 2006. All Rights Reserved. Table of Contents Executive Overview... 3 Understanding TCP/IP... 4 What is Bandwidth
Analyzing the Virtualization Deployment Advantages of Two- and Four-Socket Server Platforms
IT@Intel White Paper Intel IT IT Best Practices: Data Center Solutions Server Virtualization August 2010 Analyzing the Virtualization Deployment Advantages of Two- and Four-Socket Server Platforms Executive
OpenFlow Based Load Balancing
OpenFlow Based Load Balancing Hardeep Uppal and Dane Brandon University of Washington CSE561: Networking Project Report Abstract: In today s high-traffic internet, it is often desirable to have multiple
Reconfigurable Architecture Requirements for Co-Designed Virtual Machines
Reconfigurable Architecture Requirements for Co-Designed Virtual Machines Kenneth B. Kent University of New Brunswick Faculty of Computer Science Fredericton, New Brunswick, Canada [email protected] Micaela Serra
Interfacing Credit Card-sized PCs to Board Level Electronics
10th ICALEPCS Int. Conf. on Accelerator & Large Expt. Physics Control Systems. Geneva, 10-14 Oct 2005, PO2.062-4 (2005) Interfacing Credit Card-sized PCs to Board Level Electronics Flavio Fontanelli 1,
Discussion Paper Category 6 vs Category 5e Cabling Systems and Implications for Voice over IP Networks
Discussion Paper Category 6 vs Category 5e Cabling Systems and Implications for Voice over IP Networks By Galen Udell Belden CDT Networking 2006 Category 6 vs Category 5e Cabling Systems and Implications
Communicating with devices
Introduction to I/O Where does the data for our CPU and memory come from or go to? Computers communicate with the outside world via I/O devices. Input devices supply computers with data to operate on.
Fusionstor NAS Enterprise Server and Microsoft Windows Storage Server 2003 competitive performance comparison
Fusionstor NAS Enterprise Server and Microsoft Windows Storage Server 2003 competitive performance comparison This white paper compares two important NAS operating systems and examines their performance.
Development. Igor Sheviakov Manfred Zimmer Peter Göttlicher Qingqing Xia. AGIPD Meeting 01-02 April, 2014
Textmasterformat AGIPD Firmware/Software bearbeiten Igor Sheviakov Manfred Zimmer Peter Göttlicher Qingqing Xia AGIPD Meeting 01-02 April, 2014 Outline Textmasterformat bearbeiten Reminder: hardware set-up
Network Design. Yiannos Mylonas
Network Design Yiannos Mylonas Physical Topologies There are two parts to the topology definition: the physical topology, which is the actual layout of the wire (media), and the logical topology, which
Virtualization: TCP/IP Performance Management in a Virtualized Environment Orlando Share Session 9308
Virtualization: TCP/IP Performance Management in a Virtualized Environment Orlando Share Session 9308 Laura Knapp WW Business Consultant [email protected] Applied Expert Systems, Inc. 2011 1 Background
FORTH-ICS / TR-375 March 2006. Experimental Evaluation of QoS Features in WiFi Multimedia (WMM)
FORTH-ICS / TR-375 March 26 Experimental Evaluation of QoS Features in WiFi Multimedia (WMM) Vasilios A. Siris 1 and George Stamatakis 1 Abstract We investigate the operation and performance of WMM (WiFi
How To Analyze The Security On An Ipa Wireless Sensor Network
Throughput Analysis of WEP Security in Ad Hoc Sensor Networks Mohammad Saleh and Iyad Al Khatib iitc Stockholm, Sweden {mohsaleh, iyad}@iitc.se ABSTRACT This paper presents a performance investigation
High-Density Network Flow Monitoring
Petr Velan [email protected] High-Density Network Flow Monitoring IM2015 12 May 2015, Ottawa Motivation What is high-density flow monitoring? Monitor high traffic in as little rack units as possible
1 Storage Devices Summary
Chapter 1 Storage Devices Summary Dependability is vital Suitable measures Latency how long to the first bit arrives Bandwidth/throughput how fast does stuff come through after the latency period Obvious
VLAN for DekTec Network Adapters
Application Note DT-AN-IP-2 VLAN for DekTec Network Adapters 1. Introduction VLAN (Virtual LAN) is a technology to segment a single physical network into multiple independent virtual networks. The VLANs
Demystifying Wireless for Real-World Measurement Applications
Proceedings of the IMAC-XXVIII February 1 4, 2010, Jacksonville, Florida USA 2010 Society for Experimental Mechanics Inc. Demystifying Wireless for Real-World Measurement Applications Kurt Veggeberg, Business,
Boosting Data Transfer with TCP Offload Engine Technology
Boosting Data Transfer with TCP Offload Engine Technology on Ninth-Generation Dell PowerEdge Servers TCP/IP Offload Engine () technology makes its debut in the ninth generation of Dell PowerEdge servers,
Configuring Your Computer and Network Adapters for Best Performance
Configuring Your Computer and Network Adapters for Best Performance ebus Universal Pro and User Mode Data Receiver ebus SDK Application Note This application note covers the basic configuration of a network
Dante: Know It, Use It, Troubleshoot It.
Dante: Know It, Use It, Troubleshoot It. SymNet Composer, Symetrix next generation DSP platform, has arrived and is in full effect. SymNet Composer software manages all aspects of SymNet Dante DSP devices
Building High-Performance iscsi SAN Configurations. An Alacritech and McDATA Technical Note
Building High-Performance iscsi SAN Configurations An Alacritech and McDATA Technical Note Building High-Performance iscsi SAN Configurations An Alacritech and McDATA Technical Note Internet SCSI (iscsi)
C-GEP 100 Monitoring application user manual
C-GEP 100 Monitoring application user manual 1 Introduction: C-GEP is a very versatile platform for network monitoring applications. The ever growing need for network bandwith like HD video streaming and
Intel Data Direct I/O Technology (Intel DDIO): A Primer >
Intel Data Direct I/O Technology (Intel DDIO): A Primer > Technical Brief February 2012 Revision 1.0 Legal Statements INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION WITH INTEL PRODUCTS. NO LICENSE,
The PCIe solution from ExaSAN
The PCIe solution from ExaSAN An outlook on how PCIe can help with your media and entertainment video environment. Who we are? Since 1995 Accusys Storage Ltd. has created generations of cutting-edge and
Analysis of Effect of Handoff on Audio Streaming in VOIP Networks
Beyond Limits... Volume: 2 Issue: 1 International Journal Of Advance Innovations, Thoughts & Ideas Analysis of Effect of Handoff on Audio Streaming in VOIP Networks Shivani Koul* [email protected]
Chapter 13 Selected Storage Systems and Interface
Chapter 13 Selected Storage Systems and Interface Chapter 13 Objectives Appreciate the role of enterprise storage as a distinct architectural entity. Expand upon basic I/O concepts to include storage protocols.
Benchmarking Cassandra on Violin
Technical White Paper Report Technical Report Benchmarking Cassandra on Violin Accelerating Cassandra Performance and Reducing Read Latency With Violin Memory Flash-based Storage Arrays Version 1.0 Abstract
82557 10 100 Mbps PCI LAN Controller A Guide to 82596 Compatibility
APPLICATION NOTE 82557 10 100 Mbps PCI LAN Controller A Guide to 82596 Compatibility Technical Marketing Network Products Division November 1995 Order Number 644126-001 Information in this document is
RFC 2544 Performance Evaluation for a Linux Based Open Router
RFC 2544 Performance Evaluation for a Linux Based Open Router Raffaele Bolla, Roberto Bruschi DIST - Department of Communications, Computer and Systems Science University of Genoa Via Opera Pia 13, 16145
EVALUATING THE NETWORKING PERFORMANCE OF LINUX-BASED HOME ROUTER PLATFORMS FOR MULTIMEDIA SERVICES. Ingo Kofler, Robert Kuschnig, Hermann Hellwagner
EVALUATING THE NETWORKING PERFORMANCE OF LINUX-BASED HOME ROUTER PLATFORMS FOR MULTIMEDIA SERVICES Ingo Kofler, Robert Kuschnig, Hermann Hellwagner Institute of Information Technology (ITEC) Alpen-Adria-Universität
Evaluation of 40 Gigabit Ethernet technology for data servers
Evaluation of 40 Gigabit Ethernet technology for data servers Azher Mughal, Artur Barczyk Caltech / USLHCNet CHEP-2012, New York http://supercomputing.caltech.edu Agenda Motivation behind 40GE in Data
Introduction to PCI Express Positioning Information
Introduction to PCI Express Positioning Information Main PCI Express is the latest development in PCI to support adapters and devices. The technology is aimed at multiple market segments, meaning that
Chapter 5 Cubix XP4 Blade Server
Chapter 5 Cubix XP4 Blade Server Introduction Cubix designed the XP4 Blade Server to fit inside a BladeStation enclosure. The Blade Server features one or two Intel Pentium 4 Xeon processors, the Intel
OpenFlow Switching: Data Plane Performance
This full text paper was peer reviewed at the direction of IEEE Communications Society subject matter experts for publication in the IEEE ICC 21 proceedings OpenFlow : Data Plane Performance Andrea Bianco,
STANDPOINT FOR QUALITY-OF-SERVICE MEASUREMENT
STANDPOINT FOR QUALITY-OF-SERVICE MEASUREMENT 1. TIMING ACCURACY The accurate multi-point measurements require accurate synchronization of clocks of the measurement devices. If for example time stamps
Three Key Design Considerations of IP Video Surveillance Systems
Three Key Design Considerations of IP Video Surveillance Systems 2012 Moxa Inc. All rights reserved. Three Key Design Considerations of IP Video Surveillance Systems Copyright Notice 2012 Moxa Inc. All
ZEN LOAD BALANCER EE v3.02 DATASHEET The Load Balancing made easy
ZEN LOAD BALANCER EE v3.02 DATASHEET The Load Balancing made easy OVERVIEW The global communication and the continuous growth of services provided through the Internet or local infrastructure require to
Switch Fabric Implementation Using Shared Memory
Order this document by /D Switch Fabric Implementation Using Shared Memory Prepared by: Lakshmi Mandyam and B. Kinney INTRODUCTION Whether it be for the World Wide Web or for an intra office network, today
R2. The word protocol is often used to describe diplomatic relations. How does Wikipedia describe diplomatic protocol?
Chapter 1 Review Questions R1. What is the difference between a host and an end system? List several different types of end systems. Is a Web server an end system? 1. There is no difference. Throughout
Performance Beyond PCI Express: Moving Storage to The Memory Bus A Technical Whitepaper
: Moving Storage to The Memory Bus A Technical Whitepaper By Stephen Foskett April 2014 2 Introduction In the quest to eliminate bottlenecks and improve system performance, the state of the art has continually
Rohde & Schwarz R&S SITLine ETH VLAN Encryption Device Functionality & Performance Tests
Rohde & Schwarz R&S Encryption Device Functionality & Performance Tests Introduction Following to our test of the Rohde & Schwarz ETH encryption device in April 28 the European Advanced Networking Test
HOW TO TEST 10 GIGABIT ETHERNET PERFORMANCE
HOW TO TEST 10 GIGABIT ETHERNET PERFORMANCE March 2012 Rev. B 03/12 SPIRENT 1325 Borregas Avenue Sunnyvale, CA 94089 USA Email: Web: [email protected] www.spirent.com AMERICAS 1-800-SPIRENT +1-818-676-2683
