RouteBricks: A Fast, Software- Based, Distributed IP Router
|
|
|
- Hugh Perry
- 9 years ago
- Views:
Transcription
1 outebricks: A Fast, Software- Based, Distributed IP outer Brad Karp UCL Computer Science (with thanks to Katerina Argyraki of EPFL for slides) CS GZ03 / M th November, 2009
2 One-Day oom Change! On Monday the 23 rd of November, we will meet in Chandler House G10 at the usual time! This room change will be one time only; we will revert to this room thereafter.
3 Distributed Systems Context Many models of consistency: sequential consistency (Ivy) close-to-open consistency (NFS) eventual consistency/conflict resolution despite disconnection (Bayou) atomic appends for multiple writers (GFS) Cluster as platform: Ivy, GFS Today: distributing an IP router over a cluster of PCs, for capacity and programmability highly parallel workload: packets forwarded independently (apart from flow ordering constraint) distributed within a single PC: multiple multi-core CPUs
4 outer = N packet processing + switching Fast = high, N Programmable = any processing 4
5 Why programmable routers? New ISP services» intrusion detec.on, applica.on accelera.on Monitoring» measure link latency, track down traffic New protocols» IP traceback, Trajectory Sampling, 5
6 Today: fast O programmable Fast hardware routers» processing by specialized hardware» not programmable» aggregate throughput: Tbps Programmable soiware routers» processing by general- purpose CPUs» aggregate throughput < 10Gbps 6
7 outebricks A router out of off- the- shelf PCs» familiar programming environment» large- volume manufacturing (cheap, widely available, growing in CPU and I/O with PCs) Can we build a Tbps router out of PCs? 7
8 A hardware router N linecards linecards Processing at rate ~ per linecard 8
9 A hardware router N linecards switch fabric linecards Processing at rate ~ per linecard Switching at rate N by switch fabric 9
10 outebricks N commodity interconnect servers servers Processing at rate ~ per server Switching at rate ~ per server 10
11 outebricks N commodity interconnect servers servers Per- server processing rate: c 11
12 Outline Interconnect Server oppmizapons Performance An applicapon Conclusions 12
13 Outline Interconnect Server oppmizapons Performance An applicapon Conclusions 13
14 equirements N commodity interconnect Internal link rates < Per- server processing rate: c Per- server fanout: constant 14
15 Straw Man: Direct Full Mesh N 15
16 Straw Man: Direct Full Mesh N N external links of capacity N 2 internal links of capacity 16
17 BeWer: Valiant Load Balancing [Valiant, Brebner, 1981] N /N /N 17
18 Valiant Load Balancing /N /N N Per- server processing rate: 3» processing overhead: 50% (2 without VLB) N 2 links of capacity 2/N 18
19 VLB with Uniform Traffic /N /N /N N /N /N /N 19
20 Direct Valiant LB: OpPmize for Uniform Traffic Load /N /N N /N /N /N 20
21 Direct Valiant LB /N N If uniform traffic matrix: 2 21
22 VLB Summary /N /N N Worst- case per- server processing rate: 3 If uniform traffic matrix: 2 22
23 Per- Server Fanout? /N N ConnecPvity degree: N per server What if not enough network ports? 23
24 SoluPon #1: Increase Server Capacity 2/N N e.g., double # external ports per server Doubles data rate on internal links Cuts fanout by half 24
25 SoluPon #1: Increase Server Capacity 2 2/N N e.g., double # external ports per server Doubles data rate on internal links, processing rate per server Cuts fanout by half 25
26 Per- Server Fanout? /N N 26
27 SoluPon #2: Add Intermediate Servers /N /N N k- degree, n- stage buwerfly ( k- ary n- fly ) Per- server fanout: k Stages: n = log k N 27
28 The outebricks Interconnect: CombinaPon Assign max external ports per server Full mesh if fanout allows Extra servers otherwise 28
29 Example Assuming current servers» 5 NICs, 2 x 10G ports or 8 x 1G ports» 1 external port per server N = 32 ports: full mesh» 32 servers N = 1024 ext. ports: 16- ary 4- fly» 3072 servers (2 extra servers per port) 29
30 ecap N Valiant Load Balancing + full mesh k- ary n- fly Per- server processing rate:
31 Outline Interconnect Server oppmizapons Performance An applicapon Conclusions 31
32 First Try: a Shared- Bus Server 64- byte packets North bridge Click Ports Memory Cores» Cloverton architecture» FSB: 2 x 1.33 GHz» CPUs: 2 x Xeon 2.4GHz 4- core» NICs: 2 x Intel XFS 2x10Gbps 32
33 First Try: a Shared- Bus Server 64- byte packets North bridge Click Ports Memory Cores 820Mbps 33
34 Problem #1: the Shared Bus 64- byte packets North bridge Click Ports Memory Cores FSB address bus saturated MulP- core alone is not enough 34
35 SoluPon: NUMA architecture I/O hub Mem Ports Cores Mem» Nehalem architecture» QuickPath interconnect» CPUs: 2 x Xeon 2.4GHz 4- core» NICs: 2 x Intel XFS 2x10Gbps 35
36 SoluPon: NUMA architecture I/O hub Mem Ports Cores Mem 820Mbps 1.3Gbps 36
37 Problem #2: Per- Packet Overhead I/O hub Mem Ports Cores Mem Bookkeeping operapons» moving packet descriptors between NIC and memory» upda.ng descriptor rings 37
38 SoluPon: Batching Poll- driven batching» poll mul.ple packets at a.me» reduces updates on descriptor rings» Click already supported it NIC- driven batching» relay mul.ple packet descriptors at a.me» reduces transac.ons on PCIe and I/O buses» changed NIC driver 38
39 SoluPon: Batching I/O hub Mem Ports Cores Mem 820Mbps 1.3Gbps 3Gbps 39
40 Problem #3: Queue Access I/O hub Mem Ports Cores Mem 40
41 Problem #3: Queue Access Ports Cores 41
42 Problem #3: Queue Access ule 1: one core per port 42
43 Problem #3: Queue Access 0.6 Gbps 1.2 Gbps 1.7 Gbps ule 1: one core per port ule 2: one core per packet 43
44 Problem #3: Queue Access 0.6 Gbps Can we always enforce both rules? What about when receiving on one Gbps Gbps port? 1.7 Gbps ule 1: one core per port ule 2: one core per packet 44
45 Problem #3: Queue Access ule 1: one core per port ule 2: one core per packet 45
46 Problem #3: Queue Access ule 1: one core per port ule 2: one core per packet 46
47 SoluPon: MulP- Queue NICs ule 1: one core per port ule 2: one core per packet 47
48 Single- Server Performance I/O hub Mem Ports Cores Mem 9.7Gbps 820Mbps 1.3Gbps 3Gbps 48
49 ecap State- of- the art hardware» NUMA architecture» mul.- queue NICs Wrote NIC driver» batching» lock- free queue access Careful queue- to- core allocapon» one core per queue» one core per packet 49
50 Outline Interconnect Server oppmizapons Performance An applicapon Conclusions 50
51 Single- Server Performance eal packet trace Gbps byte packets Forwarding IP roupng IPsec 51
52 Feasible outer Line ate eal packet trace Gbps byte packets Forwarding IP roupng IPsec Per- server processing rate: 2 3 eal packet mix: = 8 12 Gbps Small packets: = 2 3 Gbps 52
53 BoWlenecks Small packets: CPU» buses far from satura.on eal packet- size mix: none» limited PCIe lanes» only for the prototype Expected evolupon?» next Nehalem: 4 x 8 = 32 cores» 4 8 PCIe2.0 slots 53
54 Projected Single- Server Performance eal packet trace Gbps byte packets Forwarding IP roupng 54
55 Projected outer Line ate eal packet trace Gbps byte packets Forwarding IP roupng eal packet mix: = Gbps Small packets: = Gbps 55
56 B4 Prototype N = 4 external ports» 1 server per port» full mesh eal packet mix: 4 x 8.75 = 35 Gbps» expected = 8 12 Gbps Small packets: 4 x 3 = 12 Gbps» expected = 2 3 Gbps 56
57 What About Packet Order? TCP cuts sending rate by ½ if packets are ever reordered by more than 3 posipons But VLB sprays packets randomly across intermediate nodes! outebricks parpal solupon:» Assign packets on same flow to same receive queue» During any 100 ms interval, VLB forwards packets from same flow to same next hop 0.15% of packets reordered with this mechanism; 5.5% without it 57
58 Latency One server: 24 microseconds Three servers: 66.4 microseconds 58
59 outebricks Summary outebricks: fast soiware router» Valiant Load- Balanced cluster of commodity servers Programmable with Click Performance:» Easily = 1Gbps, N = 100s» = 10Gbps with next- genera.on servers Programming model for more complex funcponality? 59
RouteBricks: Exploiting Parallelism To Scale Software Routers
RouteBricks: Exploiting Parallelism To Scale Software Routers Mihai Dobrescu 1 and Norbert Egi 2, Katerina Argyraki 1, Byung-Gon Chun 3, Kevin Fall 3, Gianluca Iannaccone 3, Allan Knies 3, Maziar Manesh
Performance of Software Switching
Performance of Software Switching Based on papers in IEEE HPSR 2011 and IFIP/ACM Performance 2011 Nuutti Varis, Jukka Manner Department of Communications and Networking (COMNET) Agenda Motivation Performance
RouteBricks: Exploiting Parallelism To Scale Software Routers
RouteBricks: Exploiting Parallelism To Scale Software Routers Mihai Dobrescu 1 and Norbert Egi 2, Katerina Argyraki 1, Byung-Gon Chun 3, Kevin Fall 3, Gianluca Iannaccone 3, Allan Knies 3, Maziar Manesh
MIDeA: A Multi-Parallel Intrusion Detection Architecture
MIDeA: A Multi-Parallel Intrusion Detection Architecture Giorgos Vasiliadis, FORTH-ICS, Greece Michalis Polychronakis, Columbia U., USA Sotiris Ioannidis, FORTH-ICS, Greece CCS 2011, 19 October 2011 Network
Router Architectures
Router Architectures An overview of router architectures. Introduction What is a Packet Switch? Basic Architectural Components Some Example Packet Switches The Evolution of IP Routers 2 1 Router Components
Evaluating the Suitability of Server Network Cards for Software Routers
Evaluating the Suitability of Server Network Cards for Software Routers Maziar Manesh Katerina Argyraki Mihai Dobrescu Norbert Egi Kevin Fall Gianluca Iannaccone Eddie Kohler Sylvia Ratnasamy EPFL, UCLA,
Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck
Sockets vs. RDMA Interface over 1-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Pavan Balaji Hemal V. Shah D. K. Panda Network Based Computing Lab Computer Science and Engineering
Cray Gemini Interconnect. Technical University of Munich Parallel Programming Class of SS14 Denys Sobchyshak
Cray Gemini Interconnect Technical University of Munich Parallel Programming Class of SS14 Denys Sobchyshak Outline 1. Introduction 2. Overview 3. Architecture 4. Gemini Blocks 5. FMA & BTA 6. Fault tolerance
Lecture 1: the anatomy of a supercomputer
Where a calculator on the ENIAC is equipped with 18,000 vacuum tubes and weighs 30 tons, computers of the future may have only 1,000 vacuum tubes and perhaps weigh 1½ tons. Popular Mechanics, March 1949
OpenFlow with Intel 82599. Voravit Tanyingyong, Markus Hidell, Peter Sjödin
OpenFlow with Intel 82599 Voravit Tanyingyong, Markus Hidell, Peter Sjödin Outline Background Goal Design Experiment and Evaluation Conclusion OpenFlow SW HW Open up commercial network hardware for experiment
Packet-based Network Traffic Monitoring and Analysis with GPUs
Packet-based Network Traffic Monitoring and Analysis with GPUs Wenji Wu, Phil DeMar [email protected], [email protected] GPU Technology Conference 2014 March 24-27, 2014 SAN JOSE, CALIFORNIA Background Main
Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study
White Paper Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study 2012 Cisco and/or its affiliates. All rights reserved. This
Intel DPDK Boosts Server Appliance Performance White Paper
Intel DPDK Boosts Server Appliance Performance Intel DPDK Boosts Server Appliance Performance Introduction As network speeds increase to 40G and above, both in the enterprise and data center, the bottlenecks
Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks. An Oracle White Paper April 2003
Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks An Oracle White Paper April 2003 Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building
PCI Express High Speed Networks. Complete Solution for High Speed Networking
PCI Express High Speed Networks Complete Solution for High Speed Networking Ultra Low Latency Ultra High Throughput Maximizing application performance is a combination of processing, communication, and
Tyche: An efficient Ethernet-based protocol for converged networked storage
Tyche: An efficient Ethernet-based protocol for converged networked storage Pilar González-Férez and Angelos Bilas 30 th International Conference on Massive Storage Systems and Technology MSST 2014 June
Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging
Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging In some markets and scenarios where competitive advantage is all about speed, speed is measured in micro- and even nano-seconds.
50. DFN Betriebstagung
50. DFN Betriebstagung IPS Serial Clustering in 10GbE Environment Tuukka Helander, Stonesoft Germany GmbH Frank Brüggemann, RWTH Aachen Slide 1 Agenda Introduction Stonesoft clustering Firewall parallel
Control and forwarding plane separation on an open source router. Linux Kongress 2010-10-23 in Nürnberg
Control and forwarding plane separation on an open source router Linux Kongress 2010-10-23 in Nürnberg Robert Olsson, Uppsala University Olof Hagsand, KTH Jens Låås, UU Bengt Görden, KTH SUMMARY VERSION
FPGA-based Multithreading for In-Memory Hash Joins
FPGA-based Multithreading for In-Memory Hash Joins Robert J. Halstead, Ildar Absalyamov, Walid A. Najjar, Vassilis J. Tsotras University of California, Riverside Outline Background What are FPGAs Multithreaded
TCP Servers: Offloading TCP Processing in Internet Servers. Design, Implementation, and Performance
TCP Servers: Offloading TCP Processing in Internet Servers. Design, Implementation, and Performance M. Rangarajan, A. Bohra, K. Banerjee, E.V. Carrera, R. Bianchini, L. Iftode, W. Zwaenepoel. Presented
Collecting Packet Traces at High Speed
Collecting Packet Traces at High Speed Gorka Aguirre Cascallana Universidad Pública de Navarra Depto. de Automatica y Computacion 31006 Pamplona, Spain [email protected] Eduardo Magaña Lizarrondo
Lustre Networking BY PETER J. BRAAM
Lustre Networking BY PETER J. BRAAM A WHITE PAPER FROM CLUSTER FILE SYSTEMS, INC. APRIL 2007 Audience Architects of HPC clusters Abstract This paper provides architects of HPC clusters with information
Networking Virtualization Using FPGAs
Networking Virtualization Using FPGAs Russell Tessier, Deepak Unnikrishnan, Dong Yin, and Lixin Gao Reconfigurable Computing Group Department of Electrical and Computer Engineering University of Massachusetts,
Chronicle: Capture and Analysis of NFS Workloads at Line Rate
Chronicle: Capture and Analysis of NFS Workloads at Line Rate Ardalan Kangarlou, Sandip Shete, and John Strunk Advanced Technology Group 1 Motivation Goal: To gather insights from customer workloads via
First Midterm for ECE374 03/09/12 Solution!!
1 First Midterm for ECE374 03/09/12 Solution!! Instructions: Put your name and student number on each sheet of paper! The exam is closed book. You have 90 minutes to complete the exam. Be a smart exam
Sage ERP Accpac Online
Sage ERP Accpac Online Mac Resource Guide Thank you for choosing Sage ERP Accpac Online. This Resource Guide will provide important information and instructions on how you can get started using your Mac
ECE 358: Computer Networks. Solutions to Homework #4. Chapter 4 - The Network Layer
ECE 358: Computer Networks Solutions to Homework #4 Chapter 4 - The Network Layer P 4. Consider the network below. a. Suppose that this network is a datagram network. Show the forwarding table in router
Sage 300 ERP Online. Mac Resource Guide. (Formerly Sage ERP Accpac Online) Updated June 1, 2012. Page 1
Sage 300 ERP Online (Formerly Sage ERP Accpac Online) Mac Resource Guide Updated June 1, 2012 Page 1 Table of Contents 1.0 Introduction... 3 2.0 Getting Started with Sage 300 ERP Online using a Mac....
High Performance Computing in CST STUDIO SUITE
High Performance Computing in CST STUDIO SUITE Felix Wolfheimer GPU Computing Performance Speedup 18 16 14 12 10 8 6 4 2 0 Promo offer for EUC participants: 25% discount for K40 cards Speedup of Solver
Network Traffic Monitoring & Analysis with GPUs
Network Traffic Monitoring & Analysis with GPUs Wenji Wu, Phil DeMar [email protected], [email protected] GPU Technology Conference 2013 March 18-21, 2013 SAN JOSE, CALIFORNIA Background Main uses for network
Boosting Data Transfer with TCP Offload Engine Technology
Boosting Data Transfer with TCP Offload Engine Technology on Ninth-Generation Dell PowerEdge Servers TCP/IP Offload Engine () technology makes its debut in the ninth generation of Dell PowerEdge servers,
IP SAN Fundamentals: An Introduction to IP SANs and iscsi
IP SAN Fundamentals: An Introduction to IP SANs and iscsi Updated April 2007 Sun Microsystems, Inc. 2007 Sun Microsystems, Inc., 4150 Network Circle, Santa Clara, CA 95054 USA All rights reserved. This
Capacity Planning. The purpose of capacity planning is to:
The purpose of capacity planning is to: Determine Current Solution Capacity: How close to the ceiling am I today? Estimate Growth Potential: With current growth plans, when do I upgrade hardware? Answer
Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build 164009
Performance Study Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build 164009 Introduction With more and more mission critical networking intensive workloads being virtualized
VMware vsphere 4.1 Networking Performance
VMware vsphere 4.1 Networking Performance April 2011 PERFORMANCE STUDY Table of Contents Introduction... 3 Executive Summary... 3 Performance Enhancements in vsphere 4.1... 3 Asynchronous Transmits...
Latency Considerations for 10GBase-T PHYs
Latency Considerations for PHYs Shimon Muller Sun Microsystems, Inc. March 16, 2004 Orlando, FL Outline Introduction Issues and non-issues PHY Latency in The Big Picture Observations Summary and Recommendations
Michael Kagan. [email protected]
Virtualization in Data Center The Network Perspective Michael Kagan CTO, Mellanox Technologies [email protected] Outline Data Center Transition Servers S as a Service Network as a Service IO as a Service
Exploiting Remote Memory Operations to Design Efficient Reconfiguration for Shared Data-Centers over InfiniBand
Exploiting Remote Memory Operations to Design Efficient Reconfiguration for Shared Data-Centers over InfiniBand P. Balaji, K. Vaidyanathan, S. Narravula, K. Savitha, H. W. Jin D. K. Panda Network Based
RAMCloud and the Low- Latency Datacenter. John Ousterhout Stanford University
RAMCloud and the Low- Latency Datacenter John Ousterhout Stanford University Most important driver for innovation in computer systems: Rise of the datacenter Phase 1: large scale Phase 2: low latency Introduction
Networking Driver Performance and Measurement - e1000 A Case Study
Networking Driver Performance and Measurement - e1000 A Case Study John A. Ronciak Intel Corporation [email protected] Ganesh Venkatesan Intel Corporation [email protected] Jesse Brandeburg
Achieving a High-Performance Virtual Network Infrastructure with PLUMgrid IO Visor & Mellanox ConnectX -3 Pro
Achieving a High-Performance Virtual Network Infrastructure with PLUMgrid IO Visor & Mellanox ConnectX -3 Pro Whitepaper What s wrong with today s clouds? Compute and storage virtualization has enabled
Network Traffic Monitoring and Analysis with GPUs
Network Traffic Monitoring and Analysis with GPUs Wenji Wu, Phil DeMar [email protected], [email protected] GPU Technology Conference 2013 March 18-21, 2013 SAN JOSE, CALIFORNIA Background Main uses for network
EDUCATION. PCI Express, InfiniBand and Storage Ron Emerick, Sun Microsystems Paul Millard, Xyratex Corporation
PCI Express, InfiniBand and Storage Ron Emerick, Sun Microsystems Paul Millard, Xyratex Corporation SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA. Member companies
BookKeeper. Flavio Junqueira Yahoo! Research, Barcelona. Hadoop in China 2011
BookKeeper Flavio Junqueira Yahoo! Research, Barcelona Hadoop in China 2011 What s BookKeeper? Shared storage for writing fast sequences of byte arrays Data is replicated Writes are striped Many processes
GPU-Based Network Traffic Monitoring & Analysis Tools
GPU-Based Network Traffic Monitoring & Analysis Tools Wenji Wu; Phil DeMar [email protected], [email protected] CHEP 2013 October 17, 2013 Coarse Detailed Background Main uses for network traffic monitoring
Performance Evaluation of Linux Bridge
Performance Evaluation of Linux Bridge James T. Yu School of Computer Science, Telecommunications, and Information System (CTI) DePaul University ABSTRACT This paper studies a unique network feature, Ethernet
Benchmarking Cassandra on Violin
Technical White Paper Report Technical Report Benchmarking Cassandra on Violin Accelerating Cassandra Performance and Reducing Read Latency With Violin Memory Flash-based Storage Arrays Version 1.0 Abstract
Overview of Computer Networks
Overview of Computer Networks Client-Server Transaction Client process 4. Client processes response 1. Client sends request 3. Server sends response Server process 2. Server processes request Resource
Load Balancing Mechanisms in Data Center Networks
Load Balancing Mechanisms in Data Center Networks Santosh Mahapatra Xin Yuan Department of Computer Science, Florida State University, Tallahassee, FL 33 {mahapatr,xyuan}@cs.fsu.edu Abstract We consider
Flash Memory Arrays Enabling the Virtualized Data Center. July 2010
Flash Memory Arrays Enabling the Virtualized Data Center July 2010 2 Flash Memory Arrays Enabling the Virtualized Data Center This White Paper describes a new product category, the flash Memory Array,
Performance Tuning Guidelines for PowerExchange for Microsoft Dynamics CRM
Performance Tuning Guidelines for PowerExchange for Microsoft Dynamics CRM 1993-2016 Informatica LLC. No part of this document may be reproduced or transmitted in any form, by any means (electronic, photocopying,
Question: 3 When using Application Intelligence, Server Time may be defined as.
1 Network General - 1T6-521 Application Performance Analysis and Troubleshooting Question: 1 One component in an application turn is. A. Server response time B. Network process time C. Application response
How To Build A Policy Aware Switching Layer For Data Center Data Center Servers
A Policy-aware Switching Layer for Data Centers Dilip Joseph Arsalan Tavakoli Ion Stoica University of California at Berkeley 1 Problem: Middleboxes are hard to deploy Place on network path Overload path
Performance Baseline of Hitachi Data Systems HUS VM All Flash Array for Oracle
Performance Baseline of Hitachi Data Systems HUS VM All Flash Array for Oracle Storage and Database Performance Benchware Performance Suite Release 8.5 (Build 131015) November 2013 Contents 1 System Configuration
Ethernet. Ethernet Frame Structure. Ethernet Frame Structure (more) Ethernet: uses CSMA/CD
Ethernet dominant LAN technology: cheap -- $20 for 100Mbs! first widely used LAN technology Simpler, cheaper than token rings and ATM Kept up with speed race: 10, 100, 1000 Mbps Metcalfe s Etheret sketch
PCI Express* Ethernet Networking
White Paper Intel PRO Network Adapters Network Performance Network Connectivity Express* Ethernet Networking Express*, a new third-generation input/output (I/O) standard, allows enhanced Ethernet network
ethernet alliance Overview of Requirements and Applications for 40 Gigabit and 100 Gigabit Ethernet Version 1.0 August 2007 Authors:
Overview of Requirements and Applications for 40 Gigabit and 100 Gigabit Ethernet Version 1.0 August 2007 Authors: Mark Nowell, Cisco Vijay Vusirikala, Infinera Robert Hays, Intel ethernet alliance p.o.
Resource Utilization of Middleware Components in Embedded Systems
Resource Utilization of Middleware Components in Embedded Systems 3 Introduction System memory, CPU, and network resources are critical to the operation and performance of any software system. These system
HP ProLiant BL660c Gen9 and Microsoft SQL Server 2014 technical brief
Technical white paper HP ProLiant BL660c Gen9 and Microsoft SQL Server 2014 technical brief Scale-up your Microsoft SQL Server environment to new heights Table of contents Executive summary... 2 Introduction...
D1.2 Network Load Balancing
D1. Network Load Balancing Ronald van der Pol, Freek Dijkstra, Igor Idziejczak, and Mark Meijerink SARA Computing and Networking Services, Science Park 11, 9 XG Amsterdam, The Netherlands June [email protected],[email protected],
RAID. RAID 0 No redundancy ( AID?) Just stripe data over multiple disks But it does improve performance. Chapter 6 Storage and Other I/O Topics 29
RAID Redundant Array of Inexpensive (Independent) Disks Use multiple smaller disks (c.f. one large disk) Parallelism improves performance Plus extra disk(s) for redundant data storage Provides fault tolerant
An Oracle Technical White Paper November 2011. Oracle Solaris 11 Network Virtualization and Network Resource Management
An Oracle Technical White Paper November 2011 Oracle Solaris 11 Network Virtualization and Network Resource Management Executive Overview... 2 Introduction... 2 Network Virtualization... 2 Network Resource
IP SAN Best Practices
IP SAN Best Practices A Dell Technical White Paper PowerVault MD3200i Storage Arrays THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS AND TECHNICAL INACCURACIES.
Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers
Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers White Paper rev. 2015-11-27 2015 FlashGrid Inc. 1 www.flashgrid.io Abstract Oracle Real Application Clusters (RAC)
The Bus (PCI and PCI-Express)
4 Jan, 2008 The Bus (PCI and PCI-Express) The CPU, memory, disks, and all the other devices in a computer have to be able to communicate and exchange data. The technology that connects them is called the
Storage I/O Control: Proportional Allocation of Shared Storage Resources
Storage I/O Control: Proportional Allocation of Shared Storage Resources Chethan Kumar Sr. Member of Technical Staff, R&D VMware, Inc. Outline The Problem Storage IO Control (SIOC) overview Technical Details
High-Density Network Flow Monitoring
Petr Velan [email protected] High-Density Network Flow Monitoring IM2015 12 May 2015, Ottawa Motivation What is high-density flow monitoring? Monitor high traffic in as little rack units as possible
Evaluation Report: Emulex OCe14102 10GbE and OCe14401 40GbE Adapter Comparison with Intel X710 10GbE and XL710 40GbE Adapters
Evaluation Report: Emulex OCe14102 10GbE and OCe14401 40GbE Adapter Comparison with Intel X710 10GbE and XL710 40GbE Adapters Evaluation report prepared under contract with Emulex Executive Summary As
QoS & Traffic Management
QoS & Traffic Management Advanced Features for Managing Application Performance and Achieving End-to-End Quality of Service in Data Center and Cloud Computing Environments using Chelsio T4 Adapters Chelsio
Paolo Costa [email protected]
joint work with Ant Rowstron, Austin Donnelly, and Greg O Shea (MSR Cambridge) Hussam Abu-Libdeh, Simon Schubert (Interns) Paolo Costa [email protected] Paolo Costa CamCube - Rethinking the Data Center
MoonGen. A Scriptable High-Speed Packet Generator Design and Implementation. Paul Emmerich. January 30th, 2016 FOSDEM 2016
MoonGen A Scriptable High-Speed Packet Generator Design and Implementation Paul Emmerich January 30th, 2016 FOSDEM 2016 Chair for Network Architectures and Services Department of Informatics Paul Emmerich
Performance of STAR-System
Performance of STAR-System STAR-System Application Note Stuart Mills Performance of STAR-System STAR-Dundee s new software stack, STAR-System, provides a high performance software stack for accessing STAR-Dundee
Advanced Computer Networks. High Performance Networking I
Advanced Computer Networks 263 3501 00 High Performance Networking I Patrick Stuedi Spring Semester 2014 1 Oriana Riva, Department of Computer Science ETH Zürich Outline Last week: Wireless TCP Today:
Datacenter Operating Systems
Datacenter Operating Systems CSE451 Simon Peter With thanks to Timothy Roscoe (ETH Zurich) Autumn 2015 This Lecture What s a datacenter Why datacenters Types of datacenters Hyperscale datacenters Major
Disk Storage Shortfall
Understanding the root cause of the I/O bottleneck November 2010 2 Introduction Many data centers have performance bottlenecks that impact application performance and service delivery to users. These bottlenecks
High-performance vswitch of the user, by the user, for the user
A bird in cloud High-performance vswitch of the user, by the user, for the user Yoshihiro Nakajima, Wataru Ishida, Tomonori Fujita, Takahashi Hirokazu, Tomoya Hibi, Hitoshi Matsutahi, Katsuhiro Shimano
Scalable High Resolution Network Monitoring
Scalable High Resolution Network Monitoring Open Cloud Day Wintherthur, 16 th of June, 2016 Georgios Kathareios, Ákos Máté, Mitch Gusat IBM Research GmbH Zürich Laboratory {ios, kos, mig}@zurich.ibm.com
Oracle Database Scalability in VMware ESX VMware ESX 3.5
Performance Study Oracle Database Scalability in VMware ESX VMware ESX 3.5 Database applications running on individual physical servers represent a large consolidation opportunity. However enterprises
Interconnection Networks. Interconnection Networks. Interconnection networks are used everywhere!
Interconnection Networks Interconnection Networks Interconnection networks are used everywhere! Supercomputers connecting the processors Routers connecting the ports can consider a router as a parallel
PCI Express and Storage. Ron Emerick, Sun Microsystems
Ron Emerick, Sun Microsystems SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA. Member companies and individuals may use this material in presentations and literature
How To Set Up Foglight Nms For A Proof Of Concept
Page 1 of 5 Foglight NMS Overview Foglight Network Management System (NMS) is a robust and complete network monitoring solution that allows you to thoroughly and efficiently manage your network. It is
Open-source routing at 10Gb/s
Open-source routing at Gb/s Olof Hagsand, Robert Olsson and Bengt Gördén, Royal Institute of Technology (KTH), Sweden Email: {olofh, gorden}@kth.se Uppsala University, Uppsala, Sweden Email: [email protected]
The Lagopus SDN Software Switch. 3.1 SDN and OpenFlow. 3. Cloud Computing Technology
3. The Lagopus SDN Software Switch Here we explain the capabilities of the new Lagopus software switch in detail, starting with the basics of SDN and OpenFlow. 3.1 SDN and OpenFlow Those engaged in network-related
The virtualization of SAP environments to accommodate standardization and easier management is gaining momentum in data centers.
White Paper Virtualized SAP: Optimize Performance with Cisco Data Center Virtual Machine Fabric Extender and Red Hat Enterprise Linux and Kernel-Based Virtual Machine What You Will Learn The virtualization
SUN DUAL PORT 10GBase-T ETHERNET NETWORKING CARDS
SUN DUAL PORT 10GBase-T ETHERNET NETWORKING CARDS ADVANCED PCIE 2.0 10GBASE-T ETHERNET NETWORKING FOR SUN BLADE AND RACK SERVERS KEY FEATURES Low profile adapter and ExpressModule form factors for Oracle
Scaling in a Hypervisor Environment
Scaling in a Hypervisor Environment Richard McDougall Chief Performance Architect VMware VMware ESX Hypervisor Architecture Guest Monitor Guest TCP/IP Monitor (BT, HW, PV) File System CPU is controlled
Distributed File System. MCSN N. Tonellotto Complements of Distributed Enabling Platforms
Distributed File System 1 How do we get data to the workers? NAS Compute Nodes SAN 2 Distributed File System Don t move data to workers move workers to the data! Store data on the local disks of nodes
A network is a group of devices (Nodes) connected by media links. A node can be a computer, printer or any other device capable of sending and
NETWORK By Bhupendra Ratha, Lecturer School of Library and Information Science Devi Ahilya University, Indore Email: [email protected] Network A network is a group of devices (Nodes) connected by media
A Filesystem Layer Data Replication Method for Cloud Computing
World Telecom Congress 2012 Workshop on Cloud Computing in the Telecom Environment, Bridging the Gap A Filesystem Layer Data Replication Method for Cloud Computing Masanori Itoh, Kei-ichi Yuyama, Kenjirou
Intel Xeon Processor E5-2600
Intel Xeon Processor E5-2600 Best combination of performance, power efficiency, and cost. Platform Microarchitecture Processor Socket Chipset Intel Xeon E5 Series Processors and the Intel C600 Chipset
