Network Functions Virtualisation Bob Briscoe Chief Researcher BT + Don Clarke, Pete Willis, Andy Reid, Paul Veitch (BT) + further acknowledgements within slides
Network Functions Virtualisation Approach Message Router CDN Session Border Controller WAN Acceleration Independent Software Vendors DPI Firewall Carrier Grade NAT Tester/QoE monitor Orchestrated, automatic remote install hypervisors SGSN/GGSN PE Router BRAS Radio Network Controller Generic High Volume Servers Generic High Volume Storage Classical Network Appliance Approach Generic High Volume Ethernet Switches
If price-performance is good enough, rapid deployment gains come free Mar 12: Proof of Concept testing Combined BRAS & CDN functions on Intel Xeon Processor 5600 Series HP c7000 BladeSystem using Intel 82599 10 Gigabit Ethernet Controller sidecars BRAS chosen as an acid test CDN chosen as architecturally complements BRAS BRAS created from scratch so minimal functionality: PPPoE; only PTA, priority queuing; no RADIUS, VRFs CDN COTS fully functioning commercial product acknowledge: Test Equipment & Traffic Generators Video Viewing & Internet Browsing Ethernet Switch PPPoE 10GigE Intelligent MSE BRAS & CDN 10GigE Ethernet Switch Internet IPoE Content server Systems Management IP VPN Router IPoE Intranet server Significant management stack : 1. Instantiation of BRAS & CDN modules on bare server 2. Configuration of BRAS & Ethernet switches via Tail-F 3. Configuration of CDN via VVue mgt. sys. 4. Trouble2Resolve via HP mgmt system 3
Mar 12: Proof of Concept Performance Test Results == Test Description Result Id 1.1.1 Management access Pass 1.2.1 Command line configuration: add_sp_small Pass 1.2.2 Command line configuration: add_sub_small Pass 1.2.3 Command line configuration: del_sub_small Pass 1.2.4 Command line configuration: del_sp_small Pass 1.3.1 Establish PPPoE session Pass 1.4.1 Block unauthorized access attempt: invalid Pass password 1.4.2 Block unauthorized access attempt: invalid user Pass 1.4.3 Block unauthorized access attempt: invalid VLAN Pass 1.5.1 Time to restore 1 PPPoE session after BRAS reboot Pass 1.6.1 Basic Forwarding Pass 1.7.1 Basic QoS - Premium subscriber Pass 1.7.2 Basic QoS - Economy subscriber Pass 2.1.1 Command line configuration: add_sp_medium Pass 2.1.2 Command line configuration: add_sub_medium Pass 2.2.1 Establish 288 PPPoE sessions Pass 2.3.1 Performance forwarding: downstream to 288 Pass PPPoE clients 2.3.2 Performance forwarding: upstream from 288 PPPoE Pass clients 2.3.3 Performance forwarding: upstream and downstream Pass from/to 288 PPPoE clients 2.4.1 Time to restore 288 PPPoE sessions after BRAS Pass reboot 2.5.1 Dynamic configuration: add a subscriber Pass 2.5.2 Dynamic configuration: connect new subscribers to Pass BRAS 2.5.3 Dynamic configuration: delete a subscriber Pass 2.5.4 Dynamic configuration: delete service provider Pass 2.6.1 QoS performance medium configuration Pass 3.1.1 Command line configuration: add_sp_large Pass 3.1.2 Command line configuration: add_sub_large Pass 3.2.1 Establish 1024 PPPoE sessions Pass 3.3.1 Performance forwarding: downstream to 1024 Pass PPPoE clients 3.3.2 Performance forwarding: upstream from 1024 Pass Average 3 Million Packets Per Second per Logical Core for PPPoE processing. Equivalent to 94 M PPS/97 Gbps per Blade = 1.5 G PPS/1.5 Tbps per 10 U chassis 1. Test used 1024 PPP sessions & strict priority QoS Test used an Intel Xeon E5655 @ 3.0 GHz, 8 physical cores, 16 logical cores (not all used). Scaled to 9K PPPoE sessions per vbras. Can support 3 vbras per server. Subsequent research: implemented & testing software Hierarchical QoS results so far show processing is still not the bottleneck (also tested vcdn performance & video quality) very useful performance potential to match the performance per footprint of existing BRAS equipment 1 - Using128 byte packets. A single logical core handles traffic only in one direction so figures quoted are half-duplex. 4
3 Complementary but Independent Networking Developments Creates operational flexibility Reduces Reduces CapEx, OpEx, space & power delivery time consumption Network Functions Virtualisation Open Innovation Creates competitive supply of innovative applications by third parties Software Defined Networks Creates control abstractions to foster innovation. 5
New NfV Industry Specification Group (ISG) First meeting mid-jan 2013 > 150 participants > 100 attendees from > 50 firms Engagement terms under ETSI, but open to non-members non-members sign participation agreement essentially, must declare relevant IPR and offer it under fair & reasonable terms only per-meeting fees to cover costs Deliverables White papers identifying gaps and challenges as input to relevant standardisation bodies ETSI NfV collaboration portal white paper, published deliverables how to sign up, join mail lists, etc http://portal.etsi.org/portal/server.pt/community/nfv/367 Network-operator-driven ISG Initiated by 13 carriers shown Consensus in white paper Network Operator Council offers requirements grown to 23 members so far 6
gaps & challenges examples management & orchestration infrastructure management standards multi-level identity standard resource description language applications operating systems hypervisors compute infrastructure network infrastructure switching infrastructure rack, cable, power, cooling applications network functions operating systems hypervisors compute infrastructure switching infrastructure rack, cable, power, cooling security Topology Validation & Enforcement Availability of Management Support Infrastructure Secure Boot Secure Crash Performance Isolation Tenant Service Accounting 7
Q&A and spare slides
Domain Architecture NfV Applications Domain Carrier Management NfV Container Interface Virtual Network Container Interface Infrastructure Network Domain Virtual Machine Container Interface Hypervisor Domain Compute Container Interface Compute Domain Orchestration and Management Domain 9
NVF ISG Organisation Structure Assistant Technical Manager 10
ISG Working Group Structure Technical Steering Committee Chair: Technical Manager : Don Clarke (BT) Vice Chair / Assistant Technical Manager : Diego Lopez (TF) Programme Manager : TBA NOC Chair (ISG Vice Chair) + WG Chairs + Expert Group Leaders + Others Working Group Architecture of the Virtualisation Infrastructure Steve Wright (AT&T) + Yun Chao Hu (HW) Managing Editor: Andy Reid (BT) Working Group Management & Orchestration Diego Lopez (TF) + Raquel Morera (VZ) Working Group Software Architecture Fred Feisullin (Sprint) + Marie-Paule Odini (HP) Expert Group Performance & Portability Francisco Javier Ramón Salguero (TF) Expert Group Security Bob Briscoe (BT) Additional Expert Groups can be convened at discretion of Technical Steering Committee Working Group Reliability & Availability Chair: Naseem Khan (VZ) Vice Chair: Markus Schoeller (NEC) HW = Huawei TF = Telefonica VZ = Verizon 11
NfV Hypervisor Hypervisor Hypervisor Domain... General cloud hypervisor is designed for maximum application portability Hypervisor creates Virtual CPUs Virtual NICs Hypervisor provides Virtual Ethernet switch Hypervisor fully hides real CPUs and NICs NfV Hypervisor is aimed at removing packet bottlenecks Direct binding of to core Direct communication between s and between s and NIC Virtual machine mgt and API vswitch Any performance hardware Virtual machine mgt and API NIC Sequential thread emulation instruction policing, mapping and emulation core Sequential thread emulation......... Sequential thread emulation instruction policing, mapping, and emulation... core User mode polled drivers DMA remapping SR-IOV vswitch instruction policing, mapping and emulation Many features already emerging in hypervisors NfV performance hardware NIC core core... core 12
Low level operations Intermediate level operations Service level operations Orchestration and Infrastructure Ops Domain Automated deployment of NfV applications Orchestration console Higher level carrier OSS Tools exist for automated cloud deployment vsphere Openstack Cloudstack NfV infrastructure profile for NfV application to Select host Configure host Start (s) Application profile to specify Service address assignment (mechanism) Location specific configuration NfV applications domains Infrastructure network domain Carrier OSS domain Hypervisor domain Orchestration and infrastructure ops domain Compute domain 13