Datacenter Rack Switch Redundancy Models Server Access Ethernet Switch Connectivity Options

Similar documents
RSTP to MST Spanning Tree Migration in a Live Datacenter. NANOG47 October 20, 2009 Dani Roisman droisman ~ at ~ peakwebconsulting ~ dot ~ com

Course Contents CCNP (CISco certified network professional)

Juniper / Cisco Interoperability Tests. August 2014

Troubleshooting and Maintaining Cisco IP Networks Volume 1

Network Architecture Validated designs utilizing MikroTik in the Data Center

Expert Reference Series of White Papers. Planning for the Redeployment of Technical Personnel in the Modern Data Center

RESILIENT NETWORK DESIGN

CCNP SWITCH: Implementing High Availability and Redundancy in a Campus Network

DATA CENTER. Best Practices for High Availability Deployment for the Brocade ADX Switch

Chapter 3. Enterprise Campus Network Design

Chapter 4: Spanning Tree Design Guidelines for Cisco NX-OS Software and Virtual PortChannels

CHAPTER 10 LAN REDUNDANCY. Scaling Networks

Cisco Certified Network Associate Exam. Operation of IP Data Networks. LAN Switching Technologies. IP addressing (IPv4 / IPv6)

TechBrief Introduction

: Interconnecting Cisco Networking Devices Part 2 v1.1

Layer 3 Redundancy with HSRP By Sunset Learning Instructor Andrew Stibbards

Interconnecting Cisco Networking Devices Part 2

SSVP SIP School VoIP Professional Certification

Networking and High Availability

Juniper Networks EX Series/ Cisco Catalyst Interoperability Test Results. May 1, 2009

Configuring the Transparent or Routed Firewall

Introduction about cisco company and its products (network devices) Tell about cisco offered courses and its salary benefits (ccna ccnp ccie )

OVERLAYING VIRTUALIZED LAYER 2 NETWORKS OVER LAYER 3 NETWORKS

ASM Educational Center (ASM) Est. 1992

Networking and High Availability

How To Understand and Configure Your Network for IntraVUE

Switching in an Enterprise Network

Advanced Network Services Teaming

Migrate from Cisco Catalyst 6500 Series Switches to Cisco Nexus 9000 Series Switches

"Charting the Course...

Network Virtualization and Data Center Networks Data Center Virtualization - Basics. Qin Yin Fall Semester 2013

CCNP Switch Questions/Answers Implementing High Availability and Redundancy

Disaster Recovery Design Ehab Ashary University of Colorado at Colorado Springs

VXLAN: Scaling Data Center Capacity. White Paper

Networking Topology For Your System

Implementing L3 at the Data Center Access Layer on Juniper Networks Infrastructure

Recommendations for a redundant campus network Best Practice Document

Virtual PortChannels: Building Networks without Spanning Tree Protocol

How To Learn Cisco Cisco Ios And Cisco Vlan

vsphere Networking ESXi 5.0 vcenter Server 5.0 EN

NAS 307 Link Aggregation

IMPLEMENTING CISCO SWITCHED NETWORKS V2.0 (SWITCH)

Brocade to Cisco Comparisons

Implementing L2 at the Data Center Access Layer on Juniper Networks Infrastructure

NET ACCESS VOICE PRIVATE CLOUD

Digi Certified Transport Technician Training Course (DCTT)

Networking 4 Voice and Video over IP (VVoIP)

Barracuda Link Balancer

Redundancy & the Netnod Internet Exchange Points

A Principled Technologies white paper commissioned by Dell Inc.

Simulation of High Availability Internet Service Provider s Network

AT-S60 Version Management Software for the AT-8400 Series Switch. Software Release Notes

COURSE AGENDA. Lessons - CCNA. CCNA & CCNP - Online Course Agenda. Lesson 1: Internetworking. Lesson 2: Fundamentals of Networking

Redundancy and load balancing at L3 in Local Area Networks. Fulvio Risso Politecnico di Torino

A New Approach to Developing High-Availability Server

Magnum Network Software DX

Network Virtualization

ADVANCED NETWORK CONFIGURATION GUIDE

Designing Cisco Network Service Architectures ARCH v2.1; 5 Days, Instructor-led

Cisco Data Centre: Introducing Cisco Data Center Networking

SSVVP SIP School VVoIP Professional Certification

Redundancy and load balancing at L3 in Local Area Networks. Fulvio Risso Politecnico di Torino

Juniper Networks Certified Internet Specialist Fast Track

An Oracle White Paper October How to Connect Oracle Exadata to 10 G Networks Using Oracle s Ethernet Switches

Chapter 7 Configuring Trunk Groups and Dynamic Link Aggregation

Example: Advertised Distance (AD) Example: Feasible Distance (FD) Example: Successor and Feasible Successor Example: Successor and Feasible Successor

hp ProLiant network adapter teaming

Oracle Big Data Appliance: Datacenter Network Integration

Juniper Exam JN0-343 Juniper Networks Certified Internet Specialist (JNCIS-ENT) Version: 10.1 [ Total Questions: 498 ]

How To Set Up A Virtual Network On Vsphere (Vsphere) On A 2Nd Generation Vmkernel (Vklan) On An Ipv5 Vklan (Vmklan)

Voice Over IP. MultiFlow IP Phone # 3071 Subnet # Subnet Mask IP address Telephone.

2. Are explicit proxy connections also affected by the ARM config?

Outline VLAN. Inter-VLAN communication. Layer-3 Switches. Spanning Tree Protocol Recap

VIRGINIA DEPARTMENT OF TRANSPORTATION SPECIAL PROVISION FOR NETWORK SWITCHES TEST PLAN

SCOPE DOCUMENT. Trade Name IT- Network Systems Administration Post- Secondary DATE OF DISTRIBUTION VIA WEBSITE

Read Me First for the HP ProCurve Routing Switch 9304M and Routing Switch 9308M

: Interconnecting Cisco Networking Devices Part 2 v2.0 (ICND2)

CCT vs. CCENT Skill Set Comparison

Interconnecting Cisco Networking Devices, Part 2 Course ICND2 v2.0; 5 Days, Instructor-led

Using Virtual Switches in PowerVM to Drive Maximum Value of 10 Gb Ethernet

Designing High-Availability Services

Campus Network for High Availability Design Guide

Campus Network Best Practices: Core and Edge Networks

CompTIA Exam N CompTIA Network+ certification Version: 5.1 [ Total Questions: 1146 ]

Data Center Multi-Tier Model Design

Table of Contents. Cisco How Does Load Balancing Work?

SRX High Availability Design Guide

Configuring IPS High Bandwidth Using EtherChannel Load Balancing

Layer 3 Network + Dedicated Internet Connectivity

IP Routing Configuring Static Routes

Customer Managed Connectivity - Milan

Multihoming and Multi-path Routing. CS 7260 Nick Feamster January

High-Availability Enterprise Network Design

OSPF Routing Protocol

Data Sheet. V-Net Link 700 C Series Link Load Balancer. V-NetLink:Link Load Balancing Solution from VIAEDGE

AWS Direct Connect. User Guide API Version

Walmart s Data Center. Amadeus Data Center. Google s Data Center. Data Center Evolution 1.0. Data Center Evolution 2.0

Cisco Discovery 3: Introducing Routing and Switching in the Enterprise hours teaching time

Chapter 2 Lab 2-2, Configuring EtherChannel Instructor Version

Abstract. MEP; Reviewed: GAK 10/17/2005. Solution & Interoperability Test Lab Application Notes 2005 Avaya Inc. All Rights Reserved.

Transcription:

Datacenter Rack Switch Redundancy Models Server Access Ethernet Switch Connectivity Options NANOG46 June 16, 2009 Dani Roisman droisman ~ at ~ peakwebconsulting ~ dot ~ com

Introductions Slide 1

Server Access Explained Enterprise and Content datacenters are home to racks full of servers, providing connectivity via Ethernet Typically, an Ethernet switch is located within the rack to aggregate all access connectivity for local servers, then uplinked to a network core There are multiple ways in which to provide redundant connectivity to this top-of-rack switch, each with their own benefits and drawbacks Slide 2

Who May Be Interested? Network operators responsible for server connectivity to multiple racks these design options scale from one rack switch to hundreds of rack switches in a datacenter Folks whose jobs rely upon undisrupted connectivity to servers throughout their datacenter Organizations battling with network redundancy versus stability trade-offs Slide 3

Reference: Sample Network Distribution layer is made up of Layer 3 switches Upstream Core Network If small enough, Distribution + Core may be collapsed into same chassis Distribution Switch #1 Core / Distribution Distribution Switch #2 Server Rack Rack Switch #1 All servers connect to top-of-rack switch, which in turn has upstream connections to rest of network Server(s) Slide 4

Basic Definitions Layer 1 - physical cabling infrastructure, physical link, patch panels, Ethernet cords etc. Layer 2 - switching and bridging, VLANs etc. Layer 3 - IP routing, using static or dynamic protocols such as RIP, EIGRP, OSPF, BGP, etc. Distribution - provides upstream path to rest of the network (sometimes collapsed with Core) Rack Switch a.k.a. Top-of-Rack Switch - all servers within a rack connect to this switch, and this switch in turn has uplinks to the distribution equipment (often mounted at middle of rack) Slide 5

Model 0 No Redundancy Slide 6

No Redundancy Upstream Core Network Core / Distribution Distribution Switch #1 Server Rack Rack Switch #1 Server(s) Slide 7

Model 0.5 Manual Activation Slide 8

Manual Activation Upstream Core Network Backup port is either physically disconnected or administratively disabled Distribution Switch #1 Distribution Switch #2 Core / Distribution Server Rack Rack Switch #1 One NIC per server connected to switch Server(s) Slide 9

Manual Activation Explained Pretty simple configure typical L2 redundant environment, say I m afraid of L2 loops, and then shut down one of the uplinks In response to an primary link outage, requires manual intervention to enable the backup port Do yourself a favor, shut down at the distribution side Do yourself another favor, shut down the primary before enabling the backup Do yourself a third favor, stay away from this model Slide 10

Model 1 The Cliché Layer2 + STP Slide 11

The Cliché Layer2 + STP STP Priority 8K (8192) (primary root) Upstream Core Network STP Priority 16K (16384) (secondary root) Distribution Switch #1 Core / Distribution Intentional network loop Distribution Switch #2 Server Rack Rack switch set to STP priority 40K (40960) Rack Switch #1 X STP blocks towards higher cost to root One NIC per server connected to switch Server(s) Slide 12

L2 + STP Explained Network loops are intentionally created to provide redundant paths Utilizes spanning-tree protocol, 802.1d or preferably RSTP 802.1w, may utilize MST (802.1s) depending on equipment vendor and network size STP will automatically detect and block loops during normal conditions, and will unblock to provide failover during an outage Relies upon receipt and processing of BPDUs to decide where to block loops Slide 13

L2 + STP Redundancy STP Priority 8K (8192) (primary root) Upstream Core Network STP Priority 16K (16384) (secondary root) Distribution Switch #1 Core / Distribution Distribution Switch #2 Server Rack Primary link is severed Rack switch set to STP priority 40K (40960) Rack Switch #1 STP unblocks and service continues One NIC per server connected to switch Server(s) Slide 14

L2 + STP Benefits Extremely common model, simple to use and understand Easy to verify backup connectivity and available redundancy (via STP blocking state) Most flexible option, any subnet can be extended to any server in any rack via VLAN tagging, easy server mobility (e.g. VM) Allows for centralized services to be deployed in transparent mode (SLB, firewalling, etc.) Helps conserve public IPv4 addresses Slide 15

L2 + STP Drawbacks Stability varies based vendor and device model Limitations to interoperations in a multi-vendor environment Configuration errors, cabling errors, hardware failure can cause entire datacenter shut downs Some network troubles may be difficult to trace Heavily depends on rack switch CPU health Often requires extra tweaks to achieve desired performance / stability May not be able to utilize backup capacity Slide 16

Model 2 Full Layer 3 Slide 17

Full Layer 3 Both Distribution devices learn subnet via IGP from rack switch Upstream Core Network Routing Link Distribution Switch #1 Distribution Switch #2 Core / Distribution Server Rack 10.0.1.128/30 10.0.2.128/30 Rack Switch #1 Rack switch acts as servers default gateway All servers in this rack are exclusively numbered, multiple subnets are possible Server(s) One NIC per server connected to switch Slide 18

Full L3 Explained Distribution rack switch ports are configured as Layer 3 on both sides Distribution rack switch links are point-topoint routing links Default gateway for servers is rack switch Routing protocol is run between rack switch and distribution (typically IGP such as OSPF) Redundancy is based on standard routing protocols, e.g. link state, metric/cost Slide 19

Full L3 Benefits Simple to use and understand No L2 loops, no STP or associated risks Works very well for managed hosting environments with many small customer subnets Extremely robust, configuration errors will impact only single rack in worst case Easy to verify backup connectivity and available redundancy (via routing adjacencies) Supports multi-path to take advantage of all capacity Slide 20

Full L3 Drawbacks Significantly increases cost of rack switches Racks must be exclusively numbered, which means unique subnet(s) per rack Inefficient use of public IPv4 address space, one subnet must be allocated per rack /28 = 13 servers, /27 = 29 servers, /26 = 61 servers Relocating a server (or VM) to another rack means renumbering, limits NIC teaming Additional services must run in Layer3 mode (e.g. SLB, firewall), may require fancy routing tricks, PBR, etc. Slide 21

Model 3 Layer 3 Dist with Layer 2 RS Slide 22

Layer 3 Dist with Layer 2 RS Both Distribution devices announce subnet via IGP Upstream Core Network Routing Link Distribution Distribution Switch #1 servers default gateway Switch #2 10.10.10.2 HSRP or VRRP 10.10.10.3 10.10.10.1/24 Core / Distribution Redundancy protocol for Server Rack Can use inexpensive / unmanaged* Layer 2 rack switch Rack Switch #1 *managed always recommended All servers in this rack are exclusively numbered in 10.10.10.0/24 Server(s) One NIC per server connected to switch Slide 23

L3 Dist with L2 RS Explained Distribution ports are configured as Layer 3 (a.k.a. "routed ports ) Default gateway for servers on distribution using a redundancy protocol such as HSRP or VRRP Layer 2 adjacency is formed through the rack switch (a.k.a. V-shaped ) Distribution switches announce reachability via IGP, as with standard L2 model Multiple subnets available via 802.1q tagged sub-interfaces from distribution to rack switch Slide 24

Layer 3 with Layer 2 RS Redundancy Slide 25

L3 Dist with L2 RS Benefits STP elimination as with full L3 model Lower cost than full L3 model, since rack switches are L2 only In fact, can run ultra-cheap commodity switches, e.g. not even manageable Slide 26

L3 Dist with L2 Drawbacks As with full L3 model, requires exclusive numbering, which means subnet can only exist on that one rack, as well as public IPv4 inefficiencies Cannot use upstream backup link capacity without additional configurations (e.g. multiple HSRP/VRRP groups w/alternating priorities and alternating default gateways on servers) May impose multi-netting if multiple subnets are required and 802.1q sub-if is not available Not very common or straightforward Slide 27

Model 4 Link Failover Slide 28

Link Failover Upstream Core Network Distribution Switch #1 Distribution Switch #2 Core / Distribution Server Rack Identify primary port and associate with backup port Rack Switch #1 Uplink failover feature disables traffic on (or shuts) designated backup port as long as primary port is operational One NIC per server connected to switch Server(s) Slide 29

Link Failover Explained This model uses switches connected similarly to the L2 model, but with an L1 redundancy Instead of creating a loop for backup and blocking using STP, the RS automatically disables all forwarding on (or shuts down) the backup link Many vendors support this: Cisco calls this Flex Link Foundry calls this Protected Link Groups Juniper (EX) calls this Redundant Trunk Links Force10 calls this Redundant Pairs Slide 30

Link Failover Redundancy Upstream Core Network Distribution Switch #1 Distribution Switch #2 Core / Distribution Server Rack Primary link is severed Rack Switch #1 Uplink failover feature enables traffic on designated backup port One NIC per server connected to switch Server(s) Slide 31

Link Failover Benefits Not based on health of distribution and rack switch CPUs, may be more reliable No complex state machines as with STP No loops during steady-state, backup link never passes production traffic while primary is up May provide rapid failover (depends on certain conditions) Reduced interoperability concerns, if the RS provides this feature, there is nothing needed from the distribution equipment Same server mobility & subnet flexibility as L2 Slide 32

Link Failover Drawbacks Some vendors place the backup port in a down state: link is always shown as down therefore cannot confirm interface health or guarantee redundancy without actually testing failover cannot map backup interfaces using LLDP/CDP/FDP may cause delays in failover due to STP discovery as link in distribution needs to transition to forwarding Lack of protocol (STP / IGP) limits knowledge, redundancy may not be complete surprise! Cannot load-balance traffic over redundant link STP usually disabled, may cause loops! Slide 33

Link Failover Questions Is anyone using this? Really though using it widely (all racks in a datacenter)? Why not? Sure, it s not as flexible as STP, but doesn t it provide exactly what we need at the dualuplinked rack switch? Slide 34

Model 5 Layer 2 Dist with Layer 3 RS Slide 35

Layer 2 Dist with Layer 3 RS Distribution devices are two separate switch fabrics, RSs form IGP adjacencies with other RSs and Cores Upstream Core Network Can use inexpensive / unmanaged* Distribution switch *managed always recommended Dist Switch #1 Dist Switch #2 Core / Distribution Dist #1 Routing 10.100.1.0/24 Dist #2 Routing 10.100.2.0/24 Server Rack Rack Switch #1 Rack switch acts as servers default gateway All servers in this rack are exclusively numbered, multiple subnets are possible Server(s) One NIC per server connected to switch Slide 36

L2 Dist With L3 RS Explained Each distribution switch acts as a switch fabric for a unique multi-access routing subnet, dumb switch with no routing functionality Layer 3 RSs form IGP adjacencies with all other RSs and Layer 3 cores over two diverse subnets RS uplink ports are configured as Layer 3 ports Default gateway for servers is rack switch Redundancy is based on standard routing protocols, e.g. link state, metric/cost Slide 37

Expanded L2 Dist with L3 RS ISPs etc. Layer 3 Router #1 Layer 3 Router #2 Core / Border Dist #1 Routing 10.100.1.0/24 Dist #2 Routing 10.100.2.0/24 L2 Dist Switch #1 L2 Dist Switch #2 L3 Rack Switch #1 L3 Rack Switch #2 L3 Rack Switch #3 L3 Rack Switch #4 192.168.1.0/24 192.168.2.0/24 192.168.3.0/24 192.168.4.0/24 Servers Servers Servers Servers Rack #1 Rack #2 Rack #3 Rack #4 Slide 38

L2 Dist With L3 RS Benefits All the benefits of full L3, such as no loops or STP, ability to utilize redundant uplinks for additional capacity during steady-state If application/product requirements require Layer3 access devices, will allow for some economy in the distribution hardware (don t even need VLAN support in the L2 distribution) Introduces a novel any-to-any routing over L2 switching fabric model into the datacenter Slide 39

L2 Dist With L3 RS Drawbacks As with full L3 model, more expensive RSs As with full L3 model, requires exclusive numbering, which means subnet can only exist on that one rack, as well as public IPv4 inefficiencies Since uplinks are not point-to-point, far-end outages are only detected after hold or keepalive timer expiration, may experience short-term blackhole during link failover Not very common or straightforward Slide 40

Model 6 Server-Based (Multi NIC) Slide 41

Server-Based (Multi NIC) Upstream Core Network Distribution Switch #1 Core / Distribution No loops or redundancy protocols in Distribution Distribution Switch #2 Server Rack Two switches per rack, singly connected to diverse Distribution Rack Switch #1 Rack Switch #2 X Two NICs per server connected to diverse switches, using teaming or bonding * Server(s) *recommended active-backup mode Slide 42

Server-Based Explained This model assigns redundancy burden to the servers Supply two rack switches for server access, each with diverse uplinks May be the best bet when using blade chassis with integrated switches Different methods available for servers to determine NIC usability: Link state ARP queries for default gateway Ping default gateway Slide 43

Server-Based Redundancy 1 Upstream Core Network Distribution Switch #1 Core / Distribution No loops or redundancy protocols in Distribution Distribution Switch #2 Server Rack Two switches per rack, singly connected to diverse Distribution Rack Switch #1 Rack Switch #2 Primary link is severed Server(s) NIC redundancy protocol on server detects primary NIC failure and transitions service to backup NIC Slide 44

Server-Based Redundancy 2 Upstream Core Network Distribution Switch #1 Core / Distribution No loops or redundancy protocols in Distribution Distribution Switch #2 Server Rack Uplink becomes severed Rack Switch #1 Rack Switch #2 Should have method to shut downlink ports Server(s) NIC redundancy protocol on server detects primary NIC failure and transitions service to backup NIC Slide 45

Server-Based Benefits Will work with pretty much any switch if your budget is tight you can shift some $$ from the rack access layer back to the dist / core / border layer where you need it more Same IP/subnet/VLAN flexibility as L2 models Network requires minimal attention, good for organizations that have more sysadmin skill, requiring less network engineering Network has less risk of Layer 2 or Layer 3 troubles, no worry of STP / IGP meltdowns due to misconfiguration or operator error Slide 46

Server-Based Drawbacks Increases complexity of server configuration No good centralized way of assuring 100% redundancy (have to check at each server) Uses twice the number of NICs on each server Uses twice the number of RSs (but, you may want this anyway, to eliminate RS SPOF) Should have way to signal server downlink ports of RS uplink failure, otherwise may blackhole server traffic Redundant capacity is not utilized (if configured as recommended) Slide 47

Extras Slide 48

Extra Redundancy Helpers Most failover schemes rely upon discovery of link status changes (port up / down), which sometimes may be delayed or not properly reported Remote fault signaling (link negotiation) Unidirectional Dead Link Detection (UDLD) is useful especially with fiber links, but beware buggy implementations Bidirectional Forwarding Detection (BFD) is more promising than UDLD, maybe even performed in port ASIC, but stable/reliable implementations are still few Slide 49

Extra Models There are some new models surfacing that are based on link aggregation at the rack server side, running to diverse equipment in the distribution layer: Link aggregation across different members of a switch stack (multi-vendor) Multi-chassis EtherChannel ( MEC - Cisco VSS), kind of like different members of a switch stack Virtual Port Channel ( vpc - Cisco Nexus 7000) Ink is still drying for some of these, the models included in this talk have been well-baked Slide 50

Summary Slide 51

Layer 3 versus Layer 2 Great quote from Cisco design document which summarizes the difference between Layer 3 and Layer 2: A routing protocol identifies where to send packets STP identifies where not to send frames Source: http://www.cisco.com/en/us/docs/solutions/enterprise/data_center/dc_3_0/dc-3_0_ipinfra.html Slide 52

Summary We have deployed and operated the preceding models over the past 12 years Different datacenter environments, application and security requirements, and equipment capabilities will lead to different solutions This list may not be exhaustive, come find me if you have other ways you ve done this both successes and failures! Is the L2/STP model seeing a comeback due to server virtualization? Slide 53

Any Questions? Thank you for listening Peak Web Consulting is available to assist Dani Roisman droisman ~ at ~ peakwebconsulting ~ dot ~ com