Data Center Evolution and Network Convergence. Joseph L White, Juniper Networks



Similar documents
Data Center Convergence. Ahmad Zamer, Brocade

Best Practice and Deployment of the Network for iscsi, NAS and DAS in the Data Center

Technical Overview of Data Center Networks Joseph L White, Juniper Networks

Fibre Channel over Ethernet in the Data Center: An Introduction

FIBRE CHANNEL OVER ETHERNET

Brocade One Data Center Cloud-Optimized Networks

Data Center Networking Designing Today s Data Center

Server and Storage Consolidation with iscsi Arrays. David Dale, NetApp

Expert Reference Series of White Papers. Planning for the Redeployment of Technical Personnel in the Modern Data Center

Virtualizing the SAN with Software Defined Storage Networks

Ethernet Fabrics: An Architecture for Cloud Networking

BUILDING A NEXT-GENERATION DATA CENTER

Extreme Networks: Building Cloud-Scale Networks Using Open Fabric Architectures A SOLUTION WHITE PAPER

The evolution of Data Center networking technologies

Enterasys Data Center Fabric

A Whitepaper on. Building Data Centers with Dell MXL Blade Switch

NETWORKING FOR DATA CENTER CONVERGENCE, VIRTUALIZATION & CLOUD. Debbie Montano, Chief Architect dmontano@juniper.net

全 新 企 業 網 路 儲 存 應 用 THE STORAGE NETWORK MATTERS FOR EMC IP STORAGE PLATFORMS

Network Virtualization and Data Center Networks Data Center Virtualization - Basics. Qin Yin Fall Semester 2013

The Future of Computing Cisco Unified Computing System. Markus Kunstmann Channels Systems Engineer

Cisco Datacenter 3.0. Datacenter Trends. David Gonzalez Consulting Systems Engineer Cisco

3G Converged-NICs A Platform for Server I/O to Converged Networks

Lecture 02a Cloud Computing I

VDI Optimization Real World Learnings. Russ Fellows, Evaluator Group

Data Center Evolution without Revolution

M.Sc. IT Semester III VIRTUALIZATION QUESTION BANK Unit 1 1. What is virtualization? Explain the five stage virtualization process. 2.

Juniper Networks QFabric: Scaling for the Modern Data Center

UNDERSTANDING DATA DEDUPLICATION. Tom Sas Hewlett-Packard

How To Create A Large Enterprise Cloud Storage System From A Large Server (Cisco Mds 9000) Family 2 (Cio) 2 (Mds) 2) (Cisa) 2-Year-Old (Cica) 2.5

iscsi Top Ten Top Ten reasons to use Emulex OneConnect iscsi adapters

Cloud Computing and the Internet. Conferenza GARR 2010

The Coming Decade of Data Center Networking Discontinuities

Ethernet Fabric Requirements for FCoE in the Data Center

Introducing Brocade VCS Technology

Simplifying the Data Center Network to Reduce Complexity and Improve Performance

Storage Protocol Comparison White Paper TECHNICAL MARKETING DOCUMENTATION

PCI Express Impact on Storage Architectures and Future Data Centers. Ron Emerick, Oracle Corporation

ADVANCED DEDUPLICATION CONCEPTS. Larry Freeman, NetApp Inc Tom Pearce, Four-Colour IT Solutions

Deploying Brocade VDX 6720 Data Center Switches with Brocade VCS in Enterprise Data Centers

Best Practices Guide: Network Convergence with Emulex LP21000 CNA & VMware ESX Server

Intel Ethernet Switch Converged Enhanced Ethernet (CEE) and Datacenter Bridging (DCB) Using Intel Ethernet Switch Family Switches

Ethernet, and FCoE Are the Starting Points for True Network Convergence

WHITE PAPER. Best Practices in Deploying Converged Data Centers

Global Headquarters: 5 Speen Street Framingham, MA USA P F

SAN and NAS Bandwidth Requirements

Block based, file-based, combination. Component based, solution based

Simplifying Virtual Infrastructures: Ethernet Fabrics & IP Storage

Unified Storage Networking

Technical Brief: Introducing the Brocade Data Center Fabric

Optimizing Data Center Networks for Cloud Computing

Server and Storage Virtualization with IP Storage. David Dale, NetApp

How To Design A Data Centre

David Lawler Vice President Server, Access & Virtualization Group

Juniper Update Enabling New Network Architectures. Debbie Montano Chief Architect, Gov t, Edu & Medical dmontano@juniper.

Next-Generation Federal Data Center Architecture

FCoE Enabled Network Consolidation in the Enterprise Data Center

Storage Networking Foundations Certification Workshop

Cloud-ready network architecture

Brocade Solution for EMC VSPEX Server Virtualization

HP Virtual Connect. Tarass Vercešuks / 3 rd of October, 2013

PCI Express IO Virtualization Overview

STATE OF THE ART OF DATA CENTRE NETWORK TECHNOLOGIES CASE: COMPARISON BETWEEN ETHERNET FABRIC SOLUTIONS

High Performance Computing OpenStack Options. September 22, 2015

Virtualized Converged Data Centers & Cloud how these trends are effecting Optical Networks

UNDERSTANDING DATA DEDUPLICATION. Thomas Rivera SEPATON

Next Gen Data Center. KwaiSeng Consulting Systems Engineer

Deliver Fabric-Based Infrastructure for Virtualization and Cloud Computing

Accelerating Applications and File Systems with Solid State Storage. Jacob Farmer, Cambridge Computer

I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology

Next-Gen Securitized Network Virtualization

Towards an Open Data Center with an Interoperable Network (ODIN) Volume 1: Transforming the Data Center Network Last update: May 2012

Understanding Enterprise NAS

Building the Virtual Information Infrastructure

Trends in Application Recovery. Andreas Schwegmann, HP

IP ETHERNET STORAGE CHALLENGES

CompTIA Storage+ Powered by SNIA

UNDERSTANDING DATA DEDUPLICATION. Jiří Král, ředitel pro technický rozvoj STORYFLEX a.s.

over Ethernet (FCoE) Dennis Martin President, Demartek

TRILL for Service Provider Data Center and IXP. Francois Tallet, Cisco Systems

Simplify Your Data Center Network to Improve Performance and Decrease Costs

Converged Networking Solution for Dell M-Series Blades. Spencer Wheelwright

Deploying Public, Private, and Hybrid Storage Clouds. Marty Stogsdill, Oracle

Fibre Channel over Ethernet: Enabling Server I/O Consolidation

Towards an Open Data Center with an Interoperable Network Volume 5: WAN and Ultra Low Latency Applications Last update: May 2012

HP FlexNetwork and IPv6

Building Tomorrow s Data Center Network Today

Overview of Requirements and Applications for 40 Gigabit and 100 Gigabit Ethernet

Walmart s Data Center. Amadeus Data Center. Google s Data Center. Data Center Evolution 1.0. Data Center Evolution 2.0

Transcription:

Joseph L White, Juniper Networks

SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA unless otherwise noted. Member companies and individual members may use this material in presentations and literature under the following conditions: Any slide or slides used must be reproduced in their entirety without modification The SNIA must be acknowledged as the source of any material used in the body of any document containing material from these presentations. This presentation is a project of the SNIA Education Committee. Neither the author nor the presenter is an attorney and nothing in this presentation is intended to be, or should be construed as legal advice or an opinion of counsel. If you need legal advice or a legal opinion please contact your attorney. The information presented herein represents the author's personal opinion and current understanding of the relevant issues involved. The author, the presenter, and the SNIA do not assume any responsibility or liability for damages arising out of any reliance on or use of this information. NO WARRANTIES, EXPRESS OR IMPLIED. USE AT YOUR OWN RISK. 2

Abstract FC, FCoE, NAS, iscsi, DCB, traditional LAN, internet/wan, HPC, clusters, clouds, server virtualization, storage virtualization, network virtualization, and more are all colliding in your data center. Redundancy, resiliency, security, I/O consolidation, network convergence, dynamic application distribution, and thin provisioning with high levels of service are desired at all layers and all data center sizes across a broad spectrum of use cases. You worry about operational separation, buying decisions, investment protection, cost and energy savings, and evolving standards while maintaining very high levels of service and security. Is the technology evolving to a dream come true or a nightmare? If that doesn t keep you up at night nothing will. This tutorial will untangle, define, and illustrate the main ideas and concepts behind to give context and a solid foundation for discussions with your vendors as well as for your further reading and investigation. The point of view taken for this presentation is that of the network and transport characteristics in the face of the changes taking place. 3

Topics Discussed Data Center Evolution History Influences, Trends, Drivers Data Center Map Network Convergence Techniques and Technology Complexity Scaling Tiers I/O Consolidation SAN-LAN Convergence 4

Some History First Model isolated systems direct attach storage Second Model limited networking direct attach storage Third Model Networking Explosion direct attach storage + local cluster server-server storage Third and a half Model Flexible storage via NAS and network file systems Fourth Model SANs for block storage attach plus fully entrenched NAS Fifth Model ( current one ) Server Virtualization drives first hop I/O consolidation, increased SAN attach Cloud (pooled resources of all kinds with uniform distributed access) Evolving Model Network Convergence Protocols for SAN and LAN on same infrastructure Network Scaling via virtualization and simplification (tier collapsing, distributed control planes) Increasing Scale Increasing Features 5

Recent Trends in the Data Center Consolidation Server Trends Mega DCs; 400K sq ft 4K racks, 200K servers Multi-core (8->16 >32,.128, ) Virtualization and VMs Apps Apps Guest OS Guest OS... Hypervisor (VMM) Mgt Hardware DC Scale Want Low Oversubscription Application Trends Interconnect Trends SOA, Web 2.0 MapReduce, Hadoop, Grids Banking System Web Service Request - Pay Balance (SOAP over HTTP) Make_Payment - Web Service Credit Card Systems Convergence to 10 GE Enhancements to Ethernet IP Stack Application Middleware HPC Stack FC Stack iscsi Stack Web Service Response Transaction Confirmed (SOAP over HTTP) DCB Link 10 GigE Physical Increased East-West traffic Large speed increases 10/40/100 GE 6

Explosion in infrastructure Function of new services, digital media, application design and deployment topologies Fueled by Moore s law and ever increasing price performance Driving issues of managing massive performance and scale 11.8M 15.8M 522 EB 44 EB/yr Servers in US 1 New Storage Shipped 1 (per year) Global IP Traffic (per year) 146 EB 5 EB/yr 2.6M 1 EB/yr 2.6M 54EB 1995 2000 2005 2010 Complexity increases exponentially with scale 1 Source: IDC 7

Servers and Storage Attach High End Servers with high IOPS & performance MF / Unix Mid range servers Unix / Wintel Storage is High End High IOPS Storage with Reasonable balance of capability & performance DB tier application tier SAN Attach started at the high end SAN Attach Penetrating lower Server virtualization driving towards 100% attach Rack Dense & blade servers Wintel / Linux Storage with minimal load, no persistent data, just VM images Presentation tier Server virtualization started at the low end Steadily penetrating higher Needs shared storage for VMs 8

Virtualization OF EVERYTHING Aggregate up and Virtualize down many examples such as storage arrays, servers,... avoid Accidental partitioning embrace Deliberate partitioning Aggregation Physical and Software Bring together and pool capacity with flexible connectivity Virtualization logical partitions of the aggregated systems to match actual need flexibility fungible resources everywhere Utility Infrastructure with just in time & thin provisioning THIS IS HAPPENING TO NETWORKS AS WELL 9

Virtual to Physical Server Trend Physical Server Installed Base (Millions) Logical Server Installed Base (Millions) Complexity and Operating Costs are still present Capital Savings Millions Installed Servers 80 60 40 20 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 0 Source: IDC 10

Virtualization Drives Storage Connectivity... because Data Centers are always in flux Application life cycle services introduced, updated, retired Load on servers and networks constantly changing can be unpredictable Resource management challenge Minimize the need for excess capacity Reconfigure Reclaim/Reuse Adding resources is last resort Dynamic shared resource pools address these issues Enabled by Virtualization + Full Connectivity Networks Any servers potentially needs access to any storage Drives SAN attach from 20% to near 100% If you don t converge you will end up connecting everything to everything anyway but across additional parallel networks. 11

Has Convergence already happened? For some aspects of Convergence: YES NAS Allows access to file based storage across the network iscsi Allows access to block based storage across the network SANs have been bridged across metro and wide area networks for 10 years (FCIP, ifcp, & proprietary) FCoE provides an accepted protocol for FC across an Ethernet These are good but not sufficient Data center LANs have issues at scale WAN IP SAN connections do not solve the Local Data Center problem Operational Characteristics of FC based SANs desirable and entrenched SO, we will explore the new wave of convergence 12

Network Convergence Convergence occurring along 2 major themes These are happening at the same time We will discuss them separately We will discuss how they merge together Collapsing Tiers SAN A SAN B Converging Infrastructures 13

Data Center LAN (today) Aggregation Core Routers SSL VPN Firewall IPSec VPN L2/L3 Switch L2/L3 Switch L2/L3 Switch L2/L3 Switch Servers Servers, NAS, Campus/MAN/WAN across switched network Multi-Tier 100 s to many1000 s of ports multi-link redundancy 100s of meters max diameter oversubscribed East-West Latency can be a problem Ethernet carrying predominantly IP traffic Access Firewalls and security in aggregation layer have to be distribute in the data path due to efficiency forced by oversubscription 14

Evolution of the Data Center LAN Local Area Client Server applications And so we wired the data center the same way Data Center Client Server Applications Need more redundancy: STP and disabled links OR lots of VLANs 20 years ago the Ethernet switch was introduced to solve the LAN problem And it became the basic building block of the network SOA Applications etc Up to 75% of traffic Adding in the complexity of storage networking as well, these pressures force the Ethernet network to change. 15

Bubbles of Optimal Performance Typical tree configuration Location matters in a tree architecture Local Connectivity Bubble for Optimal Performance VM One Hop 16

Appliances Create Shadows Location matters in a tree architecture Appliances and VLANs Shadow of accessibility VM 17

Simplifying the Data Center LAN Aggregate Switches Multiple physical switches that operate as a single logical device Collapse Tiers Use an aggregated switch to do the work of multiple tiers of switches Allows aggregated access to services Create A Fabric Use both techniques at the same time to build a fabric based infrastructure WE WANT: 1. Operational Simplicity 2. Lower Latency 3. Lower cost 18

Data Center SAN (FC based) Servers Storage across switched network Core Edge or Edge Core Edge in effect one level of tier collapse is done 10 s to 1000 s of ports Full Dual Rail Redundancy 100s of meters max diameter High Bandwidth, Low Latency Lossless Links Fabric Services provide Discovery, Access Control, and Change Notification Gateways and specialized extension devices provide remote access for BC/DR Attached Appliances provide data services Encryption, Block Virtualization 19

Put the SAN on the LAN What happens when you overlay the SAN picture... onto the LAN picture? SAN protocols drive the single biggest piece of the convergence story 20

C N I C N I C N I l At a high level you end up with... N I N C I disk H C B A H B H A B A N I C N I N C I C N I C N I H C B A H B H A B A nas vtl disk disk SAN disk LAN tape N I H C B A H B H A B A firewal vpn I/O Consolidation Network Convergence CNA CNA Unified storage CNA CNA CNA CNA Run multiple independent virtual networks planes across a single physical infrastructure its not JUST an edge thing -- and not JUST a SAN thing disk disk disk Unified services 21

Evolution of Network Convergence SAN A SAN B Separate Networks No overlap of operations & management separate network characteristics separate teams run them (typically) SAN SAN A B I/O Consolidation overlap confined to server and 1 st hop team consensus needed several detailed ways to accomplish transit switch L2 to access SAN gateway SAN-LAN Convergence unified management and operational model run by single team common redundancy shared bandwidth and connectivity 22

Evolution of Server Connections SAN LAN SAN LAN converged Server Rack End of Row Connections I/O Consolidation server connections become more efficient Server Rack Network Convergence server connection does not change Server Rack server guy has to deal with most of the implications of convergence during the I/O consolidation phase 23

HOW MANY NIC S IN A SERVER NIC NIC NIC NIC NIC HBA HBA HBA NIC NIC NIC NIC NIC HBA HBA HBA NIC HBA HBA HBA disk nas vtl tape firewall vpn Multiple network connections Separate subnets Separate VLANs May be separate networks May be separate switches in the same network Each configured with different QoS settings Depends on the Server file servers and/or NAS points to campus Presentation servers and intranet point to campus Application servers Database servers Backup Servers, etc etc For Example Something like this is possible: Campus Access (1+) LAN Backup NIC (1) Application Cluster NIC (1) Vmotion NIC (1) Presentation to application to database private NICs (1+) Storage Access Cards (2) I/O Consolidation reducing this to 2 NICs can save a bundle here 24

I/O Consolidation Deployments server direct to gateway server direct to FCF server to L2 to gateway server L2 to FCF 25

Making Convergence Happen Convergence needs the following Transport Convergence DCB for Ethernet is complete. Deployments happening. Switch aggregation is happening Tier Collapse is happening Protocol Evolution This has happened with iscsi refined by DCB This is happened with FCoE (FC-BB-5) refined by FC-BB-6 Operational This has started with I/O Consolidation out of server Better Connected L2 networks Collections of physical switches acting as single logical switch 26

IF we use FCoE for Network Convergence Scaling FCF at the TOR hits Domain Scaling problems FC-BB-6 intends to address this with FCF-FDF distributed domains Configuration of L2/L3 separation VLAN Virtual Fabric mapping L2 network as access between server and FCF Full FC topology overlay onto data center network Multi-hop FCoE across multiple L2 networks and several FCFs Multiple VLAN/Virtual Fabric Configuration and Management 27

IF we use iscsi for Network Convergence Services services means discovery, access control, notifications, security isns exists for these but not commonly deployed Management Tools have limited support for iscsi attach Performance to match local data center SAN requirements need DCB this is true even though TCP/IP is the transport 28

Protocol Independent Convergence Considerations End to End performance bandwidth & latency & frame processing rate congestion handling logical overlays onto physical infrastructure MAN and WAN extension for storage Under convergence the local network is Ethernet. The devices that make up the local Ethernet are not typically suited to directly support lossless distance extension. 29

Reminder: The Protocols CEE is a set of enhancements to Ethernet to allow it to better handle all the different classes of traffic in the data center Data Center Convergence CEE INCITS T11 develops the various Fibre Channel standards including the latest developments for FCoE FC Fibre Channel (T11) FCoE IEEE 802.1 IETF MANY FC-BB-5 Inter Networking Security Audio/Video Bridging DCB iscsi FC-BB-6 Shortest Path Bridging PFC TRILL ETS DCBX isns I/O Consolidation & Network Convergence TRILL (like IEEE SPB) is a mechanism bringing L3 style multi-pathing to L2 Ethernet so avoiding spanning tree issues Data Center Bridging (DCB) Task Group (TG) is the IEEE group working to provide enhancements to existing 802.1 bridge specifications to satisfy the requirements of protocols and applications in the data center. QCN EVB BPE Network Wide Congestion Management Physical/Virtual Server/Switch Interaction 30

Full Network Convergence Benefits results in logical overlays for forwarding on single, shared HW infrastructure Benefits stocking of spare FRUs combined operations fewer stranded resources better utilization lower latency better flexibility Visit the Hands-On Lab: Unified Storage Infrastructure, Unified Storage IP Solutions 31

Inhibitors and Accelerators If convergence is so great why hasn t it already happened? Actually it is in progress, but there is a balance between various factors influencing the adoption rate Existing Operational Models and Infrastructure INHIBITORS Cost of 10G/40G Infrastructure Politics / Team Issues Technology Maturity Data Center Network Convergence Development of best practices for phased deployment of convergence 10GE Performance Cost Reductions Convergence Solves Actual problems of scale, flexibility, and complexity ACCELERATORS Server Virtualization Network Evolution New Data Center Models 32

Q&A / Feedback Please send any questions or comments on this presentation to SNIA: tracknetworking@snia.org Many thanks to the following individuals for their contributions to this tutorial. - SNIA Education Committee Joseph L White Simon Gordon Gunes Aybay Charles Waters Andy Ingram 33