WHITE PAPER. Enabling 100 Gigabit Ethernet Implementing PCS Lanes



Similar documents
40 Gigabit Ethernet and 100 Gigabit Ethernet Technology Overview

WHITE PAPER. Static Load Balancers Implemented with Filters

WHITE PAPER. Extending Network Monitoring Tool Performance

WHITE PAPER. Addressing Monitoring, Access, and Control Challenges in a Virtualized Environment

White Paper. An Overview of Next-Generation 100 and 40 Gigabit Ethernet Technologies

40 Gigabit Ethernet and 100 Gigabit Ethernet Technology Overview

IxChariot Virtualization Performance Test Plan

802.3bj FEC Overview and Status. 400GbE PCS Options DRAFT. IEEE P802.3bs 400 Gb/s Ethernet Task Force. November 2014 San Antonio

WHITE PAPER. Gaining Total Visibility for Lawful Interception

EBOOK. The Network Comes of Age: Access and Monitoring at the Application Level

WHITE PAPER. Net Optics Phantom Virtual Tap Delivers Best-Practice Network Monitoring For Virtualized Server Environs

Guidebook to MEF Certification

WHITE PAPER. 50 versus 62.5 micron multimode fiber

Evaluating Wireless Broadband Gateways for Deployment by Service Provider Customers

10 Gigabit Ethernet: Scaling across LAN, MAN, WAN

WHITE PAPER. Monitoring Load Balancing in the 10G Arena: Strategies and Requirements for Solving Performance Challenges

WHITE PAPER. Tap Technology Enables Healthcare s Digital Future

WHITE PAPER. Network Traffic Port Aggregation: Improved Visibility, Security, and Efficiency

EBOOK. Software Defined Networking (SDN)

High Speed Ethernet. Dr. Sanjay P. Ahuja, Ph.D. Professor School of Computing, UNF

R2. The word protocol is often used to describe diplomatic relations. How does Wikipedia describe diplomatic protocol?

How To Test For 10 Gigabit Ethernet At 10 Gb/S

WHITE PAPER. SDN Controller Testing: Part 1

High-Speed SERDES Interfaces In High Value FPGAs

White Paper. Best Practices for 40 Gigabit Implementation in the Enterprise

What to do about Fiber Skew?

40 and 100 Gigabit Ethernet Overview

WHITE PAPER. How To Compare Virtual Devices (NFV) vs Hardware Devices: Testing VNF Performance

Introduction. Background

Optimizing Infrastructure Support For Storage Area Networks

Reduce Your Network's Attack Surface

Application Note Gigabit Ethernet Port Modes

The Key Benefits of OTN Networks

FURTHER READING: As a preview for further reading, the following reference has been provided from the pages of the book below:

25GbE: The Future of Ethernet

Latency on a Switched Ethernet Network

RFC 2544 Testing of Ethernet Services in Telecom Networks

100BASE-SX Fast Ethernet:

Using FPGAs to Design Gigabit Serial Backplanes. April 17, 2002

HOW TO TEST 10 GIGABIT ETHERNET PERFORMANCE

Understanding Ethernet and Fibre Channel Standard-Based Test Patterns An explanation of IEEE and NCITS standard test patterns By Todd Rapposelli

How To Make A Multilink Ame Relay Work Faster And More Efficiently

Migration to 40/100G in the Data Center with OM3 and OM4 Optical Connectivity

Interfaces and Payload Testing

Data Center Design for 40/100G

Lecture 1. Lecture Overview. Intro to Networking. Intro to Networking. Motivation behind Networking. Computer / Data Networks

The OSI Model: Understanding the Seven Layers of Computer Networks

10 Gigabit Ethernet MAC Core for Altera CPLDs. 1 Introduction. Product Brief Version February 2002

New WAN PHY Approach Proposals

Clearing the Way for VoIP

Communications and Computer Networks

WHITE PAPER. Best Practices for Deploying IPv6 over Broadband Access

FOUNDATION Fieldbus High Speed Ethernet Control System

Chap 4 Circuit-Switching Networks

INTRODUCTION TO 100BASE-T: FAST (AND FASTER) ETHERNET

SummitStack in the Data Center

Local Area Networks. Guest Instructor Elaine Wong. Elaine_06_I-1

Ethernet/IEEE evolution

Performance Management and Fault Management. 1 Dept. of ECE, SRM University

Chapter 2 - The TCP/IP and OSI Networking Models

Ensuring Success in a Virtual World: Demystifying SDN and NFV Migrations

SummitStack in the Data Center

Taps vs. SPAN The Forest AND the Trees: Full Visibility into Today's Networks

How To Understand The Layered Architecture Of A Network

Steve Worrall Systems Engineer.

WHITE PAPER. Best Practices in Deploying Converged Data Centers

Master Course Computer Networks IN2097

Computer Network. Interconnected collection of autonomous computers that are able to exchange information

AN OVERVIEW OF QUALITY OF SERVICE COMPUTER NETWORK

An Executive Brief for Network Security Investments

Overview of Requirements and Applications for 40 Gigabit and 100 Gigabit Ethernet

Ethernet Passive Optical Networks EPON

LAN and WAN Rate 10 GigE

Fibre Channel over Ethernet in the Data Center: An Introduction

Monitoring Service Delivery in an MPLS Environment

Virtual Private LAN Service (VPLS) Conformance and Performance Testing Sample Test Plans

WHITE PAPER. Best Practices for Eliminating Duplicate Packets

Unified Storage Networking

Auspex Support for Cisco Fast EtherChannel TM

Local Area Networks transmission system private speedy and secure kilometres shared transmission medium hardware & software

Link Layer. 5.6 Hubs and switches 5.7 PPP 5.8 Link Virtualization: ATM and MPLS

11/22/

ADTRAN NetVanta 5660

10 Gigabit Ethernet (10GbE) and 10Base-T - RoadMap

Data Link Protocols. TCP/IP Suite and OSI Reference Model

Using High Availability Technologies Lesson 12

WHITE PAPER. Packet Time Monitoring in Your Visibility Architecture

Data Center Automation - A Must For All Service Providers

WHITE PAPER. Application Performance Management and Lawful Interception

100 Gigabit Ethernet is Here!

Protocol Architecture. ATM architecture

Designing Reliable IP/MPLS Core Transport Networks

Objectives. Lecture 4. How do computers communicate? How do computers communicate? Local asynchronous communication. How do computers communicate?

Cisco Nexus Planning and Design Service

Next-Generation IP Networking

Oracle Big Data Appliance: Datacenter Network Integration

Gigabit Ethernet. Abstract. 1. Introduction. 2. Benefits of Gigabit Ethernet

Introduction to MPIO, MCS, Trunking, and LACP

Transcription:

WHITE PAPER Enabling 100 Gigabit Ethernet Implementing PCS Lanes www.ixiacom.com 915-0909-01 Rev. C, January 2014

2

Table of Contents Introduction... 4 The IEEE 802.3 Protocol Stack... 4 PCS Layer Functions... 5 Striping... 6 The Need for PCS Lanes... 7 Multiplexing of PCS Lanes... 8 Testing Challenges... 9 Receiver Testing... 9 PCS lane alignment marker lock... 9 Skew tolerance and compensation...10 Arbitrary mapping of transmitted to received PCS lanes...10 Transmit-Side Testing... 11 Ixia and PCS... 11 Summary... 11 3

Introduction Advances in data center technology and Internet usage have pushed enterprise data centers from Gigabit Ethernet links to 10 Gigabit Ethernet (GbE) links and search engines/ carriers/isps from single 10 GbE links to multiple links. Projected growth indicates the need for higher-speed Internet connections1. The IEEE 802.3ba task force has been established with the objective of standardizing 100 GbE and 40 GbE over the Internet and within the data center. Advances in data center technology and Internet usage have pushed enterprise data centers from Gigabit Ethernet links to 10 Gigabit Ethernet (GbE) links and search engines/ carriers/isps from single 10 GbE links to multiple links. If higher-speed Ethernet is to be useful in the near term, implementations must take advantage of existing copper and fiber cables, both in the data center and over the Internet. This poses an interesting problem since no technology currently exists to transport 100 Gbps or 40 Gbps as a single stream over either media. In order to transport 100 GbE over single-mode fibers, for example, 4 different wavelengths will be required using LAN wavelength-division multiplexing ( LAN WDM). Similarly, the device interfaces found in routers, switches and servers that drive these higher-speed links cannot currently handle single 100 Gbps or 40 Gbps data streams. They will be forced to resort to parallel electrical lanes to handle the flow of data; for example, 10 lanes of 10 Gbps. As technology improves, the bandwidth of fiber wavelengths and electrical lanes will increase at independent rates. Efficient 100 and 40 GbE links will need to handle a number of combinations. A technique is needed to allow efficient implementation that allows for changing numbers and bandwidths of wavelengths and electrical lanes. Within the 802.3 Ethernet specification, the protocol stack layer that performs this function is called the physical coding sublayer, or PCS. This white paper describes the baseline proposal for the physical coding sublayer (PCS) for the 100-Gbps and 40-Gbps Ethernet interfaces currently under standardization within the IEEE 802.3ba task force. The IEEE 802.3 Protocol Stack Figure 1 shows a simplified version of the standard IEEE 802.3 protocol stack The subject of this paper relates the PCS layer and the unique requirements for 100 Gbps Ethernet. Media Access Control (MAC) Reconciliation Sublayer Media Independent Interface (MII) Physical Coding Sublayer (PCS) Physical Medium Attachment (PMA) Physical Medium Dependent (PMD) Simplified IEEE 802.3 protocol stack 1) Schneider, David. An Overview of Next-Generation 100 and 40 Gigabit Ethernet Technologies. Available at http://www.ixiacom.com/library/white_papers. 4

Figure 2 shows the architectural partition for a 100 Gbps Ethernet link. The partitioning assumes: A highly integrated 100 GbE MAC/PCS chip that includes the packet interface and MAC, PCS, and PMA functions. An optical module that includes both PMA and PMD functions. A high-speed parallel electrical interface comprising n lanes between the MAC/PCS chip and the optical module. A parallel optical interface comprising m lanes, where lanes can be wavelengths or fibers. 100 GbE ASIC 100 GbE Optical Module 100 GbE Optical Module 100 GbE ASIC Packet Interface Media Access Control (MAC) Physical Coding Sublayer (PCS) Physical Medium Attachment 1 (PMA 1 ) n Physical Medium Attachment 2 (PMA 2 ) Physical Medium Dependant (PMD) m Physical Medium Dependant (PMD) Physical Medium Attachment 2 (PMA 2 ) n Physical Medium Attachment 1 (PMA 1 ) Physical Coding Sublayer (PCS) Media Access Control (MAC) Packet Interface PCS Layer Functions The requirements for the PCS layer include: Architectural partition for a 100 GbE link Provide frame delineation Control signal transport Provide the clock transitions needed by SerDes-style electrical and optical interface Bond multiple lanes together through a striping or fragmentation methodology Since the data must travel on multiple optical lanes as well as on multiple electrical lanes to the optical module, the striping mechanism should support: Low frame overhead that is independent of frame size Formatting that enables receiver-only lane deskew Evolving media and media interface widths Simple optical module implementations 5

Striping Where the number of electrical and optical lanes is the same, a simple striping of 66-bit blocks across the lanes is sufficient, as shown in Figure 3 for four electrical lanes. Incoming Stream from MII Transmit PCS 9 10 11 12 5 6 7 8 1 2 3 4 66-bit block 9 10 11 12 5 6 7 8 1 2 3 4 A A A A Outgoing Electrical Lanes Simple striping of data The transmitting PCS first performs 64/66b encoding on the incoming stream. The 66- bit blocks are distributed across the lanes on a round-robin basis. An alignment block is periodically added as a marker that allows the receive side to deskew the lanes. This allows the receiver to remove skew between the individual media lanes. Bandwidth for the alignment blocks is created by periodically deleting inter-packet gaps. Line coding is implemented once at the aggregate level in the transmitting PCS block using a single scrambler in such a way as to ensure adequate transition density on any of the individual electrical and media lanes along the path. The deskew function is also only implemented once in the receiving PCS block, and again in such a way as to compensate for skew introduced over any of the electrical or optical interfaces along the path. The advantages of this approach are obvious; it moves all of the scrambling and deskew features into a CMOS chip, where they are easy and cheap to implement, and therefore significantly simplifies the design of the optical module. 6

The Need for PCS Lanes As discussed earlier, advancing electrical and optical technologies will require the ability to handle differing and changing numbers of electrical interface lanes versus optical lanes. To handle the general case, the PCS baseline proposal calls for data to be distributed on a per-66-bit block basis to a number of PCS lanes v, where the number of PCS lanes equals the least common multiple of the number of electrical lanes n and optical lanes m. Using the virtual lane concept, the optical module can be a very simple multiplexer, which merely bit multiplexes the data from n electrical lanes down to m media lanes. The receiver s PCS block can simply demultiplex the data back into the PCS lanes and then realign the skewed data. A virtual lane is a continuous stream of 64/66b blocks. PCS lanes are created through a simple round-robin function which distributes 66-bit blocks, in order, to each virtual lane. In the case of an interface running with 20 PCS lanes, a single virtual lane would contain every 20th 66-bit block from the aggregate signal, as illustrated in Figure 4 below. In order to allow the receiver to identify and deskew the individual PCS lanes, a unique alignment block is added to each virtual lane on a periodic basis. The alignment block is a special 66-bit control signal block that is unique to each virtual lane and cannot be duplicated in the data. The current proposal under consideration calls for an alignment block once every 16,384 blocks on each virtual lane. The number of PCS lanes required depends on the electrical and optical lane combinations that need to be supported. The number of PCS lanes is the least common multiple of the number of electrical lanes and the number of optical lanes, regardless of whether the lanes are based on numbers of wavelengths, numbers of fibers, or other considerations. This constraint ensures that the total number of PCS lanes can be mapped evenly over both the number of electrical lanes and the number of optical lanes; therefore, the data from any particular virtual lane will always reside on the same electrical and media lane across the link. This guarantees that no skew can be introduced between the bits within a virtual lane, which would be impossible to remove at the receiver. As discussed earlier, advancing electrical and optical technologies will require the ability to handle differing and changing numbers of electrical interface lanes versus optical lanes. #2n+1 #n+1 #1 virtual lane 1 #2n+2 #n+2 #2 virtual lane 2 100 Gbps aggregate stream of 64/66b words #2n+3 #n+3 #3 virtual lane 3 #2n+1 #2n #n+3 #n+2 #n+1 #n #3 #2 #1 Simple 66-bit word level round robin distribution #3n #2n #n virtual lane n virtual lane markers Virtual lane data distribution 7

Multiplexing of PCS Lanes Virtual lane muxing is straightforward, consisting of bit-interleaving of all of the PCS lanes assigned to a particular electrical or optical lane and is performed in the PMA layer. Figure 6 illustrates a case with: 20 PCS lanes 10 electrical lanes 4 media lanes In Figure 5, VL stands for virtual lane and the nomenclature n.m indicates virtual lane n, bit m. The PCS layer created the 20 PCS lanes from the data source flow, as covered above. The PCS layer bit-multiplexes each pair of PCS lanes into a single lane and sends 10 electrical lanes to the PMA in the optical module. The optical device s PMA then bitmultiplexes the 10 electrical lanes into the 4 optical lanes. Virtual lane muxing is straightforward, consisting of bitinterleaving of all of the PCS lanes assigned to a particular electrical or optical lane and is performed in the PMA layer. Due to skew on the electrical interface, the random startup state of the bit multiplexers and demultiplexers is in not predictable where PCS lanes will end up in the receive PCS. The unique alignment words, however, provide the means by which the receive PCS layer identifies and reorders the individual PCS lanes. VL0 VL1 VL2 Transmit PCS VL19 0.3 1.2 0.2 1.2 0.1 1.1 0.0 1.0 2.0 3.0 4.0 5.0 6.0 7.0 8.0 9.0 10.0 11.0 12.0 13.0 14.0 15.0 16.0 17.0 18.0 19.0 1.1 0.1 1.0 0.0 3.0 2.0 5.0 4.0 7.0 6.0 9.0 8.0 11.0 10.0 Transmit PMA 13.0 12.0 15.0 14.0 17.0 16.0 19.0 18.0 Transmit PMD 0.1 2.1 4.1 6.1 13.0 15.0 17.0 19.0 5.0 7.0 9.0 11.0 16.0 18.0 1.0 3.0 8.0 10.0 12.0 14.0 0.0 2.0 4.0 6.0 1 Receive PMD Receive PMA 1.1 0.1 1.0 0.0 3.0 2.0 5.0 4.0 7.0 6.0 9.0 8.0 11.0 10.0 13.0 12.0 15.0 14.0 17.0 16.0 19.0 18.0 0.3 1.2 0.2 1.2 0.1 1.1 0.0 1.0 2.0 3.0 4.0 5.0 6.0 7.0 8.0 9.0 10.0 11.0 12.0 13.0 14.0 15.0 16.0 17.0 18.0 19.0 Receive PCS Multiplexing of PCS lanes to optical media 8

Testing Challenges As with any new technology, 100 GbE testing will be a significant challenge. Although this paper largely concentrates on PCS layer testing issues, it s important to note that layer 2 through 7 testing is as important as ever. The 100 GbE and 40 GbE links will be incorporated in switches, routers, and servers almost immediately. It s important that real-world traffic be used to properly exercise the network device/server, to determine if it can properly sustain traffic over multiple high speed links. The 100 GbE interfaces must operate in an integrated manner with lower-speed interfaces. As with every new standard, multiple companies will develop products that will conform to pre-standards and then to published standards. Proper test tools are needed to allow each vendor to independently assess their conformance to standards that will lead to high assurance of interoperability. The alternative approaches are back-to-back with one s own equipment or limited interoperability testing with other vendor s equipment. The former is self-deceptive, allowing a vendor to make the same mistakes in both transmit and receive functions. The latter is expensive and endlessly frustrating, due to multiple vendors updates. Receiver Testing Concentrating on just the receiver side of the PCS layer, major test considerations include: 64/66b word sync lock Virtual lane alignment marker lock Skew tolerance and compensation Arbitrary mapping of received to transmitted PCS lanes 64/66b Word Sync Lock Each received virtual lane must properly identify 64/66b blocks, based on the control bits. The receive logic must properly lock and come out of lock under conditions covered in the specification. In order to measure conformance, various bit sequences at the limits of the specification must be transmitted by the test equipment. The receiver of the interface under test must then be subjected to intentionally errored sync bits and out-of-sync detection must be verified. Additionally, a variety of intentionally errored patterns must be applied to also ensure that the receiver does not improperly sync. It s important that real-world traffic be used to properly exercise the network device/ server, to determine if it can properly sustain traffic over multiple high speed links. PCS lane alignment marker lock Each virtual lane must correctly lock onto alignment markers. The techniques used here are analogous to those used in 64/66b word sync lock testing. Boundary conditions need to be verified in addition to basic locking operations. For example, if the specification calls for virtual lane lock after n consecutive non-errored alignment markers, then testing must ensure that lock is not declared with any fewer alignment markers. Conversely, if virtual lane lock should be dropped with m consecutive errored alignment markers, then testing must ensure that lock is maintained with fewer errored alignment markers. 9

Skew tolerance and compensation Skew at various electrical and optical interfaces causes PCS lanes to arrive with different delays relative to other PCS lanes. The virtual lane alignment markers are designed to make it possible for received PCS lanes to identify and realign themselves properly. In order to test skew tolerance and compensation algorithms, test equipment must introduce known, specific amounts of skew amongst the PCS lanes. The receiver s PCS layer should be able compensate for the skew mandated in the standard. Test equipment must also be used to measure the actual skew tolerance when it exceeds the specifications. Arbitrary mapping of transmitted to received PCS lanes As shown in Figure 6, a particular virtual lane may be transmitted on any of a number of optical lanes and then received in any virtual lane position. The red lines trace a set of the possible paths for virtual lane 0. The virtual lane concept, in fact, does not require that the number of physical lanes on the transmit and receive paths match. To make matters worse, there is no requirement to enforce where specific PCS lanes are mapped along the transmit or receive path. The only requirement is for all transmitted bits on a given virtual lane to end up together on the same physical receive lane. When there is an incongruity, which will occur as the various vendors transmit and receive ASICs and optical modules proliferate, transmitted PCS lanes can appear on any received virtual lane. Receivers must be able to reorder and reassemble any mapping of PCS lanes into a single 100 Gbps aggregated stream. In order to exercise this reordering, test equipment must be able to transmit PCS lanes in any order to emulate differing scenarios. Transmit PCS 0 10 1 11 2 12 3 13 4 14 5 15 6 16 7 17 8 18 9 19 Electrical Lanes PMA PMD Optical Lanes PMD PMA Electrical Lanes 1:2 1:2 1:2 1:2 1:2 1:2 1:2 1:2 1:2 1:2 Virtual Lanes Lane ordering requires lane identification Receive PCS 10

Transmit-Side Testing Transmit-side testing consists of verification and monitoring or various signal attributes: 64/66b sync bits. Proper transmission of sync bits must be determined and then monitored. Counts of erroneous sync bits must be available in real time during testing. Virtual lane alignment markers. Proper transmission of alignment markers must be verified and monitored, with error counts available. Lane skew. Actual lane skew must be measured and reported. Virtual lane mapping. The actual mapping of transmitted to received PCS lanes must be identified and reported. Ixia and PCS Ixia has been actively working with the IEEE 802.3ba task force. On June 17th, 2008 Ixia demonstrated the first PCS hardware implementation at NXTcomm08. The demonstration used Ixia s proof of concept 100 GbE implementation to generate and receive line-rate traffic, which was sent round-trip from Las Vegas to Los Angeles, and to analyze the results. At Interop 2009, Ixia showed the world s first 100G IP test module with a CFP interface. The 100 GbE interface card works in conjunction and in the same manner as all other Ixia interface cards and test applications, making it possible to test all types of devices and networks from layer 2 through 7. Summary 100 GbE and 40 GbE technologies are rapidly approaching standardization and deployment. 100 GbE and 40 GbE technologies are rapidly approaching standardization and deployment. A key factor in their success will be ability to utilize existing fiber and copper media in an environment of advancing technologies. The physical coding sublayer (PCS) of the 802.3 architecture is in a perfect position to facilitate this flexibility. The current baseline proposal for PCS implementation uses a unique virtual lane concept that provides the mechanism to handle differing electrical and optical paths. Testing of this technology, however, will require refined and sophisticated test methodologies and equipment. 11

WHITE PAPER Ixia Worldwide Headquarters 26601 Agoura Rd. Calabasas, CA 91302 (Toll Free North America) 1.877.367.4942 (Outside North America) +1.818.871.1800 (Fax) 818.871.1805 www.ixiacom.com Ixia European Headquarters Ixia Technologies Europe Ltd Clarion House, Norreys Drive Maidenhead SL6 4FL United Kingdom Sales +44 1628 408750 (Fax) +44 1628 639916 Ixia Asia Pacific Headquarters 21 Serangoon North Avenue 5 #04-01 Singapore 554864 Sales +65.6332.0125 Fax +65.6332.0127 915-0909-01 Rev. C, January 2014