PCI Express Overview. And, by the way, they need to do it in less time.



Similar documents
PCI Express* Ethernet Networking

Maximizing Server Storage Performance with PCI Express and Serial Attached SCSI. Article for InfoStor November 2003 Paul Griffith Adaptec, Inc.

PCI Express Basics Ravi Budruk Senior Staff Engineer and Partner MindShare, Inc.

3 Address Spaces & Transaction Routing. The Previous Chapter. This Chapter. The Next Chapter

PCI EXPRESS TECHNOLOGY. Jim Brewer, Dell Business and Technology Development Joe Sekel, Dell Server Architecture and Technology

PCI Express: Interconnect of the future

Creating a PCI Express Interconnect AJAY V. BHATT, TECHNOLOGY AND RESEARCH LABS, INTEL CORPORATION

Appendix A. by Gordon Getty, Agilent Technologies

The proliferation of the raw processing

MSC8156 and MSC8157 PCI Express Performance

Introduction to PCI Express Positioning Information

How PCI Express Works (by Tracy V. Wilson)

PEX 8748, PCI Express Gen 3 Switch, 48 Lanes, 12 Ports

Hardware Level IO Benchmarking of PCI Express*

High Speed I/O Server Computing with InfiniBand

Fibre Channel over Ethernet in the Data Center: An Introduction

PCI Express and Storage. Ron Emerick, Sun Microsystems

The Shortcut Guide to Balancing Storage Costs and Performance with Hybrid Storage

PCI Express Impact on Storage Architectures and Future Data Centers. Ron Emerick, Oracle Corporation

Chapter 13 Selected Storage Systems and Interface

The Advantages of Multi-Port Network Adapters in an SWsoft Virtual Environment

EDUCATION. PCI Express, InfiniBand and Storage Ron Emerick, Sun Microsystems Paul Millard, Xyratex Corporation

2. THE PCI EXPRESS BUS

CCNA R&S: Introduction to Networks. Chapter 5: Ethernet

Intel PCI and PCI Express*

Universal Flash Storage: Mobilize Your Data

Computer Systems Structure Input/Output

Intel Ethernet Switch Load Balancing System Design Using Advanced Features in Intel Ethernet Switch Family

An Introduction to PCI Express

White Paper Abstract Disclaimer

PCI Express Impact on Storage Architectures and Future Data Centers. Ron Emerick, Oracle Corporation

The Bus (PCI and PCI-Express)

Accelerating High-Speed Networking with Intel I/O Acceleration Technology

How to build a high speed PCI Express bus expansion system using the Max Express product family 1

Enterprise SSD Interface Comparisons

Computer Network. Interconnected collection of autonomous computers that are able to exchange information

UMBC. ISA is the oldest of all these and today s computers still have a ISA bus interface. in form of an ISA slot (connection) on the main board.

SAN Conceptual and Design Basics

Building High-Performance iscsi SAN Configurations. An Alacritech and McDATA Technical Note

Overview of Routing between Virtual LANs

PCI Express Basic Info *This info also applies to Laptops

Serial ATA and Serial Attached SCSI technologies

IP SAN Best Practices

IP Networking. Overview. Networks Impact Daily Life. IP Networking - Part 1. How Networks Impact Daily Life. How Networks Impact Daily Life

Technology Note. PCI Express

How To Connect Virtual Fibre Channel To A Virtual Box On A Hyperv Virtual Machine

I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology

The changes in each specification and how they compare is shown in the table below. Following the table is a discussion of each of these changes.

PCI Express IO Virtualization Overview

Leased Line + Remote Dial-in connectivity

Computer Organization & Architecture Lecture #19

HBA Virtualization Technologies for Windows OS Environments

PCI Express Impact on Storage Architectures and Future Data Centers

PCI Express Gen 2 Deep Dive on Power Architecture Based Products

ESSENTIALS. Understanding Ethernet Switches and Routers. April 2011 VOLUME 3 ISSUE 1 A TECHNICAL SUPPLEMENT TO CONTROL NETWORK

OPTIMIZING SERVER VIRTUALIZATION

LAN Switching and VLANs

Architecting High-Speed Data Streaming Systems. Sujit Basu

SATA II 4 Port PCI RAID Card RC217 User Manual

Read this before starting!

Intel Data Direct I/O Technology (Intel DDIO): A Primer >

Unified Fabric: Cisco's Innovation for Data Center Networks

Serial ATA technology

Local Area Networks transmission system private speedy and secure kilometres shared transmission medium hardware & software

Broadcom Ethernet Network Controller Enhanced Virtualization Functionality

Storage Area Network Design Overview Using Brocade DCX Backbone Switches

Communication Networks. MAP-TELE 2011/12 José Ruela

Virtualizing the SAN with Software Defined Storage Networks

Interconnection Networks. Interconnection Networks. Interconnection networks are used everywhere!

Solving the Hypervisor Network I/O Bottleneck Solarflare Virtualization Acceleration

Introduction to Infiniband. Hussein N. Harake, Performance U! Winter School

PCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters. from One Stop Systems (OSS)

Achieving a High-Performance Virtual Network Infrastructure with PLUMgrid IO Visor & Mellanox ConnectX -3 Pro

Brocade Solution for EMC VSPEX Server Virtualization

Cloud-Based Apps Drive the Need for Frequency-Flexible Clock Generators in Converged Data Center Networks

Understanding PCI Bus, PCI-Express and In finiband Architecture

Enterprise Disk Storage Subsystem Directions

Virtualized Security: The Next Generation of Consolidation

SCSI vs. Fibre Channel White Paper

High-Performance IP Service Node with Layer 4 to 7 Packet Processing Features

PCI Express Impact on Storage Architectures. Ron Emerick, Sun Microsystems

Using VMWare VAAI for storage integration with Infortrend EonStor DS G7i

- Hubs vs. Switches vs. Routers -

How Router Technology Shapes Inter-Cloud Computing Service Architecture for The Future Internet

Evaluation Report: HP Blade Server and HP MSA 16GFC Storage Evaluation

A Dell Technical White Paper Dell Storage Engineering

From Bus and Crossbar to Network-On-Chip. Arteris S.A.

Gigabit Ethernet. Abstract. 1. Introduction. 2. Benefits of Gigabit Ethernet

Optimizing Data Center Networks for Cloud Computing

Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck

Frequently Asked Questions

Deploying F5 BIG-IP Virtual Editions in a Hyper-Converged Infrastructure

Local-Area Network -LAN

How To Increase Network Performance With Segmentation

Introduction to PCI Express

Using FPGAs to Design Gigabit Serial Backplanes. April 17, 2002

M.Sc. IT Semester III VIRTUALIZATION QUESTION BANK Unit 1 1. What is virtualization? Explain the five stage virtualization process. 2.

3G Converged-NICs A Platform for Server I/O to Converged Networks

Transcription:

PCI Express Overview Introduction This paper is intended to introduce design engineers, system architects and business managers to the PCI Express protocol and how this interconnect technology fits into today s environment. This technology brief will first introduce the problems that exist with interconnects today, and then discuss how the PCI Express protocol addresses these issues. We then will present an overview on the PCI Express architecture, and give an example of how PCI Express would operate. We will then wrap up with how this protocol addresses various market needs, as well as how to rationalize when to use PCI Express and Advanced Switching. The Problem Today Today s system designers are facing a challenging dilemma. They are expected to produce a higher performance solution with more features at the same cost as their current solutions. And, by the way, they need to do it in less time. Traditionally, engineers could satisfy their bandwidth requirements by migrating to the next generation PCI. This was possible because the PCI governing body, the PCI SIG, actively managed the roll out of higher performance PCI revisions to meet market demand. Further, because the next generation PCI was code compatible with legacy PCI, engineers could reuse their PCI code to reduce their design time. This was an effective strategy for several generations. Recently, however, designers have been running into a bottleneck. In order to achieve greater throughput, the PCI community defined wider busses. Including the data bus and sideband control signal, 64-bit PCI exceeds 100 pins. Even at 64-bits and 133 MHz, architectural problems remained because the higher clock rates came at the cost of reduced fanout and the bandwidth still needed to be shared over multiple devices running at ever higher data rates. As a result many, vendors began migrating to either proprietary or marginally supported protocol standards in search of a switched interconnect. While this addressed the architectural problems, it forced engineers to create new software, which slowed development. Further, because of the smaller market adoption, aggressive cost reductions were not possible. PCI Express Architecture To address these needs, the PCI SIG adopted a specification called 3GIO from the Arapahoe group, and renamed it PCI Express. The PCI Express protocol was created to address the following design goals: Support multiple market segments Boot existing OS without change Scalable performance Advanced features including QoS, Power management, and data integrity PCI Express technology is a low cost, highly scalable, switched, point-to-point, serial I/O interconnect that maintains complete software compatibility with PCI. It transfers data at 2.5 Gb/s 1 per lane, per direction, and will scale 1 Like all protocols, those using PCI Express will need to allocate some bandwidth to overhead, including management structures such as DLP (discussed below), as well as packet specific structure such as CRCs, 8b10b encoding, and header processing. For a full dissertation on this topic, please go to www.plxtech.com, and download our whitepaper on the topic.

proportionately for performance by adding lanes to the link. Explaining the Layers To achieve code compatibility with PCI, PCI Express does not modify the transaction layer. This is significant, because it allows vendors to leverage their existing PCI code to achieve not only a faster time to market, but by using their proven design, provides for a more stable and mature platform. PCI Express does, however, modify layers 1 and 2 of the OSI model. PCI Express Layers Layer 1: Physical Layer Layer 2: Data Link Layer Payload Data Layer 3: Transaction Layer To increase noise resistance, PCI Express utilizes differential current mode signaling, CML. The clock is embedded using familiar 8b10b encoding. Early on, the PCI Express architects realized that performance increases to the individual lanes would also be necessary. They therefore created the architecture so that the PHY layer could be easily swapped out for higher speeds without seriously affecting the upper layers. Layer 2 Layer 2, or the Data Link Layer, defines the data control for PCI Express. The primary task of the link layer is to provide link management and data integrity, including error detection and correction. The function of this layer is to calculate and append a Cyclic Redundancy Check (CRC) and sequence number to the information sent from the data packet. The sequence number allows proper ordering of the data packets. The CRC verifies that data from link to link has been correctly transmitted 2. Layer 1 Layer 1, or the Physical Layer, defines the electrical characteristics of PCI Express. The basic transmission unit consists of two pairs of wires, called a lane. Each pair allows for unidirectional data transmission 2.5 Gbps, so the two pairs combined provides 2.5 Gbps full duplex communication, without the risk of transmission collision. Lanes can be concatenated to provide scalable performance; so for example, combining two lanes will yield 5 Gbps throughput. The PCI Express specification allows for x1, x2, x4, x8, x12, x16, and x32 lanes, which means that the current PCI Express scales up to 80 Gbps per port in each direction. Seq 2B Layer 2 Data Sequence Numbering assures proper ordering, data receipt CRC 4B CRC assures error free link to link transmission Layer 3 Layer 3, or the transaction layer, connects the lower protocols to the upper layers. The primary value to end users is that this layer appears to the upper layers to be PCI. 2 For those customers who wish to ensure end to end data integrity, there is an optional feature in the specification to achieve this. This feature is called TLP Digest and is defined in Layer 3.

The Transaction layer packetizes and then prepends a header to the payload data. Read and write commands, as well as prior sideband signals, like interrupts and power management requests, are also included in this layer. PCI Express Packets in Action-DLLP and TLP One of the key advantages is that unlike PCI, PCI Express does not utilize sideband signaling; rather, all information is transmitted in-band which reduces pin count. This includes the clock, system status, and power management. data when the receiving device has enough resources available to accept the entire packet. In addition, PCI Express verifies data integrity at each hop. This is accomplished through packet acknowledgement, or ACK/NAK. The receiving device will, using the DLLP packets, inform the sending device whether it has correctly received the packets. The transmitting device will then resend those packets not correctly received. ACK and NAK P3 P2 P1 As mentioned earlier, the links carry both payload data (called Transaction Layer Packet, or TLP) as well as management data (called Data Link Layer Packet, or DLLP). Both payload data and management data will be interspersed within the link. Retry buffer A P3 P2 P1 NAK P2 and up P3 P2 ACK P2 and up B TLP and DLLP will be Interspersed in the Dataflow TLP TLP TLP TLP DLLP TLP TLP TLP TLP DLLP Typical Dataflow Management data contained in the DLLP includes traditional PCI functions such as power state control as well as new PCI Express specific system functions such as flow control, and packet acknowledgement. In order to maximize interconnect efficiency, the PCI Express architecture employs a mechanism called Flow Control. Using the DLLP, each link communicates the amount of resources it has available to receive data. The transmitting device will only send its The PCI Express protocol specifies a special packet to transmit payload data. This packet is called Transaction Layer Protocol or TLP. In addition to the actual data, the TLP adds a header that carries information such as packet size, message type (memory, I/O, or configuration), traffic class for QoS and any modifications to the default handling of the transaction (for example, relaxed ordering, or snooping) Applications Using PCI Express PCI Express presents a powerful value proposition. It provides not only a scalable, full featured, high performance platform, but also one designed to thrive in the high volume, rapid turning, and cost sensitive markets.

Certainly, one of the first segments to adopt PCI Express will be the desktop and server marketplaces. System integrators in this space will value PCI Express bandwidth flexibility and low pin count. Most importantly, however, these vendors will value the PCI code compatibility, which will reduce their development cost. PCI Express in the Backplane Applications include Storage, and Communications NPU NPU I/O CPU RC Server Topology Fabric I/O I/O Host Host CPU I/O or Graphics PCI Express Slots PCI-X Northbridge Memory Southbridge PCI I/O USB GbE HDD Using Non-Transparent Bridging for Multiprocessor System Multiprocessor systems have become omnipresent in today s products. Almost every application requires multiple processors, either to allow for intelligent add-in cards, or greater processing power, or higher reliability or some combination of all three. Protocol bridges will be introduced that will allow OEMs to leverage their legacy designs into the new PCI Architecture and therefore gain a time to market advantage. These benefits are not limited to desktop and servers. Other markets, including storage, communication and industrial control will also value these features. These markets will implement PCI Express, not only on the motherboard but also into the backplane. Many of these applications will migrate not only to utilize the PCI code compatibility and low pin count, but also to make use of the bandwidth, scalability, and advanced features. There are three major applications for nontransparent bridging: intelligent adapters, host failover and multiprocessor systems. Intelligent adapters are typically add in cards that employ a processor on the card to offload the host. A good example of an intelligent add in card would be a storage HBA card, but certainly almost every market niche requires that some cards would have a support microprocessor on board. In order for PCI Express to be successful, it needs to provide a mechanism for intelligent adapters. Another popular usage for multiprocessor systems would be to provide a high reliability system. High reliability is by achieved with a secondary host processor. After enumeration, the secondary host would monitor the system state, including the health of the primary processor. When the secondary host determines that the primary host

has failed, it would assume control of the system by promoting itself to the primary host, and restoring the system to an operational state. Applications that would need this include storage and communications systems, where 99.999% uptimes are required. Another popular application for multiprocessor systems is for computational bandwidth. Blade servers exemplify this need. By providing a mechanism to pass data among arrays of processors, PCI Express provides both the performance and economic features necessary for this market segment. To accommodate multihost systems, many will utilize non-transparent bridging. Nontransparent bridges logically isolate processors. This allows multiple processors to co-exist in the system. While there are many means to implement multiprocessor systems, non-transparent bridging is attractive because it is a methodology proven in the PCI space, and is non-proprietary. PCI Express and Advanced Switching There has been quite a bit of confusion on the features and capabilities of PCI Express and Advanced Switching, so it makes sense to compare and contrast the two protocols. The goal of the Advanced Switching specification is to build upon the PCI Express specification and provide new features not previously possible with the standard PCI architecture. This is accomplished by relaxing the requirement to be PCI compatible. These new features include multicast as well as congestion management, both of which are desirable as the system bandwidth utilization begins to reach 100%. Furthermore, Advanced Switching provides protocol encapsulation. Like its name implies, protocol encapsulation allows Advanced Switching to route any protocol by wrapping it in an AS packet. These features make AS an ideal protocol for those who are designing large systems, and who require those benefits not found in PCI Express. When Advanced Switching products begin to roll out, it is expected that system designers will mix AS with PCI Express in their systems. Specifically, they will employ AS only where they require the additional functionality. They will reuse their PCI Express designs to take advantage of the low cost, and proven designs. Because PCI Express and Advanced Switching share the lower layers, one can expect an easy migration. Conclusion PCI Express continues to grow in interest and maturity. This phenomenon even will accelerate past its current frenetic pace as products begin to roll out. PLX hopes that this document is a helpful introduction to this exciting technology. For more detailed information, please visit us at www.plxtech.com or send us an email at pciexpress@plxtech.com.