Power Saving Features in Mellanox Products

Similar documents
SX1024: The Ideal Multi-Purpose Top-of-Rack Switch

SX1012: High Performance Small Scale Top-of-Rack Switch

ConnectX -3 Pro: Solving the NVGRE Performance Challenge

Connecting the Clouds

InfiniBand Switch System Family. Highest Levels of Scalability, Simplified Network Manageability, Maximum System Productivity

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Long-Haul System Family. Highest Levels of RDMA Scalability, Simplified Distance Networks Manageability, Maximum System Productivity

Mellanox Accelerated Storage Solutions

InfiniBand Switch System Family. Highest Levels of Scalability, Simplified Network Manageability, Maximum System Productivity

Mellanox Academy Online Training (E-learning)

Introduction to Cloud Design Four Design Principals For IaaS

Mellanox Technologies, Ltd. Announces Record Quarterly Results

Building a Scalable Storage with InfiniBand

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

Mellanox HPC-X Software Toolkit Release Notes

Mellanox Global Professional Services

Choosing the Best Network Interface Card for Cloud Mellanox ConnectX -3 Pro EN vs. Intel XL710

Choosing the Best Network Interface Card Mellanox ConnectX -3 Pro EN vs. Intel X520

Mellanox Reference Architecture for Red Hat Enterprise Linux OpenStack Platform 4.0

Mellanox WinOF for Windows 8 Quick Start Guide

Security in Mellanox Technologies InfiniBand Fabrics Technical Overview

Mellanox OpenStack Solution Reference Architecture

Mellanox ConnectX -3 Firmware (fw-connectx3) Release Notes

Replacing SAN with High Performance Windows Share over a Converged Network

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA

Interconnecting Future DoE leadership systems

Advancing Applications Performance With InfiniBand

Juniper Networks QFabric: Scaling for the Modern Data Center

Performance Accelerated Mellanox InfiniBand Adapters Provide Advanced Levels of Data Center IT Performance, Efficiency and Scalability

Scaling 10Gb/s Clustering at Wire-Speed

RoCE vs. iwarp Competitive Analysis

Deploying Ceph with High Performance Networks, Architectures and benchmarks for Block Storage Solutions

Uncompromising Integrity. Making 100Gb/s deployments as easy as 10Gb/s

WHITE PAPER Data center consolidation: Focus on application and business services: Virtualization anywhere: Fabric convergence:

Mellanox Achieves Record Quarterly and Annual Revenue

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

HP Mellanox Low Latency Benchmark Report 2012 Benchmark Report

Large Scale Clustering with Voltaire InfiniBand HyperScale Technology

I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology

Introduction to Infiniband. Hussein N. Harake, Performance U! Winter School

The following InfiniBand products based on Mellanox technology are available for the HP BladeSystem c-class from HP:

Installing Hadoop over Ceph, Using High Performance Networking

Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 SMB Direct

SMB Direct for SQL Server and Private Cloud

SummitStack in the Data Center

InfiniBand in the Enterprise Data Center

Top Ten Reasons for Deploying Oracle Virtual Networking in Your Data Center

Open Ethernet. April

BRIDGING EMC ISILON NAS ON IP TO INFINIBAND NETWORKS WITH MELLANOX SWITCHX

State of the Art Cloud Infrastructure

Optical interconnection networks for data centers

Mellanox Academy Course Catalog. Empower your organization with a new world of educational possibilities

alcatel-lucent converged network solution The cost-effective, application fluent approach to network convergence

SummitStack in the Data Center

Block based, file-based, combination. Component based, solution based

Transform Your Business and Protect Your Cisco Nexus Investment While Adopting Cisco Application Centric Infrastructure

Interconnect Analysis: 10GigE and InfiniBand in High Performance Computing

Mellanox Achieves Record Revenue in the Third Quarter 2015

Achieving a High-Performance Virtual Network Infrastructure with PLUMgrid IO Visor & Mellanox ConnectX -3 Pro

Cloud-Based Apps Drive the Need for Frequency-Flexible Clock Generators in Converged Data Center Networks

Fabrics that Fit Matching the Network to Today s Data Center Traffic Conditions

Performance Evaluation of the RDMA over Ethernet (RoCE) Standard in Enterprise Data Centers Infrastructure. Abstract:

Simplifying Big Data Deployments in Cloud Environments with Mellanox Interconnects and QualiSystems Orchestration Solutions

Data Center Networking Designing Today s Data Center

Microsoft s Cloud Networks

Top Five Things You Need to Know Before Building or Upgrading Your Cloud Infrastructure

Cisco SFS 7000P InfiniBand Server Switch

MLNX_VPI for Windows Installation Guide

Simplifying the Data Center Network to Reduce Complexity and Improve Performance

Virtual Compute Appliance Frequently Asked Questions

Increase Simplicity and Improve Reliability with VPLS on the MX Series Routers

High Speed I/O Server Computing with InfiniBand

The following InfiniBand adaptor products based on Mellanox technologies are available from HP

Cloud Networking: A Novel Network Approach for Cloud Computing Models CQ1 2009

10GBASE-T SFP+ Transceiver Module: Get the most out of your Cat 6a Cabling

Huawei One Net Campus Network Solution

Flexible Modular Data Center Architecture Simplifies Operations

Solution Brief July All-Flash Server-Side Storage for Oracle Real Application Clusters (RAC) on Oracle Linux

LS DYNA Performance Benchmarks and Profiling. January 2009

Building Enterprise-Class Storage Using 40GbE

Photonic Switching Applications in Data Centers & Cloud Computing Networks

Storage, Cloud, Web 2.0, Big Data Driving Growth

Know your Cluster Bottlenecks and Maximize Performance

CloudEngine Series Data Center Switches. Cloud Fabric Data Center Network Solution

MS Series: Ethernet Power Study

CUTTING-EDGE SOLUTIONS FOR TODAY AND TOMORROW. Dell PowerEdge M-Series Blade Servers

ECLIPSE Performance Benchmarks and Profiling. January 2009

Business Case for NFV/SDN Programmable Networks

SDN CENTRALIZED NETWORK COMMAND AND CONTROL

White Paper Solarflare High-Performance Computing (HPC) Applications

Isilon IQ Network Configuration Guide

Cisco SFS 7000D Series InfiniBand Server Switches

Transcription:

WHITE PAPER January, 2013 Power Saving Features in Mellanox Products In collaboration with the European-Commission ECONET Project Introduction... 1 The Multi-Layered Green Fabric... 2 Silicon-Level Power Saving... 2 Link-Level Power Saving... 3 System-Level Power Saving... 4 Fabric-Wise Power Management... 5 Data Center Example... 5 Summary... 7 About ECONET... 8 About Mellanox... 8 Works Cited... 8 Introduction The growth in Cloud and Web 2.0 storage and compute requirements in recent years has led to an increase in demand for larger, stronger, and more cost efficient data centers. As data centers are growing in size, the interconnecting fabric is becoming larger and more complex, resulting in more power consuming network equipment, both in the core and on the edges. This higher power consumption increases both operational costs and carbon footprint, increasing the significance of traditionally overlooked power aspects. Furthermore, rarely all nodes in the data center are active, requiring full network connectivity to be maintained 24/7. An adaptive fabric, capable of shutting down currently unused network elements and self-optimizing its topology, will increase energy efficiency, decrease CO 2 emissions, and reduce the data center cost of operation. This paper introduces the green fabric concept, presents the Mellanox power-efficient features under development as part of the European-Commission ECONET project, displays a real-world data center scenario, and outlines additional steps to be taken toward green fabrics. The features described in this paper can reduce power consumption by up to 43%. When summed over a real-world data center scenario, a total reduction of 13% of all network components power consumption is demonstrated. This reduction can amount to millions of dollars in savings over several years.

page 2 The Multi-Layered Green Fabric Silicon-Level Power Saving An adaptive fabric, capable of dynamically shutting down unused network elements and self-optimizing its topology, will increase energy efficiency, decrease CO 2 emissions, and reduce the data center cost of operation. In order to provide an effective policy, it is important to incorporate power-saving features into all fabric levels, starting from the silicon level, through the Host Channel Adapter (HCA) and switch port, through switch systems, and finally in the Unified Fabric Manager (UFM )/Software Defined Networking (SDN) engine. Switch and HCA silicones are designed to support many combinations of network speeds and protocols. As a result, each silicon device contains components that can be shut down, scaled, or optimized for each different protocol currently being deployed in the fabric. For instance, a Phase-Locked Loop (PLL) supporting 40Gb/s Ethernet can be shut down when an HCA is working at 10GbE or InfiniBand protocols Mellanox integrated circuits incorporate several power-saving features: Unused SerDes (Serializer/Deserializer) PLLs and clock buffers closure Disabled port optimizations Closing unused PCI Express lanes (x16 -> x8 -> x4) while unsupported by board Scaling of core frequency, depending upon protocol used for interconnect Scale ASIC voltage supply These optimizations reduce the power consumption of Mellanox ConnectX -3 silicon by up to 40%, as seen in Figure 1: Figure 1. Reduction of Power in ConnectX-3 IC Due to Silicon Power-Saving Features

page 3 Link-Level Power Saving When a fabric is underutilized, many of its links are up, yet operate at a fraction of their maximum throughput. Reducing each link s speed (for example, 10GbE -> 1GbE) and width (40GbE -> 10GbE or 4xQDR -> 1xQDR) can considerably reduce power consumption for the entire fabric. The Mellanox ConnectX-3 HCA and SX6036 Top-of-Rack (ToR) switch can save power by optimizing each port s link width and speed. These optimizations will be embedded in the HCA and switch hardware and enabled via the firmware. Mellanox is also adding the ability to remotely control and manage these features upon its management software solution MLNX-OS (local device management) and to the UFM, its fabric-wise central management platform. One of the main methods in this aspect is the Width Reduction Power Saving (WRPS) 1, enabling a 40Gb/s 4xQDR port running at an effective 10Gb/s rate to shut down three inactive QDR lanes, thereby saving power. Figures 2 and 3 show the percentage of power saved due to WRPS for a SX6036 switch and a ConnectX-3 HCA, respectively. Figure 2. Width Reduction Power Saving Capability in a Mellanox SX6036 Switch Figure 3. Width Reduction Power Saving Capability in a Mellanox ConnectX-3 HCA

page 4 System-Level Power Saving Internal Port and Module Suspend/Shutdown Large switch systems are comprised of several modules, with many ports each. For instance, the SX6506 InfiniBand director switch allows up to 108 FDR ports, amounting to 12.1Tb/s aggregate switching capacity. In reality, not all ports are active all the time. Some ports are fully utilized, while some are not active at all. This results in a frequently underutilized, yet fully power-consuming switch. In such a case, some of the internal ports can be shut down without compromising the fabric performance. The Mellanox system has the ability to set a port suspension policy that enables saving power on ports that are not being used, while quickly bringing up all needed internal ports once the fabric utilization rises. Figure 4 shows a reduction of ~1% from the SX6506 power consumption for every internal switch port that is closed. Figure 4. SX6506 Power Reduction as a Function of Closed Ports With enough unutilized internal ports, a large switch can shut down an entire internal switching module, resulting in a 14% power reduction. Fan Power Savings Another effective way to save power is to reduce the system fan speed. When the switch operates at low capacity, the temperature does not require full fan Revolutions Per Minute (RPM). Mellanox is developing smart algorithms to optimize power consumption of fans. Figure 5 shows the reduction of power consumption when reducing RPM. Figure 5. Power Consumption as a Result of Fan RPM

page 5 Wake-on-LAN The ConnectX-3 host channel adapter supports the Wake-on-LAN (WOL) standard, allowing it to sleep when receiving an appropriate request from the server and quickly awaken the system once a dedicated magic packet arrives, without losing any data. Figure 6 demonstrates that more than 60% of the HCA s power consumption can be saved in comparison to a fully awake inactive HCA. Figure 6. ConnectX-3 Power Consumption with/without Wake-on-LAN Fabric-Wise Power Management Data Center Example The InfiniBand fabric is centrally managed by the Subnet Manager (SM), with advanced management and monitoring provided by Mellanox s UFM. The UFM is constantly aware of the state of nodes in the fabric, and monitors traffic patterns. Power saving features already implemented in Mellanox systems will be controllable from the UFM in the near future. The UFM will be able to make green configurations and optimizations to the fabric: modifying routing tables and allowing shut down of unused switch ports, modules, and even entire switch systems, while keeping fabric integrity and connectivity intact. This can only be achieved due to the centralized topology management of the InfiniBand fabric. Topology decisions for traditional Ethernet networks are made independently in each node, resulting in a distributed, often very power-inefficient fabric. No element exists in the network that is capable of shutting down ports, nodes, or entire switch systems, while re-engineering the fabric to route around the disabled nodes. SDN calls for unified management of an Ethernet fabric, thereby allowing all of the aforementioned power-related optimizations. Mellanox switches, silicon, and host channel adapters support both InfiniBand and Ethernet on the same platform. When combined with the future SDN-engine that will be integrated into the UFM, a complete Ethernet/InfiniBand green fabric could be built from Mellanox products, with superior performance, lower TCO (total cost of ownership), and a reduced carbon footprint. Figure 7 shows a data center of 1008 servers, interconnected via a 56Gb/s non-blocking FDR InfiniBand fabric. Such a data center is constantly underutilized, running at 10Gb/s most of the time. As a result, the Mellanox HCAs and edge switches can apply the Width Reduction feature, and the core switches can shut down internal ports.

page 6 Figure 7. Data Center with 1008 Servers Connected via a 56Gb/s InfiniBand Fabric. The Fabric Contains 56 Edge Switches and 18 Core Switches. Figure 8 depicts savings per component. Power savings features can save up to 57% from a ConnectX-3 silicon s power consumption. Figure 8. Power Saving in Each Component of the Data Center

page 7 In our example of 1008 servers (56 edge and 18 core switches), we can achieve $51,000 savings per 3 year cycle, as detailed in the following table: Regular Data Center Consumption 132.2KW Green Data Center Consumption 114.5KW Difference 13% Electricity Price $0.11/KWh TCO Reduction for 3 year cycle $51,000 Table 1. Total Power Saving for 1008-Server Data Center over a Three-Year Lifecycle (Calculated Using Average Electricity Prices in the US 2, and a Cooling Factor of 0.7 Added to Total KW) 3 Today s large Web 2.0 data centers include hundreds of thousands of servers. For a data center containing 100,000 servers, savings can amount to $5,000,000 over a three-year cycle, which is an ecological savings equivalent to 51 kilotons of CO 2, 1.3 million trees, or 10 thousand cars. Summary As the data center s power consumption and carbon footprint become increasingly critical, new powersaving methods are required. Mellanox Multi-Tier Green Fabric brings an innovative and holistic power management approach throughout all fabric layers. The multiple power-saving methods implemented in the fabric components help reduce TCO and lower the carbon footprint of the data center network. Mellanox, as a member of the European-Commission ECONET project, continues to advance toward ecofriendly networks by developing more innovative power-saving features and implementing the green fabric vision.

page 8 About ECONET The ECONET (low Energy COnsumption NETworks) project is a three-year IP project (running from October 2010 to September 2013) co-funded by the European Commission under the Framework Programme 7 (FP7). The ECONET project aims at studying and exploiting dynamic adaptive technologies (based on standby and performance scaling capabilities) for wired network devices that allow energy to be saved when a device (or part of it) is not in use. The overall idea is to introduce novel green network-specific paradigms and concepts, enabling the reduction of energy requirements of wired network equipment by 50% in the short- to mid-term (and by 80% in the long run). More information is available at https://www.econet-project.eu. Mellanox is an active partner of the ECONET project. About Mellanox Mellanox Technologies (NASDAQ: MLNX, TASE: MLNX) is a leading supplier of end-to-end InfiniBand and Ethernet interconnect solutions and services for servers and storage. Mellanox interconnect solutions increase data center efficiency by providing the highest throughput and lowest latency, delivering data faster to applications and unlocking system performance capability. Mellanox offers a choice of fast interconnect products: adapters, switches, software and silicon that accelerate application runtime and maximize business results for a wide range of markets including high performance computing, enterprise data centers, Web 2.0, cloud, storage and financial services. More information is available at www.mellanox.com. References 1 Width Reduction Power Saving technology is protected by U.S. Patent Application #12/686,401 2 http://www.eia.gov/beta/enerdat/#/topic/7?agg=2,0,1&geo=g&freq=m&start=200101&end=201209&ct ype=linechart&maptype=0&rse=0 3 http://www.apcmedia.com/salestools/vavr-5tdtef_r1_en.pdf 350 Oakmead Parkway, Suite 100, Sunnyvale, CA 94085 Tel: 408-970-3400 Fax: 408-970-3403 www.mellanox.com Copyright 2013. Mellanox Technologies. All rights reserved. Mellanox, BridgeX, ConnectX, CORE-Direct, InfiniBridge, InfiniHost, InfiniScale, PhyX, SwitchX, UFM, Virtual Protocol Interconnect and Voltaire are registered trademarks of Mellanox Technologies,Ltd. Connect-IB, CoolBox, FabricIT, Mellanox Federal Systems, Mellanox Software Defined Storage, MetroX, MLNX-OS, ScalableHPC and Unbreakable-Link are trademarks of Mellanox Technologies, Ltd. All other trademarks are property of their respective owners. 4181WP REV 1.0