Energy Efficiency Opportunities in Federal High Performance Computing Data Centers

Similar documents
Case Study: Opportunities to Improve Energy Efficiency in Three Federal Data Centers

Case Study: Innovative Energy Efficiency Approaches in NOAA s Environmental Security Computing Center in Fairmont, West Virginia

Data Center Components Overview

Recommendation For Incorporating Data Center Specific Sustainability Best Practices into EO13514 Implementation

Free Cooling in Data Centers. John Speck, RCDD, DCDC JFC Solutions

Data Center Airflow Management Retrofit

Managing Data Centre Heat Issues

Dealing with Thermal Issues in Data Center Universal Aisle Containment

CURBING THE COST OF DATA CENTER COOLING. Charles B. Kensky, PE, LEED AP BD+C, CEA Executive Vice President Bala Consulting Engineers

Benefits of. Air Flow Management. Data Center

Data Center Energy Profiler Questions Checklist

Re Engineering to a "Green" Data Center, with Measurable ROI

Improving Data Center Efficiency with Rack or Row Cooling Devices:

Optimizing Network Performance through PASSIVE AIR FLOW MANAGEMENT IN THE DATA CENTER

Cooling Capacity Factor (CCF) Reveals Stranded Capacity and Data Center Cost Savings

Control of Computer Room Air Handlers Using Wireless Sensors. Energy Efficient Data Center Demonstration Project

How To Improve Energy Efficiency Through Raising Inlet Temperatures

University of St Andrews. Energy Efficient Data Centre Cooling

National Grid Your Partner in Energy Solutions

APC APPLICATION NOTE #112

How To Improve Energy Efficiency In A Data Center

Alan Matzka, P.E., Senior Mechanical Engineer Bradford Consulting Engineers

Verizon SMARTS Data Center Design Phase 1 Conceptual Study Report Ms. Leah Zabarenko Verizon Business 2606A Carsins Run Road Aberdeen, MD 21001

DATA CENTER COOLING INNOVATIVE COOLING TECHNOLOGIES FOR YOUR DATA CENTER

General Recommendations for a Federal Data Center Energy Management Dashboard Display

Element D Services Heating, Ventilating, and Air Conditioning

- White Paper - Data Centre Cooling. Best Practice

Unified Physical Infrastructure (UPI) Strategies for Thermal Management

Liquid Cooling Solutions for DATA CENTERS - R.M.IYENGAR BLUESTAR LIMITED.

Guide to Minimizing Compressor-based Cooling in Data Centers

Presentation Outline. Common Terms / Concepts HVAC Building Blocks. Links. Plant Level Building Blocks. Air Distribution Building Blocks

Center Thermal Management Can Live Together

Data Center Power Consumption

Data Centers: How Does It Affect My Building s Energy Use and What Can I Do?

AisleLok Modular Containment vs. Legacy Containment: A Comparative CFD Study of IT Inlet Temperatures and Fan Energy Savings

Data Center Design Guide featuring Water-Side Economizer Solutions. with Dynamic Economizer Cooling

The New Data Center Cooling Paradigm The Tiered Approach

Air, Fluid Flow, and Thermal Simulation of Data Centers with Autodesk Revit 2013 and Autodesk BIM 360

Best Practices in HVAC Design/Retrofit

Guideline for Water and Energy Considerations During Federal Data Center Consolidations

Benefits of Cold Aisle Containment During Cooling Failure

White Paper. Data Center Containment Cooling Strategies. Abstract WHITE PAPER EC9001. Geist Updated August 2010

How to Build a Data Centre Cooling Budget. Ian Cathcart

European Code of Conduct for Data Centre. Presentation of the Awards 2014

Variable-Speed Fan Retrofits for Computer-Room Air Conditioners

Cooling Capacity Factor (CCF) Reveals Stranded Capacity and Data Center Cost Savings

Wireless Sensors Improve Data Center Energy Efficiency

CIBSE ASHRAE Group. Data Centre Energy Efficiency: Who, What, Why, When, Where & How

QUALITATIVE ANALYSIS OF COOLING ARCHITECTURES FOR DATA CENTERS

DOE FEMP First Thursday Seminar. Learner Guide. Core Competency Areas Addressed in the Training. Energy/Sustainability Managers and Facility Managers

NetApp Data Center Design

How To Run A Data Center Efficiently

Introduction to Datacenters & the Cloud

Blade Server & Data Room Cooling Specialists

Stanford University Server/Telecom Rooms Design Guide

Energy Efficiency Best Practice Guide Data Centre and IT Facilities

Improving Data Center Energy Efficiency Through Environmental Optimization

Data Center Rack Level Cooling Utilizing Water-Cooled, Passive Rear Door Heat Exchangers (RDHx) as a Cost Effective Alternative to CRAH Air Cooling

Hot Air Isolation Cools High-Density Data Centers By: Ian Seaton, Technology Marketing Manager, Chatsworth Products, Inc.

High Density Data Centers Fraught with Peril. Richard A. Greco, Principal EYP Mission Critical Facilities, Inc.

Data Center Leadership In the DOE Better Buildings Challenge

Analysis of data centre cooling energy efficiency

Education Evolution: Scalable Server Rooms George Lantouris Client Relationship Manager (Education) May 2009

A Comparative Study of Various High Density Data Center Cooling Technologies. A Thesis Presented. Kwok Wu. The Graduate School

IMPROVING DATA CENTER EFFICIENCY AND CAPACITY WITH AISLE CONTAINMENT

Reducing Data Center Energy Consumption

Data center upgrade proposal. (phase one)

DataCenter 2020: hot aisle and cold aisle containment efficiencies reveal no significant differences

I-STUTE Project - WP2.3 Data Centre Cooling. Project Review Meeting 8, Loughborough University, 29 th June 2015

Top 5 Trends in Data Center Energy Efficiency

GUIDE TO ICT SERVER ROOM ENERGY EFFICIENCY. Public Sector ICT Special Working Group

DATA CENTER ASSESSMENT SERVICES

Data Centre Energy Efficiency Operating for Optimisation Robert M Pe / Sept. 20, 2012 National Energy Efficiency Conference Singapore

Glossary of Heating, Ventilation and Air Conditioning Terms

The CEETHERM Data Center Laboratory

The Different Types of Air Conditioning Equipment for IT Environments

Unified Physical Infrastructure SM (UPI) Strategies for Smart Data Centers

Data Center Technology: Physical Infrastructure

Best Practices. for the EU Code of Conduct on Data Centres. Version First Release Release Public

Specialty Environment Design Mission Critical Facilities

Energy and Cost Analysis of Rittal Corporation Liquid Cooled Package

Recommendations for Measuring and Reporting Overall Data Center Efficiency

Elements of Energy Efficiency in Data Centre Cooling Architecture

Power and Cooling for Ultra-High Density Racks and Blade Servers

Data Centre Cooling Air Performance Metrics

Choosing Close-Coupled IT Cooling Solutions

White Paper #10. Energy Efficiency in Computer Data Centers

Self-benchmarking Guide for Data Center Infrastructure: Metrics, Benchmarks, Actions Sponsored by:

STULZ Water-Side Economizer Solutions

Rack Hygiene. Data Center White Paper. Executive Summary

Green Data Centre Design

Trends in Data Center Design ASHRAE Leads the Way to Large Energy Savings

DataCenter 2020: first results for energy-optimization at existing data centers

Direct Fresh Air Free Cooling of Data Centres

Great Lakes Data Room Case Study

BICSInews. plus. july/august Industrial-grade Infrastructure. + I Say Bonding, You Say Grounding + Data Center Containment.

Cooling Audit for Identifying Potential Cooling Problems in Data Centers

Transcription:

Energy Efficiency Opportunities in Federal High Performance Computing Data Centers Prepared for the U.S. Department of Energy Federal Energy Management Program By Lawrence Berkeley National Laboratory Rod Mahdavi, P.E. LEED A.P. September 2013

Contacts Rod Mahdavi, P.E. LEED AP Lawrence Berkeley National Laboratory (510) 495-2259 rmahdavi@lbl.gov For more information on FEMP: Will Lintner, P.E. Federal Energy Management Program U.S. Department of Energy (202) 586-3120 william.lintner@ee.doe.gov 2

Contents Executive Summary... 6 Overview... 7 Assessment Process... 8 Challenges...9 EEMs for HPC Data Centers... 10 Air Management Adjustment Package...10 Cooling Retrofit Package...10 Generator Block Heater Modification Package...11 Full Lighting Retrofit Package...11 Chilled Water Plant Package...11 Additional Opportunities... 11 Next Steps... 14 3

List of Figures Figure 1. Data Center Thermal Map, After Containment... 9 Figure 2. Data Center Thermal Map, After Raised Temperature... 9 Figure 3. Hoods for Power Supply Unit Air Intake... 13 4

List of Tables Table 1. HPC Sites Energy/GHG Savings... 6 Table 2. Computers and Cooling Types in HPC Sites... 7 Table 3. EEM Packages Status for HPC Data Centers... 12 Table 4. Summary of Power Losses and PUE in DOD HPC Data Centers... 14 5

Executive Summary The U.S. Department of Energy s Office of Energy Efficiency and Renewable Energy (EERE) s Federal Energy Management Program (FEMP) assessed the energy use at six Department of Defense (DOD) High Performance Computing (HPC) data centers in 2011. Table 1 provides a breakdown of the potential energy and cost savings and average payback periods for each data center. The total energy saved was estimated at more than 8,000 Megawatt-hours (MWh) with an annual greenhouse gas (GHG) emission reduction of 7,500 tons. Energy cost savings of approximately one million dollars per year were found to be possible through energy efficiency measures (EEM) that had an average payback period of less than two years. The individual data centers contained a variety of IT systems and cooling systems. Server rack power densities were measured from 1kW (non HPC racks) to 25kW. In a conventional data center with rows of sever racks, separation of the rows into hot and cold aisles followed by containment of the hot or the cold aisle can result in substantial reductions of cooling energy. This arrangement does not fit well with some of the HPC data centers because of the integrated local cooling and different cooling air path. The most effective EEM was increasing the data center temperature which made it possible to save cooling energy by raising the chilled water supply temperature. Sites Payback years Annual Energy Saving MWh Annual Greenhouse Gas Reduction (GHG) Emission Reduction Ton Site 1 2.9 1,090 940 Site 2A 2.5 3,060 2,750 Site 2B 2 2,520 2,300 Site 3 2 937 860 Site 4 2.7 520 480 Site 5 0.5 1,000 950 Table 1. HPC Sites Energy/GHG Savings 6

Overview The DOD Office of High Performance Computing Management Program and (FEMP) jointly funded assessments by personnel from the Lawrence Berkeley National Laboratory (LBNL) because of an earlier success in identifying EEMs at an HPC data center in Hawaii. This case study includes the results of the assessments, lessons learned, and recommended EEMs. Table 2 shows the IT equipment for each site with their corresponding cooling systems. High power density is the most common feature of this equipment. Power used by each rack averaged up to 10-25kW. Another common feature is the use of 480V power without intervening transformation. Since with each transformation there is power loss and heat rejection, eliminating several power transformations reduced the need to remove the generated heat. Many of the systems also have self-contained cooling using local heat exchangers in the form of refrigerant cooling or rear door heat exchangers. The advantage of this feature is that the heat removed is very close to the source enabling heat transfer to be done at higher temperatures. This by itself can make the cooling very efficient since very cold chilled water will not be needed. The result is increased of hours of air side or water side economizer operation and reduced compressor operating hours. Most of the HPC server racks are different from the usual server racks in the way cooling air flows through them. An example is the Cray unit with cooling air flowing from the bottom to the top, rather than horizontally. Site IT System Cooling Type Site 1 Cray XE6 Refrigerant Cooled using local water cooled refrigerant pumps SGI Altix 4700 Rear Door Heat Exchanger using central chilled water SGI Altix Ice 8200 Air cooled Site 2A Air cooled with integrated fan, bottom air Cray XT5 intake, top exhaust Air cooled, ready for rear door heat exchanger SGI Altix Ice 8200 Site 2B using central chilled water Linux Woodcrest Air cooled Cray XE6 Refrigerant Cooled using local water cooled refrigerant pumps Site 3 Cray XT3 Air cooled with integrated fan, bottom air intake, top exhaust Cray XT4 Air cooled with integrated fan, bottom air intake, top exhaust SGI Altix Ice 8200 Rear Door Heat Exchanger using central chilled water IBM Cluster 1600 P5 Air cooled Site 4 Cray XT5 Air cooled with integrated fan, bottom air intake, top exhaust IBM Power 6 Air cooled Site 5 Dell Power Edge M610 Air cooled Table 2. Computers and Cooling Types in HPC Sites 7

Assessment Process The first step of the assessment was to baseline the environmental conditions in the HPC data centers vs. the ASHRAE recommended and allowable thermal guidelines. LBNL installed a wireless monitoring system. The following environmental sensor points were installed throughout the data center: temperature sensors at all computer room air handlers (CRAH) or/and computer room air conditioners (CRAC) supply and return. Supply thermal nodes were installed under the raised floor just in front of the unit. Return nodes were placed over the intake filter or were strapped to the beam on the top of the intake chimney. For CRAH/CRAC with their chimneys extended to the data center ceiling, thermal nodes were placed over the ceiling tile very close to the CRAH/CRAC intake. Humidity was measured by the same thermal node. Temperature sensors were placed on front of the rack near the top, at the middle and near the base. Other temperature sensors were installed on the back of the rack, in the sub-floor, and on every third rack in a row. Pressure sensors were located throughout the data center to measure the air pressure differential between sub-floor supply plenum and the room. The same approach was used for data centers with no raised floor by installing the sensors in the ductwork. The wireless sensor network continuously sampled the data center environmental conditions and reported at five minute intervals. Power measurements (where possible), power readings from equipment (Switch gear, UPS, PDU, etc.), and power usage estimation facilitated the calculation of power usage effectiveness. Generally, power use information is critical to the assessment. IT power should be measured as close as possible to IT equipment. More recent power supply units measure power and can communicate measurements to a central monitoring system. If this is not possible, then communication or manual reading of PDU or at a higher chain level, UPS power, can be assumed as IT power usage. Loss in the electrical power chain can be measured or can be estimated while considering how efficient the UPS units are loaded. Lighting power usage can be estimated by counting the fixtures or estimated based upon watt per square foot. Cooling load measurement, including chiller power usage, can be a challenge. If reading from panel is not possible, then estimation based on theoretical plant efficiency can work. While a spot measurement at CRAH units with constant speed fans is sufficient to estimate power use, it is more advantageous to continuously monitor CRAC units because of the variable power usage of the compressors. The next step was to analyze the data and present the result of the assessment. The collected data provided empirical measures of recirculation and by-pass air mixing, and cooling system efficiency. The assessment established the data center s baseline energy utilization and identified the EEM s and their potential energy savings benefit. In some of the centers, a few low cost EEMs were completed and their impacts were observed in real time. For instance, Figure 1 is the thermal map of one the data centers that was assessed. The map shows the impact of partially contained hot aisles. The major power usage comes from the two rows on the left. The containment helped to isolate hot air to some extent. 8

Figure 1. Data Center Thermal Map, After Containment Figure 2 shows the impact of raised CRAH supply air temperature. The result is a more uniform and higher temperature within the data center space. Chilled water temperature was then increased, which reduced energy usage by the chiller plant. In this case, the IT equipment cooling air intake was in front and exhaust was at the back of the rack. Challenges Figure 2. Data Center Thermal Map, After Raised Temperature The LBNL technical team was challenged by these energy assessments, as many of the HPC data centers employed different cooling types and configurations in the same room. For example, 9

while one area used building HVAC, another system had rack based air cooling using refrigerant coils and a third system had rear door cooling heat exchangers using central chilled water system. The air intake to the IT equipment posed another challenge. While in conventional server racks air enters horizontally at the face of the rack and exits from the back, in some HPC systems (e.g., Cray) air enters from the bottom (under the raised floor) and exits at the top. With vertical air displacement the containment of aisles by itself does not impact the cooling energy use. The same holds true with in-row and in-rack cooling systems, or rear door heat exchangers, since the exhaust air can be as cold the intake air. This cold air mixes with the hot air. Temperature difference across the air handler s coils is reduced which results in inefficiencies in the cooling systems. EEMs for HPC Data Centers For the presentation of results, LBNL personnel created packages of EEMs categorized based on their cost and simple payback. Typical EEMs covered in each package with its rough cost estimate are described as follows: Air Management Adjustment Package Seal all floor leaks, Rearrange the perforated floor tiles locating them only in cold aisles for conventional front to back airflow; solid everywhere else, Contain hot air to avoid mixing with cold air, Seal spaces between and within racks, Raise the supply air temperature (SAT), Disable humidity controls and humidifiers, Control humidity only on makeup air only, and Turn off unneeded CRAH units. Typical Cost of this package is $100/kW of IT power. A typical simple payback is ~1 year. Cooling Retrofit Package Install variable speed drives for air handler fan and control fan speed by air plenum differential pressure, Install ducting from the air handler units to the ceiling to allow hot aisle exhaust to travel through the ceiling space back to the computer room air handling units, Convert computer room air handler air temperature control to rack inlet air temperature control, and Raise the chilled water supply temperature thus saving energy through better chiller efficiency. 10

Typical Cost of this package is $180/kW of IT power. Typical simple payback is 2.5years. Generator Block Heater Modification Package Equip generator s block heater with thermostat control, Seal all floor leaks, and Reduce the temperature set point for the block heater. Typical Cost of this package is $10/kW of IT power. A typical simple payback is 2.5 years. Full Lighting Retrofit Package Reposition light fixtures from above racks to above aisles, Reduce lighting, and Install occupancy sensors to control fixtures. Typical Cost of this package is $15/kW of IT power. A typical simple payback is 3 years. Chilled Water Plant Package Install water side economizer, Investigate with HPC equipment manufacturer whether chilled water supply temperature to their heat exchanger can be increased, Run two chilled water loops, one for those equipment with lower chilled water temperature requirement and the other for those with rear door heat exchangers, Install VFD on pumps, and Purchase high efficiency chillers and motors if renovation or capacity increase is planned. Typical Cost of this package is $400/kW of IT power. A typical simple payback is 3 years. Additional Opportunities The use of water cooling can lead to major reductions in power use due to the higher energy carrying capacity of liquids in comparison to air. Higher temperature water can also be used for cooling, saving additional energy. This strategy can work very well with rear door heat exchangers. This strategy can also be integrated with the use of cooling towers or dry coolers to provide cooling water directly, thus bypassing the compressor cooling. Refrigerant cooling systems can be specified to operate using higher water temperatures than the current 45-50 o F. There is already equipment that operates at 65 o F. This will increase compressor-less cooling hours. 11

Site Recommended Packages Implementation Site 1 Site 2A Site 2B Site 3 Site 4 Site 5 Air Management Adjustment Cooling Retrofit Lighting Retrofit Chilled Water Plant Air Management Adjustment Cooling Retrofit EG Block Heater Lighting Retrofit Chilled Water Plant Air Management Adjustment Cooling Retrofit EG Block Heater Lighting Retrofit Chilled Water Plant Air Management Adjustment Cooling Retrofit EG Block Heater Lighting Retrofit Chilled Water Plant Air Management Adjustment Cooling Retrofit EG Block Heater Lighting Retrofit Chilled Water Plant Air Management Adjustment Cooling Retrofit EG Block Heater Lighting Retrofit Chilled Water Plant potential Implemented Implemented Table 3. EEM Packages Status for HPC Data Centers Lessons Learned The main barrier to increasing the supply air temperature was the IT equipment and refrigerant pumping system maximum temperature requirements. In one of the sites, the refrigerant pumping system required chilled water supply temperatures of 45 o F. Granular monitoring enables temperature measurements at the server level, allowing for the implementation of some low cost simpler EEMs during the assessment without interrupting the data center operation. Chilled water supply temperature setpoint optimization can result in large energy savings, especially for the cooling systems that utilize air cooled chillers. 12

Some conventional approaches, such as sealing the floor, are applicable to HPC data centers, but some commonly applied EEMs to enterprise data centers such as hot/cold aisle isolation are not suitable to HPC data centers where the racks are cooled internally by refrigerant coils or rear door heat exchangers. Hoods are installed as is shown in Figure 3, to direct cold air to Cray power supply units, an example of site engineer s remedy to prevent mixing of cold and hot air. This is a universal problem and will work in similar situations where IT equipment has an unusual configuration. In this case, the airflow from the bottom of the racks did not reach the power supply unit so they required additional air flow from the back of the cabinet. Installing just the perforated tiles would have caused mixing of cold and hot air but with installation of the hoods on top of the perforated tiles and direct air from under raised floor to the cabinet. To avoid cold air supply release into the aisle the remainder of the perforated tiles that were exposed to the room was blanked off. Figure 3. Hoods for Power Supply Unit Air Intake Openings within the HPC create problems with overheating of cores when data center temperature is increased to save energy. Installation of internal air dams to prevent recirculation of air within the racks helped to address this problem. 13

Site Current IT Load kw/sqft Current IT Load kw Elec Dist. Loss kw Cooling Load kw Fan Load kw Other users kw Current PUE PUE Site 1 120 2,000 150 750 200 260 1.68 1.64 Site 2A 180 1,050 170 450 195 150 1.92 1.57 Site 2B 240 810 170 370 160 95 1.98 1.63 Site 3 260 1,670 100 700 125 120 1.63 1.56 Site 4 130 550 158 180 47 65 1.82 1.71 Site 5 130 510 73 265 80 33 1.88 1.65 Next Steps Table 4. Summary of Power Losses and PUE in DOD HPC Data Centers In order to implement the remaining EEMs, FEMP recommends an invest grade assessment by a firm experienced in data center efficiency improvements. If agency funds are not available for implementing the EEMs, then private sector financing mechanisms such as energy savings performance contracts (ESPC) or utilities energy savings performance contracts (UESC) may be appropriate, considering the attractive payback periods and the magnitude of savings. FEMP can assist in exploring such opportunities. In addition, the LBNL experts recommend installing metering and monitoring systems, especially with a comprehensive dashboard to present the environmental data and the energy efficiency related data, including the power use by different components and systems. The dashboard will allow the operators to make real time changes to optimize the energy efficiency. The LBNL experts also recommend that future purchases of IT equipment include a preference for water cooled systems. At a minimum, the IT equipment should be capable of operating at more than a 900F air intake temperature, operate at a high voltage (480V is preferred), and contain a variable speed server fan controlled by the server core temperature. 14

DOE/EE-0971 September 2013 Printed with a renewable-source ink on paper containing at least 50% wastepaper, including 10% post consumer waste.