Data Center Environments



Similar documents
2011 Thermal Guidelines for Data Processing Environments Expanded Data Center Classes and Usage Guidance

Increasing Energ y Efficiency In Data Centers

Guide to Minimizing Compressor-based Cooling in Data Centers

GREEN FIELD DATA CENTER DESIGN WATER COOLING FOR MAXIMUM EFFICIENCY. Shlomo Novotny, Vice President and Chief Technology Officer, Vette Corp.

The Effect of Data Centre Environment on IT Reliability & Energy Consumption

Data Center Equipment Power Trends

Challenges In Intelligent Management Of Power And Cooling Towards Sustainable Data Centre

How Does Your Data Center Measure Up? Energy Efficiency Metrics and Benchmarks for Data Center Infrastructure Systems

Improving Data Center Efficiency with Rack or Row Cooling Devices:

Chiller-less Facilities: They May Be Closer Than You Think

Data Center Rack Level Cooling Utilizing Water-Cooled, Passive Rear Door Heat Exchangers (RDHx) as a Cost Effective Alternative to CRAH Air Cooling

2008 ASHRAE Environmental Guidelines for Datacom Equipment -Expanding the Recommended Environmental Envelope-

Improving Data Center Energy Efficiency Through Environmental Optimization

BCA-IDA Green Mark for Existing Data Centres Version EDC/1.0

CURBING THE COST OF DATA CENTER COOLING. Charles B. Kensky, PE, LEED AP BD+C, CEA Executive Vice President Bala Consulting Engineers

Managing Data Centre Heat Issues

How High Temperature Data Centers and Intel Technologies Decrease Operating Costs

SURVEY RESULTS: DATA CENTER ECONOMIZER USE

Economizer Fundamentals: Smart Approaches to Energy-Efficient Free-Cooling for Data Centers

AIRAH Presentation April 30 th 2014

Trends in Data Center Design ASHRAE Leads the Way to Large Energy Savings

How To Find A Sweet Spot Operating Temperature For A Data Center

DATA CENTER COOLING INNOVATIVE COOLING TECHNOLOGIES FOR YOUR DATA CENTER

Blade Server & Data Room Cooling Specialists

Self-benchmarking Guide for Data Center Infrastructure: Metrics, Benchmarks, Actions Sponsored by:

Rittal Liquid Cooling Series

Dell s Next Generation Servers: Pushing the Limits of Data Center Cooling Cost Savings

THERMAL CLIMATE IN A DATA CENTER

Data Centre Cooling Air Performance Metrics

CIBSE ASHRAE Group. Data Centre Energy Efficiency: Who, What, Why, When, Where & How

Energy Recovery Systems for the Efficient Cooling of Data Centers using Absorption Chillers and Renewable Energy Resources

Data Center Precision Cooling: The Need For A Higher Level Of Service Expertise. A White Paper from the Experts in Business-Critical Continuity

Environmental Data Center Management and Monitoring

Defining Quality. Building Comfort. Precision. Air Conditioning

How High Temperature Data Centers & Intel Technologies save Energy, Money, Water and Greenhouse Gas Emissions

DataCenter 2020: hot aisle and cold aisle containment efficiencies reveal no significant differences

The New Data Center Cooling Paradigm The Tiered Approach

Free Cooling in Data Centers. John Speck, RCDD, DCDC JFC Solutions

Maximize energy savings Increase life of cooling equipment More reliable for extreme cold weather conditions There are disadvantages

Heat Recovery from Data Centres Conference Designing Energy Efficient Data Centres

Elements of Energy Efficiency in Data Centre Cooling Architecture

Cooling Small Server Rooms Can Be. - Jim Magallanes Computer Room Uptime: Uptime Racks:

Absolute and relative humidity Precise and comfort air-conditioning

Analysis of data centre cooling energy efficiency

White Paper. Data Center Containment Cooling Strategies. Abstract WHITE PAPER EC9001. Geist Updated August 2010

5 Reasons. Environment Sensors are used in all Modern Data Centers

Rittal White Paper 508: Economized Data Center Cooling Defining Methods & Implementation Practices By: Daniel Kennedy

DATA CENTER EFFICIENCY AND IT EQUIPMENT RELIABILITY AT WIDER OPERATING TEMPERATURE AND HUMIDITY RANGES

Case Study: Innovative Energy Efficiency Approaches in NOAA s Environmental Security Computing Center in Fairmont, West Virginia

Typical Air Cooled Data Centre Compared to Iceotope s Liquid Cooled Solution

Data Center Cooling & Air Flow Management. Arnold Murphy, CDCEP, CDCAP March 3, 2015

DataCenter 2020: first results for energy-optimization at existing data centers

Data Center 2020: Delivering high density in the Data Center; efficiently and reliably

The Benefits of Supply Air Temperature Control in the Data Centre

Data Center Design Guide featuring Water-Side Economizer Solutions. with Dynamic Economizer Cooling

Bytes and BTUs: Holistic Approaches to Data Center Energy Efficiency. Steve Hammond NREL

How High Temperature Data Centers & Intel Technologies save Energy, Money, Water and Greenhouse Gas Emissions

Recommendation For Incorporating Data Center Specific Sustainability Best Practices into EO13514 Implementation

The Use of Fresh Air in the Data Centre

Energy and Cost Analysis of Rittal Corporation Liquid Cooled Package

Australian Government Data Centre Strategy

Digital Realty Data Center Solutions Digital Chicago Datacampus Franklin Park, Illinois Owner: Digital Realty Trust Engineer of Record: ESD Architect

Case Study: Opportunities to Improve Energy Efficiency in Three Federal Data Centers

The State of Data Center Cooling

Liquid Cooling Solutions for DATA CENTERS - R.M.IYENGAR BLUESTAR LIMITED.

Energy Efficiency Opportunities in Federal High Performance Computing Data Centers

Element D Services Heating, Ventilating, and Air Conditioning

Design Guide. Retrofitting Options For HVAC Systems In Live Performance Venues

Guideline for Water and Energy Considerations During Federal Data Center Consolidations

The Data Center Dilemma - A Case Study

Evaporative Cooling for Residential and Light-Commercial

Combining Cold Aisle Containment with Intelligent Control to Optimize Data Center Cooling Efficiency

Control of Computer Room Air Handlers Using Wireless Sensors. Energy Efficient Data Center Demonstration Project

Energy Efficient Data Centre at Imperial College. M. Okan Kibaroglu IT Production Services Manager Imperial College London.

Reducing Data Center Energy Consumption

HPC TCO: Cooling and Computer Room Efficiency

Optimization of Water - Cooled Chiller Cooling Tower Combinations

Top 5 Trends in Data Center Energy Efficiency

Energy Efficient Cooling Solutions for Data Centers

SEMINAR 23: Data Center and IT Equipment Liquid Cooling: Performance Capabilities, Implementation and Emerging Technologies

Thermal Storage System Provides Emergency Data Center Cooling

COMMERCIAL HVAC CHILLER EQUIPMENT. Air-Cooled Chillers

White Paper: Free Cooling and the Efficiency of your Data Centre

Data Center Energy Profiler Questions Checklist

Selecting Rack-Mount Power Distribution Units For High-Efficiency Data Centers

7 Best Practices for Increasing Efficiency, Availability and Capacity. XXXX XXXXXXXX Liebert North America

DOE FEMP First Thursday Seminar. Learner Guide. Core Competency Areas Addressed in the Training. Energy/Sustainability Managers and Facility Managers

Rittal White Paper 507: Understanding Data Center Cooling Energy Usage & Reduction Methods By: Daniel Kennedy

A discussion of condensate removal systems for clarifier and thickener drives for water and wastewater facilities.

Best Practices. for the EU Code of Conduct on Data Centres. Version First Release Release Public

MODELLING AND OPTIMIZATION OF DIRECT EXPANSION AIR CONDITIONING SYSTEM FOR COMMERCIAL BUILDING ENERGY SAVING

Drives and motors. A guide to using variable-speed drives and motors in retail environments

Datacenter Efficiency

Transcription:

Data Center Environments ASHRAE s Evolving Thermal Guidelines By Robin A. Steinbrecher, Member ASHRAE; and Roger Schmidt, Ph.D., Member ASHRAE Over the last decade, data centers housing large numbers of servers have become significant power consumers. Manufacturers of information technology equipment () continue to increase compute capability while at the same time improving compute efficiency. On the support side, various parties including manufacturers, physical infrastructure manufacturers, data center designers and operators have focused on reducing power consumption from the non-compute part of the overall power load, which is the power and cooling infrastructure that supports the. ASHRAE Technical Committee 9.9, Mission Critical Facilities, Technology Spaces and Electronic Equipment, brings together the disparate interests in the data center industry in the area of data center cooling technologies. Thermal Guidelines for Data Processing Environments helped create a framework to enable the industry to speak a common language and better understand the implications of cooling requirements on the data center and vice versa. Since ASHRAE first issued the Thermal Guidelines in 2004 and updated them in 2008, 1 data center efficiency improvements have become an even higher priority. The original Thermal Guidelines were written by representatives from the manufacturers who understood the limits of the components in the. The objective at the time was to create a common set of environmental parameters that would ensure adequate reliability and performance within reasonable cost constraints. To address the growing concerns about energy efficiency, particularly the cooling component, TC 9.9 has been evolving the Thermal Guidelines toward wider temperature and humidity ranges. About the Authors Robin A. Steinbrecher is a server thermal architect with Intel in Dupont, Wash. Roger Schmidt, Ph.D., is an IBM fellow and chief engineer for Data Center Energy Efficiency in Poughkeepsie, N.Y., and past chair of ASHRAE TC 9.9, Mission Critical Facilities, Technology Spaces and Electronic Equipment. 42 ASHRAE Journal ashrae.org December 2011

This article was published in ASHRAE Journal, December 2011. Copyright 2011 American Society of Heating, Refrigerating and -Conditioning Engineers, Inc. Posted at www.ashrae.org. This article may not be copied and/or distributed electronically or in paper form without permission of ASHRAE. For more information about ASHRAE Journal, visit www.ashrae.org. 2008 Classes 2011 Classes 1 A1 2 A2 A3 A4 3 B 4 C Applications Information Technology Equipment Environmental Control Data Center Office, Home, Transportable Environment, etc. Point-of-Sale, Industrial, Factory, etc. Enterprise Servers, Storage Products Volume Servers, Storage Products, Personal Computers, Workstations Personal Computers, Workstations, Laptops, and Printers Point-of-Sale Equipment, Ruggedized Controllers, or Computers and PDAs Table 1: 2008 and 2011 ASHRAE environmental class definitions. Tightly Controlled Some Control Some Control, Use of Free Cooling Techniques When Allowable Some Control, Near Full-Time Usage Of Free-Cooling Techniques Minimal Control No Control Case studies in recent years have demonstrated significant savings in cooling plant energy when the data center is allowed to operate with economizers and a wider environmental envelope. The 2011 Thermal Guidelines enable these practices to transition from niche to mainstream in mixed-manufacturer data centers. Data center infrastructure energy efficiency is quantified using power usage effectiveness (PUE) as defined in the following equation: Total Facility Power PUE = IT Equipment Power The stringency of the s environmental operating parameters drives the facility power consumption. To enable improved PUEs and overall operational efficiency, ASHRAE TC 9.9 has added two new environmental classes that are more compatible with chiller-less cooling to its Thermal Guidelines, which also provide detailed guidance on how to specify an optimum environmental range for a given data center. And for the first time, TC 9.9 has created liquid-cooling environmental classes along with usage guidance, primarily to support high density data centers. Cooling Thermal Guidelines The environmental classes and their associated limits define the boundaries in which is designed to operate. The original Thermal Guidelines were created to clearly state the environmental ranges, create commonality among the manufacturers and define a metrology within the data center to ensure that the data center s cooling system was delivering properly conditioned air to the inlets of the. Data center cooling could then be designed appropriately to ensure that would be adequately cooled. The original 2004 Thermal Guidelines defined allowable and recommended temperature ranges. Compliant was designed to fully function within the allowable range, while the data center had to be capable of always meeting these parameters. Operating outside the allowable range could result in unsupported operation of the, possibly leading to data errors, reduced performance, shutdown, and permanent damage to the equipment. The recommended range defined the envelope in which would have the highest reliability. Excursions outside the recommended range were acceptable, but little information was provided on what high reliability meant or how overall data center energy consumption could be optimized. In 2008 the recommended range for temperature and humidity were expanded, enabling reduced mechanical cooling and increased economizer hours. Since 2008, innovators in data center infrastructure have created many cooling technologies that compete to successfully address the needs of data centers. The industry now recognizes that outside air can be used with economizers to vastly decrease mechanical cooling in data center implementations, that there is room to expand the application of economizers, and that there are caveats and implications associated with using outside air to cool. To address these topics, TC 9.9 has issued a white paper, 2011 Thermal Guidelines for Data Processing Environments Expanded Data Center Classes and Usage Guidance. 2 This paper defines two new environmental classes and provides guidance to data center operators on how to move their normal operating environmental range outside the recommended envelope and closer to the allowable extremes. Environmental Classes Table 1 describes the class definitions, including the previous 2008 class names. The names were changed to more easily identify those classes pertaining to data centers, the A classes. Compliance with a particular environmental class requires full operation of the equipment over the entire allowable environmental range based on non-failure conditions. Figure 1 shows the previously existing data center classes now named A1 and A2 and the two new data center classes, A3 and A4, along with the recommended range for all the classes. (Table 2 summarizes the same limits.) Most locations around the world will be able to support class A3 or A4 ranges nearly year-round using free-cooling techniques. New data centers may go chiller-less, thereby re- December 2011 ASHRAE Journal 43

ducing capital expenditure, while others may have chillers installed to ensure year-round availability. The additional classes accommodate different applications and priorities of operation. Each data center must be optimized based upon the owner s or operator s own criteria. Although there are great potential efficiency benefits associated with classes A3 and A4, one must weigh the benefits versus the detriments. Setting the Operational Envelope The intent of the recommended range is to define the range where the most reliable and power-efficient operation of the can be achieved based upon manufacturers input. The recommended range was never intended to define the absolute limits of inlet air temperature and humidity for. To clarify, the recommended range should be viewed as the starting point for the operational envelope when no evaluation has been performed to understand the optimal limits for the specific data center of interest. The upcoming third edition of Thermal Guidelines and the white paper provide guidance for performing this evaluation for different types of data centers: existing, retrofit and new construction. Operation outside of the recommended range requires a balance between the cooling system energy savings and the various deleterious effects shown in Figure 2. The evaluation to determine which Range Class Dry-Bulb Temperature Allowable Envelope Performance Corrosion Acoustical Noise 44 ASHRAE Journal ashrae.org December 2011 Power Reliability Humidity Range, Non-Condensing Operating Envelope Cost flow Maximum Dew Point Recommended All A 64.4 F to 80.6 F 41.9 F DP to 60% RH and 59 F DP 59 F Allowable class is optimal for any individual data center should include each of the following factors: Reliability. Long-term reliability of components will degrade with higher temperature. Using economizer techniques will not necessarily cause reduced reliability. The white paper describes how to evaluate long-term reliability using time and temperature weighting factors. Power. The major goal of widening the allowable ranges is to reduce overall data center power. As inlet temperature rises, fans typically speed up to keep components within their temperature limits. This increase depends upon how close the components are to their limits. Light workloads will cause minimal fan speed increase while heavy workloads cause more substantial increases. Fan power increases nonlinearly with speed. Although some opportunity exists for manufacturers to reduce or eliminate the need for higher fan airflows with increased inlet temperatures, it comes at the ex- 35 40 45 50 55 60 65 70 Figure 1: 2011 ASHRAE environmental classes for data center applications. Note: Envelopes represent conditions at IT Equipment inlet. A1 59 F to 89.6 F 20% to 80% RH 62.6 F A2 50 F to 95 F 20% to 80% RH 69.8 F A3 41 F to 104 F 10.4 F DP and 8% RH to 85% RH 75.2 F A4 41 F to 113 F 10.4 F DP & 8% RH to 90% RH 75.2 F B 41 F to 95 F 8% RH to 80% RH 82.4 F C 41 F to 104 F 8% RH to 80% RH 82.4 F Table 2: 2011 ASHRAE environmental classes for data center applications. 75 80 100% 90% 80% 70% 60% 50% 35 40 45 50 55 60 65 70 75 80 85 90 95 100 105 110 115 Dry-Bulb Temperature ( F) ESD Data Center Operational Knowledge Figure 2: Server metrics for determining data center operating environmental envelope. pense of modifying the form factors, which will increase first costs and space requirements. Also, as the temperature goes up, some silicon devices will consume more power 40% 80 75 70 65 60 55 50 45 40 35 30 Dew-Point Temperature ( F)

due to leakage currents. The white paper shows the potential power impact of increasing the inlet temperature. Performance and Allowable Envelope. Many designs include power management features that engage based upon component temperatures. For a given operating temperature, some systems may perform better than others, depending upon the s power and thermal design points. A system designed to handle any potential workload compared with one Advertisement formerly in this space. that is tuned for a specific workload may be capable of better overall performance in wider environmental ranges. Corrosion. Expanding the high end of the humidity envelope increases the opportunity for corrosion problems to occur. Moisture may exacerbate corrosion problems in data centers where corrosive gases are present. Different locations around the world have different ambient levels of potentially corrosive pollutants, and the designer must be cognizant of these pollution levels on a site-specific basis. Even when ambient levels are known to be low, the potential still exists for short-term transient conditions of potentially corrosive pollutants of which the designer must be aware. Acoustical Noise and flow. As mentioned previously, higher inlet temperatures will cause air movers to operate at higher speeds. This causes substantial increases in acoustic output. Measures to protect personnel hearing in compliance with government regulations must be taken into account. Cost. To support higher inlet temperatures, more costly thermal solutions may be required. Larger, more expensive heat sinks may be required to maintain acceptable levels of compute performance, temperature rises, and power consumption. Electrostatic discharge (ESD). Due to the lower allowable humidity ranges in classes A3 and A4, there may be increased probability of ESD events. manufacturers design their equipment based upon international standards to prevent ESD problems. Good ESD practices in the data center can eliminate most concerns. Liquid Cooling Thermal Guidelines Liquid Cooling Guidelines for Datacom Equipment Centers 3 by TC 9.9 was published in 2006, focusing mostly on the design options for liquid-cooled equipment. This book did not address the various facility water temperature ranges possible for supporting liquid cooled equipment. TC 9.9 recently posted the white paper 2011 Thermal Guidelines for Liquid Cooled Data Processing Environments 4 to describe classes for the temperature ranges of facility water supplied to liquid-cooled. In addition, this document reinforces some of the information provided in Liquid Cooling 46 ASHRAE Journal December 2011

Guidelines on the interface requirements between the and infrastructure in support of the liquid cooled. With the focus areas of performance, energy efficiency and use of the waste energy, several ranges of facility supply water temperatures have been specified to accommodate the business and technical requirements of the data center operator. Since the classes cover a wide range of facility water temperatures supplied to the, a brief description is provided for the possible infrastructure equipment that could be deployed between the liquid-cooled and the outside environment. By creating these new facility water cooling classes and not mandating the use of any one of these classes, server manufacturers can develop products for the classes depending on the customer needs and requirements for products within each class. Liquid Cooling The increasing heat density of modern electronics is stretching the ability of air to adequately cool the electronic components within servers, as well as the datacom facilities that house these servers. To meet this challenge, the use of direct water or refrigerant cooling at the rack or server level is being deployed. The ability of water and refrigerant to carry larger amounts of heat per unit volume or mass also offers tremendous advantages. Because of the operating temperatures involved with liquid cooling solutions, water-side economizers fit in well. In the past, high-performance mainframes were often water-cooled with the internal piping supplied by the manufacturer. Components are becoming available today that have similar factory-installed and leak-tested piping that can accept the water from the facility mechanical cooling system, which may also employ a water-side economizer. Increased standardization of liquid-cooled piping connection methods could also help expand the use of liquid cooling. Several implementations of liquid cooling could be deployed, such as the coolant removing a large percentage of the waste heat via a rear door heat exchanger, or a heat exchanger located above or on the side of a rack. Another implementation involves totally enclosing a rack that uses air as the working fluid and an air-to-liquid heat exchanger to transfer the heat outside the rack. Another alternative uses coolant passing through cold plates attached to components within the rack. The coolant distribution unit (CDU) can be external to the datacom rack as shown in Figure 3a, or within the datacom rack as shown in Figure 3b. Advertisement formerly in this space. December 2011 ASHRAE Journal 47

A Building Return Defined Interface External CDU Rack or Cabinet B Building Return Defined Interface Rack or Cabinet Internal CDU Foundation Foundation Figure 3: Combination air- and liquid-cooled rack or cabinet with CDU. External CDU left, internal CDU right. Liquid Cooling Classes Main Heat Rejection Equipment Typical Infrastructure Design Supplemental Cooling Equipment Facility Water Temperature W1 W2 Chiller/Cooling Tower Water-side Economizer (With Drycooler or Cooling Tower) 35.6 F to 62.6 F 35.6 F to 80.6 F W3 Cooling Tower Chiller 35.6 F to 89.6 F W4 Water-side Economizer (With Drycooler or Cooling Tower) N/A 35.6 F to 113 F W5 Building Heating System Cooling Tower >113 F Table 3: ASHRAE Liquid Cooled Guidelines. Note: See Figure 4 for corresponding classes. Figures 3a and 3b show the defined interface for a liquidcooled rack with remote heat rejection. The interface is located at the boundary of the facility water system loop and does not impact the datacom equipment cooling system loops, which will be controlled and managed by the manufacturers. However, the definition of the interface at the loop affects both the manufacturers and the facility operator. As an option, Figure 3a could be extended so that a larger buildinglevel CDU is deployed, thereby eliminating modular CDUs. Facility Water Characteristics for The facility water supplied to the is anticipated to support any liquid cooled using water, water plus additives, refrigerants, or dielectrics. Compliance with a particular environmental class requires full operation of the equipment within the class specified based on non-failure conditions. Different designs from the same class may still have different design points for the cooling components (cold plates, thermal interface materials, liquid flow rates, piping sizes, etc.). For designs that meet the higher supply temperatures as referenced by the ASHRAE classes in Table 3, enhanced thermal designs will be required to maintain the liquid cooled components within the desired temperature limits. Generally, the higher the supply water temperature, the higher the cost of the cooling solutions. Classes W1/W2. Typically, a data center that is cooled using chillers and a cooling tower, but, depending on the data center location, uses an optional water-side economizer to improve energy efficiency (Figure 4a). Class W3. For most locations, these data centers may be operated without chillers. Some locations will still require chillers (Figure 4a). Class W4. To take advantage of energy efficiency and reduce capital expense, these data centers are operated without chillers (Figure 4b). Class W5. To take advantage of energy efficiency, reduce capital expense with chiller-less operation and also make use of the waste energy, the water temperature is high enough to make use of the water exiting the for heating local buildings during the heating season and also potentially using technologies such as adsorption chillers for space cooling during the cooling season (Figure 4c). ASHRAE Classes W1 W3 were developed in concert with the Energy Efficient High Performance Computing Working Group, which is part of Lawrence Berkeley National Laboratory. Currently, no is available in liquid cooling classes W3 W5. Product availability in these ranges in the future will be based upon market demand. It is anticipated that future designs in these classes may involve trade-offs between IT cost and performance. At the same time these classes would allow lower cost data center infrastructure in some locations. The choice of the IT liquid-cooling class should involve a total cost of ownership evaluation of the combined infrastructure and IT capital and operational costs. To avoid condensation, the shown in Figures 4a and 4b would include a CDU if facility water temperature has the potential to fall below the facility dew point. 48 ASHRAE Journal ashrae.org December 2011

Operational Characteristics For Classes W1 and W2 the actual chilled water loop temperature may be set by a campus-wide operational requirement. It also may be the optimum of a balance between lower operational cost using higher temperature chilled water systems versus a lower capital cost with low temperature chilled water systems. For Classes W3, W4, and W5, the infrastructure will likely be specific to the data center and, therefore, the water temperature supplied to the water-cooled will depend on the climate zone in which the data center is located. In these classes running without an installed chiller may be required, so it becomes critical to understand the limits of the watercooled and its integration with the infrastructure designed to support the. This is important so that those extremes in temperature and humidity allow for uninterrupted operation of the data center and the liquid-cooled equipment. For Class W5, the infrastructure will be such that the waste heat from the warm water can be redirected to nearby buildings. Accommodating water temperatures nearer the upper end of the temperature range will be more critical to those applications where retrieving a large amount of waste energy is critical. The water supply temperatures for this class are specified as greater than 113 F (45 C) since the water temperature may be depend on many parameters such as the climate zone, building heating requirements, distance between data center and adjacent buildings, etc. Of course, the electronic components within the IT server need to be cooled to their temperature limits and still use the hotter water as the sink temperature. In many cases, the hotter water temperature will be a challenge to the thermal designer. Summary The thermal design of involves a delicate balance of compute performance, reliability, cost, as well as power consumption. In addition, the facility design must optimize overall power consumption between and HVAC in the data center. Understanding the trade-offs between these objectives is critical to making energy savings decisions. With increased emphasis on data center energy optimization, TC 9.9 has responded by creating expanded air-cooling environmental classes and defining liquid-cooling environmental classes. Along with these new classes, they have provided information necessary to make the critical trade-off decisions between compute performance, reliability, cost, power consumption, and HVAC power consumption. The two 2011 Thermal Guidelines white papers provide information that will enable more data center operators to achieve significant cooling plant energy savings and lower PUEs. Class W1/W2/W3 Cooling Tower Chilled Water Loop Class W4 Cooling Tower Class W5 Cooling Tower Building Heating System A B C Heat Exchanger Water Chiller Heat Exchanger Heat Exchanger Figure 4: Data center liquid cooling implementation. IT Equipment Chilled Water Loop Temperature To IT Equipment 35.6 to 62.6 F 35.6 to 80.6 F 35.6 to 89.6 F IT Equipment Temperature to IT Equipment 35.6 to 113 F IT Equipment Temperature to IT Equipment >113 F 3. TC 9.9. 2011. Thermal Guidelines for Liquid Cooled Data Processing Environments. www.tc99.ashraetcs.org. 4. ASHRAE. 2007. Liquid Cooling Guidelines for Datacom Equipment Centers. Advertisement formerly in this space. References 1. ASHRAE. 2008. Thermal Guidelines for Data Processing Environments, 2 nd Edition. 2. TC 9.9. 2011. Thermal Guidelines for Data Processing Environments. www.tc99.ashraetcs.org. December 2011 ASHRAE Journal 49