CFD SIMULATION ENHANCES A DATA CENTER S EXPANSION PROCESS



Similar documents
Verizon SMARTS Data Center Design Phase 1 Conceptual Study Report Ms. Leah Zabarenko Verizon Business 2606A Carsins Run Road Aberdeen, MD 21001

Improving Data Center Energy Efficiency Through Environmental Optimization

GUIDE TO ICT SERVER ROOM ENERGY EFFICIENCY. Public Sector ICT Special Working Group

How To Improve Energy Efficiency Through Raising Inlet Temperatures

Data Center Design Guide featuring Water-Side Economizer Solutions. with Dynamic Economizer Cooling

Impacts of Perforated Tile Open Areas on Airflow Uniformity and Air Management Performance in a Modular Data Center

Energy Efficiency Opportunities in Federal High Performance Computing Data Centers

Managing Data Centre Heat Issues

Power and Cooling for Ultra-High Density Racks and Blade Servers

IPACK DISTRIBUTED LEAKAGE FLOW IN RAISED-FLOOR DATA CENTERS

Reducing the Annual Cost of a Telecommunications Data Center

Data Center Components Overview

Effect of Rack Server Population on Temperatures in Data Centers

How to Build a Data Centre Cooling Budget. Ian Cathcart

High Density Data Centers Fraught with Peril. Richard A. Greco, Principal EYP Mission Critical Facilities, Inc.

International Telecommunication Union SERIES L: CONSTRUCTION, INSTALLATION AND PROTECTION OF TELECOMMUNICATION CABLES IN PUBLIC NETWORKS

How To Run A Data Center Efficiently

Energy Performance Optimization of Server Room HVAC System

Cooling Audit for Identifying Potential Cooling Problems in Data Centers

AisleLok Modular Containment vs. Legacy Containment: A Comparative CFD Study of IT Inlet Temperatures and Fan Energy Savings

RAISED-FLOOR DATA CENTER: PERFORATED TILE FLOW RATES FOR VARIOUS TILE LAYOUTS

Fundamentals of CFD and Data Center Cooling Amir Radmehr, Ph.D. Innovative Research, Inc.

Cooling Capacity Factor (CCF) Reveals Stranded Capacity and Data Center Cost Savings

Environmental Data Center Management and Monitoring

Dealing with Thermal Issues in Data Center Universal Aisle Containment

Air, Fluid Flow, and Thermal Simulation of Data Centers with Autodesk Revit 2013 and Autodesk BIM 360

Unified Physical Infrastructure (UPI) Strategies for Thermal Management

- White Paper - Data Centre Cooling. Best Practice

Free Cooling in Data Centers. John Speck, RCDD, DCDC JFC Solutions

Benefits of. Air Flow Management. Data Center

Benefits of Cold Aisle Containment During Cooling Failure

Green Data Centre Design

Problem Statement In order to satisfy production and storage requirements, small and medium-scale industrial

Energy Efficient Data Center Design. Can Ozcan Ozen Engineering Emre Türköz Ozen Engineering

Enclosure and Airflow Management Solution

Hot Air Isolation Cools High-Density Data Centers By: Ian Seaton, Technology Marketing Manager, Chatsworth Products, Inc.

APC APPLICATION NOTE #92

Data Centre Cooling Air Performance Metrics

Element D Services Heating, Ventilating, and Air Conditioning

Use of Computational Fluid Dynamics for Calculating Flow Rates Through Perforated Tiles in Raised-Floor Data Centers

The New Data Center Cooling Paradigm The Tiered Approach

IMPROVING DATA CENTER EFFICIENCY AND CAPACITY WITH AISLE CONTAINMENT

Modeling Rack and Server Heat Capacity in a Physics Based Dynamic CFD Model of Data Centers. Sami Alkharabsheh, Bahgat Sammakia 10/28/2013

Prediction Is Better Than Cure CFD Simulation For Data Center Operation.

Cooling Capacity Factor (CCF) Reveals Stranded Capacity and Data Center Cost Savings

BRUNS-PAK Presents MARK S. EVANKO, Principal

Data Center Power Consumption

Increasing Energ y Efficiency In Data Centers

W H I T E P A P E R. Computational Fluid Dynamics Modeling for Operational Data Centers

Our data centres have been awarded with ISO 27001:2005 standard for security management and ISO 9001:2008 standard for business quality management.

Case Study: Opportunities to Improve Energy Efficiency in Three Federal Data Centers

DataCenter 2020: hot aisle and cold aisle containment efficiencies reveal no significant differences

Data center upgrade proposal. (phase one)

Using Simulation to Improve Data Center Efficiency

CFD SIMULATION OF SDHW STORAGE TANK WITH AND WITHOUT HEATER

How Does Your Data Center Measure Up? Energy Efficiency Metrics and Benchmarks for Data Center Infrastructure Systems

CURBING THE COST OF DATA CENTER COOLING. Charles B. Kensky, PE, LEED AP BD+C, CEA Executive Vice President Bala Consulting Engineers

How Row-based Data Center Cooling Works

Analysis of the UNH Data Center Using CFD Modeling

Application of CFD modelling to the Design of Modern Data Centres

Using CFD for optimal thermal management and cooling design in data centers

IT White Paper MANAGING EXTREME HEAT: COOLING STRATEGIES FOR HIGH-DENSITY SYSTEMS

Data Center Rack Level Cooling Utilizing Water-Cooled, Passive Rear Door Heat Exchangers (RDHx) as a Cost Effective Alternative to CRAH Air Cooling

Great Lakes Data Room Case Study

DataCenter 2020: first results for energy-optimization at existing data centers

Education Evolution: Scalable Server Rooms George Lantouris Client Relationship Manager (Education) May 2009

APC APPLICATION NOTE #112

White Paper. Data Center Containment Cooling Strategies. Abstract WHITE PAPER EC9001. Geist Updated August 2010

Using Computational Fluid Dynamics (CFD) for improving cooling system efficiency for Data centers

Creating Data Center Efficiencies Using Closed-Loop Design Brent Goren, Data Center Consultant

Control Strategies for Plenum Optimization in Raised Floor Data Centers

A Comparative Study of Various High Density Data Center Cooling Technologies. A Thesis Presented. Kwok Wu. The Graduate School

Liquid Cooling Solutions for DATA CENTERS - R.M.IYENGAR BLUESTAR LIMITED.

Saving energy in the Brookhaven National Laboratory Computing Center: Cold aisle containment implementation and computational fluid dynamics modeling

Analysis of data centre cooling energy efficiency

Elements of Energy Efficiency in Data Centre Cooling Architecture

Combining Cold Aisle Containment with Intelligent Control to Optimize Data Center Cooling Efficiency

Rate of Heating Analysis of Data Centers during Power Shutdown

Measure Server delta- T using AUDIT- BUDDY

Data Centers: How Does It Affect My Building s Energy Use and What Can I Do?

COMPARATIVE ANALYSIS OF WIDELY USED AIR CONTAINMENT SYSTEMS IN COMMERCIAL DATA CENTERS FOR MAXIMUM COOLING PERFORMANCE KASTURI RANGAN RAJAGOPALAN

Sealing Gaps Under IT Racks: CFD Analysis Reveals Significant Savings Potential

Optimizing Network Performance through PASSIVE AIR FLOW MANAGEMENT IN THE DATA CENTER

Data Center Energy Profiler Questions Checklist

The Different Types of Air Conditioning Equipment for IT Environments

Data Center Temperature Rise During a Cooling System Outage

Data Center Cooling & Air Flow Management. Arnold Murphy, CDCEP, CDCAP March 3, 2015

Re Engineering to a "Green" Data Center, with Measurable ROI

GREEN FIELD DATA CENTER DESIGN WATER COOLING FOR MAXIMUM EFFICIENCY. Shlomo Novotny, Vice President and Chief Technology Officer, Vette Corp.

Airflow Simulation Solves Data Centre Cooling Problem

Energy and Cost Analysis of Rittal Corporation Liquid Cooled Package

Recommendation For Incorporating Data Center Specific Sustainability Best Practices into EO13514 Implementation

Analysis and Performance Prediction of Data Centre with Different Configurations using CFD

Unique Airflow Visualization Techniques for the Design and Validation of Above-Plenum Data Center CFD Models

Data Centers WHAT S ONTHEHORIZON FOR NR HVAC IN TITLE ? SLIDE 1

Managing Cooling Capacity & Redundancy In Data Centers Today

Data Center Optimization Services. Maximize Your Data Center

Transcription:

CFD SIMULATION ENHANCES A DATA CENTER S EXPANSION PROCESS Gang Tan 1, Kevin S. Venerable 2 1 Department of Civil and Architectural Engineering, University of Wyoming, Laramie, WY 2 IT Division, Data Centers, University of Wyoming, Laramie, WY ABSTRACT Data center s design is required to be flexible and easily upgradeable to satisfy the technology and functional changes. However, the future functional performance of an expansion process can quickly become unmanageable because of the complexity of data centers. A case study for an institutional data center is presented in this paper, including energy consumption benchmark and monitoring of temperature distribution. Computational fluid dynamics (CFD) model is used to simulate the thermal environment in the data center and the results are compared against the measurement data. Two expansion considerations have been investigated by the CFD modeling in order to predict the data center s indoor thermal environment with the expansion and thus the inappropriate proposes can be filtered out. INTRODUCTION As the world s technology shifts the paper-based economy to a digital information managed economy, data centers have become more and more common and essential support for the information and communication for business, communications, academic, and governmental systems. Data centers house large numbers of IT equipment, such as servers, data storage, and network devices to perform various functions including digital data storage, management, processing, and exchange. According to EPA s report of 2007 (EPA 2007), the energy consumed by the nation s servers and data centers is estimated about 61 billion kwh in 2006 (1.5% of total U.S. electricity consumption) for a total electricity cost of about $4.5 billion. This electricity consuming amount in 2006 is approximately more than double of the electricity that was consumed for this purpose in 2000. If based on 2006 s efficiency trends, national energy consumption by servers and data centers could nearly double again in 2011 compared to 2006, and reach more than 100 billion kwh, representing a $7.4 billions annual electricity cost. Therefore, due to the reasons of a) high energy consuming facilities, b) placing tremendous power demands on utilities, and c) increasing demand for computer resources, data centers have increasingly drawn much concern from utilities, building owners, facility managers, and IT equipment manufacturers (Sun et al 2006). Rapid growth of computer resource demand increased energy costs, emissions from electricity generation, strain on the existing power grid, and costs for expansion of data center capacity and construction of new data centers. Because of these reasons, there has been mounting interest in opportunities for energy efficiency in this sector. To capture these energy efficiency opportunities, researchers used multiple methods to investigate the indoor thermal environment in data centers. Airflow field visualization (e.g., smoke appearance and infrared camera) and computer modeling via computational fluid dynamics (CFD) are the two important methods. CFD simulation models have demonstrated their functional applications to predicting the temperature distribution and air movement in a computer room (Abanto et al 2004). In addition, researchers have investigated electrical power requirement within data centers in order to locate performance improving opportunities for air conditioning systems (Karlsson et al 2005). To effectively control thermal environment in datacom facilities, ASHRAE (2004) published guidelines for design procedures and indoor environmental settings for operating environmental parameters such as dry-bulb temperature and relative humidity based, in general, on reliability aspects of the electronic hardware. Besides controlling these environmental parameters, a good air distribution is critical to ensure that an appropriate amount of chilled air is supplied to server racks. Six types of typical air distributions have been studied using measurements and numerical simulation by Cho et.al. (2009). For the air distribution system with separation of cold/hot aisles (all server fronts face cold aisles and server backs face to each other in hot aisles), the computer room air-conditioner (CRAC) unit s supply 506

and return areas are clearly separated, which generally results in minimization of hot air re-circulating back to the server fronts (intakes). The cold/hot aisle arrangement is the most common server rack arrangement style in data centers since the cold/hot aisle configuration will minimize cold air bypass as well as hot air recirculation. The institutional data center of this paper arranges the server racks as cold/hot aisles and is cooled by several computer room air-handler units (CRAHs). To understand energy performance of this insutional data center, power demand of individual server racks has been recorded; the operation of CRAHs has been trended through energy management system (EMS); and cooling energy usage of chilled water from a central energy plant has been monitored. The temperature and humidity distribution in the data center has been measured, which is used to verify the CFD model. The CFD model is thus applied to investigate the data center s two expansion plans in this paper. FACILITY DESCRIPTION The data center is located in an office building for the IT department of University of Wyoming. The IT building is a three-floor building facing north-south direction, which houses the data center for the maority information transformation and numerical calculation for the entire university campus. The IT building also hosts offices and a few information technology related research labs that generally are computer rooms. The data center is in the eastern side of the IT building with one exterior wall facing east. The floor area of the data center is 5,904 square feet, which houses servers, data storage, and other information devices in a total power load of 116.9 kw. About 50% of the data center s space is currently occupied by information devices and the other space is unoccupied. In the next one to two years, more IT equipment will be moved into the data center and most of the space will be taken. Five uniform-size CRAHs use chilled water from the campus central energy plant to providing cooling for the data center. The CRAHs supply fans are equipped with variable speed controllers. Cold air is supplied from CRAHs to the data center space through the raised floor (plenum) and distributed from perforated titles. Cold air rises up because of velocity momentum and buoyancy forces. Air returns to CRAHs through the return air grille installed in the top part of the separation wall between the air-handlers and the data center room where no return air ducts are installed. The data center is a high ceiling space that is about 16 feet from the raised floor to ceiling. To control indoor air humidity, three humidifiers are installed in the data center s perimeter area to inect fog into the room and the fog is generated by pump increased pressure. It is acceptable to approximate the humidification process as an isothermal process. One hundred twenty T-5 fluorescent 4-foot 3-lamp linear fluorescent fixtures provide lighting for the data center space. An energy management system (EMS) is installed to manage the maor environmental control equipment for the data center and a fully-separated control office is next to the data center space. Two 360 kw UPS systems supply power for the data center. Figure 1 below sketches the floor plan of the data center. Tape library CRAH Data storage Perforated tile Server rack Figure 1 Floor plan of the data center ENERGY CONSUMPTION BENCHMARKING Humidifier To benchmark data center s energy performance, Brill (2007) defines three parameters as followings: 1. Data center energy efficiency and productivity (DC-EEP) index 2. IT productivity per embedded watt (IT-PEW) 3. Site infrastructure energy efficiency ratio (SI- EER) According to Brill, the DC-EEP index is the composite result at the data center level of multiplying two independent but interrelated ratios: IT-PEW and SI- EER. The first component is primarily the responsibility of IT, while the second is primarily the responsibility of facilities. The SI-EER ratio is power Grate grill Eastern wall 507

in to the data center as measured at the utility electric meter divided by the conditioned power out to run the IT equipment for computing. The difference between output and input are transformation losses, UPS and cooling equipment inefficiencies, and user operational choices. Brill reports that site operating data from 85 corporate members of their Institute s Site Uptime Network indicates an actual SI-EER of 2.5. The SI- EER ratio can be written as following formula: SI EER = kwh / year ( IT equipment kwh / year ( total ) ) (1) This paper focuses on examining the SI-EER ratio for the institutional data center. A. Instrumentation This study obtained power demand data of individual server racks from PDU built-in power meter, measured airflow rates through sampled perforated tiles using an airflow hood, and monitored temperature and relative humidity distribution using Onset Hobo Loggers. Table 1 provides additional information about the monitoring instruments. In total, 21 Onset Hobo Loggers have been deployed in the data center. Amongst them, 15 loggers were used to record the temperature and humidity distribution for the entire data center space while the other 6 loggers were deployed to a server rack for detailed temperature and humidify profile measurement. Table 1 Instruments for monitoring study SUBJECT INSTRUMENT SPECIFICS Power demand PDU built-in _ of server racks power meter Airflow rate Airdata Accuracy is +/- through Multimeter perforated tiles ADM-880C 3% from 100 to Temperature and RH distribution Chilled water usage Onset Hobo Loggers ONICON SYSTEM-10 BTU Meter 2000 cfm Accuracy ± 0.63 F from 32 to 122 F Differential temperature accuracy ± 0.15 o F To monitor the performance of the CRAHs, trending points have been set up through EMS to track operating conditions such as supply and return air temperatures, return humidity, cooling capacity, fan speed, static pressure difference between air in the plenum and above the raised floor, and total power (kwh) passing the UPS. Figure 2 shows the total chilled water usage of CRAHs for a period of more than 600 hours. From this figure, it can be seen that the cooling energy usage by the data center has fluctuations which may be from the variation of daily computation resource needs, variation of weekly business cycles, and variation of the UPS room s outdoor air demands. In general, computing and data transferring is more intensive in business hours than during night time, as well as higher in work days than in weekend. 50 45 40 35 30 25 Cooling Energy Rate (tons/hr) 20 0 100 200 300 400 500 600 700 Figure 2 Cooling energy rate consumed by CRAHs for the data center and UPS room B. Energy Consumption Benchmarking and Other Monitoring Results Power metering for the individual server racks indicates that IT equipment in the data center has a power factor close to 1.0, generally in the range of 0.97 to 1.0. Based on the power metering, chilled water BTU trending, as well as CRAHs monitoring, the data center is reported an average SI-EER ratio of 2.17 which is slightly better than the 85 corporate data centers SI-EER ratio of 2.5 as described by Brill (2007). Table 2 Approx. benchmark of energy consumption AVERAGE KW 1 IT equipment power 116.9 UPS and other related equipment 2.6 Lighting 13.8 Chillers 97.3 Air and water delivery (e.g., fans 22.3 and pumps) Power out to IT equipment 116.85 Power in to data center 253.32 SI-EER ratio 2.17 1 Power demands of cooling equipment (e.g., CRAHs supply fans and chillers) are based on data from one week s continuous trending. 508

Detailed temperature and relative humidity (RH) profile around a sample server rack has also been measured. The sample server rack is selected as the number 7 one in the F row (F7, see Figure 1). The server rack F7 is not fully loaded as only three IT equipments are installed inside it. Figure 3 illustrates the dimensions of server rack F7. Six loggers have been deployed to the front and back sides of this server rack. The measured temperature and RH profiles are shown in Figure 4 through Figure 7. From temperature profiles, an interesting phenomenon is observed that the logger at the mid elevation of the front side shows the lowest temperature. The reason is probably because this logger at mid-elevation location is not close to any IT equipment and the cold air directly passes this location since cold air from perforated tiles still has relatively strong momentum at the mid-elevation location. Comparing the back side (hot aisle) temperature with the front side (cold aisle) temperature, the warmest temperature in back side is 30 o F higher than the corresponding elevation of the front side. To fully understand this phenomenon, a detailed numerical study may be required to figure out the appropriate level of details in the CFD model when we model server racks. of RH % 63 62 61 60 59 58 57 56 0 500 1000 1500 2000 2500 3000 3500 4000 Point 1 Point 2 Point 3 Figure 4 Temperature profile for the front side of server rack F7 60 55 50 45 40 35 30 0 500 1000 1500 2000 2500 3000 3500 4000 Point 1 Point 2 Point 3 Figure 5 RH profile for the front side of server rack F7 94 of 92 90 88 86 84 82 80 0 500 1000 1500 2000 2500 3000 3500 4000 Point 4 Point 5 Point 6 Figure 6 Temperature profile for the back side of server rack F7 30 24 Figure 3 Sketch of server rack F7 for the detailed temperature and RH monitoring RH % 18 12 6 0 0 500 1000 1500 2000 2500 3000 3500 4000 Point 4 Point 5 Point 6 Figure 7 Relative humidity profile for the back side of server rack F7 509

CFD SIMULATION CFD simulation has been validated as a useful tool to help understand airflow and temperature distribution within a raised-floor data center without significant experimental costs. There are multiple general CFD simulation tools available in academy and in industry, such as PHOENICS (CHAM 2009) and FLUENT (ANSYS 2009), which should have the capability, in theory, to complete thermal environmental simulation for a data center. This paper uses a commercial software to model the data center since the software was purposely developed for data center thermal environment modeling (Karki et al 2003a). Built-in modules particularly for server racks/cabinets, perforated tiles, and CRACs/CRAHs have been designed in the software, which thus saves author s significant time for develping a CFD model. However, there are disadvantages to use commercial softare. For instance, turbulence models within this software are limited and no convenience for users to directly change grid system because the CFD sofware used by this paper is designed with embeded functions for computational mesh and numercial method which are done in background. In the CFD software, the mathematical model for flow in the plenum is formulated from the incompressible form of the Navier-Stokes equations, together with the k-ε turbulence model (Karki et al 2003b). The flow is assumed to be steady and isothermal. The mass and momentum equations can be expressed as: ( ρui ) = 0 i u p ( ρuiu ) = ( µ t ) + S i Here direction, and i i (2) u is the velocity component in the th coordinate component µ t is the turbulent viscosity, p is pressure, S denotes source terms for the velocity u. In the k-ε model, the turbulent viscosity is given by 2 k µ t = cµ ρ ε (3) where c µ is a constant, k is the turbulence kinetic energy, and ε is its rate of dissipation. For a perforated tile, the flow breaks up into a number of small high-velocity ets, which coalesce into a smallvelocity stream. The pressure loss is primarily due to the loss of kinetic energy of the high-velocity ets. An empirical formula for flow resistance characteristic coefficient K for describing the perforated tiles flow characters, supported by a large number of measurements, is calculaed (Idelchik 1994) K 1 0. 375 0.75 = (1 + 0.5(1 + f ) + 1.414 (1 f ) 2 f where f is the fractional open area of the tile. Therefore, the dimensionless K factor is defined as ) (4) p K = (5) 2 0.5ρv where p is the pressure drop across the tile, ρ is the fluid density, and v is the approach velocity. For the above-floor space, both the airflow pattern and temperature distribution needs to be numercially calculated with the combination of the equations governing the conservation of mass, momentum, turbulence parameters, and energy. In the CFD software, the overall solution procedure for the above-floor simulaiton is similar to that used in the under-floor simulation except that it includes the solution of the energy equation due to the exchange from electronic equipment to airflow and temperature stratifications within the space. Using the CFD software, a three-d CFD model has been created that represents the current (existing) condition of the data center as shown in Figure 8. As previously described, this data center deploys a raised floor and uses plenum to supply cooling air to IT equipment. The raised floor consists of 2ft 2 ft floor panels supported on pedestals 2 ft above the floor slab. The CRAH units mounted on the raised floor push cold air into the plenum and pressurize it. CRAH Return air grille Figure 8 The 3-D CFD model for the data center (existing condition) Server rack Perforated tile In the CFD model, data center building envelopes, including the exterior walls, are assumed adiabatic. This 510

assumption is rough but approximately accetable since the cooling load is dominated by internal heat gains for the data center facility. However, the solid walls would impact air movements by applying standard wall functions in the CFD model. The CFD software uses a non-uniform and structured Cartesian grid, with finer spacing in regions of steep gradients, like the inlets of racks. Maor inputs regarding the equipment for the CFD model are summarized in Table 3. Lighting load is not included in the CFD model in order to simplify the calculation since the lighting load doesn t significantly impact airflow pattern in the plenum and above-floor zones where IT equipment locates because all lightings are ceiling-hanging fixtures while the chilled air is supplied from under-floor plenum. The first evaluation to the CFD model is to check the calculated airflow rates passing through 30 sampled perforated tiles against the measured airflow rates using a flow hood. Figure 9 shows this comparison. There is a good agreement between the CFD model result and the measured airflow rates. Especially, the overall airflow rate for these 30 sampled perforated tiles is 16,855 cfm from measurement and the simulation result shows a total of 16,994 cfm, with a small difference of 0.8%. Therefore, the CFD model can be used to model the thermal environment for the above raised-floor space. With respect to the temperature distribution, Figure 10 gives the measured temperature and modeled temperature for 15 sampled locations. Most of the modeled temperatures agrees with the measured ones, especially for locations in front of server racks. However, temperatures of locations at the back side of server racks, some discrepancies have been observed. The four blue dots in the red circles are the CFD modeled back side temperatures that are noticeably higher than the measured temperatures. This is because the CFD model treats all server racks as a single heat resource for a rack, even if the rack is loaded with separate, dispersed IT equipment. A uniform model is used to account for the impact of the IT equipment s built-in machine exhaust fans. For a partially loaded rack, cold air may pass through rack and directly mix with warm air behind the rack. This effect is not addressed in the CFD model. From this point, a detailed CFD model shall be necessary for modeling the server racks. Although the CFD model has limitations when modeling the backside temperature distribution of server racks, the CFD simulation model can be applied to investigating two expansion plans for this data center since the front side air inlet temperature is more critical in order to guarantee safe operation of the IT equipment. Table 3 Maor inputs for the CFD model EQUIP DETAILS COMMENTS IT equipment Server cracks Power demand from individual metering Tape library Power demand varies from 0.84 to 1.83 kw Data storage (mainly IBM machines) Power demand is about 4.68 kw Perforated GrateAire perforated panel 56% Damper K=2.97 tiles open plus 90% open damper GrateAire perforated panel 56% open plus 60% open damper Damper K=7.54 GrateAire perforated panel 56% Damper K=17.57 open plus 40% open damper CRAHs Stulz CCD 2300 model with full ventilation rate of 18,000 cfm. Supply fan runs at 65% speed Return air grille Vent attached with grille: normal condition has 50% net opening area Bottom half is partially closed. cfm 1000 800 600 400 200 0 0 5 10 15 20 25 30 Measured Airflow Tiles Modeled Airflow Figure 9 Comparison of airflow rates passing through the sampled perforated tiles between measurement and CFD modeling of 95 90 85 80 75 70 65 60 55 50 0 2 4 6 8 10 12 14 16 Modeled Temperature Points Measured Temperature Figure 10 Comparison of temperature distribution at 15 sampled points between measurement and CFD modeling 511

A. Add Two 10kW Server Racks The first expansion plan proposes adding two much more power-intensive servers to the data center. The geographic department wants to put their new IT equipment into two server racks, with 10kW power load each. Crresponding changes have been made to the CFD model and the plan is to place these two server racks to southern side space of the data center. At southern side, supply air velocity in the plenum is very small and air static pressure reaches its maximum value in plenum. The perforated tiles at the front side of these two server cracks are set to fully open. Therefore, cold air supplied to the server racks will reach the maximum capacity if the CRAHs increase fan speed according to the changed total IT equipment connection load in the data center. However, the CFD simulation explains that 10kW power demand server racks will not work safely for this data center if the new racks would be placed at southern side. The inlet air temperature of these two racks rises up to 71.7 o F. Per the IT equipment manufacturer s specification and data center staff s preference, the server racks inlet air temperature should not exceed 70 o F in order to guarantee their appropriate heat dissipation. Meanwhile, air temperature at back side of these two server racks goes up to 97.8 o F as shown in Figure 11, which is too warm even for the hot aisle. There are two possible solutions to this problem: a) breaking the two 10kW loads into three, or b) reducing the warm air plume that generates by the server rack in front of the proposed new 10kW server racks. The warm air plume impact can be observed in Figure 11 that is highlightened by a red circle. The warm air plume is generated by the re-circulating warm air from back side of server racks in front of the two proposed 10 kw racks (see Figure 12 for the red-circle zone). This is partly because southern side space of the data center is far from the return air grille. To completely get rid of the re-circulating warm air, return air duct network might be necessary. However, this will bring in additional cost and interrupt data center s operation due to construction. Breaking loads into three racks is an easier solution that can also reduce the hot ailse s (at back side) temperature. Figure 11 Temperature contour in the CFD model for the two 10kW server racks case Figure 12 Airflow streamlines in the CFD model for the two 10kW server cracks case B. Add One Row of Energy Intensive Server Racks The data center plans to add one row of server racks to the row closet to the eastern wall. The question rising here is to understand what the maximum power load of each rack on average can serve without significant changes to the current air-conditioning system. This paper has examined several scenarios by changing server racks power loads and adusting the perforated tiles attached dampers net opening areas as well as CRAHs fan speed. The results are summarized in Table 4. CFD modeling shows the maximum server rack power load is about 9kW on average. This result is consistent with the findings for the two 10kW server racks case. It is wise to keep all server racks no more than 9kW power load in this data center. 512

Table 4 Scenarios of CFD modeling for adding another row of server racks next to the eastern wall SCEN. POWER (KW) PERFORATED TILES DAMPER MAX RACK INLET AIR TEMPERATURE 1 3 kw 60% open 70.1 o F 2 6 kw 90% open 67.2 o F 3 9 kw 100% open 70.4 o F DISCUSSIONS This paper presents the preliminary results of a case study that benchmarks the energy performance of an institutional data center and the CFD modeling is used to enhance the expansion consideration when extra IT equipment needs to be added to the existing data center. CFD simulation is a helpful tool for designers and operators to understand the thermal environment of data centers. CFD simulation provides detailed information about the thermal environment of data centers in a timeand cost-effective way. In this preliminary study, the performance of cold/hot aisle arragnement of server racks with under-floor air supply using perforated tiles for an institutional data center has been investigated. Effectively organizing the airflow distribution is essential to ensure that the indoor thermal environment is controlled correctlly and energy efficiently. From this study, we found a detailed model to the server rack is necessary if the temperature profile at the back side of (i.e., hot aisle) the server racks is critical for designing and/or operating a data center. Information such as the exact location of the IT equipment in the server rack and IT equipment s exhaust fan s impact should be included into the detailed CFD simulation. Further study about this topic will be carried out. When designing a data center with large space, a ducted return air path maybe not a bad idea to prevent warm air from re-circulating in locals. Without return air ducts, a good design wants warm air directly return back to CRAHs driven by buoyancy force and static pressure difference. However, it is difficult to avoid return air s recirculation within a large space or for a complex space configration. For example, in this institutional data center, the southern side space experiences locally re-circulating warm air from hot aisle which increases the inlet air temperature of neighbor racks. ANSYS, 2009, http://www.ansys.com/assets/brochures/ansysfluent-12.1.pdf ASHRAE TC9.9, 2004, Thermal Guidelines for Data Processing Environments, ISBN 1-931862-43-5. Brill K.G., 2007, Data Center Energy Efficiency and Productivity, Site Infrastructure White Paper. CHAM, 2009, An introduction to CHAM, http://www.cham.co.uk/docs/cham_intro_talk.pdf Cho J., Lim T., Kim B.S., 2009, Measurements and predictions of the air distribution systems in high compute density (Internet) data centers, Energy and Buildings, v41, pp 1107-1115. EPA, 2007, ENERGY STAR Program, Report to Congress on Server and Data Center Energy Efficiency Public Law 109-431. Idelchik I.E., 1994, Handbook of Hydraulic Resistance, CRC Press, Florida. Karki K., Karki S., Radmehr A., 2003a, Techniques for controlling airflow distribution in raised-floor data centers, proceedings of the Pacific Rim/ASME International Electronic Packaging Technical Conference and Exhibition, July 6-11, 2003, Maui, Hawaii, USA. Karki K., Radmehr A., Patankar S., 2003b, Use of computational fluid dynamics for calculating flow rates through perforated tiles in raised-floor data centers, International Journal of Heating, Ventilation, Air-Conditioning, and Refrigeration Research, v 9(2), pp. 153-166. Karlsson J.F., Moshfegh B., 2005, Investigation of indoor climate and power usage in a data center, Energy and Buildings, v37, pp1075 1083. Sun H.S., Lee S.E., 2006, Case study of data centers energy performance, Energy and Buildings, v38, pp522-533. REFERENCES: Abanto J., Barrero D., Reggio M., Ozell B., 2004, Airflow modeling in a computer room, Building and Environment, v39, pp 1393-1402. 513