1 WHITE PAPER #27 IMPACT OF VIRTUALIZATION ON DATA CENTER PHYSICAL INFRASTRUCTURE EDITOR: DENNIS BOULEY, APC BY SCHNEIDER ELECTRIC
2 PAGE 2 EXECUTIVE SUMMARY Data center virtualization, which includes storage, desktop, and server virtualization, reduces overall IT equipment electrical load through consolidation of systems. The resulting energy savings can be further maximized if IT or facilities managers adjust the power and cooling infrastructure to accommodate the reduced loads. Planning for this cycle of initial reduced load followed by load growth with IT equipment running at much higher overall utilization levels can result in the capture of a signifi cant, supplemental energy savings entitlement. This paper provides guidance on how data center professionals can maximize power and cooling equipment electrical savings in a virtualized, consolidated data center environment. INTRODUCTION A primary motivation for consolidation using virtualization is increased computing effi ciency more computing per watt of power consumed by the data center. The reduced power consumed by the consolidated IT load itself, however, is only the beginning of the savings entitlement that can be claimed when consolidating workloads to a smaller set of physical equipment. A majority of existing data centers can accrue signifi cant additional savings by rightsizing the power and cooling systems (e.g., UPS, CRAC, chillers) that support the new, now smaller, virtualized/consolidated IT load. In addition, the newly virtualized/consolidated data center will continue to grow, so the scalability of the rightsized power and cooling equipment should also be considered. This paper complements Green Grid White Paper #19, Using Virtualization to Improve Data Center Effi ciency, which addresses the IT side of virtualization and the associated effi ciency impacts. THE IMPACT OF FIXED AND PROPORTIONAL LOSS There is an IT load value at which an existing data center is tuned. Going below this point because of reductions in IT load reduces IT energy demands. However, users must be cautious that these gains are not undermined by effi ciency losses in the power and cooling subsystems. The bulk of additional savings comes from the effi ciency of a reconfi gured, streamlined power and cooling architecture that aligns more tightly with virtualization s reduced and dynamically varying demand. Delivering the right amount of conditioned power to IT equipment and running just enough cooling equipment to remove the heat generated in the data center is key to energy effi ciency. Running more power and cooling equipment than needed leads to ineffi ciency. The devices that comprise the data center physical infrastructure (e.g., power and cooling equipment) are numerous and wide ranging. Each device contains a set of internal subsystems with specifi c power requirements. These internal subsystems are characterized by a wide range of power consumption (power loss) vis-a-vis load characteristics. For some devices this power consumption behavior can be complex or even advanced. For other devices the behavior is simple and conventional (easy to understand or characterize). Regardless of the level of complexity, the power consumed by a device in the infrastructure does fall into the two easy to understand power loss categories: fi xed and proportional loss (see Figure 1). The amount of power required to keep a particular device energized and in a ready state is sometimes referred to as no load watts. This is defi ned as the measurable power at the input of the device when under no load conditions. All of the device s no load watts are usually consumed by internal functions like housekeeping / auxiliary power supplies, device communication interfaces (including front panel LCD and LEDs), cooling fans, and device logic and mode control functions (CPU s, digital signal processors, relays or contactors). None of this power is delivered to the data center IT load.
3 PAGE 3 Figure 1 Physical infrastructure loss characteristics Electrical Loss in kw (Waste due to inefficiency) PROPORTIONAL loss NO-LOAD loss SQUARE-LAW loss No Load 10% 30% 50% 70% 90% Full Load Equipment Loading Most internal functions are needed over the full load range for the device: for instance, logic and mode control functions are equally important and needed at both light load and heavy load. Many of these functions demand the same amount of power regardless of the device s output load state. The losses associated with these kinds of internal functions, being present and constant with load, are categorized as fi xed loss or noload loss. Most devices in the data center infrastructure will have some characteristic fi xed loss. Proportional losses vary with load. In general, proportional losses increase with load. Proportional losses fall into two subcategories: switching and conduction losses. When an infrastructure device assumes or carries additional load, its internal subsystems must process the load in the form of distributed currents and voltages. In other words, a larger amount of power must be processed by various components in its power path. For example, proportional losses occur from switching losses from transistors, as well as resistance losses from capacitors and inductors. It should be mentioned that similar losses occur in other internal component types. For example, magnetic devices (e.g. transformers) have their own category of conduction and switching losses in the form of copper and core losses. Internal conductors such as power wires, bus bars, contactors and circuit breakers produce losses in the form of conduction losses that are proportional to the square of the load (square law loss). Understanding the loss characteristics of the physical infrastructure is important to assessing the impact of consolidation resulting from virtualization. If an IT department has reduced the number of physical servers by 50% without making proper adjustments to the power and cooling equipment, then the overall data center effi ciency will suffer. Figure 2 provides a simple example of how fi xed losses play a role. Let s assume a data center with only two electrical loads; 100 kilowatts (kw) of physical servers and 100 kw of air conditioning fans. Assume that through virtualization the server load is reduced to 50 kw (a 50% reduction in server power). This 50% reduction translates into only 25% savings in the electric bill. Why is that? In this example, and in most existing data centers, perimeter-based air conditioners use constant-speed fans which means that the fans consume the same amount of power regardless of the IT load (fi xed loss). The two pie charts illustrate this concept by demonstrating that the fan power represents a larger proportion of the electrical bill after virtualizing. Therefore, even though the electric bill decreased (smaller pie chart) it didn t decrease by the same amount as the server power.
4 PAGE 4 Figure 2 Electricity Savings vs. Effi ciency Gains (Fans & Servers only) ENERGY EFFICIENCY METRICS The overall concept of data center effi ciency should be understood in order to better grasp the impact of physical infrastructure fi xed losses. Figure 3 illustrates a typical data center energy effi ciency curve. The effi ciency of the data center increases as the percentage of IT load increases. The effi ciency metric, known as Data Center Infrastructure Effi ciency, or DCiE (also identifi ed by its inverse known as Power Usage Effectiveness or PUE), is utilized to indicate the overall effi ciency of a data center. Figure 3 indicates that at 30% load, this data center is only about 45% effi cient. At 90% load, the data center is approximately 62% effi cient. Therefore, a data center that has just been virtualized, in all likelihood, will have reduced its overall load thereby reducing the effi ciency of its existing power and cooling infrastructure. However, this potential loss in effi ciency can be addressed by proper planning. A strategy for addressing this loss of effi ciency will be discussed later in this paper. Effi ciency measurements are important in that improvements to data center effi ciency can only be measured and documented if the proper metrics are utilized. For more information on data center energy effi ciency metrics, see The Green Grid white paper #14, The Green Grid Metrics: Data Center Infrastructure Effi ciency (DCiE) Detailed Analysis. OVERSIZING, VIRTUALIZATION AND EFFICIENCY The traditional approach to provisioning a data center often embraced the concept of oversizing. When energy costs were low and supply was abundant, oversizing of physical infrastructure equipment was a commonplace strategy for accommodating future growth. A typical UPS, for example would only be operating at 30-40% of full load. In an environment where energy costs are high and virtualization is an accepted best practice, this approach has become less sustainable. Since virtualization, and the resulting equipment consolidation, can signifi cantly reduce load, oversizing is an important effi ciency issue in a virtualized data center. Even without virtualization, oversizing has long been a primary contributor to data center ineffi ciency. Server consolidation and server power management, by reducing power and cooling loads even more, will shift effi ciency further toward the low end of the effi ciency curve (Figure 3), if power and cooling systems are not adjusted accordingly. While the electric bill will indeed go down because of the lower IT load and less air conditioning needed to cool it, the proportion of utility
5 PAGE 5 power that reaches the IT loads in other words, effi ciency as measured by The Green Grid s DCiE metric will drop, which signifi es additional power could be conserved to further reduce energy consumption. Over-provisioning power and cooling equipment is endemic to existing data centers and, although this allows room for growth, it also represents a signifi cant waste of energy. Changes in IT capacity, either up or down, also impact the effi ciency of physical infrastructure supporting IT. Figure 3 Typical infrastructure effi ciency curve, using DCiE as the metric DCiE 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% % IT Load When implementing virtualization/consolidation in an existing data center, choosing a methodology for distributing high and low density IT equipment is the single most important decision. Approaches such as spreading the heat load, adding supplemental cooling, or designating high and low density areas within the data center can help to further increase the effi ciency benefi ts of virtualization. ELEMENTS OF A VIRTUALIZATION DENSITY STRATEGY As discussed, when data centers virtualize and consolidate, overall IT power consumption goes down. This means that the remaining servers will be run at higher utilization rates, consuming more power per square foot of data center space (more kw per rack). In most cases this increases the power/cooling density of the remaining IT equipment (although, in some cases, enough unutilized white space is created and some cooling units can be eliminated). The resulting higher rack density introduces a new set of challenges. The existing cooling infrastructure will have to be examined to see if it can still effi ciently cool the load. Several approaches can be utilized when formulating a cooling strategy for the higher density. It may be that one of these strategies is best for a particular data center but in most cases a data center uses a combination of these strategies as it evolves over time. Key strategies include the following:
6 Spread the load Supplemental cooling Designation of high density and low density areas within the data center PAGE 6 SPREADING THE LOAD Spreading high density servers enables the data center infrastructure to power and cool to an average value below the peak potential enclosure value, and spreads out the load of any proposed enclosures whose load exceeds the design average value by splitting the equipment among multiple rack enclosures. Spreading the load is the most popular solution for incorporating high density equipment into today s data centers. This practice allows for a granular selection of IT equipment to fi t the existing capabilities of the data center. Fortunately, high density servers such as blade servers do not need to be installed closely spaced in the same enclosure, and can be spread out across multiple racks (see Figure 4). By splitting equipment among racks, no rack need ever exceed the design power density and consequently cooling performance is predictable. The main benefi t to this strategy is that no new power and cooling equipment is required since IT equipment was added into a physical location from which it could be powered and cooled. However, the spread the load strategy may not be viable for a number of reasons: Consumption of additional fl oor space, which may be diffi cult to justify or simply not possible Executive management perception that partially-fi lled racks are wasteful Increased cabling costs (because of longer runs) and higher infrared optical losses Increased cost and diffi culty of maintenance cabling and mounting may be intertwined with other equipment in a nonstandard manner, scattered throughout the room If the data center still relies on the pre-virtualization cooling architecture, improvement in post-virtualization DCiE/PUE may be limited Host servers inherit the same cooling availability as other less critical servers Equipment must be located in the same power, cooling and physical security redundancy domain Figure 4 Spreading the Load Spreading the load may be a good choice as a density strategy for several reasons: When the resulting average data center density (kw / rack or watts / square foot of white space) is at about the same as it was before virtualization or below 3 kw. This assumes that the cooling system was able to effectively cool the data center loads before virtualization.
7 PAGE 7 When the IT managers have complete control over where servers are placed. The freedom to place servers makes this density strategy very applicable. When U-space is available in the majority of racks. The spread the load strategy leaves a sizable amount of unused vertical space within the racks. This space must be fi lled to prevent degradation of cooling performance. Therefore blanking panels should be deployed since they prevent hot and cold air mixing which degrades the cooling effectiveness at the front of the IT equipment. SUPPLEMENTAL COOLING Supplemental cooling can be utilized to cool racks with a density greater than the design average value. This strategy may be chosen for several reasons: When the resulting post-virtualized average data center density is between 3 and 5 kw / rack When available cooling dictates where IT managers have to place IT equipment. This often requires that supplemental cooling be placed in an area with a concentration of high density servers. The supplemental cooling remains in place until the available cooling is consumed. Then a more permanent solution must be deployed. When a capital budget is limited and short-term cooling issues must be dealt with at relatively low expense. This approach requires that the installation be planned in advance. When a room has been planned with supplemental cooling in mind, a variety of techniques can be used. These include: Installation of fans to boost the cool air supply from the CRAC to an enclosure Retrofi t or installation of electronically commutated (EC) fans (capable of automatically changing speeds and adapting to temperature differences) and/or variable speed drives (VFD) Installation of specialty return ducts or fans to scavenge the hot exhaust air from an enclosure for return to the CRAC Installation of special racks (such as chimney-based racks) or rack mounted or ceiling mounted cooling devices with the capability of providing the required cooling directly at the rack Installation of row-based air conditioners between two racks to address localized hot spots A rear-door heat exchanger can also relieve the existing cooling infrastructure from added heat load. This device is attached to the rack and effi ciently transfers heat to the same cooling system utilized by the existing CRAC units. This shortens the air path (from the inside of the rack to the back of the rack) and thereby increases effi ciency. DESIGNATION OF HIGH AND LOW DENSITY AREAS WITHIN THE DATA CENTER Consolidation of high density servers and low density servers into designated areas within the data center is also an approach that can address the issue of higher densities resulting from virtualization and consolidation. This can be facilitated by deploying solutions that can address high density or low density designated areas on a rack-by-rack, row-by-row, or pod basis (see Figure 5). Segregated areas for high densities leaves the rest of the data center with the capability to power and cool to an average value below the peak enclosure value and can also ease management of the overall data center.
8 PAGE 8 This approach requires advanced planning, knowledge of the percentage of high density servers, and the ability to segregate those enclosures hosting high density servers into a special area. This strategy provides optimal space utilization and delivers predictable power and cooling density to that area. A major advantage to this strategy is that critical host servers are co-located in an area with a higher targeted availability (i.e. power redundancy, predictable redundant cooling, and physical security) and a higher cooling effi ciency. This lowers costs because the entire data center does not need to be overhauled. Figure 5 Segregation options for higher density racks SEGREGATION BY INDIVIDUAL RACK In a rack-by-rack oriented segregation approach, the CRAC units are associated with a rack and are assumed to be dedicated to that rack only. The CRAC units are directly mounted to or within the IT racks. Compared with the room-oriented or row-oriented approach, the rack-oriented airfl ow paths are even shorter and exactly defi ned, so that airfl ows are totally immune to any installation variation or room constraints. All of the rated capacity of the CRAC can be utilized, and the highest power density (up to 50 kw per rack) can be achieved. SEGREGATION BY INDIVIDUAL ROW In row oriented segregation, the CRAC units are associated with a row and are dedicated to that row. The CRAC units may be mounted among the IT racks, they may be mounted overhead, or they may be mounted under the fl oor. The row-oriented segregation approach allows one row of racks to run high density applications such as blade server, while another row satisfi es lower power density applications such as communication enclosures. SEGREGATION BY INDIVIDUAL POD A pod consists of two rows of racks that are interconnected and sealed at each end of the row and by a light weight ceiling panel. The pod ensures proper cold air distribution to servers by completely segregating cool supply and hot return air paths. High density pods may be chosen as a segregation solution for the following reasons: When the resulting post-virtualized average data center density (kw / rack or watts / square foot of white space) is greater than the average of your data center (typically 5 kw).
9 PAGE 9 When a section of the data center is available where high density racks can be located When IT managers that have complete control over where high density servers are placed When increased data center infrastructure effi ciency (improved DCiE/PUE) is a goal. When only a portion of the data center needs the higher availability offered by the high powered pod. The key operating principle of a hot aisle containment pod, for example, is the capture of all of the hot IT equipment exhaust air by enclosing the hot aisle, and then immediately cooling this air using a row-based air conditioner. The capture of the hot air combined with the short air path allows very high density cooling and also causes the system to exhibit high effi ciency (i.e. less energy needed to move the hot air). The performance of this system is completely independent of the rest of the data center (even in locations where only comfort air is available) and can be installed without a raised fl oor. When a requirement exists to keep the high density racks together, special high density power and cooling solutions are preferred. All other feasible solutions require some degree of spreading out of the high density equipment. A similar solution to hot aisle containment is the cold aisle containment. Cold aisle containment is typically deployed in traditional perimeter-based cooling environments. Traditional cooling environments use the entire room as a hot air return plenum and deliver cold air via the raised fl oor plenum to the cold aisles. Cold aisle containment encloses the cold aisle allowing the rest of the data center to become a large hot air return plenum. By containing the cold aisle, the hot / cold air streams within the data center are separated. OTHER CONSIDERATIONS The distribution of power to higher density racks will also mean more cables to manage in concentrated areas. As IT equipment is constantly refreshed in the same sets of racks, cables get swapped and sometimes abandoned. This increases the likelihood that improper cables are pulled by mistake and unintentionally disconnected thereby increasing downtime risk. The cables may not be organized, and racks may not be deep/ wide enough to handle the increased cable volume. It becomes diffi cult to keep adding new high density servers to racks that were previously designed to accommodate a lower density. Other physical constraints, such as the increased weight of high density racks, also need to be considered and accounted for. Is the raised fl oor designed to support the greater weights per rack? The existing enclosures need to support the increased physical load of the IT gear. The front and rear doors must accommodate the airfl ow required by the high density IT gear. The enclosure must accommodate space for proper cable routing in order to ensure proper cooling and reduce the potential for human error. An early plan must be formulated for how increased densities will impact the power infrastructure. For example, the addition of higher density equipment to a traditionally low density area can lead to insuffi cient breaker positions, mislabeled or missing labels on positions, and cascading panels (the daisy chaining of breaker panels). When adding new voltage levels and receptacles, staff will sometimes resort to risky hot work that can create safety and downtime issues. These situations can be avoided when a sound density strategy is put into practice. CHARACTERISTICS OF DYNAMIC LOADS While virtualization may reduce overall power consumption in the room, virtualized servers tend to be installed, operated, and grouped in ways that create localized high-density areas, or hot spots. These new densities can be an order of magnitude greater than the densities in a typical unvirtualized data center. Not
10 PAGE 10 only are densities increasing, but virtualization also means each server is running at higher utilization and power consumption levels. In addition, applications can be dynamically started, stopped and moved repeatedly the result can be loads that change both over time AND in their physical location in the room. Before virtualization brought dynamic allocation of server loads, localized high-density hot spots typically stayed associated with a physical server. If perimeter cooling and air distribution could be confi gured to adequately cool the hot spots, it would stay that way until physical servers were added, removed, or relocated. With dynamic loading, however, the load profi le of the room can shift, unseen, with no visible physical changes in equipment. Predictable and effi cient cooling calls for a system that comprehends these variations and automatically matches cooling both in location and in amount to changing power densities. The key characteristics of such a cooling system are: Short air path between cooling and load Dynamic response to power and cooling load changes Compared to traditional room-based perimeter cooling, the air paths in row-based cooling and rack-based cooling dramatically lessen mixing of supply and return air and, with containment and blanking panels, virtually eliminate mixing. On the supply side, this allows operation at a higher coil temperature, which takes less chiller energy to maintain and is much less likely to cause wasteful condensation. On the return side, it produces a higher return temperature which increases the heat removal rate. Compared to long-path cooling systems, these combined short-path effects (1) increase operational cooling effi ciency and (2) increase the cooling capacity of the heat rejection system. Cooling units located within the rows or individual racks, and instrumented to sense and respond to temperature changes, meet the above two essential criteria and are an effective and effi cient way to achieve predictable cooling of virtualized loads. Over-cooling the entire room by lowering the supply temperature of perimeter CRACs creates signifi cant ineffi ciencies and will likely still not provide adequate cooling at the highdensity locations. CONCLUSION Consolidation using virtualization is here to stay. It works it saves money by increasing computing throughput per watt of power consumed, freeing up fl oor space, providing head room for growth, facilitating load migration, and disaster recover. The virtualization savings entitlement can be multiplied if power and cooling infrastructure is modifi ed or upgraded to take advantage of the savings opportunity that virtualization offers. Given the impact of virtualization on data center power densities, three principle approaches should be considered to help manage the heat produced by higher densities: spreading the load, supplemental cooling, and designation of high density and low density areas within the data center. As each data center virtualization outcome is unique, care should be taken to examine the impact on power and cooling. The approaches and options discussed in this paper should be examined with an eye towards maximizing the overall effi ciency of the virtualized data center.
Common $en e: Solutions to Make Your Data Center Energy Effi cient Energy effi ciency is a life style. It s also common sense. In a data center, the name of the game is effi ciency and how many compute
Power and Cooling for Ultra-High Density Racks and Blade Servers White Paper #46 Introduction The Problem Average rack in a typical data center is under 2 kw Dense deployment of blade servers (10-20 kw
Virtualization: Optimized Power and Cooling to Maximize Benefits White Paper 118 Revision 2 by Suzanne Niles Click on a section to jump to it > Executive summary Data centers are routinely and unknowingly
Education Evolution: Scalable Server Rooms George Lantouris Client Relationship Manager (Education) May 2009 Agenda Overview - Network Critical Physical Infrastructure Cooling issues in the Server Room
Creating from in Data Centers and Server Rooms By Dennis Bouley White Paper #119 Executive Summary Data center professionals can rid themselves of messy racks, sub-standard under floor air distribution,
Reducing Data Center Energy Consumption By John Judge, Member ASHRAE; Jack Pouchet, Anand Ekbote, and Sachin Dixit Rising data center energy consumption and increasing energy costs have combined to elevate
Data Center Technology: Physical Infrastructure IT Trends Affecting New Technologies and Energy Efficiency Imperatives in the Data Center Hisham Elzahhar Regional Enterprise & System Manager, Schneider
Power and Cooling Capacity for Data Centers By Neil Rasmussen White Paper #150 Executive Summary High density IT equipment stresses the power density capability of modern data centers. Installation and
#112 Best Practices for Deploying the InfraStruXure InRow SC By David Roden Abstract The InfraStruXure InRow SC (ACSC100 and ACSC101) is a self-contained air conditioner for server rooms and wiring closets.
Electrical Efficiency Modeling for Data Centers By Neil Rasmussen White Paper #113 Revision 1 Executive Summary Conventional models for estimating electrical efficiency of data centers are grossly inaccurate
Dealing with Thermal Issues in Data Center Universal Aisle Containment Daniele Tordin BICSI RCDD Technical System Engineer - Panduit Europe Daniele.Tordin@Panduit.com AGENDA Business Drivers Challenges
Strategies for Deploying Blade Servers in Existing Data Centers By Neil Rasmussen White Paper #125 Revision 1 Executive Summary When blade servers are densely packed, they can exceed the power and cooling
Data Center White Paper Rack Hygiene April 14, 2011 Ed Eacueo Data Center Manager Executive Summary This paper describes the concept of Rack Hygiene, which positions the rack as an airflow management device,
The Advantages of Row and Rack- Oriented Cooling Architectures for Data Centers By Kevin Dunlap Neil Rasmussen White Paper #130 Executive Summary Room cooling is an ineffective approach for next-generation
Unified Physical Infrastructure (UPI) Strategies for Thermal Management The Importance of Air Sealing Grommets to Improving Smart www.panduit.com WP-04 August 2008 Introduction One of the core issues affecting
Raised Floor Data Centers Prepared for the Future Raised Floors Protect My Investment by Providing a Foundation for Current and Future Technologies. Creating the Future Proof Data Center Environment The
POWER USAGE EFFECTIVENESS Developing Best Practices Manual for Indian Data Centers PUE - Power usage effectiveness is Total incoming power( IT equipment + electrical And mechanical support system ) / Presentation
WHITE PAPER #6 GREEN GRID DATA CENTER POWER EFFICIENCY METRICS: PUE AND DCIE EDITORS: CHRISTIAN BELADY, MICROSOFT CONTRIBUTORS: ANDY RAWSON, AMD JOHN PFLEUGER, DELL TAHIR CADER, SPRAYCOOL PAGE 2 ABSTRACT
Guidelines for Energy-Efficient Datacenters february 16, 2007 white paper 1 Abstract In this paper, The Green Grid provides a framework for improving the energy efficiency of both new and existing datacenters.
An Improved Architecture for High-Efficiency, High-Density Data Centers By Neil Rasmussen White Paper #126 Executive Summary Data center power and cooling infrastructure worldwide wastes more than 60,000,000
WHITE PAPER #22 USAGE AND PUBLIC REPORTING GUIDELINES FOR THE GREEN GRID S INFRASTRUCTURE METRICS (PUE/DCIE) EDITORS: JON HAAS, INTEL JAMIE FROEDGE, EMERSON NETWORK POWER CONTRIBUTORS: JOHN PFLUEGER, DELL
Avoidable Mistakes that Compromise Cooling Performance in Data Centers and Network Rooms By Neil Rasmussen White Paper #49 Executive Summary Avoidable mistakes that are routinely made when installing cooling
Re Engineering to a "Green" Data Center, with Measurable ROI Alan Mamane CEO and Founder Agenda Data Center Energy Trends Benchmarking Efficiency Systematic Approach to Improve Energy Efficiency Best Practices
How Row-based Data Center Cooling Works White Paper 208 Revision 0 by Paul Lin and Victor Avelar Executive summary Row-based data center cooling is normally regarded as a cold air supply architecture that
Intelligent Containment Systems COOL Maximize cooling efficiency and easily maintain a perfectly controlled temperature. Air Management Systems Opengate solutions offer sophisticated airflow management
DATA CENTER RESOURCES WHITE PAPER IMPROVING DATA CENTER EFFICIENCY AND CAPACITY WITH AISLE CONTAINMENT BY: STEVE HAMBRUCH EXECUTIVE SUMMARY Data centers have experienced explosive growth in the last decade.
#92 Best Practices for Designing Data Centers with the InfraStruXure InRow RC By John Niemann Abstract The InfraStruXure InRow RC is designed to provide cooling at the row and rack level of a data center
A White Paper from the Experts in Business-Critical Continuity TM Data Center Cooling Assessments What They Can Do for You Executive Summary Managing data centers and IT facilities is becoming increasingly
Cooling Small Server Rooms Can Be Inexpensive, Efficient and Easy - Jim Magallanes Computer Room Uptime: www.cruptime.com Uptime Racks: www.uptimeracks.com Server Rooms Description & Heat Problem Trends
Convergence of Complexity Optimize Existing Data Center Space Measure And Allocate Energy Cost Plan for new data centers or additional capacity How to make your Data Center Last Longer, Gartner, December
Airflow Simulation Solves Data Centre Cooling Problem The owner s initial design for a data centre in China utilized 40 equipment racks filled with blade servers spread out in three rows along the length
Green Data Centre Design A Holistic Approach Stantec Consulting Ltd. Aleks Milojkovic, P.Eng., RCDD, LEED AP Tommy Chiu, EIT, RCDD, LEED AP STANDARDS ENERGY EQUIPMENT MATERIALS EXAMPLES CONCLUSION STANDARDS
GREEN IT THE GREEN DATA CENTER WHERE ECOLOGY MEETS ECONOMY We truly live in an information age. Data Centers serve a very important purpose they provide the global community with nearly unlimited access
Enabling an agile Data Centre in a (Fr)agile market Phil Dodsworth Director, Data Centre Solutions 2008 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without
Data Centers Mapping Cisco Nexus, Catalyst, and MDS Logical Architectures into PANDUIT Physical Layer Infrastructure Solutions 1 Introduction The growing speed and footprint of data centers is challenging
Ten Steps to Solving Cooling Problems Caused by High- Density Server Deployment By Peter Hannaford White Paper #42 Revision 1 Executive Summary High-density servers present a significant cooling challenge.
Ee The Efficient Enterprise Juan Carlos Londoño Data Center Projects Engineer APC by Schneider Electric Keystrokes Kilowatts Heat OUT Electricity IN Need for bandwidth exploding Going hyperbolic! 30,000
Classification of Data Center Operations Technology (OT) Management Tools White Paper by Kevin Brown and Dennis Bouley Executive summary Data centers today lack a formal system for classifying software
Cooling Solutions and Selling Strategies for Wiring Closets and Small IT Rooms Agenda Review of cooling challenge and strategies Solutions to deal with wiring closet cooling Opportunity and value Power
Management Tools for Data Center Physical Infrastructure Jack Hack Data Center Solutions BeLux Software Consultant Benelux APC by Schneider Electric X-ICT Xperience 2010 15 April Management Tools for Data
Guidelines for Specification of Data Center Power Density White Paper 120 Revision 1 by Neil Rasmussen > Executive summary Conventional methods for specifying data center density are ambiguous and misleading.
RiMatrix S Make easy. 2 RiMatrix S Rittal The System. The whole is more than the sum of its parts. The same is true of Rittal The System. With this in mind, we have bundled our innovative enclosure, power
Managing Data Centre Heat Issues Victor Banuelos Field Applications Engineer Chatsworth Products, Inc. 2010 Managing Data Centre Heat Issues Thermal trends in the data centre Hot Aisle / Cold Aisle design
PART 1 - GENERAL 1.01 OVERVIEW A. This section supplements Design Guideline Element D3041 on air handling distribution with specific criteria for projects involving design of a Data Center spaces B. Refer
Data Center Power Consumption A new look at a growing problem Fact - Data center power density up 10x in the last 10 years 2.1 kw/rack (1992); 14 kw/rack (2007) Racks are not fully populated due to power/cooling
Guidelines for Specification of Data Center Power Density By Neil Rasmussen White Paper #120 Executive Summary Conventional methods for specifying data center density are ambiguous and misleading. Describing
7 Best Practices for Increasing Efficiency, Availability and Capacity XXXX XXXXXXXX Liebert North America Emerson Network Power: The global leader in enabling Business-Critical Continuity Automatic Transfer
Optimizing Power Distribution for High-Density Computing Choosing the right power distribution units for today and preparing for the future By Michael Camesano Product Manager Eaton Corporation Executive
Unified Physical Infrastructure SM (UPI) Strategies for Smart Data Centers Deploying a Vertical Exhaust System www.panduit.com WP-09 September 2009 Introduction Business management applications and rich
WHITE PAPER Cooling Capacity Factor (CCF) Reveals Stranded Capacity and Data Center Cost Savings By Lars Strong, P.E., Upsite Technologies, Inc. Kenneth G. Brill, Upsite Technologies, Inc. 505.798.0200
Cooling Audit for Identifying Potential Cooling Problems in Data Centers By Kevin Dunlap White Paper #40 Revision 2 Executive Summary The compaction of information technology equipment and simultaneous
A Scalable, Reconfigurable, and Efficient Data Center Power Distribution Architecture By Neil Rasmussen White Paper #129 Executive Summary Significant improvements in efficiency, power density, power monitoring,
Optimizing Network Performance through PASSIVE AIR FLOW MANAGEMENT IN THE DATA CENTER Lylette Macdonald, RCDD Legrand Ortronics BICSI Baltimore 2011 Agenda: Discuss passive thermal management at the Rack
Managing Today s Data Centers Avoiding the Impending Crisis Steven Yellen Aperture Research Institute EXECUTIVE SUMMARY Organizations are facing an impending data center management crisis. The constant
Data Center Energy Profiler Questions Checklist Step 1 Case Name Date Center Company State/Region County Floor Area Data Center Space Floor Area Non Data Center Space Floor Area Data Center Support Space
Dynamic Power Variations in Data Centers and Network Rooms By Jim Spitaels White Paper #43 Revision 2 Executive Summary The power requirement required by data centers and network rooms varies on a minute
Calculating Total Power Requirements for Data Centers By Richard Sawyer White Paper #3 Executive Summary Part of data center planning and design is to align the power and cooling requirements of the IT
Energy Efficient Data Centre Cooling St. Andrews and Elsewhere Richard Lumb Consultant Engineer Future-Tech Energy Efficient Data Centre Cooling There is no one single best cooling solution for all Data
Dynamic Power Variations in Data Centers and Network Rooms White Paper 43 Revision 3 by James Spitaels > Executive summary The power requirement required by data centers and network rooms varies on a minute
Data Center Consolidation Trends & Solutions Bob Miller Vice President, Global Solutions Sales Emerson Network Power / Liebert Agenda Traditional Data Centers Drivers of a Changing Environment Alternate
Guide for Reducing Data Center Physical Infrastructure Energy Consumption in Federal Data Centers October 2010/White Paper 250 by Dennis Bouley, Senior Research Analyst Data Center Science Center and Ellen
A White Paper from the Experts in Business-Critical Continuity TM Combining Cold Aisle Containment with Intelligent Control to Optimize Data Center Cooling Efficiency Executive Summary Energy efficiency
CABINETS: ENCLOSED THERMAL MOUNTING MANAGEMENT SYSTEMS WHITE PAPER Supporting Cisco Switches In Hot Aisle/Cold Aisle Data Centers 800-834-4969 firstname.lastname@example.org www.chatsworth.com All products
WHITE PAPER #23 PROPER SIZING OF IT POWER AND COOLING LOADS WHITE PAPER CONTRIBUTORS: JOHN BEAN, APC RON BEDNAR, EMERSON NETWORK POWER RICHARD JONES, CHATSWORTH PRODUCTS ROBB JONES, CHATSWORTH PRODUCTS
Rittal Liquid Cooling Series by Herb Villa White Paper 04 Copyright 2006 All rights reserved. Rittal GmbH & Co. KG Auf dem Stützelberg D-35745 Herborn Phone +49(0)2772 / 505-0 Fax +49(0)2772/505-2319 www.rittal.de
GREEN FIELD DATA CENTER DESIGN WATER COOLING FOR MAXIMUM EFFICIENCY Shlomo Novotny, Vice President and Chief Technology Officer, Vette Corp. Overview Data centers are an ever growing part of our economy.
Best Practices for the EU Code of Conduct on Data Centres Version 1.0.0 First Release Release Public 1 Introduction This document is a companion to the EU Code of Conduct on Data Centres v0.9. This document
Data Center Operating Cost Savings Realized by Air Flow Management and Increased Rack Inlet Temperatures William Seeber Stephen Seeber Mid Atlantic Infrared Services, Inc. 5309 Mohican Road Bethesda, MD
WHITE PAPER #19 USING VIRTUALIZATION TO IMPROVE DATA CENTER EFFICIENCY EDITOR: RICHARD TALABER, VMWARE CONTRIBUTORS: TOM BREY, IBM LARRY LAMERS, VMWARE PAGE 2 ABSTRACT This paper outlines some of the advantages,
- White Paper - Data Centre Cooling Best Practice Release 2, April 2008 Contents INTRODUCTION... 3 1. AIR FLOW LEAKAGE... 3 2. PERFORATED TILES: NUMBER AND OPENING FACTOR... 4 3. PERFORATED TILES: WITH
White Paper Data Center Containment Cooling Strategies WHITE PAPER EC9001 Geist Updated August 2010 Abstract Deployment of high density IT equipment into data center infrastructure is now a common occurrence
Raised Floors vs Hard Floors for Data Center Applications White Paper 19 Revision 3 by Neil Rasmussen Executive summary Raised floors were once a standard feature of data centers, but over time a steadily
^ Business-wise, Future-drivenTM Cheat Sheets to Data Center Infrastructure Management The Brands You Trust. Table of Contents Introduction...3 How DCIM helps operations...4 How DCIM helps planning...5
GUIDE TO ICT SERVER ROOM ENERGY EFFICIENCY Public Sector ICT Special Working Group SERVER ROOM ENERGY EFFICIENCY This guide is one of a suite of documents that aims to provide guidance on ICT energy efficiency.
DataCenter 2020: hot aisle and cold aisle containment efficiencies reveal no significant differences November 2011 Powered by DataCenter 2020: hot aisle and cold aisle containment efficiencies reveal no
Green Computing: Datacentres Simin Nadjm-Tehrani Department of Computer and Information Science (IDA) Linköping University Sweden Many thanks to Jordi Cucurull For earlier versions of this course material
Green Computing: Datacentres Simin Nadjm-Tehrani Department of Computer and Information Science (IDA) Linköping University Sweden Many thanks to Jordi Cucurull For earlier versions of this course material
Data Center Energy Efficiency Looking Beyond PUE No Limits Software White Paper #4 By David Cole 2011 No Limits Software. All rights reserved. No part of this publication may be used, reproduced, photocopied,
Calculating Space and Power Density Requirements for Data Centers White Paper 155 Revision 0 by Neil Rasmussen Executive summary The historic method of specifying data center power density using a single
Energy Efficiency and Green Data Centers Overview of Recommendations ITU-T L.1300 and ITU-T L.1310 Paolo Gemma Chairman of Working Party 3 of ITU-T Study Group 5 International Telecommunication Union Agenda